Skip to content

An unofficial implementation of DeepVQE proposed by Microsoft Corp.

Notifications You must be signed in to change notification settings

Xiaobin-Rong/deepvqe

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

20 Commits
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

DeepVQE

A PyTorch implementation of DeepVQE described in DeepVQE: Real Time Deep Voice Quality Enhancement for Joint Acoustic Echo Cancellation, Noise Suppression and Dereverberation.

About DeepVQE

DeepVQE DeepVQE is a speech enhancement (SE) model proposed by Microsoft for joint echo cancellation, noise suppression and dereverberation, which outperforms the top 1 models in both 2023 DNS Challenge and 2023 AEC Challenge.

DeepVQE utilizes the U-Net architecture as backbone, while makes some improvements:

  • A new cross-attention mechanism for the microphone and far end soft alignment.
  • Add residual block for each block in encoder and decoder.
  • Use sub-pixel convolution instead of transposed convolution for up-sampling.
  • A novel mask mechanism named complex convolving mask (CCM).

Our purpose

We implement DeepVQE aiming to compare its SE performance with other two SOTA SE models, DPCRN and TF-GridNet. To this end, We modify some experimental setup in the original paper, specifically:

  • Datasets: we use DNS3 datasets in which all the utterances are sampled at 16 kHz.
  • STFT: we use a squared root Hann window of length 32 ms, a hop length of 16 ms, and an FFT length of 512.
  • Align Block: we drop the Align Block, because we do not focus on its AEC performance. Anyway, we still provide an implementation of the Align Block in our codes.

We are also interested in the inference speed presented in the paper, i.e, a relatively fast speed of 3.66 ms per frame in spite of its large complexity. So we also provide a stream version of DeepVQE, which is utilized to evaluate its inference speed.

Requirements

einops
numpy
onnx
onnxruntime
onnxsim
ptflops
torch==1.11.0

Results

1. SE performance

We are sorry to find that DeepVQE outperforms DPCRN only with a very limited margin, while requirng for much more computational resources (see below). Besides, DeepVQE lags behind TF-GridNet by a relatively large margin in terms of SE performance.

Model Param. (M) FLOPs (G)
DPCRN 0.81 3.73
TF-GridNet 1.60 22.23
DeepVQE 7.51 8.04

2. Inference speed

We are surprised to find that although DeepVQE requires for large computational resources, it achieves a relatively good real-time factor of 0.2, which corresponds to the data presented in the paper.

About

An unofficial implementation of DeepVQE proposed by Microsoft Corp.

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages