Skip to content

The official implementation of GTCRN, an ultra-lite speech enhancement model.

License

Notifications You must be signed in to change notification settings

Xiaobin-Rong/gtcrn

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

72 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

GTCRN

This repository is the official implementation of the ICASSP2024 paper: GTCRN: A Speech Enhancement Model Requiring Ultralow Computational Resources.

Audio examples are available at Audio examples of GTCRN.

About GTCRN

Grouped Temporal Convolutional Recurrent Network (GTCRN) is a speech enhancement model requiring ultralow computational resources, featuring only 23.7 K parameters and 33.0 MMACs per second. Experimental results show that our proposed model not only surpasses RNNoise, a typical lightweight model with similar computational burden, but also achieves competitive performance when compared to recent baseline models with significantly higher computational resources requirements.

Note:

  • Although the complexity mentioned in the paper is 39.6 MMACs per second, we find that it can be further reduced to 33.0 MMACs per second. This reduction can be achieved by modifying only the ERB module, specifically by substituting the invariant mapping from linear bands to ERB bands in the low-frequency dimension, from matrix multiplication to simple concatenation.
  • The explicit feature rearrangement layer in the grouped RNN, which is implemented by feature shuffle, can result in an unstreamable model. Therefore, we discard it and implicitly achieve feature rearrangement through the following FC layer in the DPGRNN.

Performance

Experiments show that GTCRN not only outperforms RNNoise by a substantial margin on the VCTK-DEMAND and DNS3 dataset, but also achieves competitive performance compared to several baseline models with significantly higher computational overhead.

Table 1: Performance on VCTK-DEMAND test set

Para. (M) MACs (G/s) SISNR PESQ STOI
Noisy - - 8.45 1.97 0.921
RNNoise (2018) 0.06 0.04 - 2.29 -
PercepNet (2020) 8.00 0.80 - 2.73 -
DeepFilterNet (2022) 1.80 0.35 16.63 2.81 0.942
S-DCCRN (2022) 2.34 - - 2.84 0.940
GTCRN (proposed) 0.02 0.04 18.83 2.87 0.940

Table 2: Performance on DNS3 blind test set.

Para. (M) MACs (G/s) DNSMOS-P.808 BAK SIG OVRL
Noisy - - 2.96 2.65 3.20 2.33
RNNoise (2018) 0.06 0.04 3.15 3.45 3.00 2.53
S-DCCRN (2022) 2.34 - 3.43 - - -
GTCRN (proposed) 0.02 0.04 3.44 3.90 3.00 2.70

Pre-trained Models

Pre-trained models are provided in checkpoints folder, which were trained on DNS3 and VCTK-DEMAND datasets, respectively.

The inference procedure is presented in infer.py.

Streaming Inference

A streaming GTCRN is provided in stream folder, which demonstrates an impressive real-time factor (RTF) of 0.07 on the 12th Gen Intel(R) Core(TM) i5-12400 CPU @ 2.50 GHz.

Related Repositories

SEtrain: A training code template for DNN-based speech enhancement.

TRT-SE: An example of how to convert a speech enhancement model into a streaming format and deploy it using ONNX or TensorRT.

About

The official implementation of GTCRN, an ultra-lite speech enhancement model.

Topics

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages