ISCA Archive Interspeech 2022
ISCA Archive Interspeech 2022

Ultra-Low-Bitrate Speech Coding with Pretrained Transformers

Ali Siahkoohi, Michael Chinen, Tom Denton, W. Bastiaan Kleijn, Jan Skoglund

Speech coding facilitates the transmission of speech over low-bandwidth networks with minimal distortion. Neural-network based speech codecs have recently demonstrated significant improvements in quality over traditional approaches. While this new generation of codecs is capable of synthesizing high-fidelity speech, their use of recurrent or convolutional layers often restricts their effective receptive fields, which prevents them from compressing speech efficiently. We propose to further reduce the bitrate of neural speech codecs through the use of pretrained Transformers, capable of exploiting long-range dependencies in the input signal due to their inductive bias. As such, we use a pretrained Transformer in tandem with a convolutional encoder, which is trained end-to-end with a quantizer and a generative adversarial net decoder. Our numerical experiments show that supplementing the convolutional encoder of a neural speech codec with Transformer speech embeddings yields a speech codec with a bitrate of 600 bps that outperforms the original neural speech codec in synthesized speech quality when trained at the same bitrate. Subjective human evaluations suggest that the quality of the resulting codec is comparable or better than that of conventional codecs operating at three to four times the rate.


doi: 10.21437/Interspeech.2022-10988

Cite as: Siahkoohi, A., Chinen, M., Denton, T., Kleijn, W.B., Skoglund, J. (2022) Ultra-Low-Bitrate Speech Coding with Pretrained Transformers. Proc. Interspeech 2022, 4421-4425, doi: 10.21437/Interspeech.2022-10988

@inproceedings{siahkoohi22_interspeech,
  author={Ali Siahkoohi and Michael Chinen and Tom Denton and W. Bastiaan Kleijn and Jan Skoglund},
  title={{Ultra-Low-Bitrate Speech Coding with Pretrained Transformers}},
  year=2022,
  booktitle={Proc. Interspeech 2022},
  pages={4421--4425},
  doi={10.21437/Interspeech.2022-10988},
  issn={2308-457X}
}