Skip to Main content Skip to Navigation

Explicit memory inclusion for efficient artificial bandwidth extension

Abstract : Most ABE algorithms exploit contextual information or memory captured via the use of static or dynamic features extracted from neighbouring speech frames. The use of memory leads to higher dimensional features and increased computational complexity. When information from look-ahead frames is also utilised, then latency also increases. Past work points toward the benefit to ABE of exploiting memory in the form of dynamic features with a standard regression model. Even so, the literature is missing a quantitative analysis of the relative benefit of explicit memory inclusion. The research presented in this thesis assesses the degree to which explicit memory is of benefit and furthermore reports a number of different techniques that allow for its inclusion without significant increases to latency and computational complexity. Benefits are shown through both a quantitative analysis with an information-theoretic measure and subjective listening tests. Key contributions relate to the preservation of computational efficiency through the use of dimensionality reduction in the form of principal component analysis, semisupervised stacked autoencoders and conditional variational auto-encoders. The two latter techniques optimise dimensionality reduction to deliver superior ABE performance.
Complete list of metadata

Cited literature [317 references]  Display  Hide  Download
Contributor : ABES STAR :  Contact
Submitted on : Monday, August 31, 2020 - 3:31:34 PM
Last modification on : Sunday, June 26, 2022 - 9:52:23 AM
Long-term archiving on: : Tuesday, December 1, 2020 - 12:40:58 PM


Version validated by the jury (STAR)


  • HAL Id : tel-02926274, version 1


Pramod Bachhav. Explicit memory inclusion for efficient artificial bandwidth extension. Signal and Image Processing. Sorbonne Université, 2019. English. ⟨NNT : 2019SORUS492⟩. ⟨tel-02926274⟩



Record views


Files downloads