The 2nd clarity enhancement challenge for hearing aid speech intelligibility enhancement: Overview and outcomes

MA Akeroyd, W Bailey, J Barker, TJ Cox… - ICASSP 2023-2023 …, 2023 - ieeexplore.ieee.org
This paper reports on the design and outcomes of the 2nd Clarity Enhancement Challenge
(CEC2), a challenge for stimulating novel approaches to hearing-aid speech intelligibility …

Remixit: Continual self-training of speech enhancement models via bootstrapped remixing

E Tzinis, Y Adi, VK Ithapu, B Xu… - IEEE Journal of …, 2022 - ieeexplore.ieee.org
We present RemixIT, a simple yet effective self-supervised method for training speech
enhancement without the need of a single isolated in-domain speech nor a noise waveform …

Dual-path mamba: Short and long-term bidirectional selective structured state space models for speech separation

X Jiang, C Han, N Mesgarani - arXiv preprint arXiv:2403.18257, 2024 - arxiv.org
Transformers have been the most successful architecture for various speech modeling tasks,
including speech separation. However, the self-attention mechanism in transformers with …

The CHiME-7 UDASE task: Unsupervised domain adaptation for conversational speech enhancement

S Leglaive, L Borne, E Tzinis, M Sadeghi… - arXiv preprint arXiv …, 2023 - arxiv.org
Supervised speech enhancement models are trained using artificially generated mixtures of
clean speech and noise signals, which may not match real-world recording conditions at test …

Efficient monaural speech enhancement with universal sample rate band-split RNN

J Yu, Y Luo - … 2023-2023 IEEE International Conference on …, 2023 - ieeexplore.ieee.org
While recent developments on the design of neural networks have greatly advanced the
state-of-the-art of speech enhancement and separation systems, practical applications of …

Universal source separation with weakly labelled data

Q Kong, K Chen, H Liu, X Du, T Berg-Kirkpatrick… - arXiv preprint arXiv …, 2023 - arxiv.org
Universal source separation (USS) is a fundamental research task for computational
auditory scene analysis, which aims to separate mono recordings into individual source …

Heterogeneous target speech separation

E Tzinis, G Wichern, A Subramanian… - arXiv preprint arXiv …, 2022 - arxiv.org
We introduce a new paradigm for single-channel target source separation where the
sources of interest can be distinguished using non-mutually exclusive concepts (eg …

[HTML][HTML] An efficient time-domain end-to-end single-channel bird sound separation network

C Zhang, Y Chen, Z Hao, X Gao - Animals, 2022 - mdpi.com
Simple Summary Automatic bird sound recognition using artificial intelligence technology
has been widely used to identify bird species recently. However, the bird sounds recorded in …

Separate but together: Unsupervised federated learning for speech enhancement from non-iid data

E Tzinis, J Casebeer, Z Wang… - 2021 IEEE Workshop …, 2021 - ieeexplore.ieee.org
We propose FedEnhance, an unsupervised federated learning (FL) approach for speech
enhancement and separation with non-IID distributed data across multiple clients. We …

On time domain conformer models for monaural speech separation in noisy reverberant acoustic environments

W Ravenscroft, S Goetze, T Hain - 2023 IEEE Automatic …, 2023 - ieeexplore.ieee.org
Speech separation remains an important topic for multispeaker technology researchers.
Convolution augmented transformers (conformers) have performed well for many speech …