Speech enhancement using self-adaptation and multi-head self-attention Y Koizumi, K Yatabe, M Delcroix, Y Masuyama, D Takeuchi ICASSP 2020-2020 IEEE International Conference on Acoustics, Speech and …, 2020 | 138 | 2020 |
Deep griffin–lim iteration Y Masuyama, K Yatabe, Y Koizumi, Y Oikawa, N Harada ICASSP 2019-2019 IEEE International Conference on Acoustics, Speech and …, 2019 | 71 | 2019 |
Griffin–Lim like phase recovery via alternating direction method of multipliers Y Masuyama, K Yatabe, Y Oikawa IEEE Signal Processing Letters 26 (1), 184-188, 2018 | 49 | 2018 |
Representation of complex spectrogram via phase conversion K Yatabe, Y Masuyama, T Kusano, Y Oikawa Acoustical Science and Technology 40 (3), 170-177, 2019 | 37 | 2019 |
Deep Griffin–Lim iteration: Trainable iterative phase reconstruction using neural network Y Masuyama, K Yatabe, Y Koizumi, Y Oikawa, N Harada IEEE Journal of Selected Topics in Signal Processing 15 (1), 37-50, 2020 | 29 | 2020 |
The chime-7 dasr challenge: Distant meeting transcription with multiple devices in diverse scenarios S Cornell, M Wiesner, S Watanabe, D Raj, X Chang, P Garcia, ... arXiv preprint arXiv:2306.13734, 2023 | 28 | 2023 |
Neural full-rank spatial covariance analysis for blind source separation Y Bando, K Sekiguchi, Y Masuyama, AA Nugraha, M Fontaine, K Yoshii IEEE Signal Processing Letters 28, 1670-1674, 2021 | 28 | 2021 |
Unsupervised training for deep speech source separation with Kullback-Leibler divergence based probabilistic loss function M Togami, Y Masuyama, T Komatsu, Y Nakagome ICASSP 2020-2020 IEEE International Conference on Acoustics, Speech and …, 2020 | 28 | 2020 |
Phase-aware harmonic/percussive source separation via convex optimization Y Masuyama, K Yatabe, Y Oikawa ICASSP 2019-2019 IEEE International Conference on Acoustics, Speech and …, 2019 | 27 | 2019 |
Low-rankness of complex-valued spectrogram and its application to phase-aware audio processing Y Masuyama, K Yatabe, Y Oikawa ICASSP 2019-2019 IEEE International Conference on Acoustics, Speech and …, 2019 | 26 | 2019 |
Phase reconstruction based on recurrent phase unwrapping with deep neural networks Y Masuyama, K Yatabe, Y Koizumi, Y Oikawa, N Harada ICASSP 2020-2020 IEEE International Conference on Acoustics, Speech and …, 2020 | 25 | 2020 |
ESPnet-SE++: Speech enhancement for robust speech recognition, translation, and understanding YJ Lu, X Chang, C Li, W Zhang, S Cornell, Z Ni, Y Masuyama, B Yan, ... arXiv preprint arXiv:2207.09514, 2022 | 20 | 2022 |
Model-based phase recovery of spectrograms via optimization on Riemannian manifolds Y Masuyama, K Yatabe, Y Oikawa 2018 16th International Workshop on Acoustic Signal Enhancement (IWAENC …, 2018 | 17 | 2018 |
Rectified linear unit can assist Griffin-Lim phase recovery K Yatabe, Y Masuyama, Y Oikawa 2018 16th international workshop on acoustic signal enhancement (IWAENC …, 2018 | 17 | 2018 |
End-to-end integration of speech recognition, dereverberation, beamforming, and self-supervised learning representation Y Masuyama, X Chang, S Cornell, S Watanabe, N Ono 2022 IEEE Spoken Language Technology Workshop (SLT), 260-265, 2023 | 14 | 2023 |
Self-supervised neural audio-visual sound source localization via probabilistic spatial modeling Y Masuyama, Y Bando, K Yatabe, Y Sasaki, M Onishi, Y Oikawa 2020 IEEE/RSJ International Conference on Intelligent Robots and Systems …, 2020 | 14 | 2020 |
Designing nearly tight window for improving time-frequency masking T Kusano, Y Masuyama, K Yatabe, Y Oikawa arXiv preprint arXiv:1811.08783, 2018 | 12 | 2018 |
Multichannel loss function for supervised speech source separation by mask-based beamforming Y Masuyama, M Togami, T Komatsu arXiv preprint arXiv:1907.04984, 2019 | 11 | 2019 |
Consistency-aware multi-channel speech enhancement using deep neural networks Y Masuyama, M Togami, T Komatsu ICASSP 2020-2020 IEEE International Conference on Acoustics, Speech and …, 2020 | 10 | 2020 |
Online phase reconstruction via DNN-based phase differences estimation Y Masuyama, K Yatabe, K Nagatomo, Y Oikawa IEEE/ACM Transactions on Audio, Speech, and Language Processing 31, 163-176, 2022 | 7 | 2022 |