Skip to content

Invited Talks

2024

  • 山岡洸瑛, “非同期分散マイクロフォンアレイの研究動向,” in 電気情報通信学会 信号処理研究会, Aug. 2024.
  • 岡本悠希, “環境音合成の研究動向,” in 電気情報通信学会 信号処理研究会, Aug. 2024.
  • 佐伯 高明 and 高道 慎之介, “最先端の予測性能を持つ合成音声品質の自動評価システム UTMOS について,” in 日本音響学会誌, Jul. 2024.

2023

  • K. Seki, “Data Selection for Text-to-speech with Feedback from Automatic Evaluation of Naturalness on Synthetic Speech,” in Joint Workshop of VoicePersonae and ASVspoof 2023 (VoiceMOS mini workshop), Nov. 2023.

2022

  • T. Nakamura, “Signal-processing-inspired deep learning,” in Seminar by IEEE NZ Signal Processing / Information Theory Joint Chapter and Acoustics Research Center, the University of Auckland, Dec. 2022.
  • Y. Saito, “Towards human-in-the-loop DNN-based speech synthesis technologies,” in Seminar by IEEE NZ Signal Processing / Information Theory Joint Chapter and Acoustics Research Center, the University of Auckland, Dec. 2022.
  • Y. Saito, “Towards human-in-the-loop speech synthesis,” in Seminar by IEEE Systems, Man and Cybernetics Singapore Chapter, Chinese and Oriental Languages Information Processing Society Teochew Doctorate Society, Singapore, and Human Language Technology Lab., National University of Singapore, Aug. 2022.
  • D. Xin, “Maintaining data consistency in speech quality assessment and speech emotion recognition,” in Seminar by IEEE Systems, Man and Cybernetics Singapore Chapter, Chinese and Oriental Languages Information Processing Society Teochew Doctorate Society, Singapore, and Human Language Technology Lab., National University of Singapore, Aug. 2022.

2021

  • H. Saruwatari, “Multichannel audio source separation based on unsupervised and semi-supervised learning,” in Proceedings of Chinese Computer Federation, Jan. 2021.

2020

  • H. Saruwatari, “Multichannel audio source separation based on unsupervised and semi-supervised learning,” in Proceedings of Asia-Pacific Signal and Information Processing Association Annual Summit and Conference (APSIPA ASC), Dec. 2020.

2019

  • Y. Takida, S. Koyama, N. Ueno, and H. Saruwatari, “Comparison of Interpolation Methods for Gridless Sound Field Decomposition Based on Reciprocity Gap Functional,” in Proceedings of International Congress on Sound and Vibration (ICSV), Montreal, Jul. 2019. (to appear)
  • S. Takamichi, “Group-delay modelling based on deep neural network with sine-skewed generalized cardioid distribution,” in Proceedings of International Conference on Soft Computing & Machine Learning (SCML), Wuhan, China, Apr. 2019. (invited)

2018

  • M. Une, Y. Saito, S. Takamichi, D. Kitamura, R. Miyazaki, and H. Saruwatari, “Generative approach using the noise generation models for DNN-based speech synthesis trained from noisy speech,” in Proceedings of Asia-Pacific Signal and Information Processing Association Annual Summit and Conference (APSIPA ASC), Hawaii, Nov. 2018.
  • S. Koyama, “Sparsity-based sound field reconstruction,” in Tohoku Universal Acoustical Communication Month, Seminar on the spatial aspects of hearing and their applications, keynote lecture, Sendai, Oct. 2018.
  • S. Takamichi, “What can GAN and GMMN do for augmented speech communication?,” in GMI workshop, Hiroshima, Aug. 2018.

2017

  • S. Takamichi, “Modulation spectrum-based speech parameter trajectory smoothing for DNN-based speech synthesis using FFT spectra,” in Proceedings of Asia-Pacific Signal and Information Processing Association Annual Summit and Conference (APSIPA ASC), Kuala Lumpur, Dec. 2017.
  • D. Kitamura, N. Ono, and H. Saruwatari, “Experimental analysis of optimal window length for independent low-rank matrix analysis,” in Proceedings of Proceedings of 25th European Signal Processing Conference, Greek island of Kos, Aug. 2017.
  • S. Koyama, N. Murata, and H. Saruwatari, “Effect of multipole dictionary in sparse sound field decomposition for super-resolution in recording and reproduction,” in Proceedings of International Congress on Sound and Vibration (ICSV), London, Jul. 2017.

2016

  • H. Nakajima, D. Kitamura, N. Takamune, S. Koyama, H. Saruwatari, Y. Takahashi, and K. Kondo, “Audio signal separation using supervised NMF with time-variant all-pole-model-based basis deformation,” in Proceedings of Asia-Pacific Signal and Information Processing Association Annual Summit and Conference (APSIPA ASC), Jeju, Dec. 2016.
  • S. Takamichi, “Speech synthesis that deceives anti-spoofing verification,” in NII Talk, Dec. 2016.
  • S. Koyama, N. Murata, and H. Saruwatari, “Super-resolution in sound field recording and reproduction based on sparse representation,” in Proceedings of 5th Joint Meeting of the Acoustical Society of America and Acoustical Society of Japan, Honolulu, Nov. 2016.
  • H. Saruwatari, K. Takata, N. Ono, and S. Makino, “Flexible microphone array based on multichannel nonnegative matrix factorization and statistical signal estimation,” in The 22nd International Congress on Acoustics (ICA2016), Sep. 2016, no. ICA2016-312.
  • S. Koyama, “Source-Location-Informed Sound Field Recording and Reproduction: A Generalization to Arrays of Arbitrary Geometry,” in Proceedings of 2016 AES International Conference on Sound Field Control, Guildford, Jul. 2016 [Online]. Available at: http://www.aes.org/e-lib/browse.cfm?elib=18303

2015

  • S. Koyama, A. Matsubayashi, N. Murata, and H. Saruwatari, “Sparse Sound Field Decomposition Using Group Sparse Bayesian Learning,” in Proceedings of Asia-Pacific Signal and Information Processing Association Annual Summit and Conference (APSIPA ASC), Dec. 2015, pp. 850–855.
  • D. Kitamura, N. Ono, H. Sawada, H. Kameoka, and H. Saruwatari, “Relaxation of rank-1 spatial constraint in overdetermined blind source separation,” in in Proceedings of The 2015 European Signal Processing Conference (EUSIPCO2015), Nice, Sep. 2015, pp. 1271–1275.
  • H. Saruwatari, “Statistical-model-based speech enhancement With musical-noise-free properties,” in in Proceedings of 2015 IEEE International Conference on Digital Signal Processing (DSP2015), Singapore, 2015.

2014

  • D. Kitamura, H. Saruwatari, S. Nakamura, Y. Takahashi, K. Kondo, and H. Kameoka, “Hybrid multichannel signal separation using supervised nonnegative matrix factorization with spectrogram restoration,” in Proceedings of Asia-Pacific Signal and Information Processing Association Annual Summit and Conference (APSIPA ASC), Siem Reap, Dec. 2014.


Preprints

2020

  • N. T. Shinnosuke Takamichi Mamoru Komachi and H. Saruwatari, “JSSS: free Japanese speech corpus for summarization and simplification,” in arXiv, Oct. 2020.