NANKAKU Yoshihiko

写真a

Affiliation Department etc.

Department of Computer Science
Department of Computer Science

Title

Associate Professor

Graduating School

  •  
    -
    2000.03

    Nagoya Institute of Technology   Faculty of Engineering   Graduated

Graduate School

  •  
    -
    2005.03

    Nagoya Institute of Technology  Graduate School, Division of Engineering  Doctor's Course  Completed

  •  
    -
    2002.03

    Nagoya Institute of Technology  Graduate School, Division of Engineering  Doctor's Course  Completed

Degree

  • Nagoya Institute of Technology -  Doctor of Engineering

  • Nagoya Institute of Technology -  Master of Engineering

Field of expertise (Grants-in-aid for Scientific Research classification)

  • Perceptual information processing

 

Research Career

  • Speech Synthesis

    (not selected)  

    Project Year:  2002.04  -  2002.04

  • Speech Recognition

    (not selected)  

    Project Year:  2001.04  -  2001.04

  • Image recognition

    (not selected)  

    Project Year:  1999.04  -  1999.04

Papers

  • PeriodNet: A Non-Autoregressive Raw Waveform Generative Model With a Structure Separating Periodic and Aperiodic Components

    Yukiya Hono, Shinji Takaki, Kei Hashimoto, Keiichiro Oura, Yoshihiko Nankaku, and Keiichi Tokuda

    IEEE Access   9   137599 - 137612   2021.10  [Refereed]

    Research paper (scientific journal)   Multiple Authorship

  • Sinsy: A Deep Neural Network-Based Singing Voice Synthesis System

    Yukiya Hono, Kei Hashimoto, Keiichiro Oura, Yoshihiko Nankaku, and Keiichi Tokuda

    IEEE/ACM Transactions on Audio, Speech and Language Processing   29   2803 - 2815   2021.08  [Refereed]

    Research paper (scientific journal)   Multiple Authorship

  • PeriodNet: A non-autoregressive waveform generation model with a structure separating periodic and aperiodic components

    Yukiya Hono, Shinji Takaki, Kei Hashimoto, Keiichiro Oura, Yoshihiko Nankaku, and Keiichi Tokuda

    Proceedings of 2021 IEEE International Conference on Acoustics, Speech, and Signal Processing (ICASSP 2021)     6049 - 6053   2021.06  [Refereed]

    Research paper (international conference proceedings)   Multiple Authorship

  • Hierarchical multi-grained generative model for expressive speech synthesis

    Yukiya Hono, KazunaTsuboi, Kei Sawada, Kei Hashimoto, Keiichiro Oura, Yoshihiko Nankaku, and Keiichi Tokuda

    Proceedings of Interspeech 2020     3441 - 3445   2020.10  [Refereed]

    Research paper (international conference proceedings)   Multiple Authorship

  • Fast and high-quality singing voice synthesis system based on convolutional neural networks

    Kazuhiro Nakamura, Shinji Takaki, Kei Hashimoto, Keiichiro Oura, Yoshihiko Nankaku, and Keiichi Tokuda

    Proceedings of 2020 IEEE International Conference on Acoustics, Speech, and Signal Processing (ICASSP 2020)     7239 - 7243   2020.05  [Refereed]

    Research paper (international conference proceedings)   Multiple Authorship

  • Semi-supervised learning based on hierarchical generative models for end-to-end speech synthesis

    Takato Fujimoto, Shinji Takaki, Kei Hashimoto, Keiichiro Oura, Yoshihiko Nankaku, and Keiichi Tokuda

    Proceedings of 2020 IEEE International Conference on Acoustics, Speech, and Signal Processing (ICASSP 2020)     7644 - 7648   2020.05  [Refereed]

    Research paper (international conference proceedings)   Multiple Authorship

  • Impacts of input linguistic feature representation on Japanese end-to-end speech synthesis

    Takato Fujimoto, Kei Hashimoto, Keiichiro Oura, Yoshihiko Nankaku, and Keiichi Tokuda

    Proceedings of 10th ISCA Speech Synthesis Workshop (SSW10)     166 - 171   2019.09  [Refereed]

    Research paper (international conference proceedings)   Multiple Authorship

  • Low computational cost speech synthesis based on deep neural networks using hidden semi-Markov model structures

    Motoki Shimada, Kei Hashimoto, Keiichiro Oura, Yoshihiko Nankaku, and Keiichi Tokuda

    Proceedings of 10th ISCA Speech Synthesis Workshop (SSW10)     177 - 182   2019.09  [Refereed]

    Research paper (international conference proceedings)   Multiple Authorship

  • Deep neural network based real-time speech vocoder with periodic and aperiodic inputs

    Keiichiro Oura, Kazuhiro Nakamura, Kei Hashimoto, Yoshihiko Nankaku, and Keiichi Tokuda

    Proceedings of 10th ISCA Speech Synthesis Workshop (SSW10)     13 - 18   2019.09  [Refereed]

    Research paper (international conference proceedings)   Multiple Authorship

  • Speaker-dependent WaveNet-based delay-free ADPCM speech coding

    Takenori Yoshimura, Kei Hashimoto, Keiichiro Oura, Yoshihiko Nankaku, and Keiichi Tokuda

    Proceedings of 2019 IEEE International Conference on Acoustics, Speech, and Signal Processing (ICASSP 2019)     7145 - 7149   2019.05  [Refereed]

    Research paper (international conference proceedings)   Multiple Authorship

display all >>

Review Papers

  • Speech Synthesis Based on Hidden Markov Models

    Keiichi Tokuda, Yoshihiko Nankaku, Tomoki Toda, Heiga Zen, Junichi, Yamagishi, and Keiichiro Oura

    Proceedings of the IEEE ( IEEE )  101 ( 5 ) 1234 - 1252   2013.05

    Introduction and explanation (international conference proceedings)   Multiple Authorship

Presentations

  • PeriodNet: A non-autoregressive waveform generation model with a structure separating periodic and aperiodic components

    Yukiya Hono, Shinji Takaki, Kei Hashimoto, Keiichiro Oura, Yoshihiko Nankaku, and Keiichi Tokuda

    2021 IEEE International Conference on Acoustics, Speech, and Signal Processing (ICASSP 2021)  2021.06  -  2021.06 

  • Hierarchical multi-grained generative model for expressive speech synthesis

    Yukiya Hono, KazunaTsuboi, Kei Sawada, Kei Hashimoto, Keiichiro Oura, Yoshihiko Nankaku, and Keiichi Tokuda

    Interspeech 2020  2020.10  -  2020.10 

  • Semi-supervised learning based on hierarchical generative models for end-to-end speech synthesis

    Takato Fujimoto, Shinji Takaki, Kei Hashimoto, Keiichiro Oura, Yoshihiko Nankaku, and Keiichi Tokuda

    2020IEEE International Conference on Acoustics, Speech, and Signal Processing (ICASSP 2020)  2020.05  -  2020.05 

  • Fast and high-quality singing voice synthesis system based on convolutional neural networks

    Kazuhiro Nakamura, Shinji Takaki, Kei Hashimoto, Keiichiro Oura, Yoshihiko Nankaku, and Keiichi Tokuda

    2020IEEE International Conference on Acoustics, Speech, and Signal Processing (ICASSP 2020)  2020.05  -  2020.05 

  • Deep neural network based real-time speech vocoder with periodic and aperiodic inputs

    Keiichiro Oura, Kazuhiro Nakamura, Kei Hashimoto, Yoshihiko Nankaku, and Keiichi Tokuda

    10th ISCA Speech Synthesis Workshop (SSW10)  2019.09  -  2019.09 

  • Impacts of input linguistic feature representation on Japanese end-to-end speech synthesis

    Takato Fujimoto, Kei Hashimoto, Keiichiro Oura, Yoshihiko Nankaku, and Keiichi Tokuda

    10th ISCA Speech Synthesis Workshop (SSW10)  2019.09  -  2019.09 

  • Low computational cost speech synthesis based on deep neural networks using hidden semi-Markov model structures

    Motoki Shimada, Kei Hashimoto, Keiichiro Oura, Yoshihiko Nankaku, and Keiichi Tokuda

    10th ISCA Speech Synthesis Workshop (SSW10)  2019.09  -  2019.09 

  • Singing voice synthesis based on generative adversarial networks

    Yukiya Hono, Kei Hashimoto, Keiichiro Oura, Yoshihiko Nankaku, and Keiichi Tokuda

    2019 IEEE International Conference on Acoustics, Speech, and Signal Processing (ICASSP 2019)  2019.05  -  2019.05 

  • Speaker-dependent WaveNet-based delay-free ADPCM speech coding

    Takenori Yoshimura, Kei Hashimoto, Keiichiro Oura, Yoshihiko Nankaku, and Keiichi Tokuda

    2019 IEEE International Conference on Acoustics, Speech, and Signal Processing (ICASSP 2019)  2019.05  -  2019.05 

  • WaveNet-based zero-delay lossless speech coding

    Takenori Yoshimura, Kei Hashimoto, Keiichiro Oura, Yoshihiko Nankaku, and Keiichi Tokuda

    2018 IEEE Workshop on Spoken Language Technology (SLT 2018)  (Athens, Greece)  2018.12  -  2018.12  IEEE

display all >>