
Tatsuya Kawahara
Refine list

refinements active!
zoomed in on ?? of ?? records
view refined list in
export refined list as
showing all ?? records
2020 – today
- 2020
- [j51]Richeng Duan
, Tatsuya Kawahara
, Masatake Dantsuji, Hiroaki Nanjo:
Cross-Lingual Transfer Learning of Non-Native Acoustic Modeling for Pronunciation Error Detection and Diagnosis. IEEE ACM Trans. Audio Speech Lang. Process. 28: 391-401 (2020) - [j50]Kouhei Sekiguchi
, Yoshiaki Bando
, Aditya Arie Nugraha
, Kazuyoshi Yoshii
, Tatsuya Kawahara
:
Fast Multichannel Nonnegative Matrix Factorization With Directivity-Aware Jointly-Diagonalizable Spatial Covariance Matrices for Blind Source Separation. IEEE ACM Trans. Audio Speech Lang. Process. 28: 2610-2625 (2020) - [c258]Tianyu Zhao, Divesh Lala, Tatsuya Kawahara:
Designing Precise and Robust Dialogue Response Evaluators. ACL 2020: 26-33 - [c257]Shuying Zhang, Tianyu Zhao, Tatsuya Kawahara:
Topic-relevant Response Generation using Optimal Transport for an Open-domain Dialog System. COLING 2020: 4067-4077 - [c256]Yicheng Du, Kouhei Sekiguchi, Yoshiaki Bando, Aditya Arie Nugraha, Mathieu Fontaine, Kazuyoshi Yoshii, Tatsuya Kawahara:
Semi-supervised Multichannel Speech Separation Based on a Phone- and Speaker-Aware Deep Generative Model of Speech Spectrograms. EUSIPCO 2020: 870-874 - [c255]Hiroshi Ishiguro, Tatsuya Kawahara, Yutaka Nakamura:
Autonomous Dialogue Technologies in Symbiotic Human-robot Interaction. HRI (Companion) 2020: 650-651 - [c254]Divesh Lala, Koji Inoue, Tatsuya Kawahara:
Prediction of Shared Laughter for Human-Robot Dialogue. ICMI Companion 2020: 62-66 - [c253]Koji Inoue, Kohei Hara, Divesh Lala, Kenta Yamamoto, Shizuka Nakamura, Katsuya Takanashi, Tatsuya Kawahara:
Job Interviewer Android with Elaborate Follow-up Question Generation. ICMI 2020: 324-332 - [c252]Han Feng, Sei Ueno, Tatsuya Kawahara:
End-to-End Speech Emotion Recognition Combined with Acoustic-to-Word ASR Model. INTERSPEECH 2020: 501-505 - [c251]Hirofumi Inaguma, Masato Mimura, Tatsuya Kawahara:
CTC-Synchronous Training for Monotonic Attention Model. INTERSPEECH 2020: 571-575 - [c250]Hirofumi Inaguma, Masato Mimura, Tatsuya Kawahara:
Enhancing Monotonic Multihead Attention for Streaming ASR. INTERSPEECH 2020: 2137-2141 - [c249]Kohei Matsuura, Masato Mimura, Shinsuke Sakai, Tatsuya Kawahara:
Generative Adversarial Training Data Adaptation for Very Low-Resource Automatic Speech Recognition. INTERSPEECH 2020: 2737-2741 - [c248]Hayato Futami, Hirofumi Inaguma, Sei Ueno, Masato Mimura, Shinsuke Sakai, Tatsuya Kawahara:
Distilling the Knowledge of BERT for Sequence-to-Sequence ASR. INTERSPEECH 2020: 3635-3639 - [c247]Viet-Trung Dang, Tianyu Zhao, Sei Ueno, Hirofumi Inaguma, Tatsuya Kawahara:
End-to-End Speech-to-Dialog-Act Recognition. INTERSPEECH 2020: 3910-3914 - [c246]Kenta Yamamoto, Koji Inoue, Tatsuya Kawahara:
Semi-Supervised Learning for Character Expression of Spoken Dialogue Systems. INTERSPEECH 2020: 4188-4192 - [c245]Kohei Matsuura, Sei Ueno, Masato Mimura, Shinsuke Sakai, Tatsuya Kawahara:
Speech Corpus of Ainu Folklore and End-to-end Speech Recognition for Ainu Language. LREC 2020: 2622-2628 - [c244]Koji Inoue, Divesh Lala, Kenta Yamamoto, Shizuka Nakamura, Katsuya Takanashi, Tatsuya Kawahara:
An Attentive Listening System with Android ERICA: Comparison of Autonomous and WOZ Interactions. SIGdial 2020: 118-127 - [i18]Kohei Matsuura, Sei Ueno, Masato Mimura, Shinsuke Sakai, Tatsuya Kawahara:
Speech Corpus of Ainu Folklore and End-to-end Speech Recognition for Ainu Language. CoRR abs/2002.06675 (2020) - [i17]Tianyu Zhao, Divesh Lala, Tatsuya Kawahara:
Designing Precise and Robust Dialogue Response Evaluators. CoRR abs/2004.04908 (2020) - [i16]Viet-Trung Dang, Tianyu Zhao, Sei Ueno, Hirofumi Inaguma, Tatsuya Kawahara:
End-to-end speech-to-dialog-act recognition. CoRR abs/2004.11419 (2020) - [i15]Hirofumi Inaguma, Masato Mimura, Tatsuya Kawahara:
CTC-synchronous Training for Monotonic Attention Model. CoRR abs/2005.04712 (2020) - [i14]Kohei Matsuura, Masato Mimura, Shinsuke Sakai, Tatsuya Kawahara:
Generative Adversarial Training Data Adaptation for Very Low-resource Automatic Speech Recognition. CoRR abs/2005.09256 (2020) - [i13]Hirofumi Inaguma, Masato Mimura, Tatsuya Kawahara:
Enhancing Monotonic Multihead Attention for Streaming ASR. CoRR abs/2005.09394 (2020) - [i12]Hayato Futami, Hirofumi Inaguma, Sei Ueno, Masato Mimura, Shinsuke Sakai, Tatsuya Kawahara:
Distilling the Knowledge of BERT for Sequence-to-Sequence ASR. CoRR abs/2008.03822 (2020) - [i11]Tianyu Zhao, Tatsuya Kawahara:
Multi-Referenced Training for Dialogue Response Generation. CoRR abs/2009.07117 (2020) - [i10]Hirofumi Inaguma, Yosuke Higuchi, Kevin Duh, Tatsuya Kawahara, Shinji Watanabe:
Orthros: Non-autoregressive End-to-end Speech Translation with Dual-decoder. CoRR abs/2010.13047 (2020) - [i9]Laurence Devillers, Tatsuya Kawahara, Roger K. Moore, Matthias Scheutz:
Spoken Language Interaction with Virtual Agents and Robots (SLIVAR): Towards Effective and Ethical Interaction (Dagstuhl Seminar 20021). Dagstuhl Reports 10(1): 1-51 (2020)
2010 – 2019
- 2019
- [j49]Yuanchao Li, Carlos Toshinori Ishi, Koji Inoue, Shizuka Nakamura, Tatsuya Kawahara:
Expressing reactive emotion based on multimodal emotion recognition for natural conversation in human-robot interaction. Adv. Robotics 33(20): 1030-1041 (2019) - [j48]Tianyu Zhao, Tatsuya Kawahara
:
Joint dialog act segmentation and recognition in human conversations using attention to dialog context. Comput. Speech Lang. 57: 108-127 (2019) - [j47]Kazuki Shimada
, Yoshiaki Bando
, Masato Mimura, Katsutoshi Itoyama
, Kazuyoshi Yoshii
, Tatsuya Kawahara
:
Unsupervised Speech Enhancement Based on Multichannel NMF-Informed Beamforming for Noise-Robust Automatic Speech Recognition. IEEE ACM Trans. Audio Speech Lang. Process. 27(5): 960-971 (2019) - [j46]Kouhei Sekiguchi
, Yoshiaki Bando
, Aditya Arie Nugraha
, Kazuyoshi Yoshii
, Tatsuya Kawahara
:
Semi-Supervised Multichannel Speech Enhancement With a Deep Speech Prior. IEEE ACM Trans. Audio Speech Lang. Process. 27(12): 2197-2212 (2019) - [c243]Kak Soky, Sheng Li
, Tatsuya Kawahara, Sopheap Seng:
Multi-lingual Transformer Training for Khmer Automatic Speech Recognition. APSIPA 2019: 1893-1896 - [c242]Hirofumi Inaguma, Kevin Duh, Tatsuya Kawahara, Shinji Watanabe
:
Multilingual End-to-End Speech Translation. ASRU 2019: 570-577 - [c241]Hirofumi Inaguma, Jaejin Cho, Murali Karthick Baskar, Tatsuya Kawahara, Shinji Watanabe
:
Transfer Learning of Language-independent End-to-end ASR with Language Model Fusion. ICASSP 2019: 6096-6100 - [c240]Sei Ueno, Masato Mimura, Shinsuke Sakai, Tatsuya Kawahara:
Multi-speaker Sequence-to-sequence Speech Synthesis for Data Augmentation in Acoustic-to-word Speech Recognition. ICASSP 2019: 6161-6165 - [c239]Divesh Lala, Koji Inoue, Tatsuya Kawahara:
Smooth Turn-taking by a Robot Using an Online Continuous Model to Generate Turn-taking Cues. ICMI 2019: 226-234 - [c238]Divesh Lala, Graham Wilcock, Kristiina Jokinen, Tatsuya Kawahara:
ERICA and WikiTalk. IJCAI 2019: 6533-6535 - [c237]Sheng Li
, Chenchen Ding, Xugang Lu, Peng Shen, Tatsuya Kawahara, Hisashi Kawai:
End-to-End Articulatory Attribute Modeling for Low-Resource Multilingual Speech Recognition. INTERSPEECH 2019: 2145-2149 - [c236]Sheng Li
, Xugang Lu, Chenchen Ding, Peng Shen, Tatsuya Kawahara, Hisashi Kawai:
Investigating Radical-Based End-to-End Speech Recognition Systems for Chinese Dialects and Japanese. INTERSPEECH 2019: 2200-2204 - [c235]Yuanchao Li, Tianyu Zhao, Tatsuya Kawahara:
Improved End-to-End Speech Emotion Recognition Using Self Attention Mechanism and Multitask Learning. INTERSPEECH 2019: 2803-2807 - [c234]Kohei Hara, Koji Inoue, Katsuya Takanashi, Tatsuya Kawahara:
Turn-Taking Prediction Based on Detection of Transition Relevance Place. INTERSPEECH 2019: 4170-4174 - [c233]Divesh Lala, Shizuka Nakamura, Tatsuya Kawahara:
Analysis of Effect and Timing of Fillers in Natural Turn-Taking. INTERSPEECH 2019: 4175-4179 - [c232]Sheng Li
, Dabre Raj, Xugang Lu, Peng Shen, Tatsuya Kawahara, Hisashi Kawai:
Improving Transformer-Based Speech Recognition Systems with Compressed Structure and Speech Attributes Augmentation. INTERSPEECH 2019: 4400-4404 - [i8]Kazuki Shimada, Yoshiaki Bando, Masato Mimura, Katsutoshi Itoyama, Kazuyoshi Yoshii, Tatsuya Kawahara:
Unsupervised Speech Enhancement Based on Multichannel NMF-Informed Beamforming for Noise-Robust Automatic Speech Recognition. CoRR abs/1903.09341 (2019) - [i7]Tianyu Zhao, Tatsuya Kawahara:
Content Word-based Sentence Decoding and Evaluating for Open-domain Neural Response Generation. CoRR abs/1905.13438 (2019) - [i6]Tianyu Zhao, Tatsuya Kawahara:
Effective Incorporation of Speaker Information in Utterance Encoding in Dialog. CoRR abs/1907.05599 (2019) - [i5]Hirofumi Inaguma, Masato Mimura, Shinsuke Sakai, Tatsuya Kawahara:
Improving OOV Detection and Resolution with External Language Models in Acoustic-to-Word ASR. CoRR abs/1909.09993 (2019) - [i4]Hirofumi Inaguma, Kevin Duh, Tatsuya Kawahara, Shinji Watanabe:
Multilingual End-to-End Speech Translation. CoRR abs/1910.00254 (2019) - 2018
- [j45]Maryam Sadat Mirzaei, Kourosh Meshgi, Tatsuya Kawahara
:
Exploiting automatic speech recognition errors to enhance partial and synchronized caption for facilitating second language listening. Comput. Speech Lang. 49: 17-36 (2018) - [j44]Toshiyuki Hagiya, Toshiharu Horiuchi, Tomonori Yazaki, Tatsuya Kawahara:
Typing Tutor: Individualized Tutoring in Text Entry for Older Adults Based on Statistical Input Stumble Detection. J. Inf. Process. 26: 362-374 (2018) - [j43]Yoshiaki Bando
, Katsutoshi Itoyama, Masashi Konyo
, Satoshi Tadokoro, Kazuhiro Nakadai
, Kazuyoshi Yoshii
, Tatsuya Kawahara
, Hiroshi G. Okuno
:
Speech Enhancement Based on Bayesian Low-Rank and Sparse Decomposition of Multichannel Magnitude Spectrograms. IEEE ACM Trans. Audio Speech Lang. Process. 26(2): 215-230 (2018) - [j42]Kousuke Itakura
, Yoshiaki Bando
, Eita Nakamura
, Katsutoshi Itoyama
, Kazuyoshi Yoshii
, Tatsuya Kawahara
:
Bayesian Multichannel Audio Source Separation Based on Integrated Source and Spatial Models. IEEE ACM Trans. Audio Speech Lang. Process. 26(4): 831-846 (2018) - [c231]Kouhei Sekiguchi, Yoshiaki Bando, Kazuyoshi Yoshii
, Tatsuya Kawahara:
Bayesian Multichannel Speech Enhancement with a Deep Speech Prior. APSIPA 2018: 1233-1239 - [c230]Kenta Yamamoto, Koji Inoue, Shizuka Nakamura, Katsuya Takanashi, Tatsuya Kawahara:
Dialogue Behavior Control Model for Expressing a Character of Humanoid Robots. APSIPA 2018: 1732-1737 - [c229]Kazuyoshi Yoshii
, Koichi Kitamura, Yoshiaki Bando, Eita Nakamura, Tatsuya Kawahara:
Independent Low-Rank Tensor Analysis for Audio Source Separation. EUSIPCO 2018: 1657-1661 - [c228]Yoshiaki Bando, Masato Mimura, Katsutoshi Itoyama, Kazuyoshi Yoshii
, Tatsuya Kawahara:
Statistical Speech Enhancement Based on Probabilistic Integration of Variational Autoencoder and Non-Negative Matrix Factorization. ICASSP 2018: 716-720 - [c227]Kazuki Shimada, Yoshiaki Bando, Masato Mimura, Katsutoshi Itoyama, Kazuyoshi Yoshii
, Tatsuya Kawahara:
Unsupervised Beamforming Based on Multichannel Nonnegative Matrix Factorization for Noisy Speech Recognition. ICASSP 2018: 5734-5738 - [c226]Sei Ueno, Hirofumi Inaguma, Masato Mimura, Tatsuya Kawahara:
Acoustic-to-Word Attention-Based Model Complemented with Character-Level CTC-Based Model. ICASSP 2018: 5804-5808 - [c225]Hirofumi Inaguma, Masato Mimura, Koji Inoue, Kazuyoshi Yoshii
, Tatsuya Kawahara:
An End-to-End Approach to Joint Social Signal Detection and Automatic Speech Recognition. ICASSP 2018: 6214-6218 - [c224]Richeng Duan, Tatsuya Kawahara, Masatake Dantsuji, Hiroaki Nanjo:
Efficient Learning of Articulatory Models Based on Multi-Label Training and Label Correction for Pronunciation Learning. ICASSP 2018: 6239-6243 - [c223]Tatsuya Kawahara, Koji Inoue, Divesh Lala, Katsuya Takanashi:
Audio-Visual Conversation Analysis by Smart Posterboard and Humanoid Robot. ICASSP 2018: 6573-6577 - [c222]Divesh Lala, Koji Inoue, Tatsuya Kawahara:
Evaluation of Real-time Deep Learning Turn-taking Models for Multiple Dialogue Scenarios. ICMI 2018: 78-86 - [c221]Koji Inoue, Divesh Lala, Katsuya Takanashi, Tatsuya Kawahara:
Engagement Recognition in Spoken Dialogue via Neural Network by Aggregating Different Annotators' Models. INTERSPEECH 2018: 616-620 - [c220]Kohei Hara, Koji Inoue, Katsuya Takanashi, Tatsuya Kawahara:
Prediction of Turn-taking Using Multitask Learning with Prediction of Backchannels and Fillers. INTERSPEECH 2018: 991-995 - [c219]Masato Mimura, Shinsuke Sakai, Tatsuya Kawahara:
Forward-Backward Attention Decoder. INTERSPEECH 2018: 2232-2236 - [c218]Sei Ueno, Takafumi Moriya, Masato Mimura, Shinsuke Sakai, Yusuke Shinohara, Yoshikazu Yamaguchi, Yushi Aono, Tatsuya Kawahara:
Encoder Transfer for Attention-based Acoustic-to-word Speech Recognition. INTERSPEECH 2018: 2424-2428 - [c217]Sheng Li
, Xugang Lu, Ryoichi Takashima, Peng Shen, Tatsuya Kawahara, Hisashi Kawai:
Improving CTC-based Acoustic Model with Very Deep Residual Time-delay Neural Networks. INTERSPEECH 2018: 3708-3712 - [c216]Toshiyuki Hagiya, Keiichiro Hoashi, Tatsuya Kawahara:
Voice Input Tutoring System for Older Adults using Input Stumble Detection. IUI 2018: 415-419 - [c215]Tatsuya Kawahara:
Spoken Dialogue System for a Human-like Conversational Robot ERICA. IWSDS 2018: 65-75 - [c214]Ryosuke Nakanishi, Koji Inoue, Shizuka Nakamura, Katsuya Takanashi, Tatsuya Kawahara:
Generating Fillers Based on Dialog Act Pairs for Smooth Turn-Taking by Humanoid Robot. IWSDS 2018: 91-101 - [c213]Koji Inoue, Divesh Lala, Katsuya Takanashi, Tatsuya Kawahara:
Latent Character Model for Engagement Recognition Based on Multimodal Behaviors. IWSDS 2018: 119-130 - [c212]Tianyu Zhao, Tatsuya Kawahara:
A Unified Neural Architecture for Joint Dialog Act Segmentation and Recognition in Spoken Dialog System. SIGDIAL Conference 2018: 201-208 - [c211]Sheng Li
, Xugang Lu, Ryoichi Takashima, Peng Shen, Tatsuya Kawahara, Hisashi Kawai:
Improving Very Deep Time-Delay Neural Network With Vertical-Attention For Effectively Training CTC-Based ASR Systems. SLT 2018: 77-83 - [c210]Hirofumi Inaguma, Masato Mimura, Shinsuke Sakai, Tatsuya Kawahara:
Improving OOV Detection and Resolution with External Language Models in Acoustic-to-Word ASR. SLT 2018: 212-218 - [c209]Masato Mimura, Sei Ueno, Hirofumi Inaguma, Shinsuke Sakai, Tatsuya Kawahara:
Leveraging Sequence-to-Sequence Speech Synthesis for Enhancing Acoustic-to-Word Speech Recognition. SLT 2018: 477-484 - [i3]Hirofumi Inaguma, Jaejin Cho, Murali Karthick Baskar, Tatsuya Kawahara, Shinji Watanabe:
Transfer learning of language-independent end-to-end ASR with language model fusion. CoRR abs/1811.02134 (2018) - 2017
- [j41]Richeng Duan, Tatsuya Kawahara
, Masatake Dantsuji, Jinsong Zhang
:
Articulatory Modeling for Pronunciation Error Detection without Non-Native Training Data Based on DNN Transfer Learning. IEICE Trans. Inf. Syst. 100-D(9): 2174-2182 (2017) - [j40]Toshiyuki Hagiya, Toshiharu Horiuchi, Tomonori Yazaki, Tsuneo Kato, Tatsuya Kawahara
:
Assistive Typing Application for Older Adults Based on Input Stumble Detection. J. Inf. Process. 25: 417-425 (2017) - [c208]Tatsuya Kawahara:
Automatic meeting transcription system for the Japanese parliament (diet). APSIPA 2017: 1006-1010 - [c207]Yuanchao Li, Carlos Toshinori Ishi, Nigel G. Ward, Koji Inoue, Shizuka Nakamura, Katsuya Takanashi, Tatsuya Kawahara:
Emotion recognition by combining prosody and sentiment analysis for expressing reactive emotion by humanoid robot. APSIPA 2017: 1356-1359 - [c206]Masato Mimura, Shinsuke Sakai, Tatsuya Kawahara:
Cross-domain speech recognition using nonparallel corpora with cycle-consistent adversarial networks. ASRU 2017: 134-140 - [c205]Sheng Li
, Xugang Lu, Peng Shen, Ryoichi Takashima, Tatsuya Kawahara, Hisashi Kawai:
Incremental training and constructing the very deep convolutional residual network acoustic models. ASRU 2017: 222-227 - [c204]Divesh Lala, Yuanchao Li, Tatsuya Kawahara:
Utterance Behavior of Users While Playing Basketball with a Virtual Teammate. ICAART (1) 2017: 28-38 - [c203]Kousuke Itakura, Yoshiaki Bando, Eita Nakamura, Katsutoshi Itoyama, Kazuyoshi Yoshii
, Tatsuya Kawahara
:
Bayesian multichannel nonnegative matrix factorization for audio source separation and localization. ICASSP 2017: 551-555 - [c202]Sheng Li
, Xugang Lu, Shinsuke Sakai, Masato Mimura, Tatsuya Kawahara
:
Semi-supervised ensemble DNN acoustic model training. ICASSP 2017: 5270-5274 - [c201]Richeng Duan, Tatsuya Kawahara
, Masatake Dantsuji, Jinsong Zhang
:
Effective articulatory modeling for pronunciation error detection of L2 learner without non-native training data. ICASSP 2017: 5815-5819 - [c200]Tianyu Zhao, Tatsuya Kawahara:
Joint Learning of Dialog Act Segmentation and Recognition in Spoken Dialog Using Neural Networks. IJCNLP(1) 2017: 704-712 - [c199]Hirofumi Inaguma, Koji Inoue, Masato Mimura, Tatsuya Kawahara:
Social Signal Detection in Spontaneous Dialogue Using Bidirectional LSTM-CTC. INTERSPEECH 2017: 1691-1695 - [c198]Shizuka Nakamura, Ryosuke Nakanishi, Katsuya Takanashi, Tatsuya Kawahara:
Analysis of the Relationship Between Prosodic Features of Fillers and its Forms or Occurrence Positions. INTERSPEECH 2017: 1726-1730 - [c197]Masato Mimura, Yoshiaki Bando, Kazuki Shimada, Shinsuke Sakai, Kazuyoshi Yoshii, Tatsuya Kawahara:
Combined Multi-Channel NMF-Based Robust Beamforming for Noisy Speech Recognition. INTERSPEECH 2017: 2451-2455 - [c196]Pierrick Milhorat, Divesh Lala, Koji Inoue, Tianyu Zhao, Masanari Ishida, Katsuya Takanashi, Shizuka Nakamura, Tatsuya Kawahara:
A Conversational Dialogue Manager for the Humanoid Robot ERICA. IWSDS 2017: 119-131 - [c195]Masaya Wake, Yoshiaki Bando, Masato Mimura, Katsutoshi Itoyama, Kazuyoshi Yoshii
, Tatsuya Kawahara:
Semi-Blind speech enhancement basedon recurrent neural network for source separation and dereverberation. MLSP 2017: 1-6 - [c194]Divesh Lala, Pierrick Milhorat, Koji Inoue, Masanari Ishida, Katsuya Takanashi, Tatsuya Kawahara:
Attentive listening system with backchanneling, response generation and flexible turn-taking. SIGDIAL Conference 2017: 127-136 - [c193]Richeng Duan, Tatsuya Kawahara, Masatake Dantsuji, Hiroaki Nanjo:
Transfer Learning based Non-native Acoustic Modeling for Pronunciation Error Detection. SLaTE 2017: 42-46 - [c192]Maryam Sadat Mirzaei, Kourosh Meshgi, Tatsuya Kawahara:
Detecting listening difficulty for second language learners using Automatic Speech Recognition errors. SLaTE 2017: 156-160 - [i2]Divesh Lala, Koji Inoue, Pierrick Milhorat, Tatsuya Kawahara:
Detection of social signals for recognizing engagement in human-robot interaction. CoRR abs/1709.10257 (2017) - [i1]Yoshiaki Bando, Masato Mimura, Katsutoshi Itoyama, Kazuyoshi Yoshii, Tatsuya Kawahara:
Statistical Speech Enhancement Based on Probabilistic Integration of Variational Autoencoder and Non-Negative Matrix Factorization. CoRR abs/1710.11439 (2017) - 2016
- [j39]Sheng Li
, Yuya Akita, Tatsuya Kawahara
:
Semi-Supervised Acoustic Model Training by Discriminative Data Selection From Multiple ASR Systems' Hypotheses. IEEE ACM Trans. Audio Speech Lang. Process. 24(9): 1524-1534 (2016) - [c191]Maryam Sadat Mirzaei, Kourosh Meshgi, Tatsuya Kawahara:
Automatic Speech Recognition Errors as a Predictor of L2 Listening Difficulties. CL4LC@COLING 2016 2016: 192-201 - [c190]Richeng Duan, Tatsuya Kawahara
, Masatake Dantsuji, Jinsong Zhang
:
Multi-lingual and multi-task DNN learning for articulatory error detection. APSIPA 2016: 1-4 - [c189]Sheng Li
, Yuya Akita, Tatsuya Kawahara
:
Data selection from multiple ASR systems' hypotheses for unsupervised acoustic model training. ICASSP 2016: 5875-5879 - [c188]Hirofumi Inaguma, Koji Inoue, Shizuka Nakamura, Katsuya Takanashi, Tatsuya Kawahara
:
Prediction of ice-breaking between participants using prosodic features in the first meeting dialogue. ASSP4MI@ICMI 2016: 11-15 - [c187]Koji Inoue, Divesh Lala, Shizuka Nakamura, Katsuya Takanashi, Tatsuya Kawahara:
Annotation and analysis of listener's engagement based on multi-modal behaviors. MA3HMI@ICMI 2016: 25-32 - [c186]Divesh Lala, Pierrick Milhorat, Koji Inoue, Tianyu Zhao, Tatsuya Kawahara
:
Multimodal interaction with the autonomous Android ERICA. ICMI 2016: 417-418 - [c185]Tatsuya Kawahara
, Takashi Yamaguchi, Koji Inoue, Katsuya Takanashi, Nigel G. Ward:
Prediction and Generation of Backchannel Form for Attentive Listening Systems. INTERSPEECH 2016: 2890-2894 - [c184]Masato Mimura, Shinsuke Sakai, Tatsuya Kawahara
:
Joint Optimization of Denoising Autoencoder and DNN Acoustic Model Based on Multi-Target Learning for Noisy Speech Recognition. INTERSPEECH 2016: 3803-3807 - [c183]Richeng Duan, Tatsuya Kawahara
, Masatake Dantsuji, Jinsong Zhang
:
Pronunciation error detection using DNN articulatory model based on multi-lingual and multi-task learning. ISCSLP 2016: 1-5 - [c182]Sheng Li
, Xugang Lu, Shinsuke Mori, Yuya Akita, Tatsuya Kawahara
:
Confidence estimation for speech recognition systems using conditional random fields trained with partially annotated data. ISCSLP 2016: 1-5 - [c181]Divesh Lala, Tatsuya Kawahara
:
Managing Dialog and Joint Actions for Virtual Basketball Teammates. IVA 2016: 355-358 - [c180]Dylan F. Glas
, Takashi Minato, Carlos Toshinori Ishi, Tatsuya Kawahara
, Hiroshi Ishiguro:
ERICA: The ERATO Intelligent Conversational Android. RO-MAN 2016: 22-29 - [c179]Koji Inoue, Pierrick Milhorat, Divesh Lala, Tianyu Zhao, Tatsuya Kawahara:
Talking with ERICA, an autonomous android. SIGDIAL Conference 2016: 212-215 - [p3]Tatsuya Kawahara
:
Smart Posterboard: Multi-modal Sensing and Analysis of Poster Conversations. Human-Harmonized Information Technology (1) 2016: 239-270 - 2015
- [j38]Koichiro Yoshino, Tatsuya Kawahara
:
Conversational system for information navigation based on POMDP with user focus tracking. Comput. Speech Lang. 34(1): 275-291 (2015) - [j37]Masato Mimura, Shinsuke Sakai, Tatsuya Kawahara
:
Reverberant speech recognition combining deep neural networks and deep autoencoders augmented with a phone-class feature. EURASIP J. Adv. Signal Process. 2015: 62 (2015) - [j36]Sheng Li
, Yuya Akita, Tatsuya Kawahara
:
Automatic Lecture Transcription Based on Discriminative Data Selection for Lightly Supervised Acoustic Model Training. IEICE Trans. Inf. Syst. 98-D(8): 1545-1552 (2015) - [c178]Yuya Akita, Nobuhiro Kuwahara, Tatsuya Kawahara
:
Automatic classification of usability of ASR result for real-time captioning of lectures. APSIPA 2015: 19-22 - [c177]Tatsuya Kawahara
, Takashi Yamaguchi, Miki Uesato, Koichiro Yoshino, Katsuya Takanashi:
Synchrony in prosodic and linguistic features between backchannels and preceding utterances in attentive listening. APSIPA 2015: 392-395 - [c176]Masato Mimura, Shinsuke Sakai, Tatsuya Kawahara
:
Deep autoencoders augmented with phone-class feature for reverberant speech recognition. ICASSP 2015: 4365-4369 - [c175]Yuya Akita, Yizheng Tong, Tatsuya Kawahara
:
Language model adaptation for academic lectures using character recognition result of presentation slides. ICASSP 2015: 5431-5435 - [c174]Masato Mimura, Shinsuke Sakai, Tatsuya Kawahara:
Speech dereverberation using long short-term memory. INTERSPEECH 2015: 2435-2439 - [c173]Sheng Li, Xugang Lu, Yuya Akita, Tatsuya Kawahara:
Ensemble speaker modeling using speaker adaptive training deep neural network for speaker adaptation. INTERSPEECH 2015: 2892-2896 - [c172]Koji Inoue, Yukoh Wakabayashi, Hiromasa Yoshimoto, Katsuya Takanashi, Tatsuya Kawahara:
Enhanced speaker diarization with detection of backchannels using eye-gaze information in poster conversations. INTERSPEECH 2015: 3086-3090 - [c171]Sheng Li, Yuya Akita, Tatsuya Kawahara:
Discriminative data selection for lightly supervised training of acoustic model using closed caption texts. INTERSPEECH 2015: 3526-3530 - [c170]Tetsuro Sasada, Shinsuke Mori, Tatsuya Kawahara
, Yoko Yamakata:
Named Entity Recognizer Trainable from Partially Annotated Data. PACLING 2015: 148-160 - [c169]Maryam Sadat Mirzaei, Tatsuya Kawahara:
ASR technology to empower partial and synchronized caption for L2 listening development. SLaTE 2015: 65-70 - [p2]