default search action
Atsunori Ogawa
Person information
Refine list
refinements active!
zoomed in on ?? of ?? records
view refined list in
export refined list as
2020 – today
- 2024
- [j16]Daiki Mori, Kengo Ohta, Ryota Nishimura, Atsunori Ogawa, Norihide Kitaoka:
Recognition of target domain Japanese speech using language model replacement. EURASIP J. Audio Speech Music. Process. 2024(1): 40 (2024) - [c101]Naohiro Tawara, Marc Delcroix, Atsushi Ando, Atsunori Ogawa:
NTT Speaker Diarization System for Chime-7: Multi-Domain, Multi-Microphone end-to-end and Vector Clustering Diarization. ICASSP 2024: 11281-11285 - [c100]William Chen, Takatomo Kano, Atsunori Ogawa, Marc Delcroix, Shinji Watanabe:
Train Long and Test Long: Leveraging Full Document Contexts in Speech Processing. ICASSP 2024: 13066-13070 - [i12]Atsunori Ogawa, Naoyuki Kamo, Kohei Matsuura, Takanori Ashihara, Takafumi Moriya, Takatomo Kano, Naohiro Tawara, Marc Delcroix:
Applying LLMs for Rescoring N-best ASR Hypotheses of Casual Conversations: Effects of Domain Adaptation and Context Carry-over. CoRR abs/2406.18972 (2024) - [i11]Kohei Matsuura, Takanori Ashihara, Takafumi Moriya, Masato Mimura, Takatomo Kano, Atsunori Ogawa, Marc Delcroix:
Sentence-wise Speech Summarization: Task, Datasets, and End-to-End Modeling with LM Knowledge Distillation. CoRR abs/2408.00205 (2024) - 2023
- [c99]Tatsunari Takagi, Atsunori Ogawa, Norihide Kitaoka, Yukoh Wakabayashi:
Streaming End-to-End ASR Using CTC Decoder and DRA for Linguistic Information Substitution. APSIPA ASC 2023: 1779-1783 - [c98]Koharu Horii, Kengo Ohta, Ryota Nishimura, Atsunori Ogawa, Norihide Kitaoka:
Language modeling for spontaneous speech recognition based on disfluency labeling and generation of disfluent text. APSIPA ASC 2023: 1851-1856 - [c97]Yuki Kitagishi, Hosana Kamiyama, Naohiro Tawara, Atsunori Ogawa, Noboru Miyazaki, Taichi Asami:
Coarse-Age Loss: A New Training Method Using Coarse-Age Labeled Data for Speaker Age Estimation. APSIPA ASC 2023: 2213-2220 - [c96]Keigo Hojo, Daiki Mori, Yukoh Wakabayashi, Kengo Ohta, Atsunori Ogawa, Norihide Kitaoka:
Combining multiple end-to-end speech recognition models based on density ratio approach. APSIPA ASC 2023: 2274-2279 - [c95]Takatomo Kano, Atsunori Ogawa, Marc Delcroix, Kohei Matsuura, Takanori Ashihara, William Chen, Shinji Watanabe:
Summarize While Translating: Universal Model With Parallel Decoding for Summarization and Translation. ASRU 2023: 1-8 - [c94]Roshan S. Sharma, William Chen, Takatomo Kano, Ruchira Sharma, Siddhant Arora, Shinji Watanabe, Atsunori Ogawa, Marc Delcroix, Rita Singh, Bhiksha Raj:
Espnet-Summ: Introducing a Novel Large Dataset, Toolkit, and a Cross-Corpora Evaluation of Speech Summarization Systems. ASRU 2023: 1-8 - [c93]Takatomo Kano, Atsunori Ogawa, Marc Delcroix, Roshan S. Sharma, Kohei Matsuura, Shinji Watanabe:
Speech Summarization of Long Spoken Document: Improving Memory Efficiency of Speech/Text Encoders. ICASSP 2023: 1-5 - [c92]Kohei Matsuura, Takanori Ashihara, Takafumi Moriya, Tomohiro Tanaka, Atsunori Ogawa, Marc Delcroix, Ryo Masumura:
Leveraging Large Text Corpora For End-To-End Speech Summarization. ICASSP 2023: 1-5 - [c91]Atsunori Ogawa, Takafumi Moriya, Naoyuki Kamo, Naohiro Tawara, Marc Delcroix:
Iterative Shallow Fusion of Backward Language Model for End-To-End Speech Recognition. ICASSP 2023: 1-5 - [c90]Takafumi Moriya, Hiroshi Sato, Tsubasa Ochiai, Marc Delcroix, Takanori Ashihara, Kohei Matsuura, Tomohiro Tanaka, Ryo Masumura, Atsunori Ogawa, Taichi Asami:
Knowledge Distillation for Neural Transducer-based Target-Speaker ASR: Exploiting Parallel Mixture/Single-Talker Speech Data. INTERSPEECH 2023: 899-903 - [c89]Yuki Kitagishi, Naohiro Tawara, Atsunori Ogawa, Ryo Masumura, Taichi Asami:
What are differences? Comparing DNN and Human by Their Performance and Characteristics in Speaker Age Estimation. INTERSPEECH 2023: 1873-1877 - [c88]Shoko Araki, Ayako Yamamoto, Tsubasa Ochiai, Kenichi Arai, Atsunori Ogawa, Tomohiro Nakatani, Toshio Irino:
Impact of Residual Noise and Artifacts in Speech Enhancement Errors on Intelligibility of Human and Machine. INTERSPEECH 2023: 2503-2507 - [c87]Kohei Matsuura, Takanori Ashihara, Takafumi Moriya, Tomohiro Tanaka, Takatomo Kano, Atsunori Ogawa, Marc Delcroix:
Transfer Learning from Pre-trained Language Models Improves End-to-End Speech Summarization. INTERSPEECH 2023: 2943-2947 - [c86]Marc Delcroix, Naohiro Tawara, Mireia Díez, Federico Landini, Anna Silnova, Atsunori Ogawa, Tomohiro Nakatani, Lukás Burget, Shoko Araki:
Multi-Stream Extension of Variational Bayesian HMM Clustering (MS-VBx) for Combined End-to-End and Vector Clustering-based Diarization. INTERSPEECH 2023: 3477-3481 - [i10]Kohei Matsuura, Takanori Ashihara, Takafumi Moriya, Tomohiro Tanaka, Atsunori Ogawa, Marc Delcroix, Ryo Masumura:
Leveraging Large Text Corpora for End-to-End Speech Summarization. CoRR abs/2303.00978 (2023) - [i9]Marc Delcroix, Naohiro Tawara, Mireia Díez, Federico Landini, Anna Silnova, Atsunori Ogawa, Tomohiro Nakatani, Lukás Burget, Shoko Araki:
Multi-Stream Extension of Variational Bayesian HMM Clustering (MS-VBx) for Combined End-to-End and Vector Clustering-based Diarization. CoRR abs/2305.13580 (2023) - [i8]Kohei Matsuura, Takanori Ashihara, Takafumi Moriya, Tomohiro Tanaka, Takatomo Kano, Atsunori Ogawa, Marc Delcroix:
Transfer Learning from Pre-trained Language Models Improves End-to-End Speech Summarization. CoRR abs/2306.04233 (2023) - [i7]Naohiro Tawara, Marc Delcroix, Atsushi Ando, Atsunori Ogawa:
NTT speaker diarization system for CHiME-7: multi-domain, multi-microphone End-to-end and vector clustering diarization. CoRR abs/2309.12656 (2023) - [i6]Atsunori Ogawa, Takafumi Moriya, Naoyuki Kamo, Naohiro Tawara, Marc Delcroix:
Iterative Shallow Fusion of Backward Language Model for End-to-End Speech Recognition. CoRR abs/2310.11010 (2023) - [i5]Atsunori Ogawa, Naohiro Tawara, Marc Delcroix, Shoko Araki:
Lattice Rescoring Based on Large Ensemble of Complementary Neural Language Models. CoRR abs/2312.12764 (2023) - [i4]Atsunori Ogawa, Naohiro Tawara, Takatomo Kano, Marc Delcroix:
BLSTM-Based Confidence Estimation for End-to-End Speech Recognition. CoRR abs/2312.14609 (2023) - 2022
- [j15]Naohiro Tawara, Atsunori Ogawa, Tomoharu Iwata, Hiroto Ashikawa, Tetsunori Kobayashi, Tetsuji Ogawa:
Multi-Source Domain Generalization Using Domain Attributes for Recurrent Neural Network Language Models. IEICE Trans. Inf. Syst. 105-D(1): 150-160 (2022) - [c85]Takatomo Kano, Atsunori Ogawa, Marc Delcroix, Shinji Watanabe:
Integrating Multiple ASR Systems into NLP Backend with Attention Fusion. ICASSP 2022: 6237-6241 - [c84]Atsunori Ogawa, Naohiro Tawara, Marc Delcroix, Shoko Araki:
Lattice Rescoring Based on Large Ensemble of Complementary Neural Language Models. ICASSP 2022: 6517-6521 - [c83]Koharu Horii, Meiko Fukuda, Kengo Ohta, Ryota Nishimura, Atsunori Ogawa, Norihide Kitaoka:
End-to-End Spontaneous Speech Recognition Using Disfluency Labeling. INTERSPEECH 2022: 4108-4112 - [i3]Ayako Yamamoto, Toshio Irino, Shoko Araki, Kenichi Arai, Atsunori Ogawa, Keisuke Kinoshita, Tomohiro Nakatani:
Subjective intelligibility of speech sounds enhanced by ideal ratio mask via crowdsourced remote experiments with effective data screening. CoRR abs/2203.16760 (2022) - 2021
- [c82]Daiki Mori, Kengo Ohta, Ryota Nishimura, Atsunori Ogawa, Norihide Kitaoka:
Advanced language model fusion method for encoder-decoder model in Japanese speech recognition. APSIPA ASC 2021: 503-510 - [c81]Koharu Horii, Meiko Fukuda, Kengo Ohta, Ryota Nishimura, Atsunori Ogawa, Norihide Kitaoka:
End-to-End Spontaneous Speech Recognition Using Hesitation Labeling. APSIPA ASC 2021: 1077-1081 - [c80]Naohiro Tawara, Atsunori Ogawa, Yuki Kitagishi, Hosana Kamiyama, Yusuke Ijima:
Robust Speech-Age Estimation Using Local Maximum Mean Discrepancy Under Mismatched Recording Conditions. ASRU 2021: 114-121 - [c79]Takatomo Kano, Atsunori Ogawa, Marc Delcroix, Shinji Watanabe:
Attention-Based Multi-Hypothesis Fusion for Speech Summarization. ASRU 2021: 487-494 - [c78]Atsunori Ogawa, Naohiro Tawara, Takatomo Kano, Marc Delcroix:
BLSTM-Based Confidence Estimation for End-to-End Speech Recognition. ICASSP 2021: 6383-6387 - [c77]Naohiro Tawara, Atsunori Ogawa, Yuki Kitagishi, Hosana Kamiyama:
Age-VOX-Celeb: Multi-Modal Corpus for Facial and Speech Estimation. ICASSP 2021: 6963-6967 - [c76]Ayako Yamamoto, Toshio Irino, Kenichi Arai, Shoko Araki, Atsunori Ogawa, Keisuke Kinoshita, Tomohiro Nakatani:
Comparison of Remote Experiments Using Crowdsourcing and Laboratory Experiments on Speech Intelligibility. Interspeech 2021: 181-185 - [i2]Ayako Yamamoto, Toshio Irino, Kenichi Arai, Shoko Araki, Atsunori Ogawa, Keisuke Kinoshita, Tomohiro Nakatani:
Comparison of remote experiments using crowdsourcing and laboratory experiments on speech intelligibility. CoRR abs/2104.10001 (2021) - [i1]Takatomo Kano, Atsunori Ogawa, Marc Delcroix, Shinji Watanabe:
Attention-based Multi-hypothesis Fusion for Speech Summarization. CoRR abs/2111.08201 (2021) - 2020
- [c75]Yosuke Higuchi, Naohiro Tawara, Atsunori Ogawa, Tomoharu Iwata, Tetsunori Kobayashi, Tetsuji Ogawa:
Noise-robust Attention Learning for End-to-End Speech Recognition. EUSIPCO 2020: 311-315 - [c74]Naohiro Tawara, Hosana Kamiyama, Satoshi Kobashikawa, Atsunori Ogawa:
Improving Speaker-Attribute Estimation by Voting Based on Speaker Cluster Information. ICASSP 2020: 6594-6598 - [c73]Naohiro Tawara, Atsunori Ogawa, Tomoharu Iwata, Marc Delcroix, Tetsuji Ogawa:
Frame-Level Phoneme-Invariant Speaker Embedding for Text-Independent Speaker Recognition on Extremely Short Utterances. ICASSP 2020: 6799-6803 - [c72]Kenichi Arai, Shoko Araki, Atsunori Ogawa, Keisuke Kinoshita, Tomohiro Nakatani, Toshio Irino:
Predicting Intelligibility of Enhanced Speech Using Posteriors Derived from DNN-Based ASR System. INTERSPEECH 2020: 1156-1160 - [c71]Atsunori Ogawa, Naohiro Tawara, Marc Delcroix:
Language Model Data Augmentation Based on Text Domain Transfer. INTERSPEECH 2020: 4926-4930
2010 – 2019
- 2019
- [j14]Michael Hentschel, Marc Delcroix, Atsunori Ogawa, Tomoharu Iwata, Tomohiro Nakatani:
Feature Based Domain Adaptation for Neural Network Language Models with Factorised Hidden Layers. IEICE Trans. Inf. Syst. 102-D(3): 598-608 (2019) - [c70]Shigeki Karita, Shinji Watanabe, Tomoharu Iwata, Marc Delcroix, Atsunori Ogawa, Tomohiro Nakatani:
Semi-supervised End-to-end Speech Recognition Using Text-to-speech and Autoencoders. ICASSP 2019: 6166-6170 - [c69]Tsubasa Ochiai, Marc Delcroix, Keisuke Kinoshita, Atsunori Ogawa, Tomohiro Nakatani:
A Unified Framework for Neural Speech Separation and Extraction. ICASSP 2019: 6975-6979 - [c68]Atsunori Ogawa, Tsutomu Hirao, Tomohiro Nakatani, Masaaki Nagata:
ILP-based Compressive Speech Summarization with Content Word Coverage Maximization and Its Oracle Performance Analysis. ICASSP 2019: 7190-7194 - [c67]Michael Hentschel, Marc Delcroix, Atsunori Ogawa, Tomoharu Iwata, Tomohiro Nakatani:
A Unified Framework for Feature-based Domain Adaptation of Neural Network Language Models. ICASSP 2019: 7250-7254 - [c66]Marc Delcroix, Shinji Watanabe, Tsubasa Ochiai, Keisuke Kinoshita, Shigeki Karita, Atsunori Ogawa, Tomohiro Nakatani:
End-to-End SpeakerBeam for Single Channel Target Speech Recognition. INTERSPEECH 2019: 451-455 - [c65]Shigeki Karita, Nelson Enrique Yalta Soplin, Shinji Watanabe, Marc Delcroix, Atsunori Ogawa, Tomohiro Nakatani:
Improving Transformer-Based End-to-End Speech Recognition with Connectionist Temporal Classification and Language Model Integration. INTERSPEECH 2019: 1408-1412 - [c64]Tsubasa Ochiai, Marc Delcroix, Keisuke Kinoshita, Atsunori Ogawa, Tomohiro Nakatani:
Multimodal SpeakerBeam: Single Channel Target Speech Extraction with Audio-Visual Speaker Clues. INTERSPEECH 2019: 2718-2722 - [c63]Atsunori Ogawa, Marc Delcroix, Shigeki Karita, Tomohiro Nakatani:
Improved Deep Duel Model for Rescoring N-Best Speech Recognition List Using Backward LSTMLM and Ensemble Encoders. INTERSPEECH 2019: 3900-3904 - [c62]Kenichi Arai, Shoko Araki, Atsunori Ogawa, Keisuke Kinoshita, Tomohiro Nakatani, Katsuhiko Yamamoto, Toshio Irino:
Predicting Speech Intelligibility of Enhanced Speech Using Phone Accuracy of DNN-Based ASR System. INTERSPEECH 2019: 4275-4279 - 2018
- [j13]Marc Delcroix, Keisuke Kinoshita, Atsunori Ogawa, Christian Huemmer, Tomohiro Nakatani:
Context Adaptive Neural Network Based Acoustic Models for Rapid Adaptation. IEEE ACM Trans. Audio Speech Lang. Process. 26(5): 895-908 (2018) - [c61]Michael Hentschel, Marc Delcroix, Atsunori Ogawa, Tomohiro Nakatani:
Feature-Based Learning Hidden Unit Contributions for Domain Adaptation of RNN-LMs. APSIPA 2018: 1692-1696 - [c60]Michael Hentschel, Marc Delcroix, Atsunori Ogawa, Tomoharu Iwata, Tomohiro Nakatani:
Factorised Hidden Layer Based Domain Adaptation for Recurrent Neural Network Language Models. APSIPA 2018: 1940-1944 - [c59]Marc Delcroix, Katerina Zmolíková, Keisuke Kinoshita, Atsunori Ogawa, Tomohiro Nakatani:
Single Channel Target Speaker Extraction and Recognition with Speaker Beam. ICASSP 2018: 5554-5558 - [c58]Shigeki Karita, Atsunori Ogawa, Marc Delcroix, Tomohiro Nakatani:
Sequence Training of Encoder-Decoder Model Using Policy Gradient for End-to-End Speech Recognition. ICASSP 2018: 5839-5843 - [c57]Tsuyoshi Morioka, Naohiro Tawara, Tetsuji Ogawa, Atsunori Ogawa, Tomoharu Iwata, Tetsunori Kobayashi:
Language Model Domain Adaptation Via Recurrent Neural Networks with Domain-Shared and Domain-Specific Representations. ICASSP 2018: 6084-6088 - [c56]Atsunori Ogawa, Marc Delcroix, Shigeki Karita, Tomohiro Nakatani:
Rescoring N-Best Speech Recognition List Based on One-on-One Hypothesis Comparison Using Encoder-Classifier Model. ICASSP 2018: 6099-6103 - [c55]Shigeki Karita, Shinji Watanabe, Tomoharu Iwata, Atsunori Ogawa, Marc Delcroix:
Semi-Supervised End-to-End Speech Recognition. INTERSPEECH 2018: 2-6 - [c54]Marc Delcroix, Shinji Watanabe, Atsunori Ogawa, Shigeki Karita, Tomohiro Nakatani:
Auxiliary Feature Based Adaptation of End-to-end ASR Systems. INTERSPEECH 2018: 2444-2448 - 2017
- [j12]Atsunori Ogawa, Takaaki Hori:
Error detection and accuracy estimation in automatic speech recognition using deep bidirectional recurrent neural networks. Speech Commun. 89: 70-83 (2017) - [c53]Michael Hentschel, Atsunori Ogawa, Marc Delcroix, Tomohiro Nakatani, Yuji Matsumoto:
Exploiting imbalanced textual and acoustic data for training prosodically-enhanced RNNLMs. APSIPA 2017: 618-621 - [c52]Hiroto Ashikawa, Naohiro Tawara, Atsunori Ogawa, Tomoharu Iwata, Tetsunori Kobayashi, Tetsuji Ogawa:
Exploiting end of sentences and speaker alternations in language modeling for multiparty conversations. APSIPA 2017: 1263-1267 - [c51]Katerina Zmolíková, Marc Delcroix, Keisuke Kinoshita, Takuya Higuchi, Atsunori Ogawa, Tomohiro Nakatani:
Learning speaker representation for neural network based multichannel speaker extraction. ASRU 2017: 8-15 - [c50]Shoko Araki, Nobutaka Ito, Marc Delcroix, Atsunori Ogawa, Keisuke Kinoshita, Takuya Higuchi, Takuya Yoshioka, Dung T. Tran, Shigeki Karita, Tomohiro Nakatani:
Online meeting recognition in noisy environments with time-frequency mask based MVDR beamforming. HSCMA 2017: 16-20 - [c49]Keisuke Kinoshita, Marc Delcroix, Atsunori Ogawa, Takuya Higuchi, Tomohiro Nakatani:
Deep mixture density network for statistical model-based feature enhancement. ICASSP 2017: 251-255 - [c48]Christian Huemmer, Marc Delcroix, Atsunori Ogawa, Keisuke Kinoshita, Tomohiro Nakatani, Walter Kellermann:
Online environmental adaptation of CNN-based acoustic models using spatial diffuseness features. ICASSP 2017: 4875-4879 - [c47]Tsubasa Ochiai, Marc Delcroix, Keisuke Kinoshita, Atsunori Ogawa, Taichi Asami, Shigeru Katagiri, Tomohiro Nakatani:
Cumulative moving averaged bottleneck speaker vectors for online speaker adaptation of CNN-based acoustic models. ICASSP 2017: 5175-5179 - [c46]Dung T. Tran, Marc Delcroix, Atsunori Ogawa, Christian Huemmer, Tomohiro Nakatani:
Feedback connection for deep neural network-based acoustic modeling. ICASSP 2017: 5240-5244 - [c45]Dung T. Tran, Marc Delcroix, Shigeki Karita, Michael Hentschel, Atsunori Ogawa, Tomohiro Nakatani:
Unfolded Deep Recurrent Convolutional Neural Network with Jump Ahead Connections for Acoustic Modeling. INTERSPEECH 2017: 1596-1600 - [c44]Shigeki Karita, Atsunori Ogawa, Marc Delcroix, Tomohiro Nakatani:
Forward-Backward Convolutional LSTM for Acoustic Modeling. INTERSPEECH 2017: 1601-1605 - [c43]Atsunori Ogawa, Keisuke Kinoshita, Marc Delcroix, Tomohiro Nakatani:
Improved Example-Based Speech Enhancement by Using Deep Neural Network Acoustic Model for Noise Robust Example Search. INTERSPEECH 2017: 1963-1967 - [c42]Katerina Zmolíková, Marc Delcroix, Keisuke Kinoshita, Takuya Higuchi, Atsunori Ogawa, Tomohiro Nakatani:
Speaker-Aware Neural Network Based Beamformer for Speaker Extraction in Speech Mixtures. INTERSPEECH 2017: 2655-2659 - [c41]Dung T. Tran, Marc Delcroix, Atsunori Ogawa, Tomohiro Nakatani:
Uncertainty Decoding with Adaptive Sampling for Noise Robust DNN-Based Acoustic Modeling. INTERSPEECH 2017: 3852-3856 - [p1]Marc Delcroix, Takuya Yoshioka, Nobutaka Ito, Atsunori Ogawa, Keisuke Kinoshita, Masakiyo Fujimoto, Takuya Higuchi, Shoko Araki, Tomohiro Nakatani:
Multichannel Speech Enhancement Approaches to DNN-Based Far-Field Speech Recognition. New Era for Robust Speech Recognition, Exploiting Deep Learning 2017: 21-49 - 2016
- [j11]Marc Delcroix, Atsunori Ogawa, Seong-Jun Hahm, Tomohiro Nakatani, Atsushi Nakamura:
Differenced maximum mutual information criterion for robust unsupervised acoustic model adaptation. Comput. Speech Lang. 36: 24-41 (2016) - [j10]Atsunori Ogawa, Takaaki Hori, Atsushi Nakamura:
Estimating Speech Recognition Accuracy Based on Error Type Classification. IEEE ACM Trans. Audio Speech Lang. Process. 24(12): 2400-2413 (2016) - [c40]Shoko Araki, Masahiro Okada, Takuya Higuchi, Atsunori Ogawa, Tomohiro Nakatani:
Spatial correlation model based observation vector clustering and MVDR beamforming for meeting recognition. ICASSP 2016: 385-389 - [c39]Marc Delcroix, Keisuke Kinoshita, Chengzhu Yu, Atsunori Ogawa, Takuya Yoshioka, Tomohiro Nakatani:
Context adaptive deep neural networks for fast acoustic model adaptation in noisy conditions. ICASSP 2016: 5270-5274 - [c38]Marc Delcroix, Keisuke Kinoshita, Atsunori Ogawa, Takuya Yoshioka, Dung T. Tran, Tomohiro Nakatani:
Context Adaptive Neural Network for Rapid Adaptation of Deep CNN Based Acoustic Models. INTERSPEECH 2016: 1573-1577 - [c37]Atsunori Ogawa, Shogo Seki, Keisuke Kinoshita, Marc Delcroix, Takuya Yoshioka, Tomohiro Nakatani, Kazuya Takeda:
Robust Example Search Using Bottleneck Features for Example-Based Speech Enhancement. INTERSPEECH 2016: 3733-3737 - [c36]Dung T. Tran, Marc Delcroix, Atsunori Ogawa, Tomohiro Nakatani:
Factorized Linear Input Network for Acoustic Model Adaptation in Noisy Conditions. INTERSPEECH 2016: 3813-3817 - 2015
- [j9]Marc Delcroix, Takuya Yoshioka, Atsunori Ogawa, Yotaro Kubo, Masakiyo Fujimoto, Nobutaka Ito, Keisuke Kinoshita, Miquel Espi, Shoko Araki, Takaaki Hori, Tomohiro Nakatani:
Strategies for distant speech recognitionin reverberant environments. EURASIP J. Adv. Signal Process. 2015: 60 (2015) - [c35]Takuya Yoshioka, Nobutaka Ito, Marc Delcroix, Atsunori Ogawa, Keisuke Kinoshita, Masakiyo Fujimoto, Chengzhu Yu, Wojciech J. Fabian, Miquel Espi, Takuya Higuchi, Shoko Araki, Tomohiro Nakatani:
The NTT CHiME-3 system: Advances in speech enhancement and recognition for mobile multi-microphone devices. ASRU 2015: 436-443 - [c34]Atsunori Ogawa, Takaaki Hori:
ASR error detection and recognition rate estimation using deep bidirectional recurrent neural networks. ICASSP 2015: 4370-4374 - [c33]Kazuo Aoyama, Atsunori Ogawa, Takashi Hattori, Takaaki Hori:
Double-layer neighborhood graph based similarity search for fast query-by-example spoken term detection. ICASSP 2015: 5216-5220 - [c32]Keisuke Kinoshita, Marc Delcroix, Atsunori Ogawa, Tomohiro Nakatani:
Text-informed speech enhancement with deep neural networks. INTERSPEECH 2015: 1760-1764 - [c31]Chengzhu Yu, Atsunori Ogawa, Marc Delcroix, Takuya Yoshioka, Tomohiro Nakatani, John H. L. Hansen:
Robust i-vector extraction for neural network adaptation in noisy environment. INTERSPEECH 2015: 2854-2857 - 2014
- [j8]Shoko Yamahata, Yoshikazu Yamaguchi, Atsunori Ogawa, Hirokazu Masataki, Osamu Yoshioka, Satoshi Takahashi:
Automatic Vocabulary Adaptation Based on Semantic and Acoustic Similarities. IEICE Trans. Inf. Syst. 97-D(6): 1488-1496 (2014) - [c30]Marc Delcroix, Takuya Yoshioka, Atsunori Ogawa, Yotaro Kubo, Masakiyo Fujimoto, Nobutaka Ito, Keisuke Kinoshita, Miquel Espi, Shoko Araki, Takaaki Hori, Tomohiro Nakatani:
Defeating reverberation: Advanced dereverberation and recognition techniques for hands-free speech recognition. GlobalSIP 2014: 522-526 - [c29]Atsunori Ogawa, Keisuke Kinoshita, Takaaki Hori, Tomohiro Nakatani, Atsushi Nakamura:
Fast segment search for corpus-based speech enhancement based on speech recognition technology. ICASSP 2014: 1557-1561 - [c28]Kazuo Aoyama, Atsunori Ogawa, Takashi Hattori, Takaaki Hori, Atsushi Nakamura:
Zero-resource spoken term detection using hierarchical graph-based similarity search. ICASSP 2014: 7093-7097 - 2013
- [j7]Satoshi Kobashikawa, Atsunori Ogawa, Taichi Asami, Yoshikazu Yamaguchi, Hirokazu Masataki, Satoshi Takahashi:
Fast unsupervised adaptation based on efficient statistics accumulation using frame independent confidence within monophone states. Comput. Speech Lang. 27(1): 369-379 (2013) - [j6]Marc Delcroix, Keisuke Kinoshita, Tomohiro Nakatani, Shoko Araki, Atsunori Ogawa, Takaaki Hori, Shinji Watanabe, Masakiyo Fujimoto, Takuya Yoshioka, Takanobu Oba, Yotaro Kubo, Mehrez Souden, Seong-Jun Hahm, Atsushi Nakamura:
Speech recognition in living rooms: Integrated speech enhancement and recognition system based on spatial, spectral and temporal modeling of sounds. Comput. Speech Lang. 27(3): 851-873 (2013) - [j5]Seong-Jun Hahm, Shinji Watanabe, Atsunori Ogawa, Masakiyo Fujimoto, Takaaki Hori, Atsushi Nakamura:
Prior-shared feature and model space speaker adaptation by consistently employing map estimation. Speech Commun. 55(3): 415-431 (2013) - [c27]Atsunori Ogawa, Takaaki Hori, Atsushi Nakamura:
Discriminative recognition rate estimation for N-best list and its application to N-best rescoring. ICASSP 2013: 6832-6836 - [c26]Tomohiro Nakatani, Mehrez Souden, Shoko Araki, Takuya Yoshioka, Takaaki Hori, Atsunori Ogawa:
Coupling beamforming with spatial and spectral feature based spectral enhancement and its application to meeting recognition. ICASSP 2013: 7249-7253 - [c25]Marc Delcroix, Atsunori Ogawa, Seong-Jun Hahm, Tomohiro Nakatani, Atsushi Nakamura:
Unsupervised discriminative adaptation using differenced maximum mutual information based linear regression. ICASSP 2013: 7888-7892 - [c24]Seong-Jun Hahm, Atsunori Ogawa, Marc Delcroix, Masakiyo Fujimoto, Takaaki Hori, Atsushi Nakamura:
Feature space variational Bayesian linear regression and its combination with model space VBLR. ICASSP 2013: 7898-7902 - [c23]Kazuo Aoyama, Atsunori Ogawa, Takashi Hattori, Takaaki Hori, Atsushi Nakamura:
Graph index based query-by-example search on a large speech data set. ICASSP 2013: 8520-8524 - [c22]Takanobu Oba, Atsunori Ogawa, Takaaki Hori, Hirokazu Masataki, Atsushi Nakamura:
Unsupervised discriminative language modeling using error rate estimator. INTERSPEECH 2013: 1223-1227 - 2012
- [j4]Atsunori Ogawa, Atsushi Nakamura:
Joint estimation of confidence and error causes in speech recognition. Speech Commun. 54(9): 1014-1028 (2012) - [j3]Takaaki Hori, Shoko Araki, Takuya Yoshioka, Masakiyo Fujimoto, Shinji Watanabe, Takanobu Oba, Atsunori Ogawa, Kazuhiro Otsuka, Dan Mikami, Keisuke Kinoshita, Tomohiro Nakatani, Atsushi Nakamura, Junji Yamato:
Low-Latency Real-Time Meeting Recognition and Understanding Using Distant Microphones and Omni-Directional Camera. IEEE Trans. Speech Audio Process. 20(2): 499-513 (2012) - [c21]Marc Delcroix, Atsunori Ogawa, Shinji Watanabe, Tomohiro Nakatani, Atsushi Nakamura:
Discriminative feature transforms using differenced maximum mutual information. ICASSP 2012: 4753-4756 - [c20]Atsunori Ogawa, Takaaki Hori, Atsushi Nakamura:
Error type classification and word accuracy estimation using alignment features from word confusion network. ICASSP 2012: 4925-4928 - [c19]Seong-Jun Hahm, Atsunori Ogawa, Masakiyo Fujimoto, Takaaki Hori, Atsushi Nakamura:
Speaker Adaptation Using Variational Bayesian Linear Regression in Normalized Feature Space. INTERSPEECH 2012: 803-806 - [c18]Shoko Yamahata, Yoshikazu Yamaguchi, Atsunori Ogawa, Hirokazu Masataki, Osamu Yoshioka, Satoshi Takahashi:
Automatic Vocabulary Adaptation Based on Semantic Similarity and Speech Recognition Confidence Measure. INTERSPEECH 2012: 2310-2313 - [c17]Marc Delcroix, Atsunori Ogawa, Tomohiro Nakatani, Atsushi Nakamura:
Dynamic variance adaptation using differenced maximum mutual information. MLSLP 2012: 9-12 - [c16]Atsunori Ogawa, Takaaki Hori, Atsushi Nakamura:
Recognition rate estimation based on word alignment network and discriminative error type classification. SLT 2012: 113-118 - 2011
- [j2]Atsunori Ogawa, Satoshi Takahashi, Atsushi Nakamura:
Efficient Combination of Likelihood Recycling and Batch Calculation for Fast Acoustic Likelihood Calculation. IEICE Trans. Inf. Syst. 94-D(3): 648-658 (2011) - [c15]Atsunori Ogawa, Satoshi Takahashi, Atsushi Nakamura:
Machine and acoustical condition dependency analyses for fast acoustic likelihood calculation techniques. ICASSP 2011: 5156-5159 - 2010
- [c14]Atsunori Ogawa, Atsushi Nakamura:
Discriminative confidence and error cause estimation for extended speech recognition function. ICASSP 2010: 4454-4457 - [c13]Atsunori Ogawa, Atsushi Nakamura:
A novel confidence measure based on marginalization of jointly estimated error cause probabilities. INTERSPEECH 2010: 242-245 - [c12]Takaaki Hori, Shoko Araki, Takuya Yoshioka, Masakiyo Fujimoto, Shinji Watanabe, Takanobu Oba, Atsunori Ogawa, Kazuhiro Otsuka, Dan Mikami, Keisuke Kinoshita, Tomohiro Nakatani, Atsushi Nakamura, Junji Yamato:
Real-time meeting recognition and understanding using distant microphones and omni-directional camera. SLT 2010: 424-429
2000 – 2009
- 2009
- [c11]Atsunori Ogawa, Satoshi Takahashi, Atsushi Nakamura:
Efficient combination of likelihood recycling and batch calculation based on conditional fast processing and acoustic back-off. ICASSP 2009: 4161-4164 - [c10]Atsunori Ogawa, Atsushi Nakamura:
Simultaneous estimation of confidence and error cause in speech recognition using discriminative model. INTERSPEECH 2009: 1199-1202 - [c9]Satoshi Kobashikawa, Atsunori Ogawa, Yoshikazu Yamaguchi, Satoshi Takahashi:
Rapid unsupervised adaptation using frame independent output probabilities of gender and context independent phoneme models. INTERSPEECH 2009: 1615-1618 - 2008
- [c8]Atsunori Ogawa, Satoshi Takahashi:
Weighted distance measures for efficient reduction of Gaussian mixture components in HMM-based acoustic model. ICASSP 2008: 4173-4176 - 2005
- [j1]Atsunori Ogawa, Yoshikazu Yamaguchi, Shoichi Matsunaga:
Children's speech recognition using elementary-school-student speech database. Syst. Comput. Jpn. 36(12): 33-42 (2005) - [c7]Satoshi Kobashikawa, Satoshi Takahashi, Yoshikazu Yamaguchi, Atsunori Ogawa:
Rapid response and robust speech recognition by preliminary model adaptation for additive and convolutional noise. INTERSPEECH 2005: 965-968 - 2003
- [c6]Shoichi Matsunaga, Atsunori Ogawa, Yoshikazu Yamaguchi, Akihiro Imamura:
Non-native English speech recognition using bilingual English lexicon and acoustic models. ICASSP (1) 2003: 340-343 - [c5]Shoichi Matsunaga, Atsunori Ogawa, Yoshikazu Yamaguchi, Akihiro Imamura:
Non-native English speech recognition using bilingual English lexicon and acoustic models. ICME 2003: 625-628 - [c4]Shoichi Matsunaga, Atsunori Ogawa, Yoshikazu Yamaguchi, Akihiro Imamura:
Speaker adaptation for non-native speakers using bilingual English lexicon and acoustic models. INTERSPEECH 2003: 3113-3116 - 2000
- [c3]Atsunori Ogawa, Yoshiaki Noda, Shoichi Matsunaga:
Novel two-pass search strategy using time-asynchronous shortest-first second-pass beam search. INTERSPEECH 2000: 290-293
1990 – 1999
- 1998
- [c2]Atsunori Ogawa, Kazuya Takeda, Fumitada Itakura:
Balancing acoustic and linguistic probabilities. ICASSP 1998: 181-184 - [c1]Kazuya Takeda, Atsunori Ogawa, Fumitada Itakura:
Estimating entropy of a language from optimal word insertion penalty. ICSLP 1998
Coauthor Index
manage site settings
To protect your privacy, all features that rely on external API calls from your browser are turned off by default. You need to opt-in for them to become active. All settings here will be stored as cookies with your web browser. For more information see our F.A.Q.
Unpaywalled article links
Add open access links from to the list of external document links (if available).
Privacy notice: By enabling the option above, your browser will contact the API of unpaywall.org to load hyperlinks to open access articles. Although we do not have any reason to believe that your call will be tracked, we do not have any control over how the remote server uses your data. So please proceed with care and consider checking the Unpaywall privacy policy.
Archived links via Wayback Machine
For web page which are no longer available, try to retrieve content from the of the Internet Archive (if available).
Privacy notice: By enabling the option above, your browser will contact the API of archive.org to check for archived content of web pages that are no longer available. Although we do not have any reason to believe that your call will be tracked, we do not have any control over how the remote server uses your data. So please proceed with care and consider checking the Internet Archive privacy policy.
Reference lists
Add a list of references from , , and to record detail pages.
load references from crossref.org and opencitations.net
Privacy notice: By enabling the option above, your browser will contact the APIs of crossref.org, opencitations.net, and semanticscholar.org to load article reference information. Although we do not have any reason to believe that your call will be tracked, we do not have any control over how the remote server uses your data. So please proceed with care and consider checking the Crossref privacy policy and the OpenCitations privacy policy, as well as the AI2 Privacy Policy covering Semantic Scholar.
Citation data
Add a list of citing articles from and to record detail pages.
load citations from opencitations.net
Privacy notice: By enabling the option above, your browser will contact the API of opencitations.net and semanticscholar.org to load citation information. Although we do not have any reason to believe that your call will be tracked, we do not have any control over how the remote server uses your data. So please proceed with care and consider checking the OpenCitations privacy policy as well as the AI2 Privacy Policy covering Semantic Scholar.
OpenAlex data
Load additional information about publications from .
Privacy notice: By enabling the option above, your browser will contact the API of openalex.org to load additional information. Although we do not have any reason to believe that your call will be tracked, we do not have any control over how the remote server uses your data. So please proceed with care and consider checking the information given by OpenAlex.
last updated on 2024-10-07 21:16 CEST by the dblp team
all metadata released as open data under CC0 1.0 license
see also: Terms of Use | Privacy Policy | Imprint