default search action
Masafumi Nishimura
Person information
- affiliation: IBM Research
Refine list
refinements active!
zoomed in on ?? of ?? records
view refined list in
export refined list as
Journal Articles
- 2021
- [j16]Yasunori Yamada, Kaoru Shinkawa, Masatomo Kobayashi, Masafumi Nishimura, Miyuki Nemoto, Eriko Tsukada, Miho Ota, Kiyotaka Nemoto, Tetsuaki Arai:
Tablet-Based Automatic Assessment for Early Detection of Alzheimer's Disease Using Speech Responses to Daily Life Questions. Frontiers Digit. Health 3: 653904 (2021) - [j15]Akihiro Nakamura, Takato Saito, Daizo Ikeda, Ken Ohta, Hiroshi Mineno, Masafumi Nishimura:
Automatic Detection of Chewing and Swallowing. Sensors 21(10): 3378 (2021) - 2018
- [j14]Takashi Fukuda, Osamu Ichikawa, Masafumi Nishimura:
Detecting breathing sounds in realistic Japanese telephone conversations and its application to automatic speech recognition. Speech Commun. 98: 95-103 (2018) - 2015
- [j13]Masayuki Suzuki, Gakuto Kurata, Masafumi Nishimura, Nobuaki Minematsu:
Discriminative re-ranking for automatic speech recognition by leveraging invariant structures. Speech Commun. 72: 208-217 (2015) - 2012
- [j12]Gakuto Kurata, Abhinav Sethy, Bhuvana Ramabhadran, Ariya Rastrow, Nobuyasu Itoh, Masafumi Nishimura:
Acoustically discriminative language model training with pseudo-hypothesis. Speech Commun. 54(2): 219-228 (2012) - [j11]Gakuto Kurata, Nobuyasu Itoh, Masafumi Nishimura, Abhinav Sethy, Bhuvana Ramabhadran:
Leveraging word confusion networks for named entity modeling and detection from conversational telephone speech. Speech Commun. 54(3): 491-502 (2012) - 2010
- [j10]Osamu Ichikawa, Takashi Fukuda, Masafumi Nishimura:
DOA Estimation with Local-Peak-Weighted CSP. EURASIP J. Adv. Signal Process. 2010 (2010) - [j9]Osamu Ichikawa, Takashi Fukuda, Masafumi Nishimura:
Dynamic Features in the Linear-Logarithmic Hybrid Domain for Automatic Speech Recognition in a Reverberant Environment. IEEE J. Sel. Top. Signal Process. 4(5): 816-823 (2010) - [j8]Takashi Fukuda, Osamu Ichikawa, Masafumi Nishimura:
Long-Term Spectro-Temporal and Static Harmonic Features for Voice Activity Detection. IEEE J. Sel. Top. Signal Process. 4(5): 834-844 (2010) - 2008
- [j7]Osamu Ichikawa, Takashi Fukuda, Masafumi Nishimura:
Local Peak Enhancement for In-Car Speech Recognition in Noisy Environment. IEICE Trans. Inf. Syst. 91-D(3): 635-639 (2008) - 2007
- [j6]Ryuki Tachibana, Tohru Nagano, Gakuto Kurata, Masafumi Nishimura, Noboru Babaguchi:
Automatic Prosody Labeling Using Multiple Models for Japanese. IEICE Trans. Inf. Syst. 90-D(11): 1805-1812 (2007) - 2006
- [j5]Tetsuya Takiguchi, Masafumi Nishimura, Yasuo Ariki:
Acoustic Model Adaptation Using First-Order Linear Prediction for Reverberant Speech. IEICE Trans. Inf. Syst. 89-D(3): 908-914 (2006) - 2005
- [j4]Osamu Ichikawa, Masafumi Nishimura:
Simultaneous Adaptation of Echo Cancellation and Spectral Subtraction for In-Car Speech Recognition. IEICE Trans. Fundam. Electron. Commun. Comput. Sci. 88-A(7): 1732-1738 (2005) - 2004
- [j3]Tetsuya Takiguchi, Masafumi Nishimura:
Improved HMM Separation for Distant-Talking Speech Recognition. IEICE Trans. Inf. Syst. 87-D(5): 1127-1137 (2004) - [j2]Osamu Ichikawa, Tetsuya Takiguchi, Masafumi Nishimura:
Sound Source Localization Using a Profile Fitting Method with Sound Reflectors. IEICE Trans. Inf. Syst. 87-D(5): 1138-1145 (2004) - 1991
- [j1]Masafumi Nishimura:
Speaker adaptation method for fenonic markov model-based speech recognition. Syst. Comput. Jpn. 22(13): 47-58 (1991)
Conference and Workshop Papers
- 2023
- [c65]Kohta Masuda, Jun Ogata, Masafumi Nishida, Masafumi Nishimura:
Multi-Self-Supervised Learning Model-Based Throat Microphone Speech Recognition. APSIPA ASC 2023: 1766-1770 - [c64]Amit Karmakar, Masafumi Nishida, Masafumi Nishimura:
Eating and Drinking Behavior Recognition Using Multimodal Fusion. GCCE 2023: 210-213 - [c63]Hibiki Takayama, Masafumi Nishida, Satoru Tsuge, Shingo Kuroiwa, Masafumi Nishimura:
Utterance-style-dependent Speaker Verification by Utilizing Emotions. GCCE 2023: 773-775 - 2022
- [c62]Kohta Masuda, Jun Ogata, Masafumi Nishida, Masafumi Nishimura:
Throat microphone speech recognition using wav2vec 2.0 and feature mapping. GCCE 2022: 395-397 - [c61]Aoi Sugita, Masafumi Nishida, Masafumi Nishimura, Yasuo Horiuchi, Shingo Kuroiwa:
Identification of vocal tract state before and after swallowing using acoustic features. GCCE 2022: 752-753 - [c60]Akihiro Nakamura, Takato Saito, Daizo Ikeda, Ken Ohta, Hiroshi Mineno, Masafumi Nishimura:
Automatic Detection of Crushing Completion Timing of Food. LifeTech 2022: 516-518 - 2021
- [c59]Kosuke Aigo, Takashi Tsunakawa, Masafumi Nishida, Masafumi Nishimura:
Question Generation using Knowledge Graphs with the T5 Language Model and Masked Self-Attention. GCCE 2021: 85-87 - [c58]Akihiro Nakamura, Takato Saito, Daizo Ikeda, Ken Ohta, Hiroshi Mineno, Masafumi Nishimura:
Automatic Detection of Chewing and Swallowing Using Attention-Based Fusion. GCCE 2021: 373-375 - [c57]Akihiro Nakamura, Takato Saito, Daizo Ikeda, Ken Ohta, Hiroshi Mineno, Masafumi Nishimura:
Automatic Detection of Chewing and Swallowing Using Multichannel Sound Information. LifeTech 2021: 173-175 - [c56]Kenta Yoshii, Masafumi Nishimura, Daiki Kimura, Akihiro Kosugi, Kaoru Shinkawa, Toshiro Takase, Masatomo Kobayashi, Yasunori Yamada, Miyuki Nemoto, Ryohei Watanabe, Eriko Tsukada, Miho Ota, Kiyotaka Nemoto, Tetsuaki Arai, Shinji Higashi:
A Study for Detecting Mild Cognitive Impairment by Analyzing Conversations with Humanoid Robots. LifeTech 2021: 347-350 - 2020
- [c55]Akihiro Nakamura, Takato Saito, Daizo Ikeda, Ken Ohta, Hiroshi Mineno, Masafumi Nishimura:
A Data Augmentation Technique for Automatic Detection of Chewing Side and Swallowing. APSIPA 2020: 578-583 - [c54]Akihiro Nakamura, Ken Ohta, Takato Saito, Hiroshi Mineno, Daizo Ikeda, Masafumi Nishimura:
Automatic Detection of Chewing and Swallowing Using Hybrid CTC/Attention. GCCE 2020: 810-812 - [c53]Haruki Fukuda, Takashi Tsunakawa, Jun Oshima, Ritsuko Oshima, Masafumi Nishida, Masafumi Nishimura:
BERT-based Automatic Text Scoring for Collaborative Learning. GCCE 2020: 917-920 - [c52]Akihiro Nakamura, Hiroshi Mineno, Masafumi Nishimura, Takato Saito, Daizo Ikeda, Ken Ohta:
Automatic Detection of the Chewing Side Using Two-channel Recordings under the Ear. LifeTech 2020: 82-83 - [c51]Muhammad Mehedi Billah, Masafumi Nishimura:
A data augmentation-based technique to classify chewing and swallowing using LSTM. LifeTech 2020: 84-85 - 2019
- [c50]Takahito Suzuki, Takashi Tsunakawa, Masafumi Nishida, Masafumi Nishimura, Jun Ogata:
Effects of Mounting Position on Throat Microphone Speech Recognition. GCCE 2019: 873-874 - [c49]Muhammad Mehedi Billah, Taiju Abe, Akihiro Nakamura, Hiroshi Mineno, Masafumi Nishimura, Takato Saito, Daizo Ikeda:
Estimation of Number of Chewing Strokes and Swallowing Events by Using LSTM-CTC and Throat Microphone. GCCE 2019: 920-921 - [c48]Takahito Suzuki, Jun Ogata, Takashi Tsunakawa, Masafumi Nishida, Masafumi Nishimura:
Knowledge Distillation for Throat Microphone Speech Recognition. INTERSPEECH 2019: 461-465 - [c47]Kaoru Shinkawa, Akihiro Kosugi, Masafumi Nishimura, Miyuki Nemoto, Kiyotaka Nemoto, Tomoko Takeuchi, Yuriko Numata, Ryohei Watanabe, Eriko Tsukada, Miho Ota, Shinji Higashi, Tetsuaki Arai, Yasunori Yamada:
Multimodal Behavior Analysis Towards Detecting Mild Cognitive Impairment: Preliminary Results on Gait and Speech. MedInfo 2019: 343-347 - 2018
- [c46]Takahito Suzuki, Jun Ogata, Takashi Tsunakawa, Masafumi Nishida, Masafumi Nishimura:
Bottleneck feature-mediated DNN-based feature mapping for throat microphone speech recognition. APSIPA 2018: 1738-1741 - [c45]Motoki Abe, Takashi Tsunakawa, Masafumi Nishida, Masafumi Nishimura:
Dialogue Breakdown Detection Based on Nonlinguistic Acoustic Information. GCCE 2018: 689-690 - 2017
- [c44]Shengke Lin, Takashi Tsunakawa, Masafumi Nishida, Masafumi Nishimura:
DNN-based feature transformation for speech recognition using throat microphone. APSIPA 2017: 596-599 - [c43]Yutaro Yamada, Takato Saito, Satoshi Kawasaki, Daizo Ikeda, Masaji Katagiri, Masafumi Nishimura, Hiroshi Mineno:
A Deep-Learning-Based Method of Estimating Water Intake. COMPSAC (2) 2017: 96-101 - [c42]Yutaro Yamada, Masafumi Nishimura, Hiroshi Mineno, Takato Saito, Satoshi Kawasaki, Daizo Ikeda, Masaji Katagiri:
Deep learning-based water-intake estimation method using second half of swallowing sound. GCCE 2017: 1-2 - 2015
- [c41]Nobuyasu Itoh, Gakuto Kurata, Ryuki Tachibana, Masafumi Nishimura:
A metric for evaluating speech recognizer output based on human-perception model. INTERSPEECH 2015: 1285-1288 - 2014
- [c40]Congying Zhang, Masayuki Suzuki, Gakuto Kurata, Masafumi Nishimura, Nobuaki Minematsu:
Leveraging phonetic context dependent invariant structure for continuous speech recognition. ChinaSIP 2014: 52-56 - [c39]Takashi Fukuda, Osamu Ichikawa, Masafumi Nishimura, Steven J. Rennie, Vaibhava Goel:
Regularized feature-space discriminative adaptation for robust ASR. INTERSPEECH 2014: 2185-2188 - 2013
- [c38]Osamu Ichikawa, Steven J. Rennie, Takashi Fukuda, Masafumi Nishimura:
Channel-mapping for speech corpus recycling. ICASSP 2013: 7160-7164 - 2012
- [c37]Osamu Ichikawa, Steven J. Rennie, Takashi Fukuda, Masafumi Nishimura:
Model-based noise reduction leveraging frequency-wise confidence metric for in-car speech recognition. ICASSP 2012: 4921-4924 - [c36]Masayuki Suzuki, Gakuto Kurata, Masafumi Nishimura, Nobuaki Minematsu:
Discriminative Reranking for LVCSR Leveraging Invariant Structure. INTERSPEECH 2012: 563-566 - 2011
- [c35]Gakuto Kurata, Nobuyasu Itoh, Masafumi Nishimura, Abhinav Sethy, Bhuvana Ramabhadran:
Named entity recognition from Conversational Telephone Speech leveraging Word Confusion Networks for training and recognition. ICASSP 2011: 5572-5575 - [c34]Gakuto Kurata, Nobuyasu Itoh, Masafumi Nishimura:
Training of error-corrective model for ASR without using audio data. ICASSP 2011: 5576-5579 - [c33]Takashi Fukuda, Osamu Ichikawa, Masafumi Nishimura:
Combining Feature Space Discriminative Training with Long-Term Spectro-Temporal Features for Noise-Robust Speech Recognition. INTERSPEECH 2011: 229-232 - [c32]Masayuki Suzuki, Gakuto Kurata, Masafumi Nishimura, Nobuaki Minematsu:
Continuous Digits Recognition Leveraging Invariant Structure. INTERSPEECH 2011: 993-996 - [c31]Gakuto Kurata, Nobuyasu Itoh, Masafumi Nishimura:
Acoustic Model Training with Detecting Transcription Errors in the Training Data. INTERSPEECH 2011: 1689-1692 - [c30]Ryoichi Takashima, Tohru Nagano, Ryuki Tachibana, Masafumi Nishimura:
Agglomerative Hierarchical Clustering of Emotions in Speech Based on Subjective Relative Similarity. INTERSPEECH 2011: 2473-2476 - [c29]Takashi Fukuda, Osamu Ichikawa, Masafumi Nishimura:
Breath-Detection-Based Telephony Speech Phrasing. INTERSPEECH 2011: 2625-2628 - 2010
- [c28]Takashi Fukuda, Osamu Ichikawa, Masafumi Nishimura:
Improved voice activity detection using static harmonic features. ICASSP 2010: 4482-4485 - [c27]Toru Nakashika, Ryuki Tachibana, Masafumi Nishimura, Tetsuya Takiguchi, Yasuo Ariki:
Speech synthesis by modeling harmonics structure with multiple function. INTERSPEECH 2010: 945-948 - 2009
- [c26]Gakuto Kurata, Nobuyasu Itoh, Masafumi Nishimura:
Acoustically discriminative training for language models. ICASSP 2009: 4717-4720 - [c25]Osamu Ichikawa, Takashi Fukuda, Ryuki Tachibana, Masafumi Nishimura:
Dynamic features in the linear domain for robust automatic speech recognition in a reverberant environment. INTERSPEECH 2009: 44-47 - [c24]Ryuki Tachibana, Zhiwei Shuang, Masafumi Nishimura:
Japanese pitch conversion for voice morphing based on differential modeling. INTERSPEECH 2009: 2651-2654 - 2008
- [c23]Tohru Nagano, Ryuki Tachibana, Nobuyasu Itoh, Masafumi Nishimura:
Improving phoneme and accent estimation by leveraging a dictionary for a stochastic TTS front-end. ICASSP 2008: 4689-4692 - [c22]Osamu Ichikawa, Takashi Fukuda, Masafumi Nishimura:
Local peak enhancement combined with noise reduction algorithms for robust automatic speech recognition in automobiles. ICASSP 2008: 4869-4872 - [c21]Takashi Fukuda, Osamu Ichikawa, Masafumi Nishimura:
Phone-duration-dependent long-term dynamic features for a stochastic model-based voice activity detection. INTERSPEECH 2008: 1293-1296 - [c20]Takashi Fukuda, Osamu Ichikawa, Masafumi Nishimura:
Short- and long-term dynamic features for robust speech recognition. INTERSPEECH 2008: 2262-2265 - 2007
- [c19]Yuta Nakashima, Ryuki Tachibana, Masafumi Nishimura, Noboru Babaguchi:
Determining Recording Location Based on Synchronization Positions of Audiowatermarking. ICASSP (2) 2007: 253-256 - [c18]Gakuto Kurata, Shinsuke Mori, Nobuyasu Itoh, Masafumi Nishimura:
Unsupervised Lexicon Acquisition from Speech and Text. ICASSP (4) 2007: 421-424 - [c17]Ryuki Tachibana, Tohru Nagano, Gakuto Kurata, Masafumi Nishimura, Noboru Babaguchi:
Preliminary experiments toward automatic generation of new TTS voices from recorded speech alone. INTERSPEECH 2007: 1917-1920 - 2006
- [c16]Gakuto Kurata, Shinsuke Mori, Masafumi Nishimura:
Unsupervised Adaptation of a Stochastic Language Model Using a Japanese Raw Corpus. ICASSP (1) 2006: 1037-1040 - [c15]Yuta Nakashima, Ryuki Tachibana, Masafumi Nishimura, Noboru Babaguchi:
Estimation of recording location using audio watermarking. MM&Sec 2006: 108-113 - 2005
- [c14]Osamu Ichikawa, Masafumi Nishimura:
Simultaneous adaptation of echo cancellation and spectral subtraction for in-car speech recognition. INTERSPEECH 2005: 2293-2296 - [c13]Tohru Nagano, Shinsuke Mori, Masafumi Nishimura:
A stochastic approach to phoneme and accent estimation. INTERSPEECH 2005: 3293-3296 - 2004
- [c12]Tetsuya Takiguchi, Masafumi Nishimura:
Acoustic model adaptation using first order prediction for reverberant speech. ICASSP (1) 2004: 869-872 - 2003
- [c11]Shinsuke Mori, Masafumi Nishimura, Nobuyasu Itoh:
Language model adaptation using word clustering. INTERSPEECH 2003: 425-428 - 2001
- [c10]Shinsuke Mori, Masafumi Nishimura, Nobuyasu Itoh:
Improvement of a structured language model: arbori-context tree. INTERSPEECH 2001: 713-716 - 2000
- [c9]Shinsuke Mori, Masafumi Nishimura, Nobuyasu Itoh, Shiho Ogino, Hideo Watanabe:
A Stochastic Parser Based on a Structural Word Prediction Model. COLING 2000: 558-564 - [c8]Nobuyasu Itoh, Masafumi Nishimura, Shinsuke Mori:
A method for style adaptation to spontaneous speech by using a semi-linear interpolation technique. INTERSPEECH 2000: 374-377 - 1998
- [c7]Shinsuke Mori, Masafumi Nishimura, Nobuyasu Itoh:
Word clustering for a word bi-gram model. ICSLP 1998 - 1989
- [c6]Masafumi Nishimura:
HMM-based speech recognition using dynamic spectral feature. ICASSP 1989: 298-301 - 1988
- [c5]Masafumi Nishimura, Kazuhide Sugawara:
Speaker adaptation method for HMM-based speech recognition. ICASSP 1988: 207-210 - 1987
- [c4]Masafumi Nishimura, Koichi Toshioka:
HMM-Based speech recognition using multi-dimensional multi-labeling. ICASSP 1987: 1163-1166 - 1986
- [c3]Kazuhide Sugawara, Masafumi Nishimura, Akihiro Kuroda:
Speaker adaptation for a hidden Markov model. ICASSP 1986: 2667-2670 - 1985
- [c2]Kazuhide Sugawara, Masafumi Nishimura, Koichi Toshioka, Masaaki Okochi, Toyohisa Kaneko:
Isolated word recognition using hidden Markov models. ICASSP 1985: 1-4 - 1984
- [c1]Yasuhiro Matsuda, Shu Tezuka, Mitsuhiko Kanoh, Masafumi Nishimura, Toyohisa Kaneko:
A method for recognizing Japanese monosyllables by using intermediate cumulative distance. ICASSP 1984: 340-343
Coauthor Index
manage site settings
To protect your privacy, all features that rely on external API calls from your browser are turned off by default. You need to opt-in for them to become active. All settings here will be stored as cookies with your web browser. For more information see our F.A.Q.
Unpaywalled article links
Add open access links from to the list of external document links (if available).
Privacy notice: By enabling the option above, your browser will contact the API of unpaywall.org to load hyperlinks to open access articles. Although we do not have any reason to believe that your call will be tracked, we do not have any control over how the remote server uses your data. So please proceed with care and consider checking the Unpaywall privacy policy.
Archived links via Wayback Machine
For web page which are no longer available, try to retrieve content from the of the Internet Archive (if available).
Privacy notice: By enabling the option above, your browser will contact the API of archive.org to check for archived content of web pages that are no longer available. Although we do not have any reason to believe that your call will be tracked, we do not have any control over how the remote server uses your data. So please proceed with care and consider checking the Internet Archive privacy policy.
Reference lists
Add a list of references from , , and to record detail pages.
load references from crossref.org and opencitations.net
Privacy notice: By enabling the option above, your browser will contact the APIs of crossref.org, opencitations.net, and semanticscholar.org to load article reference information. Although we do not have any reason to believe that your call will be tracked, we do not have any control over how the remote server uses your data. So please proceed with care and consider checking the Crossref privacy policy and the OpenCitations privacy policy, as well as the AI2 Privacy Policy covering Semantic Scholar.
Citation data
Add a list of citing articles from and to record detail pages.
load citations from opencitations.net
Privacy notice: By enabling the option above, your browser will contact the API of opencitations.net and semanticscholar.org to load citation information. Although we do not have any reason to believe that your call will be tracked, we do not have any control over how the remote server uses your data. So please proceed with care and consider checking the OpenCitations privacy policy as well as the AI2 Privacy Policy covering Semantic Scholar.
OpenAlex data
Load additional information about publications from .
Privacy notice: By enabling the option above, your browser will contact the API of openalex.org to load additional information. Although we do not have any reason to believe that your call will be tracked, we do not have any control over how the remote server uses your data. So please proceed with care and consider checking the information given by OpenAlex.
last updated on 2024-04-24 23:11 CEST by the dblp team
all metadata released as open data under CC0 1.0 license
see also: Terms of Use | Privacy Policy | Imprint