Committee |
Date Time |
Place |
Paper Title / Authors |
Abstract |
Paper # |
EA, SIP, SP, IPSJ-SLP [detail] |
2025-03-04 11:05 |
Okinawa |
|
[Poster Presentation]
Construction of a ASR model based on self-supervised learning using intermediate layer outputs Keigo Hojo, Yukoh Wakabayashi (TUT), Kengo Ohta (NITAC), Atsunori Ogawa (NTT), Norihide Kitaoka (TUT) EA2024-138 SIP2024-173 SP2024-79 |
[more] |
EA2024-138 SIP2024-173 SP2024-79 pp.369-374 |
SP, IPSJ-MUS, IPSJ-SLP [detail] |
2024-06-15 13:50 |
Tokyo |
(Primary: On-site, Secondary: Online) |
[Poster Presentation]
Improving CTC-based ASR model by weighting encoder layers using attention mechanisms Keigo Hojo, Yukoh Wakabayashi (TUT), Kengo Ohta (NITAC), Atsunori Ogawa (NTT), Norihide Kitaoka (TUT) SP2024-9 |
[more] |
SP2024-9 pp.43-48 |
SIP, SP, EA, IPSJ-SLP [detail] |
2024-03-01 09:30 |
Okinawa |
(Primary: On-site, Secondary: Online) |
Domain adaptation of speech recognition model based on multilingual SSL model with only nonparallel corpus. Takahiro Kinouchi (TUT), Atsunori Ogawa (NTT), Yukoh Wakabayashi (TUT), Kengo Ohta (NITA), Norihide Kitaoka (TUT) EA2023-100 SIP2023-147 SP2023-82 |
Automatic speech recognition (ASR) models are used in various services and businesses, and each domain’s recognition acc... [more] |
EA2023-100 SIP2023-147 SP2023-82 pp.232-237 |
SIP, SP, EA, IPSJ-SLP [detail] |
2024-03-01 09:30 |
Okinawa |
(Primary: On-site, Secondary: Online) |
Improving speech recognition system consisting of multiple speech recognition models Keigo Hojo, Yukoh Wakabayashi (TUT), Kengo Ohta (NITAC), Atsunori Ogawa (NTT), Norihide Kitaoka (TUT) EA2023-101 SIP2023-148 SP2023-83 |
[more] |
EA2023-101 SIP2023-148 SP2023-83 pp.238-243 |
SIP, SP, EA, IPSJ-SLP [detail] |
2024-03-01 10:40 |
Okinawa |
(Primary: On-site, Secondary: Online) |
Substitution of Implicit Linguistic Information in Beam Search Decoding Using CTC-based Speech Recognition Models Tatsunari Takagi, Yukoh Wakabayashi (TUT), Atsunori Ogawa (NTT), Norihide Kitaoka (TUT) EA2023-106 SIP2023-153 SP2023-88 |
The rise of neural networks in the field of automatic speech recognition has notably improved the accuracy of speech rec... [more] |
EA2023-106 SIP2023-153 SP2023-88 pp.268-273 |
PRMU, IPSJ-CVIM, IPSJ-DCC, IPSJ-CGVI |
2023-11-16 16:50 |
Tottori |
(Primary: On-site, Secondary: Online) |
Understanding level estimation using similarities between users' understanding expression patterns Yuki Kitagishi, Naohiro Tawara, Atsunori Ogawa, Taichi Asami (NTT), Tomoko Yonezawa (Kansai Univ.) PRMU2023-26 |
We define three-degree understanding levels of low/neutral/high as an audience member looks like they are understanding ... [more] |
PRMU2023-26 pp.56-61 |
SP, IPSJ-MUS, IPSJ-SLP [detail] |
2023-06-23 13:50 |
Tokyo |
(Primary: On-site, Secondary: Online) |
Streaming End-to-End speech recognition using a CTC decoder with substituted linguistic information Tatsunari Takagi (TUT), Atsunori Ogawa (NTT), Norihide Kitaoka, Yukoh Wakabayashi (TUT) SP2023-12 |
Speech recognition technology has been employed in various fields due to the enhancement of speech recognition model acc... [more] |
SP2023-12 pp.60-64 |
SP, IPSJ-MUS, IPSJ-SLP [detail] |
2023-06-24 13:50 |
Tokyo |
(Primary: On-site, Secondary: Online) |
Domain adaptation of speech recognition models based on self-supervised learning using target domain speech Takahiro Kinouchi (TUT), Atsunori Ogawa (NTT), Yuko Wakabayashi, Norihide Kitaoka (TUT) SP2023-19 |
In this study, we propose a domain adaptation method using only speech data in the target domain without using transcrib... [more] |
SP2023-19 pp.91-96 |
NLC, IPSJ-NL, SP, IPSJ-SLP [detail] |
2022-11-29 14:35 |
Tokyo |
(Primary: On-site, Secondary: Online) |
Density Ratio Approach-based multiple Encoder-Decoder ASR model integration Keigo Hojo, Daiki Mori, Yukoh Wakabayashi (TUT), Atsunori Ogawa (NTT), Norihide Kitaoka (TUT) NLC2022-10 SP2022-30 |
One of the methods to improve the performance of Encoder--Decoder speech recognition is the integration of an ASR models... [more] |
NLC2022-10 SP2022-30 pp.5-9 |
EA, SIP, SP, IPSJ-SLP [detail] |
2022-03-01 13:10 |
Okinawa |
(Primary: On-site, Secondary: Online) |
The upper limit of subjective intelligibility score of speech enhancement using IRM
-- comparison between laboratory and crowdsourcing experiments -- Ayako Yamamoto, Toshio Irino (Wakayama Univ.), Shoko Araki, Kenichi Arai, Atsunori Ogawa, Keisuke Kinoshita, Tomohiro Nakatani (NTT) EA2021-74 SIP2021-101 SP2021-59 |
We performed subjective speech intelligibility experiments in a laboratory and using crowdsourcing to get a fundamental ... [more] |
EA2021-74 SIP2021-101 SP2021-59 pp.64-69 |
SP, IPSJ-SLP, IPSJ-MUS |
2021-06-18 15:00 |
Online |
Online |
Speech Intelligibility Experiments using crowdsourcing
-- from designing Web page to Data screening -- Ayako Yamamoto, Toshio Irino (Wakayama Univ.), Kenichi Arai, Shoko Araki, Atsunori Ogawa, Keisuke Kinoshita, Tomohiro Nakatani (NTT) SP2021-5 |
Many subjective experiments have been performed to develop objective speech intelligibility measures, but the novel coro... [more] |
SP2021-5 pp.25-30 |
SP, SIP, EA |
2017-03-01 16:40 |
Okinawa |
Okinawa Industry Support Center |
[Invited Talk]
An Introduction to Example-based Speech Enhancement and Its Improvements Atsunori Ogawa, Keisuke Kinoshita, Marc Delcroix, Tomohiro Nakatani (NTT) EA2016-114 SIP2016-169 SP2016-109 |
This paper introduces example-based speech enhancement, which is a promising single-channel approach to cope with highly... [more] |
EA2016-114 SIP2016-169 SP2016-109 pp.183-188 |
SP, SIP, EA |
2017-03-02 09:00 |
Okinawa |
Okinawa Industry Support Center |
[Poster Presentation]
Use of the end of sentence and speaker-derived information in recurrent neural network language models for multiparty conversations. Hiroto Ashikawa, Naohiro Tawara (Waseda Univ.), Atsunori Ogawa, Tomoharu Iwata (NTT), Tetsuji Ogawa, Tetsunori Kobayashi (Waseda Univ.) EA2016-133 SIP2016-188 SP2016-128 |
Information on the end of sentence (EOS) and speaker alternation was exploited in recurrent neural network-based languag... [more] |
EA2016-133 SIP2016-188 SP2016-128 pp.287-290 |