Committee |
Date Time |
Place |
Paper Title / Authors |
Abstract |
Paper # |
PRMU, IBISML, IPSJ-CVIM [detail] |
2023-03-02 11:25 |
Hokkaido |
Future University Hakodate (Primary: On-site, Secondary: Online) |
Binarization of Vision Transformer with Scaling Factors Shun Sato, Shun Sawada, Hidefumi Ohmura, Kouichi katsurada (TUS) PRMU2022-83 IBISML2022-90 |
1bit neural network optimization is an optimization technique that achieves a significant increase in computational spee... [more] |
PRMU2022-83 IBISML2022-90 pp.134-139 |
SP, IPSJ-SLP, EA, SIP [detail] |
2023-02-28 09:50 |
Okinawa |
(Primary: On-site, Secondary: Online) |
End-to-End Speech Synthesis Based on Articulatory Movements Captured by Real-time MRI Yuto Otani, Shun Sawada, Hidefumi Ohmura, Kouichi Katsurada (Tokyo Univ. Sci.) EA2022-77 SIP2022-121 SP2022-41 |
We propose an end-to-end deep learning model for speech synthesis based on articulatory movements captured by real-time ... [more] |
EA2022-77 SIP2022-121 SP2022-41 pp.13-18 |
SP, WIT, IPSJ-SLP [detail] |
2022-10-22 15:40 |
Kyoto |
Kyoto University (Primary: On-site, Secondary: Online) |
Conformer based early fusion model for audio-visual speech recognition Nobukazu Aoki, Shun Sawada, Hidefumi Ohmura, Kouichi Katsurada (Tokyo Univ. of Sci.) SP2022-28 WIT2022-3 |
Previous studies of late fusion models with conformer encoders use independent encoders for both visual and audio inform... [more] |
SP2022-28 WIT2022-3 pp.8-13 |
EA, US, SP, SIP, IPSJ-SLP [detail] |
2021-03-04 17:10 |
Online |
Online |
A Vocoder-free Any-to-Many Voice Conversion using Pre-trained vq-wav2vec Takeshi Koshizuka, Hidefumi Ohmura, Kouichi Katsurada (TUS) EA2020-89 SIP2020-120 SP2020-54 |
Voice conversion (VC) is a technique that converts speaker-dependent non-linguistic information to another speaker's one... [more] |
EA2020-89 SIP2020-120 SP2020-54 pp.176-181 |
SP |
2020-01-28 16:15 |
Toyama |
|
Multi-Speaker Speech Synthesis from EMA Data Kouichi Katsurada (Tokyo Univ. Sci.), Korin Richmond (Univ. of Edinburgh) SP2019-45 |
[more] |
SP2019-45 pp.7-12 |
WIT, SP |
2019-10-27 09:00 |
Kagoshima |
Daiichi Institute of Technology |
Extraction of linguistic representation and syllable recognition from EEG signal of speech-imagery Kentaro Fukai, Hidefumi Ohmura, Kouichi Katsurada (Tokyo Univ. of Science), Satoka Hirata, Yurie Iribe (Aichi Prefectural Univ.), Mingchua Fu, Ryo Taguchi (Nagoya Inst. of Technology), Tsuneo Nitta (Waseda Univ./Toyohashi Univ. of Technology) SP2019-28 WIT2019-27 |
Speech imagery recognition from Electroencephalogram (EEG) is one of the challenging technologies for non-invasive brain... [more] |
SP2019-28 WIT2019-27 pp.63-68 |
WIT, SP |
2019-10-27 09:20 |
Kagoshima |
Daiichi Institute of Technology |
Word Recognition using word likelihood vector from speech-imagery EEG Satoka Hirata, Yurie Iribe (Aichi Prefectual Univ.), Kentaro Fukai, Kouichi Katsurada (Tokyo Univ. of Science), Tsuneo Nitta (Waseda Univ./Toyohashi Univ. of Tech.) SP2019-29 WIT2019-28 |
Previous research suggests that humans manipulate the machine using their electroencephalogram called BCI (Brain Compute... [more] |
SP2019-29 WIT2019-28 pp.69-73 |
SP |
2019-06-13 13:55 |
Kanagawa |
Tokyo Institute of Technology |
Collection of large-scale Japanese articulatory-acoustic parallel data Kohei Wakamiya, Fumiaki Taguchi, Riko Watanabe (Kyushu Univ.), Kouichi Katsurada (Tokyo Univ. of Science), Takehiko Makino (Chuo Univ.), Tokihiko Kaburagi (Kyushu Univ.) SP2019-2 |
A large-scale Japanese articulatory-acoustic parallel data is a basic data using as a natural speech synthesis using a a... [more] |
SP2019-2 pp.7-12 |
PRMU, IPSJ-CVIM, MVE [detail] |
2017-01-19 15:55 |
Kyoto |
|
ngle independent lip reading using symmetrical 3D-AAM of facial images Takuya Watanabe (TUT), Kouichi Katsurada (TUS), Yasushi Kanazawa (TUT) PRMU2016-134 MVE2016-25 |
Lip reading is a technique to recognize spoken words from only visual images of a face. There have been proposed variou... [more] |
PRMU2016-134 MVE2016-25 pp.135-140 |
PRMU, IBISML, IPSJ-CVIM [detail] |
2015-09-15 11:15 |
Ehime |
|
Evaluation of Active Appearance Models using AutoEncoder Takuya Watanabe, Kouichi Katsurada (TUT), Tsuneo Nitta (WU), Yurie Iribe (APU) PRMU2015-85 IBISML2015-45 |
Active Appearance Models (AAM) is a synthetic model of facial images, which reduces dimension of feature space construct... [more] |
PRMU2015-85 IBISML2015-45 pp.135-140 |
SP |
2015-01-22 10:00 |
Gifu |
Juroku Plaza |
Improvement of vocal sound source in singing voice synthesis using articulatory movement HMM Kotaro Irisawa, Kouichi Katsurada (TUT), Yurie Iribe (Aichi Prefectural Univ.), Tsuneo Nitta (Waseda Univ.) SP2014-127 |
Development of singing voice synthesis system has received a lot of attention along with the improvement of speech synth... [more] |
SP2014-127 pp.1-6 |
ET |
2014-03-08 10:30 |
Kochi |
Kochi National College of Technology |
Development of a Full-Text Search Module for Audio and Video Lectures with the Use of a Spoken Term Detection System Utilizing the Suffix Array Yosuke Morimoto, Kumiko Aoki (OUJ), Kouichi Katsurada, Genki Ishihara, Seiichi Miura (TUT), Yurie Iribe (Aichi Prefectural Univ.), Tsuneo Nitta (Waseda Univ.) ET2013-124 |
[more] |
ET2013-124 pp.187-192 |
SP, IPSJ-SLP |
2013-12-19 16:00 |
Tokyo |
|
Improvement of AF-based Voice Conversion for Arbitrary Speakers Narpendyah Wisjnu Ariwardhani (TUT), Yurie Iribe (Aichi Prefectureal Univ.), Kouichi Katsurada (TUT), Tsuneo Nitta (Waseda Univ.) SP2013-85 |
In this paper, we use artificial neural networks (ANNs) for articulatory feature (AF) based voice conversion. ANNs are a... [more] |
SP2013-85 pp.65-70 |
SP |
2012-03-09 14:00 |
Saitama |
Riken Brain Science Institute |
Real-time Visualization of English Pronunciation on an IPA Chart Based on Articulatory Feature Extraction Takurou Mori, Yurie Iribe, Kouichi Katsurada, Tsuneo Nitta (TUT) SP2011-169 |
CALL systems that can support Japanese students to study foreign languages have been developed in recent years. We have ... [more] |
SP2011-169 pp.77-82 |
PRMU, SP |
2012-02-09 10:30 |
Miyagi |
|
Parameter Search Improvement on Facial Expression Synthesis using Active Appearance Models Yuki Nishikiuchi, Kouichi Katsurada, Yurie Iribe, Tsuneo Nitta (Toyohashi Univ. of Tech.) PRMU2011-189 SP2011-104 |
[more] |
PRMU2011-189 SP2011-104 pp.19-24 |
PRMU, SP |
2012-02-09 11:00 |
Miyagi |
|
Toward Effect of Difference of Facial Region on Lipreading Performance Daisuke Ikeda, Kouichi Katsurada, Yurie Iribe, Tsuneo Nitta (Toyohasi Univ. of Tech.) PRMU2011-190 SP2011-105 |
(To be available after the conference date) [more] |
PRMU2011-190 SP2011-105 pp.25-30 |
SP, NLC, IPSJ-SLP [detail] |
2011-12-19 15:45 |
Tokyo |
|
Phoneme Recognition based on AF-HMMs with Optimal State Configuration Narpendyah W. Ariwardhani, Yurie Iribe, Kouichi Katsurada, Tsuneo Nitta (Toyohashi Univ. of Tech.) NLC2011-39 SP2011-84 |
Speeh recognition based on one-model of articulatory movement HMMs that are commonly applied to both speech recognition ... [more] |
NLC2011-39 SP2011-84 pp.49-54 |
SP |
2011-07-21 13:30 |
Hokkaido |
Jozankei Grand Hotel |
One-model Speech Recognition and Synthesis System Based on Articulatory Masashi Kimura, Takayuki Onoda, Yurie Iribe, Kouichi Katsurada, Tsuneo Nitta (Toyohashi Tech) SP2011-41 |
Speech recognition (SR) and speech synthesis (SS) based on one-model of articulatory movement HMMs that are commonly app... [more] |
SP2011-41 pp.1-6 |
PRMU, HIP |
2010-03-16 16:35 |
Kagoshima |
Kagoshima Univ. |
A Facial Expression Mimicking System Based on Facial Expression Elements Extraction Using AAM Ryuichi Fukui, Kouichi Katsurada, Yurie Iribe, Tsuneo Nitta (Toyohashi Univ. of Tech.) PRMU2009-320 HIP2009-205 |
We propose a method for extracting facial expression elements on AAM and develop a facial expression mimicking system. A... [more] |
PRMU2009-320 HIP2009-205 pp.513-518 |
PRMU, SP, MVE, CQ |
2010-01-22 15:20 |
Kyoto |
Kyoto Univ. |
Web-based Multimodal Interaction System based on hierarchical architecture Masashi Kudoh, Kouichi Katsurada, Yurie Iribe, Tsuneo Nitta (Toyohashi Univ. of Tech.) CQ2009-106 PRMU2009-205 SP2009-146 MVE2009-128 |
We have developed an Multi-Modal Interaction(MMI) system using a web browser as interface. However this system has a pro... [more] |
CQ2009-106 PRMU2009-205 SP2009-146 MVE2009-128 pp.351-356 |