Committee |
Date Time |
Place |
Paper Title / Authors |
Abstract |
Paper # |
SP, IPSJ-SLP, IPSJ-MUS |
2021-06-18 13:00 |
Online |
Online |
Tools and practice for supporting recommended protocol for acoustic recording of speech data for high usability
-- Application of a cascaded all-pass filters with randomized center frequencies and phase polarities -- Hideki Kawahara (Wakayama Univ.), Kohei Yatabe (Waseda Univ.), Ken-Ichi Sakakibara (Health Sci. Univ. Hokkaido), Mitsunori Mizumachi (Kyushu Inst. Tech), Masanori Morise (Meiji Univ.), Hideki Banno (Meijo Univ.), Toshio Irino (Wakayama Univ.) SP2021-1 |
Objective measurements of speech data acquisition and presentation is clucial for assureing reploducibility and reusabil... [more] |
SP2021-1 pp.1-6 |
SP, IPSJ-SLP, IPSJ-MUS |
2021-06-18 13:00 |
Online |
Online |
F0 estimation of speech based on l2-norm regularized TV-CAR analysis Keiichi Funaki (Univ. of the Ryukyus) SP2021-2 |
Linear Prediction (LP) is the most successful speech analysis in speech processing, including speech coding implemented
... [more] |
SP2021-2 pp.7-12 |
SP, IPSJ-SLP, IPSJ-MUS |
2021-06-18 15:00 |
Online |
Online |
A Beginner's Introduction to Sound Programming for Digital Stomp Boxes Naofumi Aoki (Hokkaido Univ.) SP2021-3 |
This study has developed a platform for programmable digital stomp boxes. This paper introduces the overview of our prod... [more] |
SP2021-3 pp.13-18 |
SP, IPSJ-SLP, IPSJ-MUS |
2021-06-18 15:00 |
Online |
Online |
Protection method with audio processing against Audio Adversarial Example Taisei Yamamoto, Yuya Tarutani, Yukinobu Fukusima, Tokumi Yokohira (Okayama Univ) SP2021-4 |
Machine learning technology has improved the recognition accuracy of voice recognition, and demand for voice recognition... [more] |
SP2021-4 pp.19-24 |
SP, IPSJ-SLP, IPSJ-MUS |
2021-06-18 15:00 |
Online |
Online |
Speech Intelligibility Experiments using crowdsourcing
-- from designing Web page to Data screening -- Ayako Yamamoto, Toshio Irino (Wakayama Univ.), Kenichi Arai, Shoko Araki, Atsunori Ogawa, Keisuke Kinoshita, Tomohiro Nakatani (NTT) SP2021-5 |
Many subjective experiments have been performed to develop objective speech intelligibility measures, but the novel coro... [more] |
SP2021-5 pp.25-30 |
SP, IPSJ-SLP, IPSJ-MUS |
2021-06-18 15:00 |
Online |
Online |
[Poster Presentation]
Scream detection based on deep learning using time-sequential spectral and cepstral features Takahiro Fukumori (Ritsumeikan Univ.) SP2021-6 |
Discrimination between normal speech and scream is crucial in audio surveillance and monitoring. Although deep neural ne... [more] |
SP2021-6 pp.31-36 |
SP, IPSJ-SLP, IPSJ-MUS |
2021-06-18 17:10 |
Online |
Online |
[Invited Talk]
Spoken Dialogue System for Android ERICA
-- A Multimodal Turing Test Challenge -- Koji Inoue (Kyoto Univ.) SP2021-7 |
One of the major goals for artificial intelligence research is to realize machines (or robots) that can talk like human ... [more] |
SP2021-7 p.37 |
SP, IPSJ-SLP, IPSJ-MUS |
2021-06-19 09:30 |
Online |
Online |
[Invited Talk]
Toward a Unification of Various Speech Processing Tasks Based on End-to-End Neural networks Shinji Watanabe (CMU) SP2021-8 |
This presentation will introduce the recent progress of speech processing technologies based on end-to-end neural networ... [more] |
SP2021-8 p.38 |
SP, IPSJ-SLP, IPSJ-MUS |
2021-06-19 13:00 |
Online |
Online |
Creating of Japanese Phoneme Balanced Sentences for Speech Synthesis Yuko Takai, Naofumi Aoki, Yoshinori Dobashi (Hokkaido Univ.) SP2021-9 |
When the loss of voice is inevitable due to pharyngectomy or other reasons, it has become possible to realizespeech synt... [more] |
SP2021-9 pp.39-41 |
SP, IPSJ-SLP, IPSJ-MUS |
2021-06-19 13:00 |
Online |
Online |
Verifying the Method to Generate Stage Data for Rhythm Game Using Machine Learning Atsuhito Udo, Naofumi Aoki, Yoshinori Dobashi (Hokkaido Univ.) SP2021-10 |
Recently, the video game development industry has exhibited significant growth, which resulted in more intense competiti... [more] |
SP2021-10 pp.42-45 |
SP, IPSJ-SLP, IPSJ-MUS |
2021-06-19 13:00 |
Online |
Online |
Low Loss Machine Learning for Digital Modeling of Distortion Stomp Boxes. Yuto Matsunaga, Naofumi Aoki, Yoshinori Dobashi (Hokkaido Univ.), Tetsuya Kojima (NITTC) SP2021-11 |
Distortion stomp boxes are one of the acoustic devices used on electric guitars. This device has attracted the interest ... [more] |
SP2021-11 pp.46-50 |
SP, IPSJ-SLP, IPSJ-MUS |
2021-06-19 13:00 |
Online |
Online |
A Study on Error Correction for Improving the Accuracy of Acoustic Models Saki Anazawa, Naofumi Aoki, Yoshinori Dobashi (Hokkaido Univ.) SP2021-12 |
People with ALS (amyotrophic lateral sclerosis) or dysarthria sometimes use their own voice for speech synthesis. In thi... [more] |
SP2021-12 pp.51-52 |
SP, IPSJ-SLP, IPSJ-MUS |
2021-06-19 13:00 |
Online |
Online |
A Research Related to the Fricative Sound Determination in Digital Pattern Playback Hiroki Otake, Naofumi Aoki, Kosei Ozeki, Yoshinori Dobashi (Hokkaido Univ.) SP2021-13 |
Pattern Playback (PP) is a device which converts spectrograms into speech by amplitude modulation, and Digital Pattern P... [more] |
SP2021-13 pp.53-56 |
SP, IPSJ-SLP, IPSJ-MUS |
2021-06-19 13:00 |
Online |
Online |
Study on the background cancellation system for speech privacy Jiangning Huang, Akinori Ito (Tohoku Univ.) SP2021-14 |
Evacuation centers at the time of disaster do not have sufficient sound insulation to maintain sound privacy. In this st... [more] |
SP2021-14 pp.57-62 |
SP, IPSJ-SLP, IPSJ-MUS |
2021-06-19 15:00 |
Online |
Online |
Simulation of Body-conducted Speech and Synthesis of One's Own Voice with a Sound-proof Earmuff and Bone-conduction Microphones Chen Ruiyan, Nishimura Tazuko, Minematsu Nobuaki, Saito Daisuke (UTokyo) SP2021-15 |
When one hears his/her recorded voices for the first time, s/he is probably surprised and not rarely disappointed at the... [more] |
SP2021-15 pp.63-68 |
SP, IPSJ-SLP, IPSJ-MUS |
2021-06-19 15:00 |
Online |
Online |
How logical properties in speech are processed in the brian
-- Digital Linguistics -- Kumon Tokumaru (Writer) SP2021-16 |
The digital evolution of language and intelligence of linguistic humans started 70 KA (thousand years ago) in South Afri... [more] |
SP2021-16 pp.69-74 |
SP, IPSJ-SLP, IPSJ-MUS |
2021-06-19 15:00 |
Online |
Online |
Investigation on fine-tuning with image classification networks for deep neural network-based musical instrument classification Yuki Shiroma, Yuma Kinoshita, Sayaka Shiota, Hitoshi Kiya (TMU) SP2021-17 |
In this paper, we investigate abilities of channel conversion methods for fine-tuning with image classification networks... [more] |
SP2021-17 pp.75-79 |
SP, IPSJ-SLP, IPSJ-MUS |
2021-06-19 15:00 |
Online |
Online |
Dynamic Display of Guidelines in Interactive Speech Synthesizer Daiki Goto (Hokkai Gakuen Univ.), Naofumi Aoki, Keisuke ai (Hokkaido Univ.), Kunitoshi Motoki (Hokkai Gakuen Univ.) SP2021-18 |
We are developing a speech synthesis system that can play sounds by interactive control, just like playing a musical ins... [more] |
SP2021-18 pp.80-84 |
SP, IPSJ-SLP, IPSJ-MUS |
2021-06-19 15:00 |
Online |
Online |
Preliminary study on synthesizing relaxing voices
-- from a perspective of recognized/evoked emotions and acoustic features -- Yuki Watanabe, Shuichi Sakamoto (Tohoku Univ.), Takayuki Hoshi, Yoshiki Nagatani, Manabu Nakano (Pixie Dust Technologies) SP2021-19 |
The goal of this study is to synthesize speech sound which induces relaxed emotion. As the preliminary study, we investi... [more] |
SP2021-19 pp.85-90 |
SP, IPSJ-SLP, IPSJ-MUS |
2021-06-19 15:00 |
Online |
Online |
Unseen speaker's Voice Conversion by FaderNetVC with Speaker Feature Extractor Takumi Isako, Takuya Kishida, Toru Nakashika (UEC) SP2021-20 |
In recent years, many voice conversion models using Deep Neural Network (DNN) have been proposed, and FaderNetVC is one ... [more] |
SP2021-20 pp.91-96 |
|