Kobe, Japan The First International Conference on Spoken Language Processing, ICSLP 1990, Kobe, Japan, November 18-22, 1990.
Temporal Control in the Spoken Language
- Morio Kohno, Tomoko Tanioka:
The nature of timing control in language.
- Mary E. Beckman, Maria G. Swora, Jane Rauschenberg, Kenneth de Jong:
Stress shift, stress clash, and polysyllabic shortening in a prosodically annotated discourse.
- W. Nick Campbell:
Evidence for a syllable-based model of speech timing.
- Patti Price, Colin W. Wightman, Mari Ostendorf, John Bear:
The use of relative duration in syntactic disambiguation.
- Nobuyoshi Kaiki, Kazuya Takeda, Yoshinori Sagisaka:
Statistical analysis for segmental duration rules in Japanese speech synthesis.
- Florien J. Koopmans-van Beinum:
Spectro-temporal reduction and expansion in spontaneous speech and read text: the role of focus words.
- Yasuko Nagano-Madsen:
Perception of mora in the three dialects of Japanese.
- Shihua Wang, Erdal Paksoy, Allen Gersho:
Performance of nonlinear prediction of speech.
- Ren-Hua Wang, Quan fen Guan, Hiroya Fujisaki:
A method for robust GARMA analysis of speech.
- Keiichi Tokuda, Takao Kobayashi, Satoshi Imai:
Generalized cepstral analysis of speech - unified approach to LPC and cepstral method.
- Paul J. Dix, Gerrit Bloothooft, E. J. M. van Mierlo:
A geometrical argument for imposing an additional constraint on temporal decomposition.
- Keiichi Funaki, Yukio Mitome:
A speech analysis method based on a glottal source model.
- Ki Yong Lee, Inhyok Cha, Eckho Song, Souguil Ann:
An improved method for multipulse speech analysis.
- Lu Chang, M. M. Bayoumi:
New results on theory of hidden Markov models.
Voice Source Dynamics; Facts and Models
- Ronald C. Scherer, Chwen-geng Guo:
Laryngeal modeling: translaryngeal pressure for a model with many glottal shapes.
- Shigeru Kiritani, Hiroshi Imagawa, Hajime Hirose:
Vocal cord vibration and voice source characteristics - observations by a high-speed digital image recording -.
- Bert Cranen:
Interpretation of EGG and glottal flow by means of a parametrical glottal geometry model.
- Inger Karlsson:
Voice source dynamics for female speakers.
- Takuya Koizumi, Shuji Taniguchi:
A novel model of pathological vocal cords and its application to the diagnosis of vocal cord polyp.
- Hirohisa Iijima, Nobuhiro Miki, Nobuo Nagai:
Glottal flow analysis based on a finite element simulation of a two-dimensional unsteady viscous fluid.
- Hideki Kasuya, Yuji Ando, Jinlin Lu, Osamu Komuro:
A voice source model for synthesizing speech with various voice quality variations.
- Ailbhe Ní Chasaide, Christer Gobl:
Linguistic and paralinguistic variation in the voice source.
Speech Coding and Transmission
- Paavo Alku:
Glottal-LPC based coding of telephone band vowels with simple all-pole excitation.
- Suat Yeldener, Ahmet M. Kondoz, Barry G. Evans:
Sine wave excited linear predictive coding of speech.
- Toshiki Miyano, Kazunori Ozawa:
Improvement on 8 kb/s CELP using learned codebook: LCELP.
- Samir Saoudi, Jean-Marc Boucher, Alain Le Guyader:
Optimal scalar quantization of the LSP and the LAR for speech coding.
- Shinya Takahashi, Kunio Nakajima:
4.8 kbps speech coding using frame synchronous time domain compression (FS-TDC).
- Hirohisa Tasaki, Kunio Nakajima:
Time-domain flexible matrix quantization for very-low-rate speech coding.
- Tomohiko Taniguchi, Mark Johnson, Yasuji Ohta:
Multi-vector pitch-orthogonal LPC: quality speech with low complexity at rates between 4 and 8 kbps.
- Yair Shoham, Erik Ordentlich:
Low-delay code-excited linear-predictive coding of wideband speech at 32 kbps.
- Yoshihiro Unno, Makio Nakamura, Toshifumi Sato, Toshiki Miyano, Kazunori Ozawa:
11.2 kb/s LCELP speech codec for digital cellular radio.
- Tomoyuki Ohya, Hirohito Suda, Toshio Miki, Shinji Uebayashi, Takehiro Moriya:
Revised TC-WVQ speech coder for mobile communication system.
Extraction and Processing of Voice Individuality
- Hideki Noda, Masuzo Yanagida:
Extraction of phoneme-dependent individuality using HMM-based segmentation for text-independent speaker recognition.
- J. P. Eatock, J. S. D. Mason:
Automatically focusing on good discriminating speech segments in speaker recognition.
- Tomoko Matsui, Sadaoki Furui:
Text-independent speaker recognition using vocal tract and pitch information.
- Aaron E. Rosenberg, Chin-Hui Lee, Frank K. Soong, Maureen A. McGee:
Experiments in automatic talker verification using sub-word unit hidden Markov models.
- Myoung-Wan Koo, Chong Kwan Un, Hwang Soo Lee, Jun Mo Koo, H. R. Kim:
A comparative study of speaker adaptation methods for HMM-based speech recognition.
- Hiroaki Hattori, Satoshi Nakamura, Kiyohiro Shikano, Shigeki Sagayama:
Speaker weighted training of HMM using multiple reference speakers.
- Francis Kubala, Richard M. Schwartz:
Improved speaker adaptation using multiple reference speakers.
- Masanobu Abe, Shigeki Sagayama:
Statistical study on voice individuality conversion across different languages.
- Hiroshi Matsumoto, Hirowo Inoue:
A minimum distortion spectral mapping applied to voice quality conversion.
Voice Source Characteristics and Synthesis
- Anna M. Barney, Christine H. Shadle, David W. Thomas:
Airflow measurement in a dynamic mechanical model of the vocal folds.
- Jo Estill, Noriko Kobayashi, Kiyoshi Honda, Yuki Kakita:
A study on respiratory and glottal controls in six western singing qualities: airflow and intensity measurement of professional singing.
- Satoshi Imaizumi, Hiroshi Imagawa, Shigeru Kiritani:
A model of dynamic characteristics of the voice source and formant trajectories.
- Takayuki Nakajima, Hiroshi Ohmura:
Pole-zero structure based on two-source vocal tract model, PSE inspection of continuous speech vowel part.
- Gang Wang, Nobuhiro Miki, Nobuo Nagai:
Evaluation of speech synthesis using an ARMA estimation and excitation sources.
- Kazuhiko Iwata, Yukio Mitome, Jun Kametani, Minoru Akamatsu, Seimitsu Tomotake, Kazunori Ozawa, Takao Watanabe:
A rule-based speech synthesizer using pitch controlled residual wave excitation method.
- Kenzo Itoh, Hideyuki Mizuno, Tetsuya Nomura, Hirokazu Sato:
Phoneme segment concatenation and excitation control based on spectral distortion criterion for speech synthesis.
- Stephen D. Pearson, Hector R. Javkin, Kenji Matsui, Takahiro Kamai:
Text-to-speech synthesis using a natural voice source.
- Paavo Alku, Erkki Vilkman, Unto K. Laine:
A comparison of egg and a new automatic inverse filtering method in phonation change from breathy to normal.
Speech Recognition and Enhancement
- Ki Chul Kim, Hyunsoo Yoon, Jung Wan Cho:
Enhanced parametric representation using binarized spectrum.
- Kiyoshi Asai, Shigeru Chiba:
Voiced-unvoiced classification using weighted distance measures.
- Kei Miki:
Phoneme recognition using a hierarchical time spectrum pattern.
- Susumu Sato, Takeshi Fukabayashi:
Recognition of plosive using mixed features by fisher's linear discriminant.
- Akio Ando, Kazuhiko Ozeki:
Clustering algorithms to minimize recognition error function and their applications to the vowel template learninig.
- Changfu Wang, Hiroya Fujisaki, Keikichi Hirose:
Chinese four tone recognition based on the model for process of generating F0 contours of sentences.
- Nam Soo Kim, Chong Kwan Un:
Generalized training of hidden Markov model parameters for speech recognition.
- Tatsuya Kawahara, Toru Ogawa, Shigeyoshi Kitazawa, Shuji Doshita:
Phoneme recognition by combining Bayesian linear discriminations of selected pairs of classes.
- S. Atkins, P. E. Kenne, D. Landy, S. Nulsen, Mary O'Kane:
WAL - a speech recognition programming language.
- Mario Rossi:
Automatic segmentation: why and what segments?
- Shozo Makino, Akinori Ito, Mitsuru Endo, Ken'iti Kido:
A Japanese text dictation system based on phoneme recognition using a modified LVQ2 method.
- Shinobu Mizuta, Kunio Nakajima:
An optimal discriminative training method for continuous mixture density HMMs.
- S. Datta, M. Al-Zabibi:
Discrimination of words in a large vocabulary speech recognition system.
- J. M. Koo, Chong Kwan Un, Hwang Soo Lee, H. R. Kim, Myoung-Wan Koo:
A recognition time reduction algorithm for large-vocabulary speech recognition.
- Hyung Soon Kim, Chong Kwan Un:
Speech recognition method based on the dual processing nature of speech perception.
- Koichi Shinoda, Ken-ichi Iso, Takao Watanabe:
Speaker adaptation for demi-syllable based speech recognition using continuous HMM.
- Toby E. Skinner:
Speech signal processing on a neurocomputer.
- Shigeru Ono:
Syllable structure parsing for continuous speech recognition.
- Hiroyuki Tsuboi, Hiroshi Kanazawa, Yoichi Takebayashi:
An accelerator for high-speed spoken word-spotting and noise immunity learning system.
- Zainul Abidin Md. Sharrif, Masuri Othman, Mohammad Ibrahim A. K. B. Maiden:
Recognition of standard malaysian language pronunciation.
- M. Djoudi, Jean Paul Haton:
The SAPHA acoustic-phonetic decoder system for standard Arabic.
- Markus Bodden:
A concept for a cocktail-party-processor.
- Tsuyoshi Usagawa, Yuji Morita, Masanao Ebata:
Remote control system using speech-reduction of known noise.
- Yumi Takizawa, Masahiro Hamada:
Lombard speech recognition by formant-frequency-shifted LPC cepstrum.
- Hiroshi Matsumoto, Hirokazu Mitsui:
A robust distance measure based on group delay difference weighted by power spectra.
- B. Yegnanarayana, Hema A. Murthy, V. R. Ramachandran:
Speech enhancement using group delay functions.
- Hong Wang, Fumitada Itakura:
Recovery of reverberated speech using multi-microphone sub-band envelope estimation.
- Alain Marchal, Marie-Hélène Casanova, P. Gavarry, M. Avon:
DISPE: a divers' speech data-base.
Synthesis of Spoken Language
- Rolf Carlson, Björn Granström, Sheri Hunnicutt:
Lexical components in rule-based speech systems.
- Ken Ceder, Bertil Lyberg:
The integration of linguistic levels in a text-to-speech conversion system.
- Tohru Shimizu, Norio Higuchi, Hisashi Kawai, Seiichi Yamamoto:
The linguistic processing module for Japanese text-to-speech system.
- Yukiko Yamaguchi, Tatsuro Matsumoto:
A neural network approach to multi-language text-to-speech system.
- Hiroya Fujisaki, Keikichi Hirose, Yasuharu Asano:
Proposal and evaluation of a new type of terminal analog speech synthesizer.
- Bathsheba J. Malsheen, Mariscela Amador-Hernandez:
The interrelationship of intelligibility and naturalness in text-to-speech.
- Tomohisa Hirokawa, Kazuo Hakoda:
Segment selection and pitch modification for high quality speech synthesis using waveform segments.
- Kazuya Takeda, Katsuo Abe, Yoshinori Sagisaka:
On the unit search criteria and algorithms for speech synthesis using non-uniform units.
- Katsuhiko Shirai, Y. Sato, Kazuo Hashimoto:
Speech synthesis using superposition of sinusoidal waves generated by synchronized oscillators.
- David Rainton, S. J. Young:
Time-frequency spectral analysis of speech.
- Bert Van Coile:
Inductive learning of grapheme-to-phoneme rules.
- Yoichi Yamashita, Hiroyuki Fujiwara, Yasuo Nomura, Nobuyoshi Kaiki, Riichiro Mizoguchi:
A support environment based on rule interpreter for synthesis by rule.
- Jung-Chul Lee, Yong-Ju Lee, Hee-il Han, Eung-Bae Kim, Chang-Joo Kim, Kyung-Tae Kim:
Speech synthesis using demisyllables for Korean: a preliminary system.
- Seung-Kwon Ahn, Koeng-Mo Sung:
The rules in a Korean text-to-speech system.
- Chi-Shi Liu, Wern-Jun Wang, Shiow-Min Yu, Hsiao-Chuan Wang:
Mandarin speech synthesis by the unit of coarticulatory demi-syllable.
- Ryunen Teranishi:
A study on various prosody styles in Japanese speech synthesizable with the text-to-speech system.
- Hiroki Kamanaka, Takashi Yazu, Keiichi Chihara, Makoto Morito:
Japanese text-to-speech conversion system.
- Yasushi Ishikawa, Kunio Nakajima:
Neural network based concatenation method of synthesis units for synthesis by rule.
- Norio Higuchi, Hisashi Kawai, Tohru Shimizu, Seiichi Yamamoto:
Improvement of the synthetic speech quality of the formant-type speech synthesizer and its subjective evaluation.
- Thierry Galas, Xavier Rodet:
A parametric model of speech signals: application to high quality speech synthesis by spectral and prosodic modifications.
- Tomoki Hamagami, Shinichiro Hashimoto:
The improved source model for high-quality synthetic speech sound.
- Kazuo Hakoda, Shin'ya Nakajima, Tomohisa Hirokawa, Hideyuki Mizuno:
A new Japanese text-to-speech synthesizer based on COC synthesis method.
- G. M. Asher, K. M. Curtis, J. R. Andrews, J. Burniston:
A parallel multialgorithmic approach for an accurate and fast English text to speech transcriber.
- K. M. Curtis, G. M. Asher, S. E. Pack, J. R. Andrews:
A highly programmable formant speech synthesiser utilising parallel processors.
- Kris Maeda, Yasuki Yamashita, Yoichi Takebayashi:
Enhancement of human-computer interaction through the synthesis of nonverbal expressions.
- W. Nick Campbell, Stephen D. Isard, Alex I. C. Monaghan, J. Verhoeven:
Duration, pitch and diphones in the CSTR TTS system.
- Sin-Horng Chen, Su-Min Lee, Saga Chang:
A Chinese fundamental frequency synthesizer based on a statistical model.
- Cinzia Avesani:
A contribution to the synthesis of Italian intonation.
- Kazuhiko Iwata, Yukio Mitome, Takao Watanabe:
Pause rule for Japanese text-to-speech conversion using pause insertion probability.
- Hiroya Fujisaki, Keikichi Hirose, Pierre Halle, Haitao Lei:
Analysis and modeling of tonal features in polysyllabic words and sentences of the standard Chinese.
- Akira Yamamura, Hiroharu Kunizawa, Noboru Ueji, Hiroshi Itoyama, Osamu Kakusho:
Voice response unit embedded in factory automation systems.
- Klaus Wothke:
Tetos - a text-to-speech system for German.
- Michel Divay:
A written text processing expert system for text to phoneme conversion.
- Mikio Yamaguchi:
Trial production of a module for speech synthesis by rule.
- Katsuhiko Shirai, Naoki Hosaka, E. Kitagawa, T. Endo:
Speaker adaptable phoneme recognition selecting reliable acoustic features based on mutual information.
- Claude Montacié, Marie-José Caraty, Xavier Rodet:
Experiments in the use of an automatic learning system for acoustic-phonetic decoding.
- Shigeki Sagayama, Shigeru Honrna:
Estimation of unknown context using a phoneme environment clustering algorithm.
- Yves Laprie, Jean Paul Haton, Jean-Marie Pierrel:
Phonetic triplets in knowledge based approach of acoustic-phonetic decoding.
- Y. Ariki, Andrew M. Sutherland, Mervyn A. Jack:
Optimisation of English phoneme recognition based on HMM.
- Horacio Franco, António Serralheiro:
A new discriminative training algorithm for hidden Markov models.
- Yoshimitsu Hirata, Seiichi Nakagawa:
Speaker adaptation of continuous parameter HMM.
- Tatsuya Hirahara, Hitoshi Iwamida:
Auditory spectrograms in HMM phoneme recognition.
Recent Progress in Speech Perception Research
- Sieb G. Nooteboom, P. Scharpff, Vincent J. van Heuven:
Effects of several pausing strategies on the recognizability of words in synthetic speech.
- Yoshinori Kitahara, Yoh'ichi Tohkura:
The role of temporal structure of speech in word perception and spoken language understanding.
- Judith C. Goodman, Howard C. Nusbaum, Lisa Lee, Kevin Broihier:
The effects of syntactic and discourse variables on the segmental intelligibility of speech.
- Shigeaki Amano:
Lexical and coarticulatory effects on phoneme monitoring before and after a word identification point in spoken Japanese words.
- David B. Pisoni, Ellen E. Garber:
Lexical memory in visual and auditory modalities: the case for a common mental lexicon.
- John J. Ohala, Elizabeth Shriberg:
Hypercorrection in speech perception.
- Howard C. Nusbaum:
The role of learning and attention in speech perception.
- Dominic W. Massaro, Michael M. Cohen:
The joint influence of stimulus information and context in speech perception.
- Hiroya Fujisaki, Keikichi Hirose, Sumio Ohno, Nobuaki Minematsu:
Influence of context and knowledge on the perception of continuous speech.
Speech Production, Prosody and Analysis
- Arne Kjell Foldvik, O. Husby, Jorn Kvaerness, I. C. Nordli, Peter A. Rinck:
MRI (magnetic resonance imaging) film of articulatory movements.
- Masafumi Matsumura, Atsushi Sugiura:
Modeling of 3-dimensional vocal tract shapes obtained by magnetic resonance imaging for speech synthesis.
- Tokihiko Kaburagi, Masaaki Honda:
Ultrasonic measurement of tongue motion.
- Kunitoshi Motoki, Nobuhiro Miki, Nobuo Nagai:
Measurement of sound wave characteristics in the vocal tract.
- Hisayoshi Suzuki, Takayoshi Nakai, Jianwu Dang, Chengxiang Lu:
Speech production model involving subglottal structure and oral-nasal coupling through closed velum.
- Yorinobu Sonoda, Keisuke Mori, Tetsuaki Kuriyama:
Articulatory characteristics of lip shape during the production of Japanese.
- Naoki Kusakawa, Kiyoshi Honda, Yuki Kakita:
Sequential control model of speech articulation in producing word utterance.
- Zyun'ici B. Simada, Satoshi Horiguchi, Seiji Niimi, Hajime Hirose:
Sternohyoid muscle activity and pitch control at the onset of utterances.
- Junichi Azuma, Yoshimasa Tsukuma:
Prosodic features marking the major syntactic boundary of Japanese: a study on syntactically ambiguous sentences of the kinki dialect.
- H. D. Wang, Gérard Bailly, D. Tuffelli:
Automatic segmentation and alignment of continuous speech based on temporal decomposition model.
- Hee-Il Hahn, Minsoo Hahn:
Voiced/unvoiced/silence classification of spoken Korean.
- E. Angderi, M. Barsotti, L. Mazzei, L. Vttrano, R. Volpentesta:
Vocal pauses in teaching: statistical analysis and applications.
- Shubha Kadambe, Gloria Faye Boudreaux-Bartels:
A pitch detector based on event detection using the dyadic wavelet tranform.
- Hiroya Fujisaki, Keikichi Hirose, Shigenobu Seto:
Proposal and evaluation of a new scheme for reliable pitch extraction of speech.
- Masahide Sugiyama:
Spectral interpolation using distortion geodesic lines.
- Hirofumi Yogo, Naoki Inagaki:
Adaptive speech processing using an accelerated stochastic approximation method.
The Role of Prosody in Production and Perception of Spoken Language
- Fikret S. Gürgen, Shigeki Sagayama, Sadaoki Furui:
Line spectrum pair frequency - based distance measures for speech recognition.
- Hiroshi Shimodaira, Yoshio Horiuchi, Masayuki Kimura:
Speaker independent isolated word recognition using local and global structural features.
- Jorge A. Gurlekian, Horacio Franco, Miguel Santagada:
Speaker independent recognition of isolated Spanish digits.
- Nobuo Sugi, Jun'ichi Iwasaki, Hiroshi Matsu'ura, Tsuneo Nitta, Akira Fukumine, Akira Nakayama:
Speaker independent word recognition system based on the structured transition network of phonetic segments.
- Akihiro Imamura, Yoshitake Suzuki:
Speaker-independent word spotting and a transputer-based implementation.
- Jin Yul Kim, Yun-Seok Cho, Soon Young Yoon, Hwang Soo Lee, Chong Kwan Un:
An efficient viterbi scoring architecture for HMM-based isolated word recognition systems.
- Tatsuo Matsuoka:
Word spotting using context-dependent phoneme-based HMMs.
- V. Vittorelli, Gilles Adda, Roberto Billi, Lou Boves, Mervyn A. Jack, E. Vivalda:
POLYGLOT: multilingual speech recognition and synthesis.
- Satoshi Takahashi, Shoichi Matsunaga, Shigeki Sagayama:
Isolated word recognition using pitch pattern information.
Perception of Spoken Language
- Makio Kashino:
Distribution of perceptual cues for Japanese intervocalic stop consonants.
- W. Datscheweit:
Frication noise and formant-onset frequency as independent cues for the perception of /f/, /s/ and /// in vowel-fricative-vowel stimuli.
- Minoru Tsuzaki, Jorge A. Gurlekian:
Effects of different standards on the within-category discrimination of synthesized /ABA/ sequences: comparison between Japanese and Spanish.
- Masato Akagi:
Contextual effect models and psycho acoustic evidence for the models.
- Sumi Shigeno:
Vowel-contingent anchoring effects on the perception of stop consonants.
- Dominic W. Massaro:
Process and connectionist models of speech perception.
- Anne Cutler, Dennis Norris, Brit van Ooyen:
Vowels as phoneme detection targets.
- Noriko Uosaki, Morio Kohno:
Perception of rhythm: a comparison between americans and Japanese.
- Sotaro Sekimoto:
Perceptual frequency normalization of frequency compressed or expanded voiceless consonants.
Perception, Impairments/Aids, Phonetics in Language Teaching and Speech Coding
- Akiko Hayashi, Satoshi Imaizumi, Takehiko Harada, Hideaki Seki, Hiroshi Hosoi:
Effects of temporal factors on the speech perception of the hearing impaired.
- Shinobu Masaki, Itaru F. Tatsumi, Sumiko Sasanuma:
Analysis of temporal coordination between articulatory movements and pitch control in the realization of Japanese word accent by a patient with apraxia of speech.
- Brian C. J. Moore, Jeannette Seloover Johnson, Vincent Pluvinage, Teresa M. Clark:
Multiband dynamic range compression sound processing for hearing impaired patients: effect on intelligibility of speech in background noise.
- Takao Mizutani, Kiyoshi Hashimoto, Masahiko Wakumoto, Ken-ich Michi, Hareo Hamada, Tanetoshi Miura:
New graphical expression of the high-speed palatographic data in study of the articulatory behaviors of the tongue.
- Makoto Kariyasu, Kukiko Maruyama:
Aging in the rate and regularity of maximum syllable repetition under bite-block.
- Minje Zhi, Yong-Ju Lee:
Vowel quantity contrast in Korean: production and perception.
- Jan-Olof Svantesson:
Phonetic correlates of stress in mongolian.
- Ray Iwata, Hajime Hirose, Seiji Niimi, Masayuki Sawashima, Satoshi Horiguchi:
Syllable final stops LN east asian languages: southern Chinese, Thai and Korean.
- Seiji Niimi, Qun Yan, Satoshi Horiguchi, Hajime Hirose:
An electromyographic study on laryngeal adjustment for production of the light tone in Mandarin Chinese.
- Jingxu Cui, Shuichi Itahashi:
A comparison of the articulation of the Chinese /i, l, l/ by Chinese and Japanese speakers.
- Hirotake Nakashima, Masao Yamaguchi:
The durations of Japanese long vowels and geminated consonants uttered by indonesian.
- Izumi Saita:
On phrasing of Japanese language learners.
- PROTS (pronunciation training system) - Kawai Musical Instruments.
- Yair Shoham:
Constrained-stochastic excitation coding of speech at 4.8 kb/s.
- Fumie Hazu, Akihiko Sugiyama, Masahiro Iwadare, Takao Nishitani:
Adaptive transform coding with an adaptive block size using a modified DCT.
- Takehiro Moriya:
Medium-delay 8 kbit/s speech coder based on conditional pitch prediction.
- Sung Ro Lee, Hwang Soo Lee, Chong Kwan Un:
A low rate VQ speech coding algorithm with variable transmission frame length.
Neural Networks for Speech Processing I, II
- Ken-ichi Iso, Takao Watanabe:
Speech recognition using demi-syllable neural prediction model.
- Frédéric Bimbot, Gérard Chollet, Jean-Pierre Tubach:
Phonetic features extraction using time-delay neural networks.
- Masami Nakamura, Shinichi Tamura:
Vowel recognition by phoneme filter neural networks.
- Kari Torkkola, Mikko Kokkonen:
A comparison of two methods to transcribe speech into phonemes: a rule-based method vs. back-propagation.
- Jun-ichi Takami, Shigeki Sagayama:
Phoneme recognition by pairwise discriminant TDNNs.
- Yasuyuki Masai, Hiroshi Matsu'ura, Tsuneo Nitta:
Speaker independent speech recognition based on neural networks of each category with embedded eigenvectors.
- Kiyoaki Aikawa, Alexander H. Waibel:
Speech recognition using sub-phoneme recognition neural network.
- Li-Qun Xu, Tie-Cheng Yu, G. D. Tattersall:
Speech recognition based on the integration of FSVQ and neural network.
- Samir I. Sayegh:
Fast text-to-speech learning.
- Nelson Morgan, Chuck Wooters, Hervé Bourlard, Michael Cohen:
Continuous speech recognition on the resource management database using connectionist probability estimation.
- Eiichi Tsuboka, Yoshihiro Takada, Hisashi Wakita:
Neural predictive hidden Markov model.
- Yasuhiro Minami, Toshiyuki Hanazawa, Hitoshi Iwamida, Erik McDermott, Kiyohiro Shikano, Shigeru Katagiri, Masaona Kagawa:
On the robustness of HMM and ANN speech recognition algorithms.
- Hidefumi Sawai:
The TDNN-LR large-vocabulary and continuous speech recognition system.
- Rémy Bulot, Henri Meloni, Pascal Nocera:
Rule-driven neural networks for acoustic-phonetic decoding.
- Franck Poirier:
Knowledge-based segmentation and feature maps for speech recognition.
- Mark A. Fanty, Ronald A. Cole:
Speaker-independent English alphabet recognition: experiments with the e-set.
- Pinaki Poddar, P. V. S. Rao:
Neural network based segmentation of continuous speech.
- Tomio Takara, Motonori Tamaki:
A normalization of coarticulation of connected vowels using neural network.
- Tomio Watanabe, Masaki Kohda:
Lip-reading of Japanese vowels using neural networks.
- H. Lucke, Frank Fallside:
Application of the compositional representation to lexical access using neural networks.
- Abdul Mobin, S. S. Agrawal, Anil Kumar, K. D. Pavate:
A voice input-output system using isolated words.
- Tatiana Slama-Cazacu:
A psycholinguistic model of first and second language learning.
Continuous Speech Recognition
- Yunxin Zhao, Hisashi Wakita:
Experiments with a speaker-independent continuous speech recognition system on the timit database.
- Walter Weigel:
Continuous speech recognition with vowel-context-independent hidden-Markov-models for demisyllables.
- Satoru Hayamizu, Kai-Fu Lee, Hsiao-Wuen Hon:
Description of acoustic variations by tree-based phone modeling.
- Frank K. Soong, Eng-Fong Huang:
A tree-trellis based fast search for finding the n best sentence hypotheses in continuous speech recognition.
- Fabio Gabrieli, A. Dimundo, A. Rizzi, G. Colangelit, A. Stagni:
Modeling vocabularies for a connected speech recognizer.
- Takeshi Kawabata, Toshiyuki Hanazawa, Katsunobu Itou, Kiyohiro Shikano:
Japanese phonetic typewriter using HMM phone units and syllable trigrams.
- Minoru Shigenaga, Yoshihiro Sekiguchi, Toshihiko Hanagata, Takehiro Yamaguchi, Ryouta Masuda:
A large vocabulary continuous speech recognition system with high prediction capability.
- Yutaka Kobayashi, Yasuhisa Niimi:
Evaluation of a speech understanding system - suskit-2.
- Patti Price, Victor Abrash, Douglas E. Appelt, John Bear, Jared Bernstein, Bridget Bly, John Butzberger, Michael Cohen, Eric Jackson, Robert C. Moore, Douglas B. Moran, Hy Murveit, Mitchel Weintraub:
Spoken language system integration and development.
Modeling of First and Second Language Acquisition
Application of Speech Recognition / Synthesis Technologies
- D. A. Berkley, James L. Flanagan:
Integration of speech recognition, text-to-speech synthesis, and talker verification into a hands-free audio/image teleconferencing system (humanet).
- G. Velius, Candace A. Kamm, Mary Jo Altom, T. C. Feustel, Marian J. Macchi, Murray F. Spiegel:
Bellcore efforts in applying speech technology to telephone network services.
- Fumihiro Yato, Kazuki Katagisi, Norio Higuchi:
Extension number guidance system.
- Hirokazu Sato:
Japanese text-to-speech equipment: current applications and trends.
- Mariscela Amador-Hernandez, Bathsheba J. Malsheen:
The synthesis of dialectal variation in English and Spanish.
- Hiroyoshi Saito, Motoshi Kurihara, Ken-ichiro Kobayashi, Yoshiyuki Hara, Naritoshi Saito:
A Japanese text-to-speech system for electronic mail.
- Tsuneo Nitta, Nobuo Sugi:
Issues concerning voice input applications.
- Toshiaki Tsuboi, Noboru Sugamura:
A prototype for a speech-to-text transcription system.
- Masahiro Hamada, Yumi Takizawa, Takeshi Norimatsu:
A noise robust speech recognition system.
- A. Corazzat, Renato de Mori, Roberto Gretter, Giorgio Satta:
Computation of probabilities for island-driven parsers.
- Keh-Yih Su, Tung-Hui Chiang, Yi-Chung Lin:
A unified probabilistic score function for integrating speech and language information in spoken language processing.
- Kenji Kita, Toshiyuki Takezawa, Junko Hosaka, Terumasa Ehara, Tsuyoshi Morimoto:
Continuous speech recognition using two-level LR parsing.
- Hiroaki Saito:
Gap-filling LR parsing for noisy spoken input: towards interactive speech recognition.
- S. Bornerand, Francoise Neel, Gérard Sabah:
Semantic weights derived from syntax-directed understanding in DTW-based spoken language processing.
- Hiroaki Kitano, Tetsuya Higuchi, Masaru Tomita:
Massively parallel spoken language processing using a parallel associative processor IXM2.
- Tsuyoshi Morimoto, Kiyohiro Shikano, Hitoshi Iida, Akira Kurematsu:
Integration of speech recognition and language processing in spoken language translation system (SL-TRANS).
- Toshiya Sakano, Tsuyoshi Morimoto:
Design principle of language model for speech recognition.
- Shoichi Matsunaga, Shigeki Sagayama:
Sentence speech recognition using semantic dependency analysis.
Phonetics and Phonology
- Leigh Lisker:
Distinctive, redundant, predictable, neotssary, sufficffint accounting for English /bdg/-/ptk/.
- Rob Kassel, Victor W. Zue:
An information theoretic approach to the study of phoneme collocational constraints.
- Bruce L. Derwing, Terrance M. Nearey:
Real-time effects of some intrasyllabic collocational constraints in English.
- Paul Dalsgaard, William J. Barry:
Acoustic-phonetic features in the framework of neural-network multi-lingual label alignment.
- James Hieronymus:
Preliminary study of vowel coarticulation in british English.
- Caroline B. Huang:
Effects of context, stress, and speech style on american vowels.
- M. Djoudi, H. Aouizerat, Jean Paul Haton:
Phonetic study and recognition of standard Arabic emphatic consonants.
- Daniel Recasens, Edda Farnetani:
Articulatory and acoustic properties of different allophones of /l/ in american English, catalan and Italian.
- Hiroshi Suzuki, Ghen Ohyama, Shigeru Kiritani:
In search of a method to improve the prosodic features of English spoken by Japanese.
Assessment / Human Factors, Database and Neural Networks
- Zinny S. Bond, Thomas J. Moore:
A note on loud and lombard speech.
- Ute Jekosch:
A weighted intelligibility measure for speech assessment.
- Shinji Hayashi:
Improvements in binaural articulation score by simulated localization using head-related transfer functions.
- Kim E. A. Silverman, Sara Basson, Suzi Levas:
Evaluating synthesiser performance: is segmental intelligibility enough?
- Fumio Maehara, Masamichi Nakagawa, Kunio Nobori, Toshiyuki Maeda, Tsutomu Mori, Makoto Fujimoto:
Media conversion into language and voice for intelligent communication.
- Rolf Carlson, Björn Granström, Lennart Nord:
Segmental intelligibility of synthetic and natural speech in real and nonsense words.
- Chorkin Chan, Ren-Hua Wang:
The HKU-USTC speech corpus.
- Torbjørn Svendsen, Knut Kvale:
Automatic alignment of phonemic labels with continuous speech.
- D. Tuffelli, H. D. Wang:
TELS: a speech time-expansion labelling system.
- Kazuhiro Arai, Yoichi Yamashita, Tadahiro Kitahashi, Riichiro Mizoguchi:
A speech labeling system based on knowledge processing.
- Hans G. Tillmann, Maximilian Hadersbeck, Hans Georg Piroth, Barbara Eisen:
Development and experimental use of phonwork a new phonetic workbench.
- Hiroyuki Chimoto, Hideaki Shinchi, Hideki Hashimoto, Shinya Amano:
A speech recognition research environment based on large-scale word and concept dictionaries.
- Benjamin Chigier, Judith Spitz:
Are laboratory databases appropriate for training and testing telephone speech recognizers?
- Sven W. Danielsen:
Standardisation of speech input assessment within the SAM esprit project.
- Hiroshi Irii, Kenzo Itoh, Nobuhiko Kitawaki:
Multilingual speech data base for evaluating quality of digitized speech.
- Lizhong Wu, Frank Fallside:
The optimal gain sequence for fastest learning in connectionist vector quantiser design.
- Tony Robinson, John Holdsworth, Roy Patterson, Frank Fallside:
A comparison of preprocessors for the cambridge recurrent error propagation network speech recognition system.
- Robert B. Allen, Candace A. Kamm, S. B. James:
A recurrent neural network for word identification from phoneme sequences.
- Lieven Depuydt, Jean-Pierre Martens, Luc Van Immerseel, Nico Weymaere:
Improved broad phonetic classification and segmentation with a neural network and a new auditory model.
- Kazuaki Obara, Hideyuki Takagi:
Formant extraction model by neural networks and auditory model based on signal processing theory.
- Noboru Kanedera, Tetsuo Funada:
/b, d, g/ recognition with elliptic discrimination neural units.
- Helen M. Meng, Victor W. Zue:
A comparative study of acoustic representations of speech for vowel classification using multi-layer perceptrons.
- Yong Duk Cho, Ki Chul Kim, Hyunsoo Yoon, Seung Ryoul Maeng, Jung Wan Cho:
Extended elman's recurrent neural network for syllable recognition.
- Hong C. Leung, James R. Glass, Michael S. Phillips, Victor W. Zue:
Detection and classification of phonemes using context-independent error back-propagation.
- Shigeru Chiba, Kiyoshi Asai:
A new method of consonant detection and classification using neural networks.
- Shigeyoshi Kitazawa, Masahiro Serizawa:
An artificial neural network for the burst point detection.
- Claude Lefebvre, Dariusz A. Zwierzynski:
The use of discriminant neural networks in the integration of acoustic cues for voicing into a continuous-word recognition system.
- Kouichi Yamaguchi, Kenji Sakamoto, Toshio Akabane, Yoshiji Fujimoto:
A neural network for speaker-independent isolated word recognition.
Speech I/O Assessment and Database I, II
- Shuichi Itahashi:
Recent speech database projects in Japan.
- Joon-Hyuk Choi, Kyung-Tae Kim:
Construction of a large Korean speech database and its management system in ETRI.
- Yoshinori Sagisaka, Kazuya Takeda, M. Abel, Shigeru Katagiri, T. Umeda, Hisao Kuwabara:
A large-scale Japanese speech database.
- Terumasa Ehara, Kentaro Ogura, Tsuyoshi Morimoto:
ATR dialogue database.
- Jean-Luc Gauvain, Lori Lamel, Maxine Eskenazi:
Design considerations and text selection for BREF, a large French read-speech corpus.
- Kazuyo Tanaka, Satoru Hayamizu, Kozo Ohta:
The ETL speech database for speech analysis and recognition research.
- Michal Soclof, Victor W. Zue:
Collection and analysis of spontaneous and read corpora for spoken language system development.
- Shozo Makino, Toshihiko Shirokaze, Ken'iti Kido:
A distributed speech database with an automatic acquisition system of speech information.
- J. Bruce Millar, Phillip Dermody, M. Harrington, Julie Vonwiller:
A national database of spoken language: concept, design, and implementation.
- Giuseppe Castagneri, Kyriaki Vagges:
The Italian national database for speech recognition.
- Louis C. W. Pols:
How useful are speech databases for rule synthesis development and assessment?
- William J. Hardcastle, Alain Marchal:
Eur-accor: a multi-lingual articulatory and acoustic database.
Speech Recognition in Noisy Environments
- B. H. Juang:
Recent developments in speech recognition under adverse conditions.
- Brian A. Hanson, Ted H. Applebaum:
Features for noise-robust speaker-independent word recognition.
- Alejandro Acero, Richard M. Stern:
Acoustical pre-processing for robust spoken language systems.
- John H. L. Hansen, Oscar N. Bria:
Lombard effect compensation for robust automatic speech recognition in noise.
- Tadashi Kitamura, Etsuro Hayahara, Yasuhiko Simazciki:
Speaker-independent word recogniton in noisy environments using dynamic and averaged spectral features based on a two-dimensional mel-cepstrum.
- A. Noll:
Problems of speech recognition in mobile environments.
- L. Fissore, Pietro Laface, M. Codogno, G. Venuti:
HMM modeling for voice-activated mobile-radio system.
- Yoshio Nakadai, Noboru Sugamura:
A speech recognition method for noise environments using dual inputs.
- Shuji Morii, Toshiyuki Morii, Masakatsu Hoshimi, Shoji Hiraoka, Taisuke Watanabe, Katsuyuki Niyada:
Noise robustness in speaker independent speech recognition.
- Kaoru Gyoutoku, Hidefumi Kobatake:
Maximum likelihood estimation of speech waveform under nonstationary noise environments.
Foreign Language Teaching
Continuous Speech Recognition and Speaker Recognition
- Yoshiharu Abe, Kunio Nakajima:
Vocabulary independent phrase recognition with a linear phonetic context model.
- Y. Ariki, Mervyn A. Jack:
Phoneme probability presentation of continuous speech.
- Haiyan Ye, Jean Caelen:
Duration constraints for the speech input interface in the MULTIWORKS project.
- Zhi-ping Hu, Satoshi Imai:
Chinese continuous speech recognition system using the state transition models both of phonemes and words.
- Jade Goldstein, Akio Amano, Hideki Murayama, Mariko Izawa, Akira Ichikawa:
A new training method for multi-phone speech units for use in a hidden Markov model speech recognition system.
- Yoshio Ueda, Seiichi Nakagawa:
Diction for phoneme/syllable/word-category and identification of language using HMM.
- Takashi Otsuki, Shozo Makino, Toshio Sone, Ken'iti Kido:
Performance evaluation in speech recognition system using transition probability between linguistic units.
- Isao Murase, Seiichi Nakagawa:
Sentence recognition method using word cooccurrence probability and its evaluation.
- Yanghai Lu, Beiqian Dai:
A knowledge-based understanding system for the Chinese spoken language.
- Akio Komatsu, Eiji Oohira, Akira Ichikawa:
Conversational speech understanding based on cooperative problem solving.
- Michio Okada:
A one-pass search algorithm for continuous speech recognition directed by context-free phrase structure grammar.
- Andrea Di Carlo, Rino Falcone:
A blackboard architecture for a word hypothesizer and a chart parser interaction in an ASR system.
- P. Mousel, Jean-Marie Pierrel, A. Roussanaly:
Heuristic search problems in a natural language task oriented spoken man-machine dialogue system.
- Hiroaki Kitano:
The making of a speech-to-speech translation system: some findings from the dmdialog project.
- Kyung-ho Loken-Kim, Yasuhiro Nara, Shinta Kimura:
Using high level knowledge sources as a means of recovering DLL-formed Japanese sentences distorted by ambient noise.
- Anders Baekgaard, Paul Dalsgaard:
Tools for designing dialogues in speech understanding interfaces.
- Osamu Takizawa, Masuzo Yanagida:
A method for expressing associative relations using fuzzy concepts -aiming at advanced speech recognition-.
- Jean-Pierre Tubach, Raymond Descout, Pierre Isabelle:
Bilingual speech interface for a bidirectional machine translation system.
- Yves Laprie:
Optimum spectral peak track interpretation in terms of formants.
- Spriet Thierry:
A speech understanding system.
- Seiichiro Hangai, Kazuhiro Miyauchi:
Speaker based on multipulse excitation and UPC vocal-tract model.
- I-Chang Jou, Su-Ling Lee, Min-Tau Lin, Chih-Yuan Tseng, Shih-Shien You, Yuh-Juain Tsay:
A neural network based speaker verification system.
- Hujun Yin, Tong Zhou:
Speaker recognition using static and dynamic CEPSTRAL feature by a learning neural network.
Dialogue Modeling and Processing
- Naotoshi Osaka:
Conversational turn-taking model using PETRI net.
- Tetsuya Yamamoto, Yoshikazu Ohta, Yoichi Yamashita, Riichiro Mizoguchi:
Dialog management system mascots in speech understanding system.
- Sharon L. Oviatt, Philip R. Cohen, Ann Podlozny:
Spoken language in interpreted telephone dialogues.
- Tsuyoshi Morimoto, Toshiyuki Takezawa:
Linguistic knowledge for spoken dialogue processing.
- Harald Höge:
SPICOS II - a speech understanding dialogue system.
- Victor W. Zue, James R. Glass, Dave Goddeau, David Goodine, Hong C. Leung, Michael K. McCandless, Michael S. Phillips, Joseph Polifroni, Stephanie Seneff, Dave Whitney:
Recent progress on the MIT VOYAGER spoken language system.
Last update Sat May 25 18:31:38 2013
CET by the DBLP Team — Data released under the ODC-BY 1.0 license — See also our legal information page
- Gunnar Fant:
The speech code. segmental and prosodic features.
- David B. Pisoni:
Effects of talker variability on speech perception: implications for current research and theory.
- Fumitada Itakura:
Early developments of LPC speech coding techniques.