SPEECH ENHANCEMENT BY FORMANT SHARPENING IN THE CEPSTRAL DOMAIN
|
|
- Alexia Patrick
- 5 years ago
- Views:
Transcription
1 SPEECH ENHANCEMENT BY FORMANT SHARPENING IN THE CEPSTRAL DOMAIN David Cole and Sridha Sridharan Speech Research Laboratory, School of Electrical and Electronic Systems Engineering, Queensland University of Technology ABSTRACT: This paper presents a method for enhancing speech signals in the root cepstral domain, typically in conjunction with the cepstral subtraction technique. The effect of the processing is similar to the spectral sharpening method performed in the time domain, but with a much lower computational requirement when combined with cepstral subtraction and with better performance. The portions of the signal corresponding to frequency formant peaks are further amplified, while formant valleys are attenuated, with the aims of further reducing noise in those regions and improving speech quality. The cepstral procedure maintains the spectral tilt of all speech segments, unlike the time domain approach which only attempts to maintain the long-term average spectral tilt. The scheme was devised as part of an enhancement system designed for forensic speech enhancement, where speech intelligibility is an important consideration. INTRODUCTION Speech enhancement is a common requirement in speech processing systems, either as preprocessing for procedures such as speech or speaker recognition, or for improving the quality or intelligibility of speech for human audition. Thus speech enhancement has varying goals, depending on the use of the enhanced speech output and the procedures used as pre-processing for speech or speaker recognition might be quite different from those used for speech enhancement intended for improved human reception of the speech. The enhancement technique described in this paper was designed as an addition to the noise reduction system described by Fisher and Sridharan (1995), whose purpose was forensic speech enhancement. They used a combination of spectral subtraction and cepstral subtraction to produce enhancement results suitable for use in such an application. This requires that intelligibility not be impaired, and preferably improved. The procedure manipulates the cepstrum of the signal to provide an enhancement of the formant structure of the speech signal, with negligible computational cost when the signal cepstrum is already available. The procedure has similar aims to the time domain spectral sharpening technique, but has improved spectral tilt performance over that method. SPECTRAL SHARPENING The use of spectral sharpening for speech enhancement and noise reduction was suggested by Schaub and Straum (1991). The technique originally was used for adaptive post-filtering in speech coding schemes using linear prediction (Ramamoorthy and Jayant 1984). The basis is the linear predictive filter defined by A(z) = a K 1 2 m 1z + a 2z + + amz These coefficients are used for post-filtering of the decoded signal using the transfer function with filter parameters 0 < β < γ < 1. H (z) β = γ Accepted after full review page 244
2 The result of this post-filtering is a sharpening of the formant structure of the speech signal, with amplified formant peaks and attenuated formant valleys effectively due to the poles of the linear predictive resynthesis filter being shifted closer to the unit circle. As well as the enhancement application described by Schaub & Straub, this post-filtering technique has also been used to improve the quality of synthesized speech (Dines et al 2001). In the enhancement scheme of Schaub & Straub, high-pass filtering is applied to the speech signal before the spectral sharpening filter in an attempt to compensate for the general spectral tilt of the long-term speech spectrum. This is to avoid overemphasis of the spectral tilt characteristic. As will be shown here, this is not ideal, as the tilt of the speech spectrum in the short term often differs markedly from the long-term characteristic. A block diagram of the scheme is shown in Figure 1. LP analysis a 1 a m 1 ( z ) αz 1 β γ Figure 1. Spectral sharpening in the time domain using linear prediction. Figure 2 illustrates the effect of the time domain spectral sharpening scheme. This shows the smoothed (10 th order LPC) frequency spectrum of the vowel /a/ before (solid line) and after (dashed line) processing. The high pass pre-emphasis filter parameters were adjusted to maintain approximately the spectral tilt of the original spectrum for this segment. Figure 2. Smoothed spectrum of phoneme /a/ (a) before time domain processing (solid line) (b) after time domain processing (dashed line) Accepted after full review page 245
3 Figure 3 shows another segment of the same utterance, corresponding to the phoneme /s/. The original spectrum (solid line) shows the rising spectral tilt typical of this phoneme. After processing (with identical parameters to those for Figure 2), the smoothed spectrum of the processed segment shows the problem inherent in this scheme the pre-emphasis is compensating for an expected falling spectral tilt and this, combined with the rising spectral tilt of the post-filter, produces an overemphasis of the higher frequencies. Figure 3. Smoothed spectrum of phoneme /s/ (a) before time domain processing (solid line) (b) after time domain processing (dashed line) ROOT CEPSTRAL PROCESSING OF SPEECH SIGNALS The use of the root cepstrum in speech processing is generally associated with cepstral subtraction. The techniques of spectral subtraction and cepstral subtraction are both well known and widely used for noise reduction, where the speech signal s(n) is corrupted additively by noise z(n) to produce the observed noisy speech signal y(n): y (n) = s(n) + z(n) where n is the discrete time index. The discrete frequency domain equivalent is Y (k) = S(k) + Z(k) Spectral subtraction involves calculating an estimate Z ~ (k) of the noise spectral magnitude and subtracting this from the observed spectral magnitude Y (k). The resultant estimate of the clean signal spectral magnitude S ~ (k) is recombined with the noisy signal phase and inverse Fourier transformed to yield the estimated clean speech signal ~ s(n) : s(n) = F {( Y(k) Z ~ (k) ) e } j Y(k) ~ -1 Accepted after full review page 246
4 Generally, it is attempted to calculate the noise estimate during speech pauses, although a common approach is to use an auto-regressive average of the noisy speech signal as the noise estimate. The original cepstrum calculation used the inverse Fourier transform of the logarithm of the spectral magnitude. This is impractical in low noise conditions, where the undefined log(0) condition can arise. Thus, the root cepstrum is generally used for speech cepstral subtraction. The root cepstrum of the observed signal is: ŷ(n) -1 = F 1 p { Y(k) } with p typically 2 or 4. Cepstral subtraction follows the same form as spectral subtraction, using an accumulated cepstral noise estimate subtracted from the observed cepstrum. Despite its lack of a mathematical basis (since subtraction in the cepstral domain equates to deconvolution, not subtraction, in the time domain) cepstral subtraction is accepted as superior in performance to spectral subtraction for additive noise removal (Wu et al 1991), producing a less distorted output. Fisher and Sridharan (1994) concluded that spectral subtraction outperformed cepstral subtraction for low signal to noise ratios, and so devised a tandem system using both techniques which outperformed either used singly. The spectral sharpening procedure described below was developed to further improve the output quality of this system. SPECTRAL SHARPENING IN THE ROOT CEPSTRAL DOMAIN The use of the cepstrum in speech processing is not confined to speech enhancement. It has also been used widely in speech and speaker recognition fields because of its ability to separate vocal tract and excitation information. For vocal tract characterisation, the cepstrum is similar to linear prediction coefficients, although the information contained in the two sets of parameters is not identical. Linear prediction provides formant location and bandwidth defined by the roots of the LPC equation. In contrast, the cepstral coefficients can be used to build up a smoothed spectrum using the sinusoidal basis functions of the Fourier transform. Thus it is common for speech or speaker recognition systems to use either linear predictive coefficients or (low order) cepstral coefficients to parameterise vocal tract characteristics. The use of the root cepstrum for spectral sharpening utilises the encoding of the smoothed speech spectrum in the low order cepstral coefficients. By increasing the magnitude of selected cepstral coefficients, it is possible to increase the amplitude of the corresponding smoothed spectral envelope of the reconstructed signal. By considering the nature of the Fourier transformation, it is clear that the main contribution to the spectral tilt of a speech segment will be contained in the low order cepstral coefficients: n = 1 and possibly also 2. Thus to maintain the spectral tilt of a speech segment, these lowest order cepstral coefficients are unmodified (as well as the 0 th coefficient, which represents the energy level of the segment.) Expressing this scheme mathematically, we operate on the input cepstrum ŷ (n) to produce output cepstrum qˆ (n) as follows: qˆ(n) = ŷ(n) k n k n = b k n = 1 n < n < n l otherwise h Typically the boost factor b, which is greater than 1, is applied (for a frame size of 256 samples) for n=3 20 or thereabouts, which is adequate to represent the speech formant structure. Clearly, the procedure requires very little computational effort when the cepstrum has already been calculated for a cepstral subtraction procedure. Accepted after full review page 247
5 Figures 4 and 5 show the performance of the scheme for the same two speech segments used previously. The /a/ segment output is almost identical to that produced by the time domain scheme, and the /s/ segment output is quite clearly superior, maintaining the overall spectral tilt well while enhancing the formant structure. Figure 4. Smoothed spectrum of phoneme /a/ (a) before cepstral processing (solid line) (b) after cepstral processing (dashed line) Figure 5. Smoothed spectrum of phoneme /s/ (a) before cepstral processing (solid line) (b) after cepstral processing (dashed line) Accepted after full review page 248
6 OUTPUT QUALITY The maintenance of spectral tilt for individual segments, as shown in the above figures, results in an audible improvement in quality of the output of the cepstrally enhanced speech as compared with the time domain enhanced version. The need to use fixed pre-emphasis in the time domain scheme means that, in general, the spectral tilt of individual segments will be altered, adversely affecting the quality of the speech output. Usually, the result is a general increase in high frequency content, producing an unpleasant, harsh output. In contrast, the cepstral enhancement method maintains segmental spectral tilt, producing more natural and higher quality speech. These quality comparisons have not yet been quantified, but audio demonstrations that show these differences may be found at REFERENCES Fisher, A. & Sridharan, S. (1994) Speech Enhancement for Forensic Applications, International Conference on Speech Science and Technology, SST-94, Dines, J. & Sridharan, S. (2001) Trainable speech synthesis with hidden Markov models, Proceedings of International Conference on Acoustics, Speech and Signal Processing (ICASSP-2001), Ramamoorthy, V. & Jayant, N. (1984) Enhancement of ADPCM speech by adaptive post-filtering, Technical Report Technical Journal of AT&T, V63-I8, Schaub, A. & Straub, P. (1991). Spectral sharpening for speech enhancement / noise reduction, Proceedings of International Conference on Acoustics, Speech and Signal Processing (ICASSP-91), Wu, C.S., Nguyen, V.V., Sabrin, H., Kushner, W. & Damoulakis, J. (1991). Fast self-adapting broadband noise removal in the cepstral domain, Proceedings of International Conference on Acoustics, Speech and Signal Processing (ICASSP-91), Accepted after full review page 249
Speaker Identification by Comparison of Smart Methods. Abstract
Journal of mathematics and computer science 10 (2014), 61-71 Speaker Identification by Comparison of Smart Methods Ali Mahdavi Meimand Amin Asadi Majid Mohamadi Department of Electrical Department of Computer
More informationSpeech Synthesis in Noisy Environment by Enhancing Strength of Excitation and Formant Prominence
INTERSPEECH September,, San Francisco, USA Speech Synthesis in Noisy Environment by Enhancing Strength of Excitation and Formant Prominence Bidisha Sharma and S. R. Mahadeva Prasanna Department of Electronics
More informationHuman Emotion Recognition From Speech
RESEARCH ARTICLE OPEN ACCESS Human Emotion Recognition From Speech Miss. Aparna P. Wanare*, Prof. Shankar N. Dandare *(Department of Electronics & Telecommunication Engineering, Sant Gadge Baba Amravati
More informationRobust Speech Recognition using DNN-HMM Acoustic Model Combining Noise-aware training with Spectral Subtraction
INTERSPEECH 2015 Robust Speech Recognition using DNN-HMM Acoustic Model Combining Noise-aware training with Spectral Subtraction Akihiro Abe, Kazumasa Yamamoto, Seiichi Nakagawa Department of Computer
More informationDesign Of An Automatic Speaker Recognition System Using MFCC, Vector Quantization And LBG Algorithm
Design Of An Automatic Speaker Recognition System Using MFCC, Vector Quantization And LBG Algorithm Prof. Ch.Srinivasa Kumar Prof. and Head of department. Electronics and communication Nalanda Institute
More informationAUTOMATIC DETECTION OF PROLONGED FRICATIVE PHONEMES WITH THE HIDDEN MARKOV MODELS APPROACH 1. INTRODUCTION
JOURNAL OF MEDICAL INFORMATICS & TECHNOLOGIES Vol. 11/2007, ISSN 1642-6037 Marek WIŚNIEWSKI *, Wiesława KUNISZYK-JÓŹKOWIAK *, Elżbieta SMOŁKA *, Waldemar SUSZYŃSKI * HMM, recognition, speech, disorders
More informationAnalysis of Emotion Recognition System through Speech Signal Using KNN & GMM Classifier
IOSR Journal of Electronics and Communication Engineering (IOSR-JECE) e-issn: 2278-2834,p- ISSN: 2278-8735.Volume 10, Issue 2, Ver.1 (Mar - Apr.2015), PP 55-61 www.iosrjournals.org Analysis of Emotion
More informationNoise-Adaptive Perceptual Weighting in the AMR-WB Encoder for Increased Speech Loudness in Adverse Far-End Noise Conditions
26 24th European Signal Processing Conference (EUSIPCO) Noise-Adaptive Perceptual Weighting in the AMR-WB Encoder for Increased Speech Loudness in Adverse Far-End Noise Conditions Emma Jokinen Department
More informationSpeech Emotion Recognition Using Support Vector Machine
Speech Emotion Recognition Using Support Vector Machine Yixiong Pan, Peipei Shen and Liping Shen Department of Computer Technology Shanghai JiaoTong University, Shanghai, China panyixiong@sjtu.edu.cn,
More informationSpeech Segmentation Using Probabilistic Phonetic Feature Hierarchy and Support Vector Machines
Speech Segmentation Using Probabilistic Phonetic Feature Hierarchy and Support Vector Machines Amit Juneja and Carol Espy-Wilson Department of Electrical and Computer Engineering University of Maryland,
More informationWHEN THERE IS A mismatch between the acoustic
808 IEEE TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 14, NO. 3, MAY 2006 Optimization of Temporal Filters for Constructing Robust Features in Speech Recognition Jeih-Weih Hung, Member,
More informationPhonetic- and Speaker-Discriminant Features for Speaker Recognition. Research Project
Phonetic- and Speaker-Discriminant Features for Speaker Recognition by Lara Stoll Research Project Submitted to the Department of Electrical Engineering and Computer Sciences, University of California
More informationSpeaker recognition using universal background model on YOHO database
Aalborg University Master Thesis project Speaker recognition using universal background model on YOHO database Author: Alexandre Majetniak Supervisor: Zheng-Hua Tan May 31, 2011 The Faculties of Engineering,
More informationA Comparison of DHMM and DTW for Isolated Digits Recognition System of Arabic Language
A Comparison of DHMM and DTW for Isolated Digits Recognition System of Arabic Language Z.HACHKAR 1,3, A. FARCHI 2, B.MOUNIR 1, J. EL ABBADI 3 1 Ecole Supérieure de Technologie, Safi, Morocco. zhachkar2000@yahoo.fr.
More informationA comparison of spectral smoothing methods for segment concatenation based speech synthesis
D.T. Chappell, J.H.L. Hansen, "Spectral Smoothing for Speech Segment Concatenation, Speech Communication, Volume 36, Issues 3-4, March 2002, Pages 343-373. A comparison of spectral smoothing methods for
More informationThe NICT/ATR speech synthesis system for the Blizzard Challenge 2008
The NICT/ATR speech synthesis system for the Blizzard Challenge 2008 Ranniery Maia 1,2, Jinfu Ni 1,2, Shinsuke Sakai 1,2, Tomoki Toda 1,3, Keiichi Tokuda 1,4 Tohru Shimizu 1,2, Satoshi Nakamura 1,2 1 National
More informationClass-Discriminative Weighted Distortion Measure for VQ-Based Speaker Identification
Class-Discriminative Weighted Distortion Measure for VQ-Based Speaker Identification Tomi Kinnunen and Ismo Kärkkäinen University of Joensuu, Department of Computer Science, P.O. Box 111, 80101 JOENSUU,
More informationSpeaker Recognition. Speaker Diarization and Identification
Speaker Recognition Speaker Diarization and Identification A dissertation submitted to the University of Manchester for the degree of Master of Science in the Faculty of Engineering and Physical Sciences
More informationAutomatic segmentation of continuous speech using minimum phase group delay functions
Speech Communication 42 (24) 429 446 www.elsevier.com/locate/specom Automatic segmentation of continuous speech using minimum phase group delay functions V. Kamakshi Prasad, T. Nagarajan *, Hema A. Murthy
More informationVoice conversion through vector quantization
J. Acoust. Soc. Jpn.(E)11, 2 (1990) Voice conversion through vector quantization Masanobu Abe, Satoshi Nakamura, Kiyohiro Shikano, and Hisao Kuwabara A TR Interpreting Telephony Research Laboratories,
More informationQuarterly Progress and Status Report. VCV-sequencies in a preliminary text-to-speech system for female speech
Dept. for Speech, Music and Hearing Quarterly Progress and Status Report VCV-sequencies in a preliminary text-to-speech system for female speech Karlsson, I. and Neovius, L. journal: STL-QPSR volume: 35
More informationInternational Journal of Computational Intelligence and Informatics, Vol. 1 : No. 4, January - March 2012
Text-independent Mono and Cross-lingual Speaker Identification with the Constraint of Limited Data Nagaraja B G and H S Jayanna Department of Information Science and Engineering Siddaganga Institute of
More informationADVANCES IN DEEP NEURAL NETWORK APPROACHES TO SPEAKER RECOGNITION
ADVANCES IN DEEP NEURAL NETWORK APPROACHES TO SPEAKER RECOGNITION Mitchell McLaren 1, Yun Lei 1, Luciana Ferrer 2 1 Speech Technology and Research Laboratory, SRI International, California, USA 2 Departamento
More informationSegregation of Unvoiced Speech from Nonspeech Interference
Technical Report OSU-CISRC-8/7-TR63 Department of Computer Science and Engineering The Ohio State University Columbus, OH 4321-1277 FTP site: ftp.cse.ohio-state.edu Login: anonymous Directory: pub/tech-report/27
More informationLearning Methods in Multilingual Speech Recognition
Learning Methods in Multilingual Speech Recognition Hui Lin Department of Electrical Engineering University of Washington Seattle, WA 98125 linhui@u.washington.edu Li Deng, Jasha Droppo, Dong Yu, and Alex
More informationA study of speaker adaptation for DNN-based speech synthesis
A study of speaker adaptation for DNN-based speech synthesis Zhizheng Wu, Pawel Swietojanski, Christophe Veaux, Steve Renals, Simon King The Centre for Speech Technology Research (CSTR) University of Edinburgh,
More informationModeling function word errors in DNN-HMM based LVCSR systems
Modeling function word errors in DNN-HMM based LVCSR systems Melvin Jose Johnson Premkumar, Ankur Bapna and Sree Avinash Parchuri Department of Computer Science Department of Electrical Engineering Stanford
More informationLikelihood-Maximizing Beamforming for Robust Hands-Free Speech Recognition
MITSUBISHI ELECTRIC RESEARCH LABORATORIES http://www.merl.com Likelihood-Maximizing Beamforming for Robust Hands-Free Speech Recognition Seltzer, M.L.; Raj, B.; Stern, R.M. TR2004-088 December 2004 Abstract
More informationOn the Formation of Phoneme Categories in DNN Acoustic Models
On the Formation of Phoneme Categories in DNN Acoustic Models Tasha Nagamine Department of Electrical Engineering, Columbia University T. Nagamine Motivation Large performance gap between humans and state-
More informationBody-Conducted Speech Recognition and its Application to Speech Support System
Body-Conducted Speech Recognition and its Application to Speech Support System 4 Shunsuke Ishimitsu Hiroshima City University Japan 1. Introduction In recent years, speech recognition systems have been
More informationSpeech Recognition using Acoustic Landmarks and Binary Phonetic Feature Classifiers
Speech Recognition using Acoustic Landmarks and Binary Phonetic Feature Classifiers October 31, 2003 Amit Juneja Department of Electrical and Computer Engineering University of Maryland, College Park,
More informationUTD-CRSS Systems for 2012 NIST Speaker Recognition Evaluation
UTD-CRSS Systems for 2012 NIST Speaker Recognition Evaluation Taufiq Hasan Gang Liu Seyed Omid Sadjadi Navid Shokouhi The CRSS SRE Team John H.L. Hansen Keith W. Godin Abhinav Misra Ali Ziaei Hynek Bořil
More informationInternational Journal of Advanced Networking Applications (IJANA) ISSN No. :
International Journal of Advanced Networking Applications (IJANA) ISSN No. : 0975-0290 34 A Review on Dysarthric Speech Recognition Megha Rughani Department of Electronics and Communication, Marwadi Educational
More informationDigital Signal Processing: Speaker Recognition Final Report (Complete Version)
Digital Signal Processing: Speaker Recognition Final Report (Complete Version) Xinyu Zhou, Yuxin Wu, and Tiezheng Li Tsinghua University Contents 1 Introduction 1 2 Algorithms 2 2.1 VAD..................................................
More informationModeling function word errors in DNN-HMM based LVCSR systems
Modeling function word errors in DNN-HMM based LVCSR systems Melvin Jose Johnson Premkumar, Ankur Bapna and Sree Avinash Parchuri Department of Computer Science Department of Electrical Engineering Stanford
More informationEli Yamamoto, Satoshi Nakamura, Kiyohiro Shikano. Graduate School of Information Science, Nara Institute of Science & Technology
ISCA Archive SUBJECTIVE EVALUATION FOR HMM-BASED SPEECH-TO-LIP MOVEMENT SYNTHESIS Eli Yamamoto, Satoshi Nakamura, Kiyohiro Shikano Graduate School of Information Science, Nara Institute of Science & Technology
More informationVimala.C Project Fellow, Department of Computer Science Avinashilingam Institute for Home Science and Higher Education and Women Coimbatore, India
World of Computer Science and Information Technology Journal (WCSIT) ISSN: 2221-0741 Vol. 2, No. 1, 1-7, 2012 A Review on Challenges and Approaches Vimala.C Project Fellow, Department of Computer Science
More informationSpeech Recognition at ICSI: Broadcast News and beyond
Speech Recognition at ICSI: Broadcast News and beyond Dan Ellis International Computer Science Institute, Berkeley CA Outline 1 2 3 The DARPA Broadcast News task Aspects of ICSI
More informationModule 12. Machine Learning. Version 2 CSE IIT, Kharagpur
Module 12 Machine Learning 12.1 Instructional Objective The students should understand the concept of learning systems Students should learn about different aspects of a learning system Students should
More informationIEEE TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 17, NO. 3, MARCH
IEEE TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 17, NO. 3, MARCH 2009 423 Adaptive Multimodal Fusion by Uncertainty Compensation With Application to Audiovisual Speech Recognition George
More informationProceedings of Meetings on Acoustics
Proceedings of Meetings on Acoustics Volume 19, 2013 http://acousticalsociety.org/ ICA 2013 Montreal Montreal, Canada 2-7 June 2013 Speech Communication Session 2aSC: Linking Perception and Production
More informationVoiceless Stop Consonant Modelling and Synthesis Framework Based on MISO Dynamic System
ARCHIVES OF ACOUSTICS Vol. 42, No. 3, pp. 375 383 (2017) Copyright c 2017 by PAN IPPT DOI: 10.1515/aoa-2017-0039 Voiceless Stop Consonant Modelling and Synthesis Framework Based on MISO Dynamic System
More informationAuthor's personal copy
Speech Communication 49 (2007) 588 601 www.elsevier.com/locate/specom Abstract Subjective comparison and evaluation of speech enhancement Yi Hu, Philipos C. Loizou * Department of Electrical Engineering,
More informationSemi-Supervised GMM and DNN Acoustic Model Training with Multi-system Combination and Confidence Re-calibration
INTERSPEECH 2013 Semi-Supervised GMM and DNN Acoustic Model Training with Multi-system Combination and Confidence Re-calibration Yan Huang, Dong Yu, Yifan Gong, and Chaojun Liu Microsoft Corporation, One
More informationLecture 9: Speech Recognition
EE E6820: Speech & Audio Processing & Recognition Lecture 9: Speech Recognition 1 Recognizing speech 2 Feature calculation Dan Ellis Michael Mandel 3 Sequence
More informationA Privacy-Sensitive Approach to Modeling Multi-Person Conversations
A Privacy-Sensitive Approach to Modeling Multi-Person Conversations Danny Wyatt Dept. of Computer Science University of Washington danny@cs.washington.edu Jeff Bilmes Dept. of Electrical Engineering University
More informationBODY LANGUAGE ANIMATION SYNTHESIS FROM PROSODY AN HONORS THESIS SUBMITTED TO THE DEPARTMENT OF COMPUTER SCIENCE OF STANFORD UNIVERSITY
BODY LANGUAGE ANIMATION SYNTHESIS FROM PROSODY AN HONORS THESIS SUBMITTED TO THE DEPARTMENT OF COMPUTER SCIENCE OF STANFORD UNIVERSITY Sergey Levine Principal Adviser: Vladlen Koltun Secondary Adviser:
More informationGDP Falls as MBA Rises?
Applied Mathematics, 2013, 4, 1455-1459 http://dx.doi.org/10.4236/am.2013.410196 Published Online October 2013 (http://www.scirp.org/journal/am) GDP Falls as MBA Rises? T. N. Cummins EconomicGPS, Aurora,
More informationA NOVEL SCHEME FOR SPEAKER RECOGNITION USING A PHONETICALLY-AWARE DEEP NEURAL NETWORK. Yun Lei Nicolas Scheffer Luciana Ferrer Mitchell McLaren
A NOVEL SCHEME FOR SPEAKER RECOGNITION USING A PHONETICALLY-AWARE DEEP NEURAL NETWORK Yun Lei Nicolas Scheffer Luciana Ferrer Mitchell McLaren Speech Technology and Research Laboratory, SRI International,
More informationPython Machine Learning
Python Machine Learning Unlock deeper insights into machine learning with this vital guide to cuttingedge predictive analytics Sebastian Raschka [ PUBLISHING 1 open source I community experience distilled
More informationDyslexia/dyslexic, 3, 9, 24, 97, 187, 189, 206, 217, , , 367, , , 397,
Adoption studies, 274 275 Alliteration skill, 113, 115, 117 118, 122 123, 128, 136, 138 Alphabetic writing system, 5, 40, 127, 136, 410, 415 Alphabets (types of ) artificial transparent alphabet, 5 German
More informationAutoregressive product of multi-frame predictions can improve the accuracy of hybrid models
Autoregressive product of multi-frame predictions can improve the accuracy of hybrid models Navdeep Jaitly 1, Vincent Vanhoucke 2, Geoffrey Hinton 1,2 1 University of Toronto 2 Google Inc. ndjaitly@cs.toronto.edu,
More informationBAUM-WELCH TRAINING FOR SEGMENT-BASED SPEECH RECOGNITION. Han Shu, I. Lee Hetherington, and James Glass
BAUM-WELCH TRAINING FOR SEGMENT-BASED SPEECH RECOGNITION Han Shu, I. Lee Hetherington, and James Glass Computer Science and Artificial Intelligence Laboratory Massachusetts Institute of Technology Cambridge,
More informationHOLMER GREEN SENIOR SCHOOL CURRICULUM INFORMATION
HOLMER GREEN SENIOR SCHOOL CURRICULUM INFORMATION Subject: Mathematics Year Group: 7 Exam Board: (For years 10, 11, 12 and 13 only) Assessment requirements: Students will take 3 large assessments during
More informationAnalysis of Speech Recognition Models for Real Time Captioning and Post Lecture Transcription
Analysis of Speech Recognition Models for Real Time Captioning and Post Lecture Transcription Wilny Wilson.P M.Tech Computer Science Student Thejus Engineering College Thrissur, India. Sindhu.S Computer
More informationRhythm-typology revisited.
DFG Project BA 737/1: "Cross-language and individual differences in the production and perception of syllabic prominence. Rhythm-typology revisited." Rhythm-typology revisited. B. Andreeva & W. Barry Jacques
More informationStatewide Framework Document for:
Statewide Framework Document for: 270301 Standards may be added to this document prior to submission, but may not be removed from the framework to meet state credit equivalency requirements. Performance
More informationA Hybrid Text-To-Speech system for Afrikaans
A Hybrid Text-To-Speech system for Afrikaans Francois Rousseau and Daniel Mashao Department of Electrical Engineering, University of Cape Town, Rondebosch, Cape Town, South Africa, frousseau@crg.ee.uct.ac.za,
More informationAGS THE GREAT REVIEW GAME FOR PRE-ALGEBRA (CD) CORRELATED TO CALIFORNIA CONTENT STANDARDS
AGS THE GREAT REVIEW GAME FOR PRE-ALGEBRA (CD) CORRELATED TO CALIFORNIA CONTENT STANDARDS 1 CALIFORNIA CONTENT STANDARDS: Chapter 1 ALGEBRA AND WHOLE NUMBERS Algebra and Functions 1.4 Students use algebraic
More informationDIDACTIC MODEL BRIDGING A CONCEPT WITH PHENOMENA
DIDACTIC MODEL BRIDGING A CONCEPT WITH PHENOMENA Beba Shternberg, Center for Educational Technology, Israel Michal Yerushalmy University of Haifa, Israel The article focuses on a specific method of constructing
More informationUnvoiced Landmark Detection for Segment-based Mandarin Continuous Speech Recognition
Unvoiced Landmark Detection for Segment-based Mandarin Continuous Speech Recognition Hua Zhang, Yun Tang, Wenju Liu and Bo Xu National Laboratory of Pattern Recognition Institute of Automation, Chinese
More informationMandarin Lexical Tone Recognition: The Gating Paradigm
Kansas Working Papers in Linguistics, Vol. 0 (008), p. 8 Abstract Mandarin Lexical Tone Recognition: The Gating Paradigm Yuwen Lai and Jie Zhang University of Kansas Research on spoken word recognition
More informationSTUDIES WITH FABRICATED SWITCHBOARD DATA: EXPLORING SOURCES OF MODEL-DATA MISMATCH
STUDIES WITH FABRICATED SWITCHBOARD DATA: EXPLORING SOURCES OF MODEL-DATA MISMATCH Don McAllaster, Larry Gillick, Francesco Scattone, Mike Newman Dragon Systems, Inc. 320 Nevada Street Newton, MA 02160
More informationUNIDIRECTIONAL LONG SHORT-TERM MEMORY RECURRENT NEURAL NETWORK WITH RECURRENT OUTPUT LAYER FOR LOW-LATENCY SPEECH SYNTHESIS. Heiga Zen, Haşim Sak
UNIDIRECTIONAL LONG SHORT-TERM MEMORY RECURRENT NEURAL NETWORK WITH RECURRENT OUTPUT LAYER FOR LOW-LATENCY SPEECH SYNTHESIS Heiga Zen, Haşim Sak Google fheigazen,hasimg@google.com ABSTRACT Long short-term
More informationEdinburgh Research Explorer
Edinburgh Research Explorer Personalising speech-to-speech translation Citation for published version: Dines, J, Liang, H, Saheer, L, Gibson, M, Byrne, W, Oura, K, Tokuda, K, Yamagishi, J, King, S, Wester,
More informationStatistical Parametric Speech Synthesis
Statistical Parametric Speech Synthesis Heiga Zen a,b,, Keiichi Tokuda a, Alan W. Black c a Department of Computer Science and Engineering, Nagoya Institute of Technology, Gokiso-cho, Showa-ku, Nagoya,
More informationStrategies for Solving Fraction Tasks and Their Link to Algebraic Thinking
Strategies for Solving Fraction Tasks and Their Link to Algebraic Thinking Catherine Pearn The University of Melbourne Max Stephens The University of Melbourne
More informationA Cross-language Corpus for Studying the Phonetics and Phonology of Prominence
A Cross-language Corpus for Studying the Phonetics and Phonology of Prominence Bistra Andreeva 1, William Barry 1, Jacques Koreman 2 1 Saarland University Germany 2 Norwegian University of Science and
More informationAutomatic Pronunciation Checker
Institut für Technische Informatik und Kommunikationsnetze Eidgenössische Technische Hochschule Zürich Swiss Federal Institute of Technology Zurich Ecole polytechnique fédérale de Zurich Politecnico federale
More informationCalibration of Confidence Measures in Speech Recognition
Submitted to IEEE Trans on Audio, Speech, and Language, July 2010 1 Calibration of Confidence Measures in Speech Recognition Dong Yu, Senior Member, IEEE, Jinyu Li, Member, IEEE, Li Deng, Fellow, IEEE
More informationPerceptual scaling of voice identity: common dimensions for different vowels and speakers
DOI 10.1007/s00426-008-0185-z ORIGINAL ARTICLE Perceptual scaling of voice identity: common dimensions for different vowels and speakers Oliver Baumann Æ Pascal Belin Received: 15 February 2008 / Accepted:
More informationTITLE 23: EDUCATION AND CULTURAL RESOURCES SUBTITLE A: EDUCATION CHAPTER I: STATE BOARD OF EDUCATION SUBCHAPTER b: PERSONNEL PART 25 CERTIFICATION
ISBE 23 ILLINOIS ADMINISTRATIVE CODE 25 TITLE 23: EDUCATION AND CULTURAL RESOURCES : EDUCATION CHAPTER I: STATE BOARD OF EDUCATION : PERSONNEL Section 25.10 Accredited Institution PART 25 CERTIFICATION
More informationA Neural Network GUI Tested on Text-To-Phoneme Mapping
A Neural Network GUI Tested on Text-To-Phoneme Mapping MAARTEN TROMPPER Universiteit Utrecht m.f.a.trompper@students.uu.nl Abstract Text-to-phoneme (T2P) mapping is a necessary step in any speech synthesis
More informationOn Developing Acoustic Models Using HTK. M.A. Spaans BSc.
On Developing Acoustic Models Using HTK M.A. Spaans BSc. On Developing Acoustic Models Using HTK M.A. Spaans BSc. Delft, December 2004 Copyright c 2004 M.A. Spaans BSc. December, 2004. Faculty of Electrical
More informationMathematics. Mathematics
Mathematics Program Description Successful completion of this major will assure competence in mathematics through differential and integral calculus, providing an adequate background for employment in
More informationDocument number: 2013/ Programs Committee 6/2014 (July) Agenda Item 42.0 Bachelor of Engineering with Honours in Software Engineering
Document number: 2013/0006139 Programs Committee 6/2014 (July) Agenda Item 42.0 Bachelor of Engineering with Honours in Software Engineering Program Learning Outcomes Threshold Learning Outcomes for Engineering
More informationClass Meeting Time and Place: Section 3: MTWF10:00-10:50 TILT 221
Math 155. Calculus for Biological Scientists Fall 2017 Website https://csumath155.wordpress.com Please review the course website for details on the schedule, extra resources, alternate exam request forms,
More informationCOMPUTER INTERFACES FOR TEACHING THE NINTENDO GENERATION
Session 3532 COMPUTER INTERFACES FOR TEACHING THE NINTENDO GENERATION Thad B. Welch, Brian Jenkins Department of Electrical Engineering U.S. Naval Academy, MD Cameron H. G. Wright Department of Electrical
More informationAn Empirical Analysis of the Effects of Mexican American Studies Participation on Student Achievement within Tucson Unified School District
An Empirical Analysis of the Effects of Mexican American Studies Participation on Student Achievement within Tucson Unified School District Report Submitted June 20, 2012, to Willis D. Hawley, Ph.D., Special
More informationBUILDING CONTEXT-DEPENDENT DNN ACOUSTIC MODELS USING KULLBACK-LEIBLER DIVERGENCE-BASED STATE TYING
BUILDING CONTEXT-DEPENDENT DNN ACOUSTIC MODELS USING KULLBACK-LEIBLER DIVERGENCE-BASED STATE TYING Gábor Gosztolya 1, Tamás Grósz 1, László Tóth 1, David Imseng 2 1 MTA-SZTE Research Group on Artificial
More informationEvolutive Neural Net Fuzzy Filtering: Basic Description
Journal of Intelligent Learning Systems and Applications, 2010, 2: 12-18 doi:10.4236/jilsa.2010.21002 Published Online February 2010 (http://www.scirp.org/journal/jilsa) Evolutive Neural Net Fuzzy Filtering:
More informationSegmental Conditional Random Fields with Deep Neural Networks as Acoustic Models for First-Pass Word Recognition
Segmental Conditional Random Fields with Deep Neural Networks as Acoustic Models for First-Pass Word Recognition Yanzhang He, Eric Fosler-Lussier Department of Computer Science and Engineering The hio
More informationInvestigation on Mandarin Broadcast News Speech Recognition
Investigation on Mandarin Broadcast News Speech Recognition Mei-Yuh Hwang 1, Xin Lei 1, Wen Wang 2, Takahiro Shinozaki 1 1 Univ. of Washington, Dept. of Electrical Engineering, Seattle, WA 98195 USA 2
More informationAlgebra 1, Quarter 3, Unit 3.1. Line of Best Fit. Overview
Algebra 1, Quarter 3, Unit 3.1 Line of Best Fit Overview Number of instructional days 6 (1 day assessment) (1 day = 45 minutes) Content to be learned Analyze scatter plots and construct the line of best
More informationPREDICTING SPEECH RECOGNITION CONFIDENCE USING DEEP LEARNING WITH WORD IDENTITY AND SCORE FEATURES
PREDICTING SPEECH RECOGNITION CONFIDENCE USING DEEP LEARNING WITH WORD IDENTITY AND SCORE FEATURES Po-Sen Huang, Kshitiz Kumar, Chaojun Liu, Yifan Gong, Li Deng Department of Electrical and Computer Engineering,
More informationControl Tutorials for MATLAB and Simulink
Control Tutorials for MATLAB and Simulink Last updated: 07/24/2014 Author Information Prof. Bill Messner Carnegie Mellon University Prof. Dawn Tilbury University of Michigan Asst. Prof. Rick Hill, PhD
More informationRadius STEM Readiness TM
Curriculum Guide Radius STEM Readiness TM While today s teens are surrounded by technology, we face a stark and imminent shortage of graduates pursuing careers in Science, Technology, Engineering, and
More informationWiggleWorks Software Manual PDF0049 (PDF) Houghton Mifflin Harcourt Publishing Company
WiggleWorks Software Manual PDF0049 (PDF) Houghton Mifflin Harcourt Publishing Company Table of Contents Welcome to WiggleWorks... 3 Program Materials... 3 WiggleWorks Teacher Software... 4 Logging In...
More informationMTH 215: Introduction to Linear Algebra
MTH 215: Introduction to Linear Algebra Fall 2017 University of Rhode Island, Department of Mathematics INSTRUCTOR: Jonathan A. Chávez Casillas E-MAIL: jchavezc@uri.edu LECTURE TIMES: Tuesday and Thursday,
More informationMathematics subject curriculum
Mathematics subject curriculum Dette er ei omsetjing av den fastsette læreplanteksten. Læreplanen er fastsett på Nynorsk Established as a Regulation by the Ministry of Education and Research on 24 June
More informationMajor Milestones, Team Activities, and Individual Deliverables
Major Milestones, Team Activities, and Individual Deliverables Milestone #1: Team Semester Proposal Your team should write a proposal that describes project objectives, existing relevant technology, engineering
More informationContent Language Objectives (CLOs) August 2012, H. Butts & G. De Anda
Content Language Objectives (CLOs) Outcomes Identify the evolution of the CLO Identify the components of the CLO Understand how the CLO helps provide all students the opportunity to access the rigor of
More informationSpeech Recognition by Indexing and Sequencing
International Journal of Computer Information Systems and Industrial Management Applications. ISSN 215-7988 Volume 4 (212) pp. 358 365 c MIR Labs, www.mirlabs.net/ijcisim/index.html Speech Recognition
More informationDOMAIN MISMATCH COMPENSATION FOR SPEAKER RECOGNITION USING A LIBRARY OF WHITENERS. Elliot Singer and Douglas Reynolds
DOMAIN MISMATCH COMPENSATION FOR SPEAKER RECOGNITION USING A LIBRARY OF WHITENERS Elliot Singer and Douglas Reynolds Massachusetts Institute of Technology Lincoln Laboratory {es,dar}@ll.mit.edu ABSTRACT
More informationUsing Articulatory Features and Inferred Phonological Segments in Zero Resource Speech Processing
Using Articulatory Features and Inferred Phonological Segments in Zero Resource Speech Processing Pallavi Baljekar, Sunayana Sitaram, Prasanna Kumar Muthukumar, and Alan W Black Carnegie Mellon University,
More informationUniversityy. The content of
WORKING PAPER #31 An Evaluation of Empirical Bayes Estimation of Value Added Teacher Performance Measuress Cassandra M. Guarino, Indianaa Universityy Michelle Maxfield, Michigan State Universityy Mark
More informationAffective Classification of Generic Audio Clips using Regression Models
Affective Classification of Generic Audio Clips using Regression Models Nikolaos Malandrakis 1, Shiva Sundaram, Alexandros Potamianos 3 1 Signal Analysis and Interpretation Laboratory (SAIL), USC, Los
More informationGuidelines for blind and partially sighted candidates
Revised August 2006 Guidelines for blind and partially sighted candidates Our policy In addition to the specific provisions described below, we are happy to consider each person individually if their needs
More informationThe Good Judgment Project: A large scale test of different methods of combining expert predictions
The Good Judgment Project: A large scale test of different methods of combining expert predictions Lyle Ungar, Barb Mellors, Jon Baron, Phil Tetlock, Jaime Ramos, Sam Swift The University of Pennsylvania
More information