Voice conversion through vector quantization

Size: px
Start display at page:

Download "Voice conversion through vector quantization"

Transcription

1 J. Acoust. Soc. Jpn.(E)11, 2 (1990) Voice conversion through vector quantization Masanobu Abe, Satoshi Nakamura, Kiyohiro Shikano, and Hisao Kuwabara A TR Interpreting Telephony Research Laboratories, Sanpeidani, Inuidani, Seika-cho, Souraku-gun, Kyoto, Japan (Received 13 May 1989) A new voice conversion technique through vector quantization and spectrum mapping is proposed. This technique is based on mapping codebooks which represent the cor respondencebetween different speakers' codebooks. The mapping codebooks for spectrum parameters, power values, and pitch frequencies are separately generated using training utterances. This technique makes it possible to precisely control voice individuality. The performance of this technique is confirmed by spectrum distortion and pitch frequency difference. To evaluate the overall performance of this technique, listening tests are carried out on two kinds of voice conversions:one between male and female speakers, the other between male speakers. In the male-to-female conversion experiment, all converted utterances are judged as female, and in the male-to-male conversion, 57% of them are identified as the target speaker. PACS number: Ja 1. INTRODUCTION In daily communication, voice individuality is one of the most important aspects of human speech. It is especially important for identifying other person in a telephone conversation. A technique to control speech individuality, therefore, has an important role and offers many applications. Our present study is concerned with converting voice quality from one speaker to another and developing a technique which enables us to give individuality to synthesized speech. One system goal we can imagine is shown in Fig. 1. Using a voice conversion system as a post-proces sorfor a synthesis-by-rule system, various kinds of speech, such as a particular person's voice, a child likevoice, husky voice, etc., can be synthesized. For voice conversion, as shown in Fig. 1, it is neces saryto have a database of voice individuality. Speech individuality generally consists of two major factors:acoustic features and prosodic fea tures.as the first step in this research, we are trying to control the acoustic features. 1, 2) According to previous studies, the acoustic features that con tributeto speech individuality are distributed among various parameters, such as formant frequencies, formant bandwidths, spectral tilt, and glottal wave forms.3, 4) Because speech individuality is deter minedby all of these, it is difficult to control it by modifying each parameter independently. On the other hand, codebooks used in vector quantization represent all of these parameters altogether. There fore,speech individuality of a speaker is represented by the code-vectors in a codebook of the speaker. A conversion of acoustic features from one speaker to another is reduced to the problem of finding a correspondence between the codebooks of the two speakers. The basic problem is, therefore, to find mapping function from one codebook to another, which we will call a 'mapping codebook.' This is the basic idea of this conversion technique. In our proposed technique, the mapping codebooks which represent the correspondence between different speakers' codebooks provide the database of speech individuality in Fig. 1. In section 2, a method of making mapping code booksand a synthesis procedure are described. In section 3, the mapping codebooks are evaluated by

2 J. Acoust. Soc. Jpn.(E)11, 2 (1990) Fig. 1 Synthesis-by-rule system with voice conversion. measuring distortion. In section 4, the overall performance of this technique is evaluated by listen ingtests. 2. VOICE CONVERSION THROUGH VECTOR QUANTIZATION Our voice conversion technique consists of two steps:a learning step and a conversion-synthesis step. The learning step generates the mapping codebooks, and the conversion-synthesis step synthe sizesspeech using the mapping codebooks. 2.1 Learning Step The mapping codebooks are codebooks that describe a mapping function between the vector spaces of two speakers. Figure 2 illustrates the block diagram of the procedure for generating a mapping codebook for spectrum parameters. 1. Two speakers, A and B, pronounce a learning word set. Using these learning utterances, a codebook is generated for each speaker. Next, a word uttered by speaker A is vector quantized using his/her codebook. The same word uttered by speaker B is also vector quantized by the same way. 2. The correspondence between vectors of the same words from the two speakers is deter minedusing Dynamic Time Warping (DTW). 3. The vector correspondences between two speakers are accumulated as histograms. Histograms of vector correspondence are made by applying the same procedure to other learning words. 4. Using the histogram for each codevector of speaker A as a weighting function, a mapping codebook from speaker A to B is defined as a linear combination of speaker B's vectors. 5. The codevectors of speaker A are replaced with codevectors of a mapping codebook. 6. If the decrease in average distortion through DTW over all training words is not converged, steps 2, 3, 4 and 5 are repeated to refine the mapping codebook. Pitch frequencies and power values contribute a great deal to speech individuality. Mapping code booksfor these parameters are also generated at the same time using almost the same procedure men tionedabove. The differences are: 1. Pitch frequencies and power values are each scalar-quantized. 2. The mapping codebook for pitch frequencies is defined based on the maximum occurrence in the histogram. 2.2 Conversion-synthesis Step Figure 3 shows the block diagram of the conver sion-synthesis step. First speaker A's speech is analyzed by the linear prediction method. Then the spectrum parameters are vector-quantized using his/her codebook, and parameters for pitch frequencies and power values are scalar-quantized using his/her codebooks. Next, all parameters are decoded using mapping codebooks between speakers A and B. Finally, speech is synthesized by LPC vocoder. The output speech will have the voice individuality of speaker B. Fig. 2 Method for generating a mapping codebook.

3 M. ABE et al.:voice CONVERSION THROUGH VECTOR QUANTIZATION Fig. 3 Block diagram of voice conversion from speaker A to speaker B. 3. CONVERSION EXPERIMENTS To evaluate the performance of the conversion technique, distortion measurements were carried out for the spectrum parameters as well as the pitch frequencies. 3.1 Spectrum Conversion Experiments Experiment conditions are listed in Table 1. A set of 100 phonetically-balanced words was used for learning to produce mapping codebooks. Spectrum conversions were made between female Table 1 Experiment conditions. Table 2 Spectrum distortion. and male voices, between male and male and between female and female voices. Six speakers (3 male and 3 female speakers, all professional announcers) were used to provide speech material. Table 2 shows the results of the open test. After vector quantization, two kinds of spectrum distortions between two speech samples were cal culated:between the input and target speaker's (before conversion in Table 2), and the converted and target speaker's speech (after conversion in Table 2). In the female-to-female conversion, the distortion decreased by 27% compared to non con version,the male-to-male conversion by 49%, and the male-to-female conversion by 66%. These results show that this conversion technique is highly effective when there is a large enough difference between two speakers's voices. 3.2 Pitch Frequency Conversion Experiments Pitch frequency conversion was also carried out through the same process described in 3.1. The experiment results are shown in Fig. 4. This figure shows the relationship between the number of learning words and the average pitch frequency differences after conversion. The value at the point where the number of learning words is 0 shows the natural average pitch frequency dif ferencebetween the two speakers. According to this figure, 60 words are considered to be enough to make a mapping codebook for pitch frequency regardless of speaker combinations, and the average pitch frequency difference decreases to less than 15 Hz. 4. EVALUATION BY LISTENING TEST To evaluate overall performance of this technique, three kinds of listening tests were carried out. The first experiment deals with the male-to-female con versionand the other two experiments deal with the

4 J.Acoust.Soc.Jpn.(E) 11, 2 (1990) loud speaker in a sound proof room. Twelve listeners were asked to rate the similarity of each pair in five categories:"similar," "slightly similar," "difficult to decide," "slightly dissimilar," "dis similar." Experiment 2 Experiment 2 was designed to evaluate the con versionbetween two male speakers by the so called ABX method. Stimuli A and B are vector-quantiz edoriginal speech tokens for speakers N and M respectively. X takes either the converted token (N M or M N) or the vector-quantized original token (N or M). Four different words were used for the conversions and each triad was a combina tionof 3 different words. A total of 96 speech triads were presented to the listeners as described number of learning words number of learning words 1. male male conversion 2. female conversion Fig. 4 Pitch frequency differences for num berof learning words. male-to-male conversion. 4.1 Experiment Procedure Experiment 1 Experiment 1 was designed to evaluate the voice quality for male-to-female voice conversion by a pair-comparison listening test. In addition to the fully converted speech, conversion was also done for pitch and spectrum parameters separately in order to examine the individual contribution of these parameters to speech individuality. The following is a list of 5 different speech conversions performed in this experiment. 1. vector-quantized original male speech (m) 2. male-to-female converted speech:pitch fre quencyconversion only (mp-fp) 3. male-to-female converted speech:spectrum conversion only (ms-fs) 4. male-to-female converted speech:all parame ters(m-f) 5. vector-quantized original female speech which is the target for the conversions (f) In order to avoid unnecessary cues for the judge mentof voice quality, 2 different words were used to make speech pairs for the listening test. A set of speech pairs consists of all possible combinations of stimuli from the 5 different conversions, 40 in total. They were presented to listeners through a above. The listeners were required to select the stimulus (A or B) which more closely resembled the stimulus X Experiment 3 Experiment 3 was designed to evaluate the con versionbetween male speakers in the same way as in However, conversions for pitch frequencies alone and spectrum parameters alone were excluded. The following is a list of the 4 conversions. 1. vector-quantized male speech (male 1) 2. same as 1 but for another male speaker (male 2) 3. converted speech from male 1 to male 2 (male 1 male 2) 4. converted speech from male 2 to male 1 (male 2 male 1) A total of 72 speech pairs were generated using the same procedures in experiment Experiment Results Evaluation of male-to-female conversion (Results of Experiment 1) Hayashi's fourth method of quantification 5) was applied to the experimental data obtained by the listening test. This method places stimuli in a space according to the similarities between every two stimuli. Its formulation minimizes the measure Q, Q=-SUM[e(i,j){x(i)-x(j)}2] where e(i, j) denotes the similarity between stimuli i and j, and x(i) represents the location of stimulus i in the space. The projection onto a two-dimensional space is shown in Fig. 5.

5 M.ABE et al.:voice CONVERSION THROUGH VECTOR QUANTIZATION correctly identify the speaker even if the original speaker's speech is used as stimuli X; i.e. the correct answer is about 85% in the male 1-male 2 pair, and about 60% in the male 1-male 3 pair. Judging from these scores, the average score 72.2% in the male 1-male 2 conversions shows satisfactory per formance.as total performance, 57% of the con vertedutterance is judged correctly. A relatively poor performance for the male 1-male 3 conversion stems from the fact that male l's voice quality is very close to male 3's voice. To confirm this, the "distance" between the original speaker and target speaker is shown in Fig. 6. This figure shows spectrum distance and pitch frequency distance before and after conversion. From this figure, the pitch frequency range of male Fig. 5 Distribution of psychological dis tancesfor the male-to-female voice con version. This figure represents the relative similarity distancebetween stimuli. In this figure, converted 1 and male 3 is found to be very close. Figure 7 represents the result for Experiment 3 analyzed by the same method as in It is observed that the converted speech samples, "male 1 male 2" and "male 2 male 1," are both placed closer to their target speech. This indicates speech "m f" is placed very close to the speech "f." This indicates that this technique properly converted the male speech to the target female speech. Judging from the positions of "mp fp" and "ms fs," it is observed that the first and second axes roughly correspond to pitch frequency and spectrum differences, respectively. The result in dicatesthat neither pitch frequency nor spectrum carries enough information about speech indi viduality,and both are necessary Evaluation of male-to-male conversion (Re sultsof Experiments 2 and 3) The results of Experiment 2 are shown in Table 3. spectrum distortion Fig. 6 Speaker distance in spectrum distor tionand pitch frequency difference. The numbers in this table represent the percentage of responses where stimuli X are judged correctly. According to the table, listeners can't always Table 3 Percentage of correct responses. Fig. 7 Distribution of psychological dis tancesfor the male-to-male voice con version.

6 J.Acoust. Soc.Jpn.(E) 11, 2 (1990) that the proposed technique can also convert speech individuality between same-sex speakers. 5. CONCLUSION A new voice conversion technique through vector quantization and spectrum mapping was proposed. The advantage of this technique is summarized as follows: 1. The mapping codebooks which make it possi bleto give an individuality to synthesized speech are generated from a limited number of word utterances. 2. The mapping codebooks enable voice con versionof high quality between any two speakers. 3. The synthesis process requires few computa tionand produces speech in real time. The performance of this technique is confirmed by spectrum distortion and pitch frequency dif ference.the spectrum distortion between original speech and target speech decreased by a range of 27% to 66%. Pitch frequency difference decreased to less than 15 Hz. The overall performance of this technique is also confirmed by listening tests. It can be concluded that the converted speech has a voice quality very close to the target speaker's. ACKNOWLEDGMENTS We are grateful to Dr. Kurematsu, president of ATR Interpreting Telephony Research Laboratories, for his continuous support of this work. We wish to also thank Dr. Tohkura, head of Hearing& Speech Perception Department, for helpful discus sion. REFERENCES 1 ) K.Shikano, K.Lee, and R.Reddy,"Speaker ad aptationthrough vector quantization," ICASSP 86, (1986). 2 ) S.Nakamura and K.Shikano,"Spectrogram nor malizationbased on vector quantization," Tech. Rep. Speech Acoust. Soc.Jpn.SP87-17, 9-16 (1987)(in Japanese). 3 ) H.Kuwabara and T.Takagi,"Quality control of speech by modifying formant frequencies and bandwidth," 11th Int.Congr. Phonetic Sciences, , August (1987). 4 ) D.G.Childers, B.Yegnanarayana, and K.Wu, "Voice conversion:factors responsible for quality," ICASSP 85, (1985). 5 ) C.Hayashi,"Recent theoretical and methodological developments in multidimensional scaling and its related methods in Japan," Behaviormetrika No.18 (1985).

Mandarin Lexical Tone Recognition: The Gating Paradigm

Mandarin Lexical Tone Recognition: The Gating Paradigm Kansas Working Papers in Linguistics, Vol. 0 (008), p. 8 Abstract Mandarin Lexical Tone Recognition: The Gating Paradigm Yuwen Lai and Jie Zhang University of Kansas Research on spoken word recognition

More information

Design Of An Automatic Speaker Recognition System Using MFCC, Vector Quantization And LBG Algorithm

Design Of An Automatic Speaker Recognition System Using MFCC, Vector Quantization And LBG Algorithm Design Of An Automatic Speaker Recognition System Using MFCC, Vector Quantization And LBG Algorithm Prof. Ch.Srinivasa Kumar Prof. and Head of department. Electronics and communication Nalanda Institute

More information

Eli Yamamoto, Satoshi Nakamura, Kiyohiro Shikano. Graduate School of Information Science, Nara Institute of Science & Technology

Eli Yamamoto, Satoshi Nakamura, Kiyohiro Shikano. Graduate School of Information Science, Nara Institute of Science & Technology ISCA Archive SUBJECTIVE EVALUATION FOR HMM-BASED SPEECH-TO-LIP MOVEMENT SYNTHESIS Eli Yamamoto, Satoshi Nakamura, Kiyohiro Shikano Graduate School of Information Science, Nara Institute of Science & Technology

More information

Speech Emotion Recognition Using Support Vector Machine

Speech Emotion Recognition Using Support Vector Machine Speech Emotion Recognition Using Support Vector Machine Yixiong Pan, Peipei Shen and Liping Shen Department of Computer Technology Shanghai JiaoTong University, Shanghai, China panyixiong@sjtu.edu.cn,

More information

Human Emotion Recognition From Speech

Human Emotion Recognition From Speech RESEARCH ARTICLE OPEN ACCESS Human Emotion Recognition From Speech Miss. Aparna P. Wanare*, Prof. Shankar N. Dandare *(Department of Electronics & Telecommunication Engineering, Sant Gadge Baba Amravati

More information

A study of speaker adaptation for DNN-based speech synthesis

A study of speaker adaptation for DNN-based speech synthesis A study of speaker adaptation for DNN-based speech synthesis Zhizheng Wu, Pawel Swietojanski, Christophe Veaux, Steve Renals, Simon King The Centre for Speech Technology Research (CSTR) University of Edinburgh,

More information

Class-Discriminative Weighted Distortion Measure for VQ-Based Speaker Identification

Class-Discriminative Weighted Distortion Measure for VQ-Based Speaker Identification Class-Discriminative Weighted Distortion Measure for VQ-Based Speaker Identification Tomi Kinnunen and Ismo Kärkkäinen University of Joensuu, Department of Computer Science, P.O. Box 111, 80101 JOENSUU,

More information

Speech Recognition at ICSI: Broadcast News and beyond

Speech Recognition at ICSI: Broadcast News and beyond Speech Recognition at ICSI: Broadcast News and beyond Dan Ellis International Computer Science Institute, Berkeley CA Outline 1 2 3 The DARPA Broadcast News task Aspects of ICSI

More information

Quarterly Progress and Status Report. VCV-sequencies in a preliminary text-to-speech system for female speech

Quarterly Progress and Status Report. VCV-sequencies in a preliminary text-to-speech system for female speech Dept. for Speech, Music and Hearing Quarterly Progress and Status Report VCV-sequencies in a preliminary text-to-speech system for female speech Karlsson, I. and Neovius, L. journal: STL-QPSR volume: 35

More information

Speaker recognition using universal background model on YOHO database

Speaker recognition using universal background model on YOHO database Aalborg University Master Thesis project Speaker recognition using universal background model on YOHO database Author: Alexandre Majetniak Supervisor: Zheng-Hua Tan May 31, 2011 The Faculties of Engineering,

More information

A Comparison of DHMM and DTW for Isolated Digits Recognition System of Arabic Language

A Comparison of DHMM and DTW for Isolated Digits Recognition System of Arabic Language A Comparison of DHMM and DTW for Isolated Digits Recognition System of Arabic Language Z.HACHKAR 1,3, A. FARCHI 2, B.MOUNIR 1, J. EL ABBADI 3 1 Ecole Supérieure de Technologie, Safi, Morocco. zhachkar2000@yahoo.fr.

More information

Body-Conducted Speech Recognition and its Application to Speech Support System

Body-Conducted Speech Recognition and its Application to Speech Support System Body-Conducted Speech Recognition and its Application to Speech Support System 4 Shunsuke Ishimitsu Hiroshima City University Japan 1. Introduction In recent years, speech recognition systems have been

More information

Unvoiced Landmark Detection for Segment-based Mandarin Continuous Speech Recognition

Unvoiced Landmark Detection for Segment-based Mandarin Continuous Speech Recognition Unvoiced Landmark Detection for Segment-based Mandarin Continuous Speech Recognition Hua Zhang, Yun Tang, Wenju Liu and Bo Xu National Laboratory of Pattern Recognition Institute of Automation, Chinese

More information

International Journal of Computational Intelligence and Informatics, Vol. 1 : No. 4, January - March 2012

International Journal of Computational Intelligence and Informatics, Vol. 1 : No. 4, January - March 2012 Text-independent Mono and Cross-lingual Speaker Identification with the Constraint of Limited Data Nagaraja B G and H S Jayanna Department of Information Science and Engineering Siddaganga Institute of

More information

Speech Synthesis in Noisy Environment by Enhancing Strength of Excitation and Formant Prominence

Speech Synthesis in Noisy Environment by Enhancing Strength of Excitation and Formant Prominence INTERSPEECH September,, San Francisco, USA Speech Synthesis in Noisy Environment by Enhancing Strength of Excitation and Formant Prominence Bidisha Sharma and S. R. Mahadeva Prasanna Department of Electronics

More information

Perceptual scaling of voice identity: common dimensions for different vowels and speakers

Perceptual scaling of voice identity: common dimensions for different vowels and speakers DOI 10.1007/s00426-008-0185-z ORIGINAL ARTICLE Perceptual scaling of voice identity: common dimensions for different vowels and speakers Oliver Baumann Æ Pascal Belin Received: 15 February 2008 / Accepted:

More information

Noise-Adaptive Perceptual Weighting in the AMR-WB Encoder for Increased Speech Loudness in Adverse Far-End Noise Conditions

Noise-Adaptive Perceptual Weighting in the AMR-WB Encoder for Increased Speech Loudness in Adverse Far-End Noise Conditions 26 24th European Signal Processing Conference (EUSIPCO) Noise-Adaptive Perceptual Weighting in the AMR-WB Encoder for Increased Speech Loudness in Adverse Far-End Noise Conditions Emma Jokinen Department

More information

WHEN THERE IS A mismatch between the acoustic

WHEN THERE IS A mismatch between the acoustic 808 IEEE TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 14, NO. 3, MAY 2006 Optimization of Temporal Filters for Constructing Robust Features in Speech Recognition Jeih-Weih Hung, Member,

More information

The Perception of Nasalized Vowels in American English: An Investigation of On-line Use of Vowel Nasalization in Lexical Access

The Perception of Nasalized Vowels in American English: An Investigation of On-line Use of Vowel Nasalization in Lexical Access The Perception of Nasalized Vowels in American English: An Investigation of On-line Use of Vowel Nasalization in Lexical Access Joyce McDonough 1, Heike Lenhert-LeHouiller 1, Neil Bardhan 2 1 Linguistics

More information

AUTOMATIC DETECTION OF PROLONGED FRICATIVE PHONEMES WITH THE HIDDEN MARKOV MODELS APPROACH 1. INTRODUCTION

AUTOMATIC DETECTION OF PROLONGED FRICATIVE PHONEMES WITH THE HIDDEN MARKOV MODELS APPROACH 1. INTRODUCTION JOURNAL OF MEDICAL INFORMATICS & TECHNOLOGIES Vol. 11/2007, ISSN 1642-6037 Marek WIŚNIEWSKI *, Wiesława KUNISZYK-JÓŹKOWIAK *, Elżbieta SMOŁKA *, Waldemar SUSZYŃSKI * HMM, recognition, speech, disorders

More information

Quarterly Progress and Status Report. Voiced-voiceless distinction in alaryngeal speech - acoustic and articula

Quarterly Progress and Status Report. Voiced-voiceless distinction in alaryngeal speech - acoustic and articula Dept. for Speech, Music and Hearing Quarterly Progress and Status Report Voiced-voiceless distinction in alaryngeal speech - acoustic and articula Nord, L. and Hammarberg, B. and Lundström, E. journal:

More information

Automatic Pronunciation Checker

Automatic Pronunciation Checker Institut für Technische Informatik und Kommunikationsnetze Eidgenössische Technische Hochschule Zürich Swiss Federal Institute of Technology Zurich Ecole polytechnique fédérale de Zurich Politecnico federale

More information

Analysis of Emotion Recognition System through Speech Signal Using KNN & GMM Classifier

Analysis of Emotion Recognition System through Speech Signal Using KNN & GMM Classifier IOSR Journal of Electronics and Communication Engineering (IOSR-JECE) e-issn: 2278-2834,p- ISSN: 2278-8735.Volume 10, Issue 2, Ver.1 (Mar - Apr.2015), PP 55-61 www.iosrjournals.org Analysis of Emotion

More information

Lip reading: Japanese vowel recognition by tracking temporal changes of lip shape

Lip reading: Japanese vowel recognition by tracking temporal changes of lip shape Lip reading: Japanese vowel recognition by tracking temporal changes of lip shape Koshi Odagiri 1, and Yoichi Muraoka 1 1 Graduate School of Fundamental/Computer Science and Engineering, Waseda University,

More information

Speech Segmentation Using Probabilistic Phonetic Feature Hierarchy and Support Vector Machines

Speech Segmentation Using Probabilistic Phonetic Feature Hierarchy and Support Vector Machines Speech Segmentation Using Probabilistic Phonetic Feature Hierarchy and Support Vector Machines Amit Juneja and Carol Espy-Wilson Department of Electrical and Computer Engineering University of Maryland,

More information

A comparison of spectral smoothing methods for segment concatenation based speech synthesis

A comparison of spectral smoothing methods for segment concatenation based speech synthesis D.T. Chappell, J.H.L. Hansen, "Spectral Smoothing for Speech Segment Concatenation, Speech Communication, Volume 36, Issues 3-4, March 2002, Pages 343-373. A comparison of spectral smoothing methods for

More information

Intra-talker Variation: Audience Design Factors Affecting Lexical Selections

Intra-talker Variation: Audience Design Factors Affecting Lexical Selections Tyler Perrachione LING 451-0 Proseminar in Sound Structure Prof. A. Bradlow 17 March 2006 Intra-talker Variation: Audience Design Factors Affecting Lexical Selections Abstract Although the acoustic and

More information

Rhythm-typology revisited.

Rhythm-typology revisited. DFG Project BA 737/1: "Cross-language and individual differences in the production and perception of syllabic prominence. Rhythm-typology revisited." Rhythm-typology revisited. B. Andreeva & W. Barry Jacques

More information

Developing a College-level Speed and Accuracy Test

Developing a College-level Speed and Accuracy Test Brigham Young University BYU ScholarsArchive All Faculty Publications 2011-02-18 Developing a College-level Speed and Accuracy Test Jordan Gilbert Marne Isakson See next page for additional authors Follow

More information

Revisiting the role of prosody in early language acquisition. Megha Sundara UCLA Phonetics Lab

Revisiting the role of prosody in early language acquisition. Megha Sundara UCLA Phonetics Lab Revisiting the role of prosody in early language acquisition Megha Sundara UCLA Phonetics Lab Outline Part I: Intonation has a role in language discrimination Part II: Do English-learning infants have

More information

A Cross-language Corpus for Studying the Phonetics and Phonology of Prominence

A Cross-language Corpus for Studying the Phonetics and Phonology of Prominence A Cross-language Corpus for Studying the Phonetics and Phonology of Prominence Bistra Andreeva 1, William Barry 1, Jacques Koreman 2 1 Saarland University Germany 2 Norwegian University of Science and

More information

OPTIMIZATINON OF TRAINING SETS FOR HEBBIAN-LEARNING- BASED CLASSIFIERS

OPTIMIZATINON OF TRAINING SETS FOR HEBBIAN-LEARNING- BASED CLASSIFIERS OPTIMIZATINON OF TRAINING SETS FOR HEBBIAN-LEARNING- BASED CLASSIFIERS Václav Kocian, Eva Volná, Michal Janošek, Martin Kotyrba University of Ostrava Department of Informatics and Computers Dvořákova 7,

More information

Analysis of Speech Recognition Models for Real Time Captioning and Post Lecture Transcription

Analysis of Speech Recognition Models for Real Time Captioning and Post Lecture Transcription Analysis of Speech Recognition Models for Real Time Captioning and Post Lecture Transcription Wilny Wilson.P M.Tech Computer Science Student Thejus Engineering College Thrissur, India. Sindhu.S Computer

More information

UNIDIRECTIONAL LONG SHORT-TERM MEMORY RECURRENT NEURAL NETWORK WITH RECURRENT OUTPUT LAYER FOR LOW-LATENCY SPEECH SYNTHESIS. Heiga Zen, Haşim Sak

UNIDIRECTIONAL LONG SHORT-TERM MEMORY RECURRENT NEURAL NETWORK WITH RECURRENT OUTPUT LAYER FOR LOW-LATENCY SPEECH SYNTHESIS. Heiga Zen, Haşim Sak UNIDIRECTIONAL LONG SHORT-TERM MEMORY RECURRENT NEURAL NETWORK WITH RECURRENT OUTPUT LAYER FOR LOW-LATENCY SPEECH SYNTHESIS Heiga Zen, Haşim Sak Google fheigazen,hasimg@google.com ABSTRACT Long short-term

More information

Module 12. Machine Learning. Version 2 CSE IIT, Kharagpur

Module 12. Machine Learning. Version 2 CSE IIT, Kharagpur Module 12 Machine Learning 12.1 Instructional Objective The students should understand the concept of learning systems Students should learn about different aspects of a learning system Students should

More information

Dimensions of Classroom Behavior Measured by Two Systems of Interaction Analysis

Dimensions of Classroom Behavior Measured by Two Systems of Interaction Analysis Dimensions of Classroom Behavior Measured by Two Systems of Interaction Analysis the most important and exciting recent development in the study of teaching has been the appearance of sev eral new instruments

More information

Instructor: Mario D. Garrett, Ph.D. Phone: Office: Hepner Hall (HH) 100

Instructor: Mario D. Garrett, Ph.D.   Phone: Office: Hepner Hall (HH) 100 San Diego State University School of Social Work 610 COMPUTER APPLICATIONS FOR SOCIAL WORK PRACTICE Statistical Package for the Social Sciences Office: Hepner Hall (HH) 100 Instructor: Mario D. Garrett,

More information

Major Milestones, Team Activities, and Individual Deliverables

Major Milestones, Team Activities, and Individual Deliverables Major Milestones, Team Activities, and Individual Deliverables Milestone #1: Team Semester Proposal Your team should write a proposal that describes project objectives, existing relevant technology, engineering

More information

Robot manipulations and development of spatial imagery

Robot manipulations and development of spatial imagery Robot manipulations and development of spatial imagery Author: Igor M. Verner, Technion Israel Institute of Technology, Haifa, 32000, ISRAEL ttrigor@tx.technion.ac.il Abstract This paper considers spatial

More information

Probabilistic Latent Semantic Analysis

Probabilistic Latent Semantic Analysis Probabilistic Latent Semantic Analysis Thomas Hofmann Presentation by Ioannis Pavlopoulos & Andreas Damianou for the course of Data Mining & Exploration 1 Outline Latent Semantic Analysis o Need o Overview

More information

PROJECT MANAGEMENT AND COMMUNICATION SKILLS DEVELOPMENT STUDENTS PERCEPTION ON THEIR LEARNING

PROJECT MANAGEMENT AND COMMUNICATION SKILLS DEVELOPMENT STUDENTS PERCEPTION ON THEIR LEARNING PROJECT MANAGEMENT AND COMMUNICATION SKILLS DEVELOPMENT STUDENTS PERCEPTION ON THEIR LEARNING Mirka Kans Department of Mechanical Engineering, Linnaeus University, Sweden ABSTRACT In this paper we investigate

More information

Acoustic correlates of stress and their use in diagnosing syllable fusion in Tongan. James White & Marc Garellek UCLA

Acoustic correlates of stress and their use in diagnosing syllable fusion in Tongan. James White & Marc Garellek UCLA Acoustic correlates of stress and their use in diagnosing syllable fusion in Tongan James White & Marc Garellek UCLA 1 Introduction Goals: To determine the acoustic correlates of primary and secondary

More information

Phonological and Phonetic Representations: The Case of Neutralization

Phonological and Phonetic Representations: The Case of Neutralization Phonological and Phonetic Representations: The Case of Neutralization Allard Jongman University of Kansas 1. Introduction The present paper focuses on the phenomenon of phonological neutralization to consider

More information

Cross Language Information Retrieval

Cross Language Information Retrieval Cross Language Information Retrieval RAFFAELLA BERNARDI UNIVERSITÀ DEGLI STUDI DI TRENTO P.ZZA VENEZIA, ROOM: 2.05, E-MAIL: BERNARDI@DISI.UNITN.IT Contents 1 Acknowledgment.............................................

More information

Grade 6: Correlated to AGS Basic Math Skills

Grade 6: Correlated to AGS Basic Math Skills Grade 6: Correlated to AGS Basic Math Skills Grade 6: Standard 1 Number Sense Students compare and order positive and negative integers, decimals, fractions, and mixed numbers. They find multiples and

More information

The NICT/ATR speech synthesis system for the Blizzard Challenge 2008

The NICT/ATR speech synthesis system for the Blizzard Challenge 2008 The NICT/ATR speech synthesis system for the Blizzard Challenge 2008 Ranniery Maia 1,2, Jinfu Ni 1,2, Shinsuke Sakai 1,2, Tomoki Toda 1,3, Keiichi Tokuda 1,4 Tohru Shimizu 1,2, Satoshi Nakamura 1,2 1 National

More information

Learning Structural Correspondences Across Different Linguistic Domains with Synchronous Neural Language Models

Learning Structural Correspondences Across Different Linguistic Domains with Synchronous Neural Language Models Learning Structural Correspondences Across Different Linguistic Domains with Synchronous Neural Language Models Stephan Gouws and GJ van Rooyen MIH Medialab, Stellenbosch University SOUTH AFRICA {stephan,gvrooyen}@ml.sun.ac.za

More information

Using Articulatory Features and Inferred Phonological Segments in Zero Resource Speech Processing

Using Articulatory Features and Inferred Phonological Segments in Zero Resource Speech Processing Using Articulatory Features and Inferred Phonological Segments in Zero Resource Speech Processing Pallavi Baljekar, Sunayana Sitaram, Prasanna Kumar Muthukumar, and Alan W Black Carnegie Mellon University,

More information

Expressive speech synthesis: a review

Expressive speech synthesis: a review Int J Speech Technol (2013) 16:237 260 DOI 10.1007/s10772-012-9180-2 Expressive speech synthesis: a review D. Govind S.R. Mahadeva Prasanna Received: 31 May 2012 / Accepted: 11 October 2012 / Published

More information

Robust Speech Recognition using DNN-HMM Acoustic Model Combining Noise-aware training with Spectral Subtraction

Robust Speech Recognition using DNN-HMM Acoustic Model Combining Noise-aware training with Spectral Subtraction INTERSPEECH 2015 Robust Speech Recognition using DNN-HMM Acoustic Model Combining Noise-aware training with Spectral Subtraction Akihiro Abe, Kazumasa Yamamoto, Seiichi Nakagawa Department of Computer

More information

Journal of Phonetics

Journal of Phonetics Journal of Phonetics 41 (2013) 297 306 Contents lists available at SciVerse ScienceDirect Journal of Phonetics journal homepage: www.elsevier.com/locate/phonetics The role of intonation in language and

More information

Semi-Supervised GMM and DNN Acoustic Model Training with Multi-system Combination and Confidence Re-calibration

Semi-Supervised GMM and DNN Acoustic Model Training with Multi-system Combination and Confidence Re-calibration INTERSPEECH 2013 Semi-Supervised GMM and DNN Acoustic Model Training with Multi-system Combination and Confidence Re-calibration Yan Huang, Dong Yu, Yifan Gong, and Chaojun Liu Microsoft Corporation, One

More information

A Neural Network GUI Tested on Text-To-Phoneme Mapping

A Neural Network GUI Tested on Text-To-Phoneme Mapping A Neural Network GUI Tested on Text-To-Phoneme Mapping MAARTEN TROMPPER Universiteit Utrecht m.f.a.trompper@students.uu.nl Abstract Text-to-phoneme (T2P) mapping is a necessary step in any speech synthesis

More information

Perceived speech rate: the effects of. articulation rate and speaking style in spontaneous speech. Jacques Koreman. Saarland University

Perceived speech rate: the effects of. articulation rate and speaking style in spontaneous speech. Jacques Koreman. Saarland University 1 Perceived speech rate: the effects of articulation rate and speaking style in spontaneous speech Jacques Koreman Saarland University Institute of Phonetics P.O. Box 151150 D-66041 Saarbrücken Germany

More information

Lecture 9: Speech Recognition

Lecture 9: Speech Recognition EE E6820: Speech & Audio Processing & Recognition Lecture 9: Speech Recognition 1 Recognizing speech 2 Feature calculation Dan Ellis Michael Mandel 3 Sequence

More information

Houghton Mifflin Online Assessment System Walkthrough Guide

Houghton Mifflin Online Assessment System Walkthrough Guide Houghton Mifflin Online Assessment System Walkthrough Guide Page 1 Copyright 2007 by Houghton Mifflin Company. All Rights Reserved. No part of this document may be reproduced or transmitted in any form

More information

Artificial Neural Networks written examination

Artificial Neural Networks written examination 1 (8) Institutionen för informationsteknologi Olle Gällmo Universitetsadjunkt Adress: Lägerhyddsvägen 2 Box 337 751 05 Uppsala Artificial Neural Networks written examination Monday, May 15, 2006 9 00-14

More information

BAUM-WELCH TRAINING FOR SEGMENT-BASED SPEECH RECOGNITION. Han Shu, I. Lee Hetherington, and James Glass

BAUM-WELCH TRAINING FOR SEGMENT-BASED SPEECH RECOGNITION. Han Shu, I. Lee Hetherington, and James Glass BAUM-WELCH TRAINING FOR SEGMENT-BASED SPEECH RECOGNITION Han Shu, I. Lee Hetherington, and James Glass Computer Science and Artificial Intelligence Laboratory Massachusetts Institute of Technology Cambridge,

More information

NCEO Technical Report 27

NCEO Technical Report 27 Home About Publications Special Topics Presentations State Policies Accommodations Bibliography Teleconferences Tools Related Sites Interpreting Trends in the Performance of Special Education Students

More information

Python Machine Learning

Python Machine Learning Python Machine Learning Unlock deeper insights into machine learning with this vital guide to cuttingedge predictive analytics Sebastian Raschka [ PUBLISHING 1 open source I community experience distilled

More information

Learners Use Word-Level Statistics in Phonetic Category Acquisition

Learners Use Word-Level Statistics in Phonetic Category Acquisition Learners Use Word-Level Statistics in Phonetic Category Acquisition Naomi Feldman, Emily Myers, Katherine White, Thomas Griffiths, and James Morgan 1. Introduction * One of the first challenges that language

More information

Learning Methods in Multilingual Speech Recognition

Learning Methods in Multilingual Speech Recognition Learning Methods in Multilingual Speech Recognition Hui Lin Department of Electrical Engineering University of Washington Seattle, WA 98125 linhui@u.washington.edu Li Deng, Jasha Droppo, Dong Yu, and Alex

More information

STUDIES WITH FABRICATED SWITCHBOARD DATA: EXPLORING SOURCES OF MODEL-DATA MISMATCH

STUDIES WITH FABRICATED SWITCHBOARD DATA: EXPLORING SOURCES OF MODEL-DATA MISMATCH STUDIES WITH FABRICATED SWITCHBOARD DATA: EXPLORING SOURCES OF MODEL-DATA MISMATCH Don McAllaster, Larry Gillick, Francesco Scattone, Mike Newman Dragon Systems, Inc. 320 Nevada Street Newton, MA 02160

More information

AGENDA LEARNING THEORIES LEARNING THEORIES. Advanced Learning Theories 2/22/2016

AGENDA LEARNING THEORIES LEARNING THEORIES. Advanced Learning Theories 2/22/2016 AGENDA Advanced Learning Theories Alejandra J. Magana, Ph.D. admagana@purdue.edu Introduction to Learning Theories Role of Learning Theories and Frameworks Learning Design Research Design Dual Coding Theory

More information

AGS THE GREAT REVIEW GAME FOR PRE-ALGEBRA (CD) CORRELATED TO CALIFORNIA CONTENT STANDARDS

AGS THE GREAT REVIEW GAME FOR PRE-ALGEBRA (CD) CORRELATED TO CALIFORNIA CONTENT STANDARDS AGS THE GREAT REVIEW GAME FOR PRE-ALGEBRA (CD) CORRELATED TO CALIFORNIA CONTENT STANDARDS 1 CALIFORNIA CONTENT STANDARDS: Chapter 1 ALGEBRA AND WHOLE NUMBERS Algebra and Functions 1.4 Students use algebraic

More information

(Includes a Detailed Analysis of Responses to Overall Satisfaction and Quality of Academic Advising Items) By Steve Chatman

(Includes a Detailed Analysis of Responses to Overall Satisfaction and Quality of Academic Advising Items) By Steve Chatman Report #202-1/01 Using Item Correlation With Global Satisfaction Within Academic Division to Reduce Questionnaire Length and to Raise the Value of Results An Analysis of Results from the 1996 UC Survey

More information

Calibration of Confidence Measures in Speech Recognition

Calibration of Confidence Measures in Speech Recognition Submitted to IEEE Trans on Audio, Speech, and Language, July 2010 1 Calibration of Confidence Measures in Speech Recognition Dong Yu, Senior Member, IEEE, Jinyu Li, Member, IEEE, Li Deng, Fellow, IEEE

More information

Modeling function word errors in DNN-HMM based LVCSR systems

Modeling function word errors in DNN-HMM based LVCSR systems Modeling function word errors in DNN-HMM based LVCSR systems Melvin Jose Johnson Premkumar, Ankur Bapna and Sree Avinash Parchuri Department of Computer Science Department of Electrical Engineering Stanford

More information

DOMAIN MISMATCH COMPENSATION FOR SPEAKER RECOGNITION USING A LIBRARY OF WHITENERS. Elliot Singer and Douglas Reynolds

DOMAIN MISMATCH COMPENSATION FOR SPEAKER RECOGNITION USING A LIBRARY OF WHITENERS. Elliot Singer and Douglas Reynolds DOMAIN MISMATCH COMPENSATION FOR SPEAKER RECOGNITION USING A LIBRARY OF WHITENERS Elliot Singer and Douglas Reynolds Massachusetts Institute of Technology Lincoln Laboratory {es,dar}@ll.mit.edu ABSTRACT

More information

Statistical Parametric Speech Synthesis

Statistical Parametric Speech Synthesis Statistical Parametric Speech Synthesis Heiga Zen a,b,, Keiichi Tokuda a, Alan W. Black c a Department of Computer Science and Engineering, Nagoya Institute of Technology, Gokiso-cho, Showa-ku, Nagoya,

More information

10.2. Behavior models

10.2. Behavior models User behavior research 10.2. Behavior models Overview Why do users seek information? How do they seek information? How do they search for information? How do they use libraries? These questions are addressed

More information

Self-Supervised Acquisition of Vowels in American English

Self-Supervised Acquisition of Vowels in American English Self-Supervised Acquisition of Vowels in American English Michael H. Coen MIT Computer Science and Artificial Intelligence Laboratory 32 Vassar Street Cambridge, MA 2139 mhcoen@csail.mit.edu Abstract This

More information

Trend Survey on Japanese Natural Language Processing Studies over the Last Decade

Trend Survey on Japanese Natural Language Processing Studies over the Last Decade Trend Survey on Japanese Natural Language Processing Studies over the Last Decade Masaki Murata, Koji Ichii, Qing Ma,, Tamotsu Shirado, Toshiyuki Kanamaru,, and Hitoshi Isahara National Institute of Information

More information

Radius STEM Readiness TM

Radius STEM Readiness TM Curriculum Guide Radius STEM Readiness TM While today s teens are surrounded by technology, we face a stark and imminent shortage of graduates pursuing careers in Science, Technology, Engineering, and

More information

Measurement. When Smaller Is Better. Activity:

Measurement. When Smaller Is Better. Activity: Measurement Activity: TEKS: When Smaller Is Better (6.8) Measurement. The student solves application problems involving estimation and measurement of length, area, time, temperature, volume, weight, and

More information

arxiv: v1 [math.at] 10 Jan 2016

arxiv: v1 [math.at] 10 Jan 2016 THE ALGEBRAIC ATIYAH-HIRZEBRUCH SPECTRAL SEQUENCE OF REAL PROJECTIVE SPECTRA arxiv:1601.02185v1 [math.at] 10 Jan 2016 GUOZHEN WANG AND ZHOULI XU Abstract. In this note, we use Curtis s algorithm and the

More information

Segregation of Unvoiced Speech from Nonspeech Interference

Segregation of Unvoiced Speech from Nonspeech Interference Technical Report OSU-CISRC-8/7-TR63 Department of Computer Science and Engineering The Ohio State University Columbus, OH 4321-1277 FTP site: ftp.cse.ohio-state.edu Login: anonymous Directory: pub/tech-report/27

More information

Course Law Enforcement II. Unit I Careers in Law Enforcement

Course Law Enforcement II. Unit I Careers in Law Enforcement Course Law Enforcement II Unit I Careers in Law Enforcement Essential Question How does communication affect the role of the public safety professional? TEKS 130.294(c) (1)(A)(B)(C) Prior Student Learning

More information

Constructing a support system for self-learning playing the piano at the beginning stage

Constructing a support system for self-learning playing the piano at the beginning stage Alma Mater Studiorum University of Bologna, August 22-26 2006 Constructing a support system for self-learning playing the piano at the beginning stage Tamaki Kitamura Dept. of Media Informatics, Ryukoku

More information

Lesson M4. page 1 of 2

Lesson M4. page 1 of 2 Lesson M4 page 1 of 2 Miniature Gulf Coast Project Math TEKS Objectives 111.22 6b.1 (A) apply mathematics to problems arising in everyday life, society, and the workplace; 6b.1 (C) select tools, including

More information

VIEW: An Assessment of Problem Solving Style

VIEW: An Assessment of Problem Solving Style 1 VIEW: An Assessment of Problem Solving Style Edwin C. Selby, Donald J. Treffinger, Scott G. Isaksen, and Kenneth Lauer This document is a working paper, the purposes of which are to describe the three

More information

Self-Supervised Acquisition of Vowels in American English

Self-Supervised Acquisition of Vowels in American English Self-Supervised cquisition of Vowels in merican English Michael H. Coen MIT Computer Science and rtificial Intelligence Laboratory 32 Vassar Street Cambridge, M 2139 mhcoen@csail.mit.edu bstract This paper

More information

AP Statistics Summer Assignment 17-18

AP Statistics Summer Assignment 17-18 AP Statistics Summer Assignment 17-18 Welcome to AP Statistics. This course will be unlike any other math class you have ever taken before! Before taking this course you will need to be competent in basic

More information

Building Text Corpus for Unit Selection Synthesis

Building Text Corpus for Unit Selection Synthesis INFORMATICA, 2014, Vol. 25, No. 4, 551 562 551 2014 Vilnius University DOI: http://dx.doi.org/10.15388/informatica.2014.29 Building Text Corpus for Unit Selection Synthesis Pijus KASPARAITIS, Tomas ANBINDERIS

More information

MULTILINGUAL INFORMATION ACCESS IN DIGITAL LIBRARY

MULTILINGUAL INFORMATION ACCESS IN DIGITAL LIBRARY MULTILINGUAL INFORMATION ACCESS IN DIGITAL LIBRARY Chen, Hsin-Hsi Department of Computer Science and Information Engineering National Taiwan University Taipei, Taiwan E-mail: hh_chen@csie.ntu.edu.tw Abstract

More information

Ansys Tutorial Random Vibration

Ansys Tutorial Random Vibration Ansys Tutorial Random Free PDF ebook Download: Ansys Tutorial Download or Read Online ebook ansys tutorial random vibration in PDF Format From The Best User Guide Database Random vibration analysis gives

More information

Speaker Recognition. Speaker Diarization and Identification

Speaker Recognition. Speaker Diarization and Identification Speaker Recognition Speaker Diarization and Identification A dissertation submitted to the University of Manchester for the degree of Master of Science in the Faculty of Engineering and Physical Sciences

More information

Evaluation of a Simultaneous Interpretation System and Analysis of Speech Log for User Experience Assessment

Evaluation of a Simultaneous Interpretation System and Analysis of Speech Log for User Experience Assessment Evaluation of a Simultaneous Interpretation System and Analysis of Speech Log for User Experience Assessment Akiko Sakamoto, Kazuhiko Abe, Kazuo Sumita and Satoshi Kamatani Knowledge Media Laboratory,

More information

Experiments with SMS Translation and Stochastic Gradient Descent in Spanish Text Author Profiling

Experiments with SMS Translation and Stochastic Gradient Descent in Spanish Text Author Profiling Experiments with SMS Translation and Stochastic Gradient Descent in Spanish Text Author Profiling Notebook for PAN at CLEF 2013 Andrés Alfonso Caurcel Díaz 1 and José María Gómez Hidalgo 2 1 Universidad

More information

Edinburgh Research Explorer

Edinburgh Research Explorer Edinburgh Research Explorer Personalising speech-to-speech translation Citation for published version: Dines, J, Liang, H, Saheer, L, Gibson, M, Byrne, W, Oura, K, Tokuda, K, Yamagishi, J, King, S, Wester,

More information

A NOVEL SCHEME FOR SPEAKER RECOGNITION USING A PHONETICALLY-AWARE DEEP NEURAL NETWORK. Yun Lei Nicolas Scheffer Luciana Ferrer Mitchell McLaren

A NOVEL SCHEME FOR SPEAKER RECOGNITION USING A PHONETICALLY-AWARE DEEP NEURAL NETWORK. Yun Lei Nicolas Scheffer Luciana Ferrer Mitchell McLaren A NOVEL SCHEME FOR SPEAKER RECOGNITION USING A PHONETICALLY-AWARE DEEP NEURAL NETWORK Yun Lei Nicolas Scheffer Luciana Ferrer Mitchell McLaren Speech Technology and Research Laboratory, SRI International,

More information

ADVANCES IN DEEP NEURAL NETWORK APPROACHES TO SPEAKER RECOGNITION

ADVANCES IN DEEP NEURAL NETWORK APPROACHES TO SPEAKER RECOGNITION ADVANCES IN DEEP NEURAL NETWORK APPROACHES TO SPEAKER RECOGNITION Mitchell McLaren 1, Yun Lei 1, Luciana Ferrer 2 1 Speech Technology and Research Laboratory, SRI International, California, USA 2 Departamento

More information

Using EEG to Improve Massive Open Online Courses Feedback Interaction

Using EEG to Improve Massive Open Online Courses Feedback Interaction Using EEG to Improve Massive Open Online Courses Feedback Interaction Haohan Wang, Yiwei Li, Xiaobo Hu, Yucong Yang, Zhu Meng, Kai-min Chang Language Technologies Institute School of Computer Science Carnegie

More information

Introduction to Information System

Introduction to Information System Spring Quarter 2015-2016 Meeting day/time: N/A at Online Campus (Distance Learning). Location: Use D2L.depaul.edu to access the course and course materials Instructor: Miranda Standberry-Wallace Office:

More information

EQuIP Review Feedback

EQuIP Review Feedback EQuIP Review Feedback Lesson/Unit Name: On the Rainy River and The Red Convertible (Module 4, Unit 1) Content Area: English language arts Grade Level: 11 Dimension I Alignment to the Depth of the CCSS

More information

SARDNET: A Self-Organizing Feature Map for Sequences

SARDNET: A Self-Organizing Feature Map for Sequences SARDNET: A Self-Organizing Feature Map for Sequences Daniel L. James and Risto Miikkulainen Department of Computer Sciences The University of Texas at Austin Austin, TX 78712 dljames,risto~cs.utexas.edu

More information

UTD-CRSS Systems for 2012 NIST Speaker Recognition Evaluation

UTD-CRSS Systems for 2012 NIST Speaker Recognition Evaluation UTD-CRSS Systems for 2012 NIST Speaker Recognition Evaluation Taufiq Hasan Gang Liu Seyed Omid Sadjadi Navid Shokouhi The CRSS SRE Team John H.L. Hansen Keith W. Godin Abhinav Misra Ali Ziaei Hynek Bořil

More information

LEGO MINDSTORMS Education EV3 Coding Activities

LEGO MINDSTORMS Education EV3 Coding Activities LEGO MINDSTORMS Education EV3 Coding Activities s t e e h s k r o W t n e d Stu LEGOeducation.com/MINDSTORMS Contents ACTIVITY 1 Performing a Three Point Turn 3-6 ACTIVITY 2 Written Instructions for a

More information

Unit Selection Synthesis Using Long Non-Uniform Units and Phonemic Identity Matching

Unit Selection Synthesis Using Long Non-Uniform Units and Phonemic Identity Matching Unit Selection Synthesis Using Long Non-Uniform Units and Phonemic Identity Matching Lukas Latacz, Yuk On Kong, Werner Verhelst Department of Electronics and Informatics (ETRO) Vrie Universiteit Brussel

More information