VARIATION BETWEEN PALATAL VOICED FRICATIVE AND PALATAL APPROXIMANT IN URDU SPOKEN LANGUAGE

Similar documents
Quarterly Progress and Status Report. VCV-sequencies in a preliminary text-to-speech system for female speech

Linguistics 220 Phonology: distributions and the concept of the phoneme. John Alderete, Simon Fraser University

Universal contrastive analysis as a learning principle in CAPT

The Perception of Nasalized Vowels in American English: An Investigation of On-line Use of Vowel Nasalization in Lexical Access

CLASSIFICATION OF PROGRAM Critical Elements Analysis 1. High Priority Items Phonemic Awareness Instruction

Mandarin Lexical Tone Recognition: The Gating Paradigm

SEGMENTAL FEATURES IN SPONTANEOUS AND READ-ALOUD FINNISH

Quarterly Progress and Status Report. Voiced-voiceless distinction in alaryngeal speech - acoustic and articula

Speech Recognition using Acoustic Landmarks and Binary Phonetic Feature Classifiers

Unvoiced Landmark Detection for Segment-based Mandarin Continuous Speech Recognition

Consonants: articulation and transcription

Phonological and Phonetic Representations: The Case of Neutralization

Speech Segmentation Using Probabilistic Phonetic Feature Hierarchy and Support Vector Machines

Arabic Orthography vs. Arabic OCR

Learning Methods in Multilingual Speech Recognition

1. REFLEXES: Ask questions about coughing, swallowing, of water as fast as possible (note! Not suitable for all

Design Of An Automatic Speaker Recognition System Using MFCC, Vector Quantization And LBG Algorithm

Segregation of Unvoiced Speech from Nonspeech Interference

STUDIES WITH FABRICATED SWITCHBOARD DATA: EXPLORING SOURCES OF MODEL-DATA MISMATCH

Intra-talker Variation: Audience Design Factors Affecting Lexical Selections

ELA/ELD Standards Correlation Matrix for ELD Materials Grade 1 Reading

Phonological Processing for Urdu Text to Speech System

Florida Reading Endorsement Alignment Matrix Competency 1

Phonetics. The Sound of Language

Rachel E. Baker, Ann R. Bradlow. Northwestern University, Evanston, IL, USA

have to be modeled) or isolated words. Output of the system is a grapheme-tophoneme conversion system which takes as its input the spelling of words,

Speech Recognition at ICSI: Broadcast News and beyond

Speech Synthesis in Noisy Environment by Enhancing Strength of Excitation and Formant Prominence

A Cross-language Corpus for Studying the Phonetics and Phonology of Prominence

Coast Academies Writing Framework Step 4. 1 of 7

1 st Quarter (September, October, November) August/September Strand Topic Standard Notes Reading for Literature

Proceedings of Meetings on Acoustics

Rhythm-typology revisited.

Speaker Recognition. Speaker Diarization and Identification

Learners Use Word-Level Statistics in Phonetic Category Acquisition

age, Speech and Hearii

THE RECOGNITION OF SPEECH BY MACHINE

First Grade Curriculum Highlights: In alignment with the Common Core Standards

English Language and Applied Linguistics. Module Descriptions 2017/18

AUTOMATIC DETECTION OF PROLONGED FRICATIVE PHONEMES WITH THE HIDDEN MARKOV MODELS APPROACH 1. INTRODUCTION

Speaker recognition using universal background model on YOHO database

An Acoustic Phonetic Account of the Production of Word-Final /z/s in Central Minnesota English

Program Matrix - Reading English 6-12 (DOE Code 398) University of Florida. Reading

Analysis of Emotion Recognition System through Speech Signal Using KNN & GMM Classifier

Demonstration of problems of lexical stress on the pronunciation Turkish English teachers and teacher trainees by computer

Pobrane z czasopisma New Horizons in English Studies Data: 18/11/ :52:20. New Horizons in English Studies 1/2016

Class-Discriminative Weighted Distortion Measure for VQ-Based Speaker Identification

Opportunities for Writing Title Key Stage 1 Key Stage 2 Narrative

Body-Conducted Speech Recognition and its Application to Speech Support System

Eli Yamamoto, Satoshi Nakamura, Kiyohiro Shikano. Graduate School of Information Science, Nara Institute of Science & Technology

Taught Throughout the Year Foundational Skills Reading Writing Language RF.1.2 Demonstrate understanding of spoken words,

Perceptual scaling of voice identity: common dimensions for different vowels and speakers

Speech Emotion Recognition Using Support Vector Machine

Books Effective Literacy Y5-8 Learning Through Talk Y4-8 Switch onto Spelling Spelling Under Scrutiny

To appear in the Proceedings of the 35th Meetings of the Chicago Linguistics Society. Post-vocalic spirantization: Typology and phonetic motivations

Understanding and Supporting Dyslexia Godstone Village School. January 2017

The NICT/ATR speech synthesis system for the Blizzard Challenge 2008

Automatic segmentation of continuous speech using minimum phase group delay functions

Houghton Mifflin Reading Correlation to the Common Core Standards for English Language Arts (Grade1)

ABSTRACT. Some children with speech sound disorders (SSD) have difficulty with literacyrelated

A comparison of spectral smoothing methods for segment concatenation based speech synthesis

CEFR Overall Illustrative English Proficiency Scales

REVIEW OF CONNECTED SPEECH

SOUND STRUCTURE REPRESENTATION, REPAIR AND WELL-FORMEDNESS: GRAMMAR IN SPOKEN LANGUAGE PRODUCTION. Adam B. Buchwald

Problems of the Arabic OCR: New Attitudes

Fix Your Vowels: Computer-assisted training by Dutch learners of Spanish

The analysis starts with the phonetic vowel and consonant charts based on the dataset:

Cambridgeshire Community Services NHS Trust: delivering excellence in children and young people s health services

**Note: this is slightly different from the original (mainly in format). I would be happy to send you a hard copy.**

The Indian English of Tibeto-Burman language speakers*

California Department of Education English Language Development Standards for Grade 8

Markedness and Complex Stops: Evidence from Simplification Processes 1. Nick Danis Rutgers University

Christine Mooshammer, IPDS Kiel, Philip Hoole, IPSK München, Anja Geumann, Dublin

Stages of Literacy Ros Lugg

Language Acquisition by Identical vs. Fraternal SLI Twins * Karin Stromswold & Jay I. Rifkin

Journal of Phonetics

A Neural Network GUI Tested on Text-To-Phoneme Mapping

Phonetic- and Speaker-Discriminant Features for Speaker Recognition. Research Project

Parallel Evaluation in Stratal OT * Adam Baker University of Arizona

THE PENNSYLVANIA STATE UNIVERSITY SCHREYER HONORS COLLEGE DEPARTMENT OF MATHEMATICS ASSESSING THE EFFECTIVENESS OF MULTIPLE CHOICE MATH TESTS

Phonology Revisited: Sor3ng Out the PH Factors in Reading and Spelling Development. Indiana, November, 2015

The Effect of Discourse Markers on the Speaking Production of EFL Students. Iman Moradimanesh

OCR for Arabic using SIFT Descriptors With Online Failure Prediction

Richardson, J., The Next Step in Guided Writing, Ohio Literacy Conference, 2010

Dyslexia/dyslexic, 3, 9, 24, 97, 187, 189, 206, 217, , , 367, , , 397,

Radical CV Phonology: the locational gesture *

Modeling function word errors in DNN-HMM based LVCSR systems

Voice conversion through vector quantization

Consonant-Vowel Unity in Element Theory*

On the Formation of Phoneme Categories in DNN Acoustic Models

The Bruins I.C.E. School

Modeling function word errors in DNN-HMM based LVCSR systems

English for Life. B e g i n n e r. Lessons 1 4 Checklist Getting Started. Student s Book 3 Date. Workbook. MultiROM. Test 1 4

Speaker Identification by Comparison of Smart Methods. Abstract

TEKS Comments Louisiana GLE

Lecturing Module

Different Task Type and the Perception of the English Interdental Fricatives

AGS THE GREAT REVIEW GAME FOR PRE-ALGEBRA (CD) CORRELATED TO CALIFORNIA CONTENT STANDARDS

On Developing Acoustic Models Using HTK. M.A. Spaans BSc.

SARDNET: A Self-Organizing Feature Map for Sequences

Transcription:

46 VARIATION BETWEEN PALATAL VOICED FRICATIVE AND PALATAL APPROXIMANT IN URDU SPOKEN LANGUAGE SHERAZ BASHIR 1. INTRODUCTION Urdu is the national language of Pakistan. It has most of the common vocalic and consonantal sounds present in other languages. There is not much research done on distribution of Urdu sounds. There is no official record of allophones, which exists in Urdu, and consequently no data available about the distribution of Urdu allophones. This paper is proposed to find the distribution between the palatal voiced consonants (palatal voiced fricative [Z] and palatal approximant [j]). 2. LITERATURE REVIEW This part explains the conditions necessary for two sounds to be allophones. And if they are allophones than how their distribution can be determined. complementary distribution (Laver, 1994, pg. 38). 2.2 Allophones Allophones are the predictable variants of a phoneme. The different phones that represent or are derived from one phone are called allophones of the phoneme. Allophones have to be phonetically similar like [h] and [ŋ] (Clark and Yallop, 1992, pg.127-128). Now consider phoneme /p/ as an example. English has two P sounds [p] and [p h ] (i.e. unaspirated and aspirated). These two sounds are phonetically similar. Native speakers of English tells us immediately that [p] and [p h ] are actually just two varieties of the same sound, that is, they are allophones of /p/. Concrete evidence for this intuition would be the complete absence in English of any minimal pairs differing on [p] and [p h ]. These allophones of /p/ can be visually represented as 2.1 Phonological Distribution phoneme /p/ For any given consonant or vowel, the range of possible locations is constrained by the phonology of the language. This range of permitted locations constitutes the phonological distribution of that consonant or vowel. The distributional range of contextual occurrences of a given consonant or vowel in an accent is said to constitute its phonotactic range. The distribution of two consonants is often compared, for purposes of describing the phonological function of the segments in question. When two consonants or two vowels share the same distribution exactly (that is, when they can be said to have exactly the same phonotactic range), they are said to be in parallel distribution. When they share no locations, they are said to be in allophones [p] [p h ] 2.3 Complementary Distribution When two or more sounds never occur in the same environment, they are said to be in complementary distribution (e.g. aspirated voiceless stops vs. unaspirated voiceless stops). When two or more sounds are in complementary distribution, we know which sound occurs in which environment. That is, the distribution of those sounds is predictable. For those sounds in complementary distribution to be allophones of a single phoneme, they have to be

47 phonetically similar. (Clark and Yallop, 1992, pg. 130). In fact, when we look carefully, we discover something interesting about the distribution of [p] and [ph] in English. Consider the listings below: [p h ] [p] port sport peace soup append apple By inspecting the two listings above, we discover the condition under which the aspirated variety of /p/ appears: Only when /p/ appears as the first sound in a stressed syllable does aspiration show up. In all other positions we find the usual, unaspirated variety of /p/. More generally, when we find that allophones of a single phoneme appear in different environments, we speak of their being in complementary distribution. The allophone that appears in the larger number of positions is then said to meet the elsewhere condition. We can thus add more information to the visual representation we made before for [p] and [p h ]. phoneme /p/ allophones [p] [p h ] elsewhere Allophones of a single phoneme in complementary distribution are thus ones where the phonetic environments are not the same. The allophone that appears in the greatest number of environments meets the elsewhere condition. 2.4 Free Variation first sound in stressed syllables Sometimes we observe allophones in exactly the same environment. A good example of this phenomenon involves, once again, our model-organism phoneme /p/. Before we noted that this phoneme has two allophones, the unaspirated [p] and the aspirated [p h ]. In both of these sounds, the blockage of the airflow is released before the following sound. There is, however, another allophone of the phoneme /p/ in English, however. This one is the unreleased variety, which we represent with a diacritic: [p ]. This allophone of /p/ appears when we do not release the blockage of the airflow (i.e., lips not open) at the end of its articulation. Consider the following: a. heap [hiyp ] or [hiyp] b. rope [rowp ] or [rowp] c. laptop [læp top] d. popcorn [p h op krn] e. hippy [hıpiy] As these data indicate, [p ] appears at the end of words, including words that are used together with other words to form what we call compounds. Crucial, however, are the examples in (a) and (b) above, which show that the allophones [p] and [p ] can also appear in exactly the same environment. When allophones of a single phoneme appear in exactly the same environment in this way, we speak of overlapping distribution (as opposed to complementary distribution), and we conclude that these allophones are in free variation. 3. PROBLEM STATEMENT Palatal voiced fricative and palatal approximant are allophones of each other because they are phonetically similar and there are no existing minimal pairs for both the consonants (These conditions are stated in the previous section.) The question is about its phonological distribution, which is discussed in the paper with the help of experiments. The scope of this paper is to show whether these two consonants share the same distribution exactly (that is, when they can be said to have exactly the same phonotactic range) or they share no locations.

48 4. METHODOLOGY According to (Fromkin V. 2000, pg.537) here is a procedure for examining the target sounds in the data. For each sound, we compile all of the contexts in which it appears. Since the conditioning environment for allophones is usually (though not always) found in the immediately preceding and following segments. So the Urdu into English dictionary (Qureshi B. 1992) was used to find all the words of Urdu containing [j] or [Z] in the syllable initial, final or at both positions. Than those words were selected which had the same syllable context or at least the same vowel following or preceding both the target consonants as displayed in 4.1. For example the words d.var (meaning root ) and j.dq.be.za (meaning luminous hand ) have the same vowel in the first syllable following the target consonants. (4.1) ZaZ nonsense t.jag.na renounce Z d.var root j.dq.be.za luminous hand Zo.li.da confused k.jo.ti mixture 4.1 Speakers Recordings were obtained from 5 speakers, all males. All subjects were native speakers of Urdu and they knew about the purpose of the experiment. All the speakers belonged to age group of 20 to 22 years and they were normal in speech and hearing. 4.2 Recording Procedure All the speakers were given words to be recorded written in Urdu script and transcribed phonetically so that they pronounce them correctly. All the words were written in a sequential manner as shown in 4.2. All speakers were asked to rehearse on these words before recording so that they don t fumble any word during recording. (4.2) ZaZ Z dvar Zolida t jagna j dqbeza k joti The speakers were also asked to speak at a natural rate and loudness. After the speakers were ready their voice was recorded using the ESPS (Entropic Signal Processing Software) version 5.3. The same sequence of words was recorded three times by each speaker in order to have better results. Recorded speech was than filtered using the lowpass filter of Entropic to remove the noise added by the recording machine. 4.3 Analysis Recorded speech was analyzed using the xwaves (Speech data Visualizing facility by Entropic). The time waveform was marked between word s start and end places using the xmark facility in xwaves. Then the spectrogram of each marked word was drawn using the spectrogram facility of the xwaves. Then the target consonant was marked in the spectrogram of the word and align & rescale facility was used to align the selection with the time waveform. 4.4 Measurements The formant utility of xwaves was used to calculate F0 and the Formants of the recorded speech. Than the cursor was moved between the selected area and the value of fundamental frequency and the first formant was recorded at that point. Spectrograms are not ideal for examination of the detailed spectral features of fricatives. For this purpose, it is preferable to use spectra determined by methods such as DFT or LPC (Kent & Read, 1992, pg.123). So DFT was taken to calculate the amplitude of the F0 and Formants. DFT was taken of the selected signal using xspectrum utility of xwaves. The parameters were set according to 4.3. The amplitudes of the F0 and first formant (calculated previously) were recorded. The ratios of both amplitudes were calculated to find whether both the target consonants show similar acoustic behavior in some context or are they totally different from each other.

49 (4.3.) Parameter (value) Analysis type (DFT) Window type (Hanning) Window limits from cursor (+ - 2 size / 2) Reemphasis coeff. (0.000000) Integration coefficient (0.9900) Horizontal cursor (Off) Reticle (On) Harmonic Cursors (Off) Formants (On) Plot scale (log pwr (db)) Filtering (Lowpass) 5. RESULTS After compiling the words collected by the dictionary all of the contexts for the target sounds are listed in the chart 5.1. [ syllable and ] syllable represents the syllable start and syllable end respectively. The line represents the location of the target consonant. (5.1) [ ] [j] / [ syllable e / [ syllable / [ syllable A / [ syllable A / [ syllable / [ syllable e / [ syllable o / [ syllable o / ] syllable / [ syllable u / [ syllable i / [ syllable I / [ syllable ç / [ syllable u) / [ syllable a) Words Containing Target Consonant F0 (db) F1 (db) a 83 76.66667 d.var 81.33333 78.33333 o.li.da 85.66667 75.66667 t.ja.na 85.66667 99.66667 j.d.be.za 84.66667 85.33333 k.jo.ti 88 92.33333 TABLE 1.1 Mean of the three repetitions recorded by the speaker MS. The results show that the intensity of the first formant is higher than the intensity of the fundamental frequency in case of the consonant [j] and is lower in the case of consonant [Z]. To find the relations between the two consonants, ratio between the intensities of the two frequencies was calculated and the mean of all the resultant ratios is tabulated in Table 1.2. Words Containing Mean of ratios Target Consonant (F0/F1) ZaZ 1.102055 Z d9.var 1.044995 Zo.li.d9a 1.073304 t.jag.na 0.846639 The purpose of the experiment was to measure the relationship between intensity of the fundamental frequency and the first formant of target consonant. The results shown in Table 1.1 are the mean of the three repetitions taken by the speech recorded by the speaker MS. They are somewhere in between all other readings tabulated in the Appendix A. j.d9.be.za 0.979536 k.jo.ti 0.879285 TABLE 1.2 Mean of all ratios calculated on the data of three repetitions by each of the five speakers. 6. DISCUSSION The results can be described both phonetically and phonemically according to the data available to us.

50 6.1 Phonetic Description Mean of the ratios between intensities of the fundaments frequency and the first formant for the target consonant [Z] are higher than one as shown in Table 1.2. This indicates that denominator (intensity of the first formant) is less than the numerator (intensity of the fundamental formant). So [Z] is more fricative like. But the same ratios for the target consonant [j] are less than one. This indicates that denominator (intensity of the first formant) is greater than the numerator (intensity of the fundamental formant). So [j] is more approximant like. As there is never a case when one of them converges to the other one, which proves that they are phonetically distinct and they are not allophones but different phonemes. Kent & Read. 1992. The Acoustic Analysis of Speech. Singular Publishing Group, INC. San Diego, California. Laver. 1994. Principles of Phonetics. Cambridge University Press, Cambridge U.K. Qureshi B. 1992. Standard Twentieth Century Dictionary (Urdu into English). Educational Publishing House, Delhi. 6.2 Phonemic Description Analysis of the data collected from the dictionary (Qureshi B. 1992.) for the two target sounds (listed in Appendix B) shows that there are no minimal or near minimal pairs existing for the two target sounds in Urdu, which indicates that they are allophones in Urdu. They share same places like in the words Zo.li.d9a and k.jo.ti so they are not in complementary distribution but they are in free variation. 6.3 Concluding the discussion Phonemic analysis cannot be relied upon because Urdu language has very few words containing the phone [Z], which is the main reason for non-existing minimal pairs. Relying on the phonetic analysis it can be concluded that the two target sounds [Z] and [j] are phones of two different phonemes /Z and /j/ in Urdu. 7. REFERENCES Clark & Yallop. 1992. An Introduction To Phonetics And Phonology. Dotesios, Wiltshire Fromkin V. (ed.). 2000 Linguistics. Blackwell Publishers Inc. Massachusetts USA.