Pitch shifts do not alter the amount of song produced

Similar documents
Build on students informal understanding of sharing and proportionality to develop initial fraction concepts.

Grade 6: Correlated to AGS Basic Math Skills

Quarterly Progress and Status Report. VCV-sequencies in a preliminary text-to-speech system for female speech

Mathematics Success Level E

Grade 2: Using a Number Line to Order and Compare Numbers Place Value Horizontal Content Strand

Class-Discriminative Weighted Distortion Measure for VQ-Based Speaker Identification

Speech Recognition at ICSI: Broadcast News and beyond

AUTOMATIC DETECTION OF PROLONGED FRICATIVE PHONEMES WITH THE HIDDEN MARKOV MODELS APPROACH 1. INTRODUCTION

Lecture 2: Quantifiers and Approximation

Speech Synthesis in Noisy Environment by Enhancing Strength of Excitation and Formant Prominence

Design Of An Automatic Speaker Recognition System Using MFCC, Vector Quantization And LBG Algorithm

Mandarin Lexical Tone Recognition: The Gating Paradigm

OPTIMIZATINON OF TRAINING SETS FOR HEBBIAN-LEARNING- BASED CLASSIFIERS

Speech Segmentation Using Probabilistic Phonetic Feature Hierarchy and Support Vector Machines

Statewide Framework Document for:

Modeling function word errors in DNN-HMM based LVCSR systems

Contents. Foreword... 5

Instructor: Mario D. Garrett, Ph.D. Phone: Office: Hepner Hall (HH) 100

Pre-Algebra A. Syllabus. Course Overview. Course Goals. General Skills. Credit Value

Analysis of Emotion Recognition System through Speech Signal Using KNN & GMM Classifier

Segregation of Unvoiced Speech from Nonspeech Interference

J j W w. Write. Name. Max Takes the Train. Handwriting Letters Jj, Ww: Words with j, w 321

Introducing the New Iowa Assessments Mathematics Levels 12 14

Proceedings of Meetings on Acoustics

Large Kindergarten Centers Icons

Mathematics Success Grade 7

Many instructors use a weighted total to calculate their grades. This lesson explains how to set up a weighted total using categories.

Human Emotion Recognition From Speech

Correspondence between the DRDP (2015) and the California Preschool Learning Foundations. Foundations (PLF) in Language and Literacy

Common Core State Standards

Understanding and Interpreting the NRC s Data-Based Assessment of Research-Doctorate Programs in the United States (2010)

AGS THE GREAT REVIEW GAME FOR PRE-ALGEBRA (CD) CORRELATED TO CALIFORNIA CONTENT STANDARDS

FOR TEACHERS ONLY. The University of the State of New York REGENTS HIGH SCHOOL EXAMINATION PHYSICAL SETTING/PHYSICS

United States Symbols First Grade By Rachel Horen. Featured Selection:

Dyslexia/dyslexic, 3, 9, 24, 97, 187, 189, 206, 217, , , 367, , , 397,

Unraveling symbolic number processing and the implications for its association with mathematics. Delphine Sasanguie

Modeling function word errors in DNN-HMM based LVCSR systems

GDP Falls as MBA Rises?

Massachusetts Department of Elementary and Secondary Education. Title I Comparability

Higher Education Six-Year Plans

Curriculum Design Project with Virtual Manipulatives. Gwenanne Salkind. George Mason University EDCI 856. Dr. Patricia Moyer-Packenham

Perceptual scaling of voice identity: common dimensions for different vowels and speakers

THE RECOGNITION OF SPEECH BY MACHINE

CONSTRUCTION OF AN ACHIEVEMENT TEST Introduction One of the important duties of a teacher is to observe the student in the classroom, laboratory and

Multiplication of 2 and 3 digit numbers Multiply and SHOW WORK. EXAMPLE. Now try these on your own! Remember to show all work neatly!

STUDENT MOODLE ORIENTATION

End-of-Module Assessment Task

WHEN THERE IS A mismatch between the acoustic

Algebra 1, Quarter 3, Unit 3.1. Line of Best Fit. Overview

1. REFLEXES: Ask questions about coughing, swallowing, of water as fast as possible (note! Not suitable for all

Learning Structural Correspondences Across Different Linguistic Domains with Synchronous Neural Language Models

9.85 Cognition in Infancy and Early Childhood. Lecture 7: Number

Speech Recognition using Acoustic Landmarks and Binary Phonetic Feature Classifiers

STT 231 Test 1. Fill in the Letter of Your Choice to Each Question in the Scantron. Each question is worth 2 point.

WiggleWorks Software Manual PDF0049 (PDF) Houghton Mifflin Harcourt Publishing Company

Teacher Quality and Value-added Measurement

Creating a Test in Eduphoria! Aware

International Journal of Computational Intelligence and Informatics, Vol. 1 : No. 4, January - March 2012

Functional Skills Mathematics Level 2 assessment

OCR for Arabic using SIFT Descriptors With Online Failure Prediction

Using Web Searches on Important Words to Create Background Sets for LSI Classification

A comparison of spectral smoothing methods for segment concatenation based speech synthesis

Reflective Teaching KATE WRIGHT ASSOCIATE PROFESSOR, SCHOOL OF LIFE SCIENCES, COLLEGE OF SCIENCE

Noise-Adaptive Perceptual Weighting in the AMR-WB Encoder for Increased Speech Loudness in Adverse Far-End Noise Conditions

Content Language Objectives (CLOs) August 2012, H. Butts & G. De Anda

Appendix L: Online Testing Highlights and Script

Quantitative analysis with statistics (and ponies) (Some slides, pony-based examples from Blase Ur)

Ricopili: Postimputation Module. WCPG Education Day Stephan Ripke / Raymond Walters Toronto, October 2015

THE PENNSYLVANIA STATE UNIVERSITY SCHREYER HONORS COLLEGE DEPARTMENT OF MATHEMATICS ASSESSING THE EFFECTIVENESS OF MULTIPLE CHOICE MATH TESTS

Ohio s Learning Standards-Clear Learning Targets

Georgia Department of Education Georgia Standards of Excellence Framework GSE Sophisticated Shapes Unit 1

Rhythm-typology revisited.

Trend Survey on Japanese Natural Language Processing Studies over the Last Decade

Genevieve L. Hartman, Ph.D.

Characteristics of Functions

Speaker Identification by Comparison of Smart Methods. Abstract

First Grade Standards

Focus of the Unit: Much of this unit focuses on extending previous skills of multiplication and division to multi-digit whole numbers.

Sight Word Assessment

Introduction to Causal Inference. Problem Set 1. Required Problems

Acoustic correlates of stress and their use in diagnosing syllable fusion in Tongan. James White & Marc Garellek UCLA

Extending Place Value with Whole Numbers to 1,000,000

Using Proportions to Solve Percentage Problems I

GOLD Objectives for Development & Learning: Birth Through Third Grade

Left, Left, Left, Right, Left

Speaker recognition using universal background model on YOHO database

Backwards Numbers: A Study of Place Value. Catherine Perez

EDCI 699 Statistics: Content, Process, Application COURSE SYLLABUS: SPRING 2016

GCSE. Mathematics A. Mark Scheme for January General Certificate of Secondary Education Unit A503/01: Mathematics C (Foundation Tier)

PROJECT MANAGEMENT AND COMMUNICATION SKILLS DEVELOPMENT STUDENTS PERCEPTION ON THEIR LEARNING

Learning Disability Functional Capacity Evaluation. Dear Doctor,

Speaker Recognition. Speaker Diarization and Identification

Florida Reading Endorsement Alignment Matrix Competency 1

Session 2B From understanding perspectives to informing public policy the potential and challenges for Q findings to inform survey design

Unvoiced Landmark Detection for Segment-based Mandarin Continuous Speech Recognition

Increasing Student Engagement

12- A whirlwind tour of statistics

6 Financial Aid Information

GCSE Mathematics B (Linear) Mark Scheme for November Component J567/04: Mathematics Paper 4 (Higher) General Certificate of Secondary Education

Objectives. Chapter 2: The Representation of Knowledge. Expert Systems: Principles and Programming, Fourth Edition

Transcription:

Pitch shifts do not alter the amount of song produced Supplementary Figure 1: The number of syllables produced does not change significantly over the course of the shift epoch. a, Number of syllables produced during each day's analysis window (10am-noon) during the baseline and shift epochs. Color and symbol shape conventions as in Figure 3a in the main text. b, Mean +/- S.D. number of syllables, combined across experiments. No significant correlation between time and the amount of singing was observed (dashed red line, p=0.54).

Pitch shifts do not affect the syntax of song We compared the syntax of song before and 14 days after the onset of pitch shifts to determine whether shifted auditory feedback led to changes in the ordering of song syllables. We found that the gross structure of song was not altered by pitch shifts, and that when more precise measures of syntax were used there were no greater changes in response to pitch shifts than were present in control conditions. As shown in Supplementary Figure 2, we constructed a syllable transition matrix for each bird's song both in the baseline epoch and at the end of the shift epoch. Each entry in the matrix describes the probability of one syllable following another. For example, in the baseline epoch in the example shown below, probability of syllable 'B' being followed by syllable 'C' was 0.61. We then subjected the transition matrices to two different analyses. To asses any changes to the gross structure of song, we first asked whether the most common (or "dominant") transition in the baseline condition was also the dominant transition following adaptation to a pitch shift. In each matrix shown below, the dominant transition (i.e. the transition in each column that has the highest probability) is shown in bold. We found that in all cases (45/45 syllables analyzed) the dominant transition in the baseline epoch was also the dominant transition following the feedback shift. This result indicates that on a gross level song syntax is not changed as a result of shifts in the pitch of auditory feedback. To investigate subtler syntax changes, we also compared the prevalence of changes in transition probability in control birds to that observed in birds experiencing pitch shifts. To do this, we asked whether each transition probability (e.g. each entry in the transition matrices shown below) changed significantly between the baseline and shift epochs (using a z-test for proportions at p<0.05 and a Bonferroni correction for multiple comparisons), considering every case in which a transition had a nonzero probability in either epoch. In a group of control birds that did not wear headphones, we found that 3% (3/106) of transition probabilities changed over a 14-day period. This proportion was not significantly different from that observed in pitch-shifted birds (p=0.46, z-test for proportions), in which 6% (7/124) of transition probabilities changed over the 14-day shift epoch. We therefore conclude that syntax is not significantly affected by the shifted auditory feedback used in our experiments. Finally, among the 7 significant changes in transition probability, we asked whether a significant majority reflected increases or decreases in sequence entropy. Deafening adult birds and introducing auditory perturbations have both been shown to increase the entropy of syllable transitions (Okanoya and Yamaguchi, 1997; Woolley and Rubel, 1997; Sakata and Brainard, 2006). For each transition, if the quantity [-p shift log(p shift )] [-p baseline log(p baseline )] was greater than zero, then the entropy of that transition increased, and values less than zero represent decreases in entropy. We found that 29% (2/7) of changes in transition probabilities represented entropy increases and 71% (5/7) were entropy decreases. The

ratio of increases to decreases was not significantly different from equality (chi-squared test, p=0.26). Supplementary Figure 2: Technique for quantifying changes in syntax. a, Segment of song showing the 8 syllables (IABCDEFG) produced by one of the birds in our study. b, Matrices describing the probability of transitions between syllables in the baseline epoch (left) and the last day of the shift epoch (right). The row and column of each matrix describes the probability with which the syllable indicated at the top of each column will be followed by the syllable indicated to the left of each row. For example, in the baseline epoch syllable "A" has a probability of 0.41 of being followed by syllable "B". In each column, the transition with the greatest probability is shown in boldface type. References: Okanoya, K. & Yamaguchi, A. Adult Bengalese finches (Lonchura striata var. domestica) require real-time auditory feedback to produce normal song syntax. Journal of Neurobiology 33, 343-356 (1997).

Sakata JT, Brainard MS Real-time contributions of auditory feedback to avian vocal motor control. J Neurosci 26:9619-9628 (2006). Woolley, S. M. & Rubel, E. W. Bengalese finches Lonchura Striata domestica depend upon auditory feedback for the maintenance of adult song. J Neurosci 17, 6380-6390 (1997).

Pitch shifts do not consistently change spectral emphasis The analysis described in the main text shows that birds shift the fundamental frequency of syllables in the direction opposite the imposed feedback shift. However, birds might also change the relative acoustic power of each syllable's harmonics ("spectral emphasis"), presumably by changing the filtering properties of the upper vocal tract (for example, by emphasizing the higher harmonics in order to compensate for a downshift in the pitch of auditory feedback). Our technique for measuring such changes is illustrated in Supplementary Figure 3 below. For each spectral frame, we computed the mean normalized power spectrum both during the baseline epoch and at the end of the shift epoch (black and red lines, Supplementary Fig. 3b). We then computed the ratio of the peak power at the highest and lowest measurable harmonics (dashed lines in Supplementary Fig. 3b). The example in Supplementary Figure 3 shows the change in mean power spectrum before and two weeks after the onset of a 100 cent downward shift in the pitch of auditory feedback. The ratio of peak power at the high and low harmonics increased 26% over the shift epoch, corresponding to an increased emphasis on the high harmonic. This datum is indicated with a red arrow in Supplementary Figure 3c. When this analysis was performed for all syllables, we found no significant difference in the change in spectral emphasis following upward or downward shifts in the pitch of auditory feedback (p=0.16, 1-tailed t-test). Together with our demonstration of significant adaptive changes in the pitch of song syllables, this result suggests that our feedback manipulations primarily drive changes in syringeal control as opposed to changes in upper vocal tract filtering.

Supplementary Figure 3: Changes in spectral emphasis. a, Spectrogram of an example syllable. The black arrowhead indicates the spectral frame (measurement time) used to calculate the power spectra. b, Mean normalized power spectra of the example syllable in the baseline epoch (black) and during shift days 12-14 during a downward shift in the pitch of auditory feedback (red). As described in the text, we quantified the ratio of peak power (dashed lines) at the highest and lowest measurable harmonic (ratios shown in inset) and then quantified the change in this measure between the baseline and shift epochs. In this example, the "harmonic ratio" increased by 26%, corresponding to an increase in power at the high harmonic. c, Change in harmonic ratio for all spectral features, combined across all downward (red) and upward (blue) shifts in the pitch of auditory feedback. No significant difference was found between the effects of upward and downward shifts (p=0.16, 1-tailed t-test), and the mean changes in both groups (dashed red and blue lines) were not significantly different from zero (p>0.12 in both cases, 1-tailed t-tests). The red arrow indicates the value of the example shown in b. Plotting conventions as in Figure 3b in the main text.

Fundamental frequency does not predict learning across harmonic features Supplementary Figure 4: Adaptive pitch change as a function of fundamental frequency. For each spectral frame, we plot the mean fundamental frequency in the baseline condition against the adaptive change in pitch (i.e. change in the direction opposite the imposed pitch shift). Symbol shapes identify individual birds as in Figure 3a in the main text. Data from upward and downward shifts in the pitch of auditory feedback (+/- 100 cents) are plotted in blue and red, respectively. There was no significant correlation between fundamental frequency and adaptive pitch change when data from upward and downward shifts were pooled (dashed black line, p=0.84) or when data from upward and downward shifts were considered separately (p=0.22 and p=0.96, respectively).

Supplementary Figure 5: Gradual shifts in feedback pitch drive large changes in the pitch of song. In three birds the pitch of auditory feedback was shifted gradually (by 35 cents every 6 days, as indicated by red and blue lines labeled to indicate the imposed shift as a function of time). Two birds experienced a sequence of upward shifts (filled and unfilled blue symbols) and one bird experienced a sequence of downward shifts (red symbols). Other plotting conventions are as in Figure 3a in the main text.

Time course data for 0-shift control birds Supplementary Figure 6: Stability of song in the absence of pitch shift. Data from two control birds who wore headphones but were not subjected to shifts in the pitch of auditory feedback (plotted in black and gray for visual clarity). Other plotting conventions as in Figure 3a in the main text.

Adaptive pitch learning in evening song In the results presented in the main text, only songs recorded in the late morning (between 10 am and noon) were analyzed. This sampling technique leaves open the question of whether evening song is similarly adaptable. This question is especially important given that prior work has shown significant differences between morning and evening song in young birds, and further demonstrated that the size of this difference predicts the final similarity to the song model (Deregnaucort et al. 2005). We addressed this issue by repeating the analysis shown in Figure 3b in the main text on song data recorded in the evening (between 6pm and 8pm). As shown in Supplementary Figure 7a below, the pitch of evening song changed in the adaptive direction (i.e. opposite the imposed pitch shift) in all 6 experiments. We then compared the magnitude of learning assayed in the evening window with that quantified in the original morning window on an experiment-by-experiment basis. As shown in Supplementary Figure 7b below, there was no significant difference between adaptive pitch changes during the two windows (p=0.72, paired-sample t-test). These results indicate that the manuscript s key finding is obtained independently of whether morning or evening song is used for data analysis. Supplementary Figure 7: Comparison of adaptive song changes when measured in the morning and evening. a, Mean +/- S.E. changes in the pitch of song during shift days 12-14 when pitch is measured in a window from 6:00-8:00 pm, rather than the 10:00 amnoon window used in the main analysis. All other plotting conventions as in Figure 3b in the main text. b, Comparison of the magnitude of adaptive pitch changes (mean +/- S.E.) when measured in the morning and evening. Adaptive changes (i.e. changes in the direction opposite the imposed shift in auditory feedback) were not significantly different when measured in the morning or evening windows (p=0.72, paired-sample t-test). Deregnaucourt S, Mitra PP, Feher O, Maul KK, Lints TJ, Tchernichovski O (2004) Song development: in search of the error-signal. Ann N Y Acad Sci 1016:364-376.