TY - JOUR
T1 - Decoding time for the identification of musical key
AU - Farbood, Morwaread M.
AU - Rowland, Jess
AU - Marcus, Gary
AU - Ghitza, Oded
AU - Poeppel, David
N1 - Funding Information:
This work is supported by NIH 2R01 05660 awarded to DP.
Publisher Copyright:
© 2014, The Psychonomic Society, Inc.
PY - 2014
Y1 - 2014
N2 - This study examines the decoding times at which the brain processes structural information in music and compares them to timescales implicated in recent work on speech. Combining an experimental paradigm based on Ghitza and Greenberg (Phonetica, 66(1-2), 113–126, 2009) for speech with the approach of Farbood et al. (Journal of Experimental Psychology: Human Perception and Performance, 39(4), 911–918, 2013) for musical key-finding, listeners were asked to judge the key of short melodic sequences that were presented at a highly a compressed rate with varying durations of silence inserted in a periodic manner in the audio signal. The distorted audio signals comprised signal-silence alternations showing error rate curves that identify peak performance centered around an event rate of 5–7 Hz (143–200 ms interonset interval; 300–420 beats/min), where event rate is defined as the average rate of pitch change. The data support the hypothesis that the perceptual analysis of music entails the processes of parsing the signal into chunks of the appropriate temporal granularity and decoding the signal for recognition. The music-speech comparison points to similarities in how auditory processing builds on the specific temporal structure of the input, and how that structure interacts with the internal temporal dynamics of the neural mechanisms underpinning perception.
AB - This study examines the decoding times at which the brain processes structural information in music and compares them to timescales implicated in recent work on speech. Combining an experimental paradigm based on Ghitza and Greenberg (Phonetica, 66(1-2), 113–126, 2009) for speech with the approach of Farbood et al. (Journal of Experimental Psychology: Human Perception and Performance, 39(4), 911–918, 2013) for musical key-finding, listeners were asked to judge the key of short melodic sequences that were presented at a highly a compressed rate with varying durations of silence inserted in a periodic manner in the audio signal. The distorted audio signals comprised signal-silence alternations showing error rate curves that identify peak performance centered around an event rate of 5–7 Hz (143–200 ms interonset interval; 300–420 beats/min), where event rate is defined as the average rate of pitch change. The data support the hypothesis that the perceptual analysis of music entails the processes of parsing the signal into chunks of the appropriate temporal granularity and decoding the signal for recognition. The music-speech comparison points to similarities in how auditory processing builds on the specific temporal structure of the input, and how that structure interacts with the internal temporal dynamics of the neural mechanisms underpinning perception.
KW - Brain rhythms
KW - Key finding
KW - Music structure
KW - Neuronal oscillations
KW - Speech rate
KW - Tonal induction
UR - http://www.scopus.com/inward/record.url?scp=84925534012&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=84925534012&partnerID=8YFLogxK
U2 - 10.3758/s13414-014-0806-0
DO - 10.3758/s13414-014-0806-0
M3 - Article
C2 - 25487869
AN - SCOPUS:84925534012
SN - 1943-3921
VL - 77
SP - 28
EP - 35
JO - Attention, Perception, and Psychophysics
JF - Attention, Perception, and Psychophysics
IS - 1
ER -