web analytics
Categories
Auditory Cortex Auditory Neuroscience Auditory Perception EEG / MEG Neural Oscillations Neural Phase Papers Perception Publications

New paper in press: Her­rmann, Hen­ry, Grigutsch & Obleser, The Jour­nal of Neu­ro­science [Update]

Oscil­la­to­ry Phase Dynam­ics in Neur­al Entrain­ment Under­pin Illu­so­ry Per­cepts of Time

Nat­ur­al sounds like speech and music inher­ent­ly vary in tem­po over time. Yet, con­tex­tu­al fac­tors such as vari­a­tions in the sound’s loud­ness or pitch influ­ence per­cep­tion of tem­po­ral rate change towards slow­ing down or speed­ing up.

A new MEG study by Björn Her­rmann, Mol­ly Hen­ry, Maren Grigutsch and Jonas Obleser asked for the neur­al oscil­la­to­ry dynam­ics that under­pin con­text-induced illu­sions in tem­po­ral rate change and found illu­so­ry per­cepts to be linked to changes in the neur­al phase pat­terns of entrained oscil­la­tions while the exact fre­quen­cy of the oscil­la­to­ry response was relat­ed to veridi­cal percepts.

The paper is in press and forth­com­ing in The Jour­nal of Neuroscience.

 

Update:

Paper is avail­able online.

Ref­er­ences

  • Her­rmann B, Hen­ry MJ, Grigutsch M, Obleser J. Oscil­la­to­ry phase dynam­ics in neur­al entrain­ment under­pin illu­so­ry per­cepts of time. J Neu­rosci. 2013 Oct 2;33(40):15799–809. PMID: 24089487. [Open with Read]
Categories
Auditory Neuroscience Auditory Perception Auditory Working Memory Executive Functions fMRI Papers Perception Publications

New paper has been pub­lished in Cere­bral Cor­tex by Hen­ry, Her­rmann, & Obleser

When we lis­ten to sounds like speech and music, we have to make sense of dif­fer­ent acoustic fea­tures that vary simul­ta­ne­ous­ly along mul­ti­ple time scales. This means that we, as lis­ten­ers, have to selec­tive­ly attend to, but at the same time selec­tive­ly ignore, sep­a­rate but inter­twined fea­tures of a stimulus.

Brain regions associated with selective attending to and selective ignoring of temporal stimulus features.
Brain regions asso­ci­at­ed with selec­tive attend­ing to and selec­tive ignor­ing of tem­po­ral stim­u­lus fea­tures. (more)

A new­ly pub­lished fMRI study by Mol­ly Hen­ry, Björn Her­rmann, and Jonas Obleser found a net­work of brain regions that respond­ed oppo­site­ly to iden­ti­cal stim­u­lus char­ac­ter­is­tics depend­ing on whether they were rel­e­vant or irrel­e­vant, even when both stim­u­lus fea­tures involved atten­tion to time and tem­po­ral features.

You can check out the arti­cle here:

http://cercor.oxfordjournals.org/content/early/2013/08/23/cercor.bht240.full

Ref­er­ences

  • Hen­ry MJ, Her­rmann B, Obleser J. Selec­tive Atten­tion to Tem­po­ral Fea­tures on Nest­ed Time Scales. Cereb Cor­tex. 2013 Aug 26. PMID: 23978652. [Open with Read]
Categories
Auditory Cortex Auditory Neuroscience Auditory Perception Clinical relevance EEG / MEG Evoked Activity Hearing Loss Papers Psychology Publications

New paper in press: Her­rmann et al., Hear­ing Research [Update]

Audi­to­ry fil­ter width affects response mag­ni­tude but not fre­quen­cy speci­fici­ty in audi­to­ry cortex

This is fan­tas­tic news on a fri­day morn­ing: Obleser lab Post­doc Björn Her­rmann teamed up with his fel­low Post­docs Math­ias Scharinger and Mol­ly Hen­ry to study how spec­tral analy­sis in the audi­to­ry periph­ery (termed fre­quen­cy selec­tiv­i­ty) relates to pro­cess­ing in audi­to­ry cor­tex (termed fre­quen­cy speci­fici­ty; see also Björns paper in J Neu­ro­phys­i­ol 2013).

Giv­ing this an age­ing and hear­ing loss per­spec­tive and build­ing on the con­cept of audi­to­ry fil­ters in the cochlea (Moore et al.), Björn found that  the over­all N1 ampli­tude of lis­ten­ers, but not their fre­quen­cy-spe­cif­ic neur­al adap­ta­tion pat­terns, is cor­re­lat­ed with the pass-band of the audi­to­ry filter.

This sug­gests that widened audi­to­ry fil­ters are com­pen­sat­ed for by a response gain in fre­quen­cy-spe­cif­ic areas of audi­to­ry cor­tex; the paper is in press and forth­com­ing in Hear­ing Research.

 

Update:

Paper is avail­able online.

Ref­er­ences

  • Her­rmann B, Hen­ry MJ, Scharinger M, Obleser J. Audi­to­ry fil­ter width affects response mag­ni­tude but not fre­quen­cy speci­fici­ty in audi­to­ry cor­tex. Hear Res. 2013 Oct;304:128–36. PMID: 23876524. [Open with Read]
Categories
Auditory Cortex Auditory Neuroscience Auditory Perception Auditory Speech Processing Degraded Acoustics Executive Functions fMRI Noise-Vocoded Speech Papers Perception Publications Speech

New paper out: Erb, Hen­ry, Eis­ner & Obleser — Jour­nal of Neuroscience

We are proud to announce that PhD stu­dent Julia Erb just came out with a paper issued in Jour­nal  of Neu­ro­science:

The Brain Dynam­ics of Rapid Per­cep­tu­al Adap­ta­tion to Adverse Lis­ten­ing Conditions

Effects of adaptation

Grab it here:

Abstract:

Lis­ten­ers show a remark­able abil­i­ty to quick­ly adjust to degrad­ed speech input. Here, we aimed to iden­ti­fy the neur­al mech­a­nisms of such short-term per­cep­tu­al adap­ta­tion. In a sparse-sam­pling, car­diac-gat­ed func­tion­al mag­net­ic res­o­nance imag­ing (fMRI) acqui­si­tion, human lis­ten­ers heard and repeat­ed back 4‑band-vocod­ed sentences 

Ref­er­ences

  • Erb J, Hen­ry MJ, Eis­ner F, Obleser J. The brain dynam­ics of rapid per­cep­tu­al adap­ta­tion to adverse lis­ten­ing con­di­tions. J Neu­rosci. 2013 Jun 26;33(26):10688–97. PMID: 23804092. [Open with Read]
Categories
Auditory Neuroscience Auditory Perception Editorial Notes EEG / MEG Neural Oscillations Neural Phase Papers Perception Publications

New paper out: Hen­ry & Her­rmann, Jour­nal of Neuroscience

Proud to announce that our post­docs Mol­ly Hen­ry and Björn Her­rmann just came out with a review/op piece in the Jour­nal of Neu­ro­science “jour­nal club” sec­tion, where only grad stu­dents or post­docs are allowed to author short review pieces.

A Pre­clud­ing Role of Low-Fre­quen­cy Oscil­la­tions for Audi­to­ry Per­cep­tion in a Con­tin­u­ous Pro­cess­ing Mode

The Jour­nal of Neu­ro­science, 5 Decem­ber 2012, 32(49): 17525–17527; doi: 10.1523/​JNEUROSCI.4456–12.2012

Mol­ly and Björn review (and com­ment on) an impor­tant paper by our friends and col­leagues Christoph Kayser and Benedikt Ng in the same jour­nal. Essen­tial­ly, they argue for the dis­tinc­tion of a con­tin­u­ous from an oscil­la­to­ry pro­cess­ing mode in lis­ten­ing, and pro­vide ten­ta­tive expla­na­tions of why some­times miss­es might be more mod­u­lat­ed by neur­al oscil­la­to­ry phase than hits. Con­grats, guys!

Ref­er­ences

  • Hen­ry MJ, Her­rmann B. A pre­clud­ing role of low-fre­quen­cy oscil­la­tions for audi­to­ry per­cep­tion in a con­tin­u­ous pro­cess­ing mode. J Neu­rosci. 2012 Dec 5;32(49):17525–7. PMID: 23223276. [Open with Read]
Categories
Auditory Perception EEG / MEG Events Evoked Activity Posters Publications Speech

Poster Pre­sen­ta­tions at SFN

There will be two poster pre­sen­ta­tions at SFN in Wash­ing­ton, DC., on the top­ic of audi­to­ry pre­dic­tions in speech per­cep­tion. The first poster, authored by Alexan­dra Ben­dix­en, Math­ias Scharinger, and Jonas Obleser, sum­ma­rizes as follows:

Speech sig­nals are often com­pro­mised by dis­rup­tions orig­i­nat­ing from exter­nal (e.g., mask­ing noise) or inter­nal (e.g., slug­gish artic­u­la­tion) sources. Speech com­pre­hen­sion thus entails detect­ing and replac­ing miss­ing infor­ma­tion based on pre­dic­tive and restora­tive mech­a­nisms. The nature of the under­ly­ing neur­al mech­a­nisms is not yet well under­stood. In the present study, we inves­ti­gat­ed the detec­tion of miss­ing infor­ma­tion by occa­sion­al­ly omit­ting the final con­so­nants of the Ger­man words “Lachs” (salmon) or “Latz” (bib), result­ing in the syl­la­ble “La” (no seman­tic mean­ing). In three dif­fer­ent con­di­tions, stim­u­lus pre­sen­ta­tion was set up so that sub­jects expect­ed only the word “Lachs” (con­di­tion 1), only the word “Latz” (con­di­tion 2), or the words “Lachs” or “Latz” with equal prob­a­bil­i­ty (con­di­tion 3). Thus essen­tial­ly, the final seg­ment was pre­dictable in con­di­tions 1 and 2, but unpre­dictable in con­di­tion 3. Stim­uli were pre­sent­ed out­side the focus of atten­tion while sub­jects were watch­ing a silent video. Brain respons­es were mea­sured with mul­ti-chan­nel elec­troen­cephalo­gram (EEG) record­ings. In all con­di­tions, an omis­sion response was elicit­ed from 125 to 165 ms after the expect­ed onset of the final seg­ment. The omis­sion response shared char­ac­ter­is­tics of the omis­sion mis­match neg­a­tiv­i­ty (MMN) with gen­er­a­tors in audi­to­ry cor­ti­cal areas. Crit­i­cal­ly, the omis­sion response was enhanced in ampli­tude in the two pre­dictable con­di­tions (1, 2) com­pared to the unpre­dictable con­di­tion (3). Vio­lat­ing a strong pre­dic­tion thus elicit­ed a more pro­nounced omis­sion response. Con­sis­tent with a pre­dic­tive cod­ing account, the pro­cess­ing of miss­ing lin­guis­tic infor­ma­tion appears to be mod­u­lat­ed by pre­dic­tive context.

The sec­ond poster looks at sim­i­lar mate­r­i­al, but con­trasts coro­nal [t] with dor­sal [k], yield­ing inter­est­ing asym­me­tries in MMN responses:

Research in audi­to­ry neu­ro­science has lead to a bet­ter under­stand­ing of the neur­al bases of speech per­cep­tion, but the rep­re­sen­ta­tion­al nature of speech sounds with­in words is still a mat­ter of debate. Elec­tro­phys­i­o­log­i­cal research on sin­gle speech sounds pro­vid­ed evi­dence for abstract rep­re­sen­ta­tion­al units that com­prise infor­ma­tion about both acoustic struc­ture and artic­u­la­tor con­fig­u­ra­tion (Phillips et al., 2000), there­by refer­ring to phono­log­i­cal cat­e­gories. Here, we test the pro­cess­ing of word-final con­so­nants dif­fer­ing in their place of artic­u­la­tion (coro­nal [ts] vs. dor­sal [ks]) and acoustic struc­ture, as seen in the time-vary­ing for­mant (res­o­nance) fre­quen­cies. The respec­tive con­so­nants dis­tin­guish between the Ger­man nouns Latz (bib) and Lachs (salmon), record­ed from a female native speak­er. Ini­tial con­so­nant-vow­el sequences were aver­aged across the two nouns in order to avoid coar­tic­u­la­to­ry cues before the release of the con­so­nants. Latz and Lachs served as stan­dard and deviant in a pas­sive odd­ball par­a­digm, while the EEG from 20 par­tic­i­pants was record­ed. The change from stan­dard [ts] to deviant [ks] and vice ver­sa was accom­pa­nied by a dis­cernible Mis­match Neg­a­tiv­i­ty (MMN) response (Näätä­nen et al., 2007). This response showed an intrigu­ing asym­me­try, as seen in a main effect con­di­tion (deviant Latz vs. deviant Lachs, F(1,1920) = 291.84, p < 0.001) of an omnibus mixed-effect mod­el. Cru­cial­ly, the MMN for the deviant Latz was on aver­age more neg­a­tive than the MMN for the deviant Lachs from 135 to 185 ms post deviance onset (p < 0.001). We inter­pret these find­ings as reflect­ing a dif­fer­ence in phono­log­i­cal speci­fici­ty: Fol­low­ing Eulitz and Lahiri, 2004, we assume coro­nal seg­ments ([ts]) to have less spe­cif­ic (‘fea­t­u­ral­ly under­spec­i­fied’) rep­re­sen­ta­tions than dor­sal seg­ments ([ks]). While in stan­dard posi­tion, Lachs acti­vat­ed a mem­o­ry trace with a more spe­cif­ic final con­so­nant for which the deviant pro­vid­ed a stronger mis­match than vice ver­sa, i.e. when Latz acti­vat­ed a mem­o­ry trace with a less spe­cif­ic final con­so­nant. Our results sup­port a mod­el of speech per­cep­tion where sen­so­ry infor­ma­tion is processed in terms of dis­crete units inde­pen­dent of high­er lex­i­cal prop­er­ties, as the asym­me­try can­not be explained by dif­fer­ences in lex­i­cal sur­face fre­quen­cies between Latz and Lachs (both log-fre­quen­cies of 0.69). We can also rule out a fre­quen­cy effect on the seg­men­tal lev­el. Thus, it appears that speech per­cep­tion involves a lev­el of pro­cess­ing where indi­vid­ual seg­men­tal rep­re­sen­ta­tions with­in words are evaluated.

Categories
Auditory Perception Auditory Speech Processing EEG / MEG Evoked Activity Linguistics Papers Place of Articulation Features Publications Speech

New paper out in Jour­nal of Speech, Lan­guage, & Hear­ing Research [Update]

We are hap­py to announce that our paper “Asym­me­tries in the pro­cess­ing of vow­el height” will be appear­ing in the Jour­nal of Speech, Lan­guage, & Hear­ing Research, authored by Philip Mon­a­han, William Idsar­di and Math­ias Scharinger. A short sum­ma­ry is giv­en below:

Pur­pose: Speech per­cep­tion can be described as the trans­for­ma­tion of con­tin­u­ous acoustic infor­ma­tion into dis­crete mem­o­ry rep­re­sen­ta­tions. There­fore, research on neur­al rep­re­sen­ta­tions of speech sounds is par­tic­u­lar­ly impor­tant for a bet­ter under­stand­ing of this trans­for­ma­tion. Speech per­cep­tion mod­els make spe­cif­ic assump­tions regard­ing the rep­re­sen­ta­tion of mid vow­els (e.g., [{varepsilon}]) that are artic­u­lat­ed with a neu­tral posi­tion in regard to height. One hypoth­e­sis is that their rep­re­sen­ta­tion is less spe­cif­ic than the rep­re­sen­ta­tion of vow­els with a more spe­cif­ic posi­tion (e.g., [æ]).

Method: In a mag­ne­toen­cephalog­ra­phy study, we test­ed the under­spec­i­fi­ca­tion of mid vow­el in Amer­i­can Eng­lish. Using a mis­match neg­a­tiv­i­ty (MMN) par­a­digm, mid and low lax vow­els ([{varepsilon}]/[æ]), and high and low lax vow­els ([I]/[æ]), were opposed, and M100/N1 dipole source para­me­ters as well as MMN laten­cy and ampli­tude were examined.

Results: Larg­er MMNs occurred when the mid vow­el [{varepsilon}] was a deviant to the stan­dard [æ], a result con­sis­tent with less spe­cif­ic rep­re­sen­ta­tions for mid vow­els. MMNs of equal mag­ni­tude were elicit­ed in the high–low com­par­i­son, con­sis­tent with more spe­cif­ic rep­re­sen­ta­tions for both high and low vow­els. M100 dipole loca­tions sup­port ear­ly vow­el cat­e­go­riza­tion on the basis of lin­guis­ti­cal­ly rel­e­vant acoustic–phonetic features.

Con­clu­sion: We take our results to reflect an abstract long-term rep­re­sen­ta­tion of vow­els that do not include redun­dant spec­i­fi­ca­tions at very ear­ly stages of pro­cess­ing the speech sig­nal. More­over, the dipole loca­tions indi­cate extrac­tion of dis­tinc­tive fea­tures and their map­ping onto rep­re­sen­ta­tion­al­ly faith­ful cor­ti­cal loca­tions (i.e., a fea­ture map).

[Update]

The paper is avail­able here.

Ref­er­ences

  • Scharinger M, Mon­a­han PJ, Idsar­di WJ. Asym­me­tries in the pro­cess­ing of vow­el height. J Speech Lang Hear Res. 2012 Jun;55(3):903–18. PMID: 22232394. [Open with Read]
Categories
Auditory Neuroscience Auditory Perception fMRI Linguistics Papers Publications Speech

New paper out: “Upstream del­e­ga­tion” for pro­cess­ing of com­plex syn­tax under degrad­ed acoustics

A new paper is about to appear in Neu­roim­age on the inter­ac­tion of syn­tac­tic com­plex­i­ty and acoustic degradation.

It is writ­ten by myself, PhD stu­dent Lars Mey­er, and Angela Friederi­ci. In a way, the paper brings togeth­er one of Angela’s main research ques­tions (which brain cir­cuits medi­ate the pro­cess­ing of syn­tax?) with a long-stand­ing inter­est of mine, that is, how do adverse lis­ten­ing sit­u­a­tions affect the com­pre­hen­sion of speech.

The paper is entitled

Dynam­ic assign­ment of neur­al resources in audi­to­ry com­pre­hen­sion of com­plex sentences

The paper first estab­lish­es that acoustic vari­ants of increas­ing­ly com­plex sen­tences essen­tial­ly behave like writ­ten ver­sions of these sentences.
The data then neat­ly show that pro­cess­ing chal­leng­ing (but legal) syn­tax under var­i­ous lev­els of degra­da­tion has a very dif­fer­ent effect on the neur­al cir­cuits involved than prof­it­ing from seman­tics: While the lat­ter has been shown pre­vi­ous­ly to involve more wide­spread, het­ero­modal brain areas, the dou­ble demand of increas­ing­ly com­plex syn­tax and an increas­ing­ly degrad­ed speech sig­nal (from which the com­plex syn­tax has to be parsed) elic­it an “upstream” shift of acti­va­tion back to less abstract pro­cess­ing areas in the supe­ri­or tem­po­ral and prefrontal/frontal cortex.

We ten­ta­tive­ly have termed this process “upstream del­e­ga­tion”. We have also tried and estab­lished a slight­ly unusu­al method to do jus­tice to the fMRI acti­va­tion data: We have includ­ed all z‑scores gath­ered along cer­tain spa­tial dimen­sions, irre­spec­tive of whether they were sub- or suprathresh­old, and have treat­ed them as dis­tri­b­u­tions. Check it out and let us know what you think.

Ref­er­ences

  • Obleser J, Mey­er L, Friederi­ci AD. Dynam­ic assign­ment of neur­al resources in audi­to­ry com­pre­hen­sion of com­plex sen­tences. Neu­roim­age. 2011 Jun 15;56(4):2310–20. PMID: 21421059. [Open with Read]