Showing posts with label semantic. Show all posts
Showing posts with label semantic. Show all posts

Monday, January 14, 2008

Final discussion

Robert Daland points out that amplitude is another cue to sentence boundaries - Amanda answers that indeed there are many others (such as initial strengthening), but the point remains that younger babies were less able to attend to single cues than older babies.
Jerry Jeager points out Levelt's model does not address actual production. Lisa G agrees that Levelt's model is useful, and further clarifies that she was trying to imply that there's a need to get converge in terms of categories with articulatory phonology.
Matt notes that S. Blumstein also found in L2 more transfer for non-words
* what is the nature of the mechanism governing attention?
* Lisa D. is asked whether she makes a distinction between 'lexical' and 'semantic'. She answers that, for the purposes of the task, it doesn't matter: in fact, it is possible that anything that 'made them care' would have helped.
Amanda (?) asks whether it is an explicit or implicit task - and whether an active versus a passive task would have made a difference.
Mary Beckman points out that perhaps the answer is having a *symbolic* category.
Jennifer Cole notes that we have to factor in that it is language - not any arbitrary cue can be used, since some things are unlearnable [I think she may have referred to tasks where you had to learn that a given rule applied only to female speakers, but not male ones - on this note, doesn't that actually happen in speech, that an alternation occurs only in one sociolect?]
Lisa D. points the effects of telling participants that they are about to hear a foreign language (e.g. Russian) actually improves their performance.
[I think there was a return to the question of frication versus vocalic information]
Matt wonders also about the fact that in speech social information is coded primarily in the vowels leading people to attend more to them, but notes that it is also in encoded consonants.
Amanda suggests that perhaps inventory size is a reason for it.
Lisa D. confirms this with her experience with Bengali speakers who... ??
Mary Beckman wonders how do we know why we attend to something - Jennifer Cole agrees: in fact, it seems so hard conceptually that we are able to converge on the right cues given the multidimensionality of speech. Jessica points out that we couldn't just throw out information that is not adjacent to the sound in question in cases of long distance dependencies such as vowel harmony. How does one determine the limits, the distance at which one will stop paying attention.
Ken Dejong (?) points out that we have been using the term 'attention' without really defining it. Attention could mean recruitment of resources and usage of information. There are two issues attached to this: we tend to underestimate people's ability to use something and, for development, we need to address how we learn to focus attention.
Matt asks whether it is just interference of categories that allows to focus attention. For instance, Dell did a study where people had to learn that engma appeared in syllable initial position, and it was really hard.
Mary Beckman relates this to babies' inability to learn that a nasalized vowel is followed by t, given that this sequence is illegal in English.
Amanda answers that perhaps it isn't given that sometimes the nasalized vowel is the only cue to the presence of a nasal consonant and this one is dropped.
Jessica remembers that when she heard the stimuli used in that study they sounded like vowel+nt sequences to her.
?? Ken Dejong asks Lisa G whether it is a question of salience - Lisa D adds whether it is mirroring
LG answers that when a novel word was at the end of the sentence, the novel word was not more variable, but the rest of the sentence was.

Friday, November 30, 2007

Some questions on the ideas put forward by y'all

Hello everyone,

I was re-reading everyone's abstracts and thought that there were some interesting coincidences and intriguing mismatches.

For example, I might be reading Susan's wrong, but it seems that she predicts that, with a 1-to-1 mapping from articulation to acoustics, the magnitude of the change in the articulation predicts basically two levels of structure: one corresponding to slower and larger movements and one to more abrupt changes. Only the latter would, according to my interpretation of her abstract, provide acoustic cues. This matches Ying and Jeff's finding that an automatic feature extractor operating on the raw acoustic signal is good at finding manner features, but not so good with place features. On the contrary, motor cues are most useful for place of articulation, but not so great with manner.

Susan also suggests that it is not only the level of fine acoustic cues that affects speech perception, but also the other level. May this be related to Jessica, Robert and Matt (JRM)'s learner of phonotactic constraints and segments? If I get their point right, both infants and adults would rely on acoustic cues in context to discover/perceive phonological categories, and this process may go from a holistic chunk (perhaps syllable-based?) to phone-sized elements. In relation to this, I was thinking that while one may expect that articulatory experience would aid the formation of prosodic categories (like syllabic/word templates), it was at first surprising for me that it is also important for learning of segments. Now I see that this fits really well with JRM's proposal that we may learn phonotactics and segments at the same time, don't you think?

Further, this learning strategy (of phonotactics+segments) must be flexible and still active in adulthood, given that Lisa D's adults are able to learn a minimal pair of words that relies on a non-native phonotactic contrast, but her results also suggest that the presence of a minimal pair (which, through semantic cues, forces listeners to focus on the relevant acoustic cues) is a necessary condition for learning in adults. On the other hand, the presence of minimal pairs cannot be a necessary condition in toddlers (cf. Pater et al. 2004), as Ying and Jeff point out, while already in childhood semantic information is actually helpful, according to Lisa G's results. Is there a developmental trend here? May it be geared by a difference in processing abilities? (If so, would we predict that adults in a high-processing load task would behave like Pater's toddlers and have a *harder* time with the minimal pair than with the non-minimal pair?)

Another question that follows from Susan's hypothesis that these two levels interact in perception is how language experience affects this interaction. Grant's paper seems to suggest that this interaction is mediated by specific experience, such that sheer experience with one of the categories (AmEng speakers must have experience with retroflex sounds, and even with the retroflex fricative in words like 'shrill') is not enough, but speakers must have been exposed to the particular contrast that relies on the acoustic/higher level features under study. This is particularly important for early category formation: Chandan's corpus analysis looks at the correlation of pitch and VOT, which could be interpreted as respective examples of the larger and finer categories proposed by Susan. If IDS takes advantage of the apparent predilection of infants for pitch, will it have its voiced segments be primarily marked with pitch? However, given young infants' very limited articulatory abilities, we'd expect them to rely primarily on acoustic-feature based contrasts, which -now following Ying and Jeff - would predict a primacy of VOT.

Monday, July 30, 2007

Incentive to focus: Word learning helps listeners distinguish native and non-native sequences



By Lisa Davidson

Previous research in cross-language perception has shown that non-native listeners often assimilate both single phonemes and phonotactic sequences to native language categories (e.g., Best 1995, Kuhl and Iverson 1995, Dupoux et al. 1999, Flege et al. 2003). These findings suggest that it would be difficult for second language learners to overcome these phonetic barriers to learning new sounds or sequences. To study whether higher-level cues can assist learners, two experiments examined whether associating meaning with unfamiliar words assists listeners in distinguishing the phonotactically possible and unattested sequences (see also Hayes- Harb 2007 for phoneme discrimination).

In Experiment 1, American English listeners were trained on word-picture pairings of words containing a phonological contrast between CC and CəC sequences, but which were not minimal pairs (e.g., [ftake], [fətalu]). In Experiment 2, the word-picture pairings specifically consisted of minimal pairs (e.g., [ftake], [fətake]). In the test phase, listeners saw a picture and
heard words spoken by a new speaker and had to indicate which of the words matched the picture. Results showed that participants chose the accurate CC or CəC form more often when they learned minimal pairs as opposed to phonological contrast alone. Nevertheless, there was a significant asymmetry in accuracy in Experiment 1: listeners were more accurate on CəC word- picture pairings than on CC pairings. Subsequent investigation of individual listeners revealed that the participants could be divided into a high performing and a low performing group: the high performers were much more capable of learning the contrast between native and non-native words, while the low performers remained at chance. Results for high performers are shown in Figure 1 (attached).

These findings suggest that at least for high performers, learning minimal pairs provides greater incentive to distinguish non-native sequences like CC from native counterparts like CəC. These experiments can be compared to a previous AX discrimination task using the same stimuli which did not include any training on the stimuli beforehand (Davidson to appear). In the AX task, listeners were at chance in discriminating between CC and CəC tokens. Unlike evidence from the infant literature suggesting that infants may encounter processing limitations in tasks requiring them to assign meaning to contrastive sounds (Pater, Stager and Werker 2004), adults have ample resources allowing them to use meaning cues to better learn the distinction between native and non-native sequences. Furthermore, greater accuracy on phonotactically legal CəC sequences may be due to the ability to establish a more robust phonetic representation. For words learned as CC, participants seem to accept a greater variety of productions, suggesting that the native language phonological prohibition on the CC sequences used in this study hampers a detailed phonetic encoding of these items.

Attention to cues and phonological categorization: Motor contributions

By Lisa Goffman


The acquisition of phonological units relies on perceptual and motor experiences and biases. Perceptual factors have been the emphasis of much research, with infants responding to prosodic and segmental cues in the input. Such perceptual sensitivities are thought to provide a mechanism for parsing important language units, such as words and sentences.
Motor contributions to the acquisition of production units have been less well investigated than perceptual. MacNeilage and Davis (2000) have argued that motor primitives associated with the syllable emerge in the context of canonical babbling and that the segmental content of these oscillatory open-close movements is detailed over the course of development. In adult speakers, Browman and Goldstein (1992) have developed the theory of Articulatory Phonology, in which movement primitives associated with each individual articulator (e.g., tongue body, lips, velum) are coordinated, with phonological units emerging from this coordination of gestures. Current models of language production (e.g., Levelt et al., 1999) include Articulation as an element, but do not develop how this processing level interacts with the phonological level, especially during acquisition.
The research discussed in this paper attempts to bridge this gap by measuring articulatory movement output as children and adults produce various language units. The working hypothesis is that motor capacities interact with phonological units and that these interactions change over the course of development. I incorporate methodologies from speech motor control and from psycholinguistics to assess how grammatical, lexical, and phonological processing levels are linked to articulatory output.
In this work I recorded oral articulatory movements (using the Optotrak system) while children and adults produce different segments, syllables, words, and sentences. The patterning of oral movements is assessed across these different speech production tasks, as well as the variability of such patterning. Three groups of results (i.e., segmental, prosodic, and semantic) are summarized. In the segmental domain, changes in a single phoneme influence the oral movement patterning of a single articulator (Goffman & Smith, 1999). Further, changes in a single segment exert broad coarticulatory effects that cross word and phrase boundaries. Another group of findings focuses on prosody, demonstrating that children produce late developing (in English) prosodic structures with relatively stable and small and short articulatory movements. Early developing trochees are, counter-intuitively, produced with relatively poorly controlled and variable articulatory patterns (e.g., Goffman, 1999; Goffman, Heisler, & Chakraborty, 2006). That is, stress patterns that are acquired relatively early for English-learning children are actually produced with less motor precision. It may be that children either omit syllables or produce equal stress in their attempts to produce more precise iambs. Finally, data will be presented showing that 4-year-old children produce more stable articulatory movement patterns when a novel phonetic form is provided with a visual or functional referent. All of this work together provides clues into how articulatory cues contribute to children’s developing phonological units.