L12: Syntax

0.0(0)
studied byStudied by 0 people
learnLearn
examPractice Test
spaced repetitionSpaced Repetition
heart puzzleMatch
flashcardsFlashcards
Card Sorting

1/21

encourage image

There's no tags or description

Looks like no tags are added yet.

Study Analytics
Name
Mastery
Learn
Test
Matching
Spaced

No study sessions yet.

22 Terms

1
New cards

What are the rules that govern sentence form?

Moving from fixed forms (e.g. ‘apple’) to derivational forms:

  • play → play, played, playing

  • I, you, admire → “I admire you”

Morphology and syntax

• In all languages, the formation of words and sentences follows highly regular patterns

• How are the regulations and exceptions represented?

The study and analysis of language production in children reveals common and persistent patterns

2
New cards

What is the Wug Test (Berko, 1958)

  • Experiment for Morphology

<ul><li><p>Experiment for Morphology</p></li></ul><p></p>
3
New cards

How was children’s performance on the Wug Test

  • showed young children have a natural abstract morphology thinking

<ul><li><p>showed young children have a natural abstract morphology thinking</p></li></ul><p></p>
4
New cards

Describe the two aspects of the case study: Learning English Past Tense

The problem of English past tense formation:

  • Regular formation: stem + ‘ed’

  • Irregulars do show some patterns

    • No-change: hit → hit

    • Vowel-change: ring → rang, sing → sang

Over-regularizations are common: goed

  • These errors often occur after the child has already produced the correct irregular form: went

5
New cards

Describe the learning curve of children learning morphology and the 4 aspects

Observed U-shaped learning curve:

  • Imitation: an early phase of conservative language use

  • Generalization: general regularities are applied to new forms

  • Overgeneralization: occasional misapplication of general patterns

  • Recovery: over time, overgeneralization errors cease to happen

Lack of Negative Evidence

  • Children do not receive reliable corrective feedback from

6
New cards

A Symbolic Account of English Past Tense - Model

Describe 2 Predictions of the Dual-Route Account model

Prediction:

  • Errors result from transition from rote learning to rule-governed

  • Recovery occurs after sufficient exposure to irregulars

<p>Prediction:</p><ul><li><p>Errors result from transition from rote learning to rule-governed</p></li><li><p>Recovery occurs after sufficient exposure to irregulars</p></li></ul><p></p>
7
New cards

A Connectionist Account of Learning English Past - Model

Describe 2 properties of the connectionist model

Properties:

  • Early in training, the model shows tendency to overgeneralize; by the end of training, it exhibits near perfect performance

  • U-shaped performance is achieved using a single learning mechanism, but depends on sudden change in the training size

<p>Properties:</p><ul><li><p>Early in training, the model shows tendency to overgeneralize; by the end of training, it exhibits near perfect performance</p></li><li><p>U-shaped performance is achieved using a single learning mechanism, but depends on sudden change in the training size</p></li></ul><p></p>
8
New cards

Describe the innateness of Language - Syntax

  • Central claim: humans have innate knowledge of language

    • Assumption: all languages have a common structural basis

  • Argument from the Poverty of the Stimulus (Chomsky 1965)

    • Linguistic experience of children is not sufficiently rich for learning the grammar of the language, hence they must have some innate specification of grammar

    • Assumption: knowing a language involves knowing a grammar

  • Universal Grammar (UG)

    • A set of rules which organize language in the human brain

9
New cards

What are 3 principles and parameters of universal grammar?

  • A finite set of fundamental principles that are common to all languages

    • E.g., “a sentence must have a subject”

  • A finite set of parameters that determine syntactic variability amongst languages

    • E.g., a binary parameter that determines whether the subject of a sentence must be overtly pronounced

  • Learning involves identifying the correct grammar

    • I.e., setting UG parameters to proper values for the current language

10
New cards

Describe the general approach of computational implementation of the principles and parameters of universal grammar

  • Formal parameter setting models for a small set of grammars

    • Clark 1992, Gibson & Wexler 1994, Niyogi & Berwick 1996, Briscoe 2000

  • General approach:

    • Analyze current input string and set the parameters accordingly

    • Set a parameter when receiving evidence from an example which exhibits that parameter (trigger)

  • Representative models:

    • Triggering Learning Algorithm or TLA

11
New cards
<p>Describe the 3 represenative models for computational implementation of the principles and parameters of universal grammarrepresentative</p>

Describe the 3 represenative models for computational implementation of the principles and parameters of universal grammarrepresentative

  • TLA: randomly modifies a parameter value if it cannot parse the input

  • STL: learns sub-trees (treelets) as parameter values

  • VL: assigns a weight to each parameter, and rewards or penalizes these weights depending on parsing success

<ul><li><p><strong>TLA</strong>: randomly modifies a parameter value if it cannot parse the input</p></li><li><p><strong>STL</strong>: learns sub-trees (treelets) as parameter values</p></li><li><p><strong>VL</strong>: assigns a weight to each parameter, and rewards or penalizes these weights depending on parsing success</p></li></ul><p></p>
12
New cards
<p>Describe the 3 representative models for <strong>computational implementation </strong>of the principles and parameters of universal grammar <strong><u>BUT with an ambiguous trigger</u></strong></p>

Describe the 3 representative models for computational implementation of the principles and parameters of universal grammar BUT with an ambiguous trigger

  • TLA: chooses one of the possible interpretations of the ambiguous trigger

  • STL: ignores ambiguous triggers and waits for unambiguous ones

  • VL: each interpretation is parsed and the parameter weights

13
New cards

What are the Computational Challenges of the principles and parameters of universal grammar (4)

  • Practical limitations:

    • Formalizing a UG that covers existing languages is a challenge

    • Learning relies on well-formed sentences as input

  • P&P framework predicts a huge space of possible grammars

    • 20 binary parameters lead to > 1 million grammars

  • Search spaces for a grammar contain local maxima

    • I.e. learner may converge to an incorrect grammar

  • Most of the P&P models are psychologically implausible

    • They predict that a child may repeatedly revisit the same

14
New cards

Describe Usage-based Accounts of Language Acquisition

  • Main claims:

    • Children learn language regularities from input alone, without guidance from innate principles

    • Mechanisms of language learning are not domain-specific

  • Verb Island Hypothesis (Tomasello, 1992)

    • Children build their linguistic knowledge around individual items rather than adjusting general grammar rules they already possess

    • Children use cognitive processes to gradually categorize the syntactic structure of their item-based constructions

15
New cards

Distributional Representation of Grammar

  • Knowing a language is not equated with knowing a grammar

    • Knowledge of language is developed to perform communicative tasks of comprehension and production

  • Neural networks for language representation and acquisition

    • Different levels of linguistic representation are emergent structures that a network develops in the course of learning

16
New cards

Case study: Elman (1990)

<p></p>
17
New cards

Case study: Elman (1990)

*The model learns to categorize words into appropriate classes, and expect the correct order between them in a sentence

<p>*The model learns to categorize words into appropriate classes, and expect the correct order between them in a sentence</p>
18
New cards
term image
19
New cards

Describe the Case Study: MOSAIC

MOSAIC (Model Of Syntax Acquisition In Children

  • Learns from raw text, and produces utterances similar to what children produce using a discrimination network

Underlying mechanisms

  • Learning: expand the network based on input data

  • production: traverse the network and output contents of the nodes

Generalization

  • Similarity links allow limited generalization abilities

  • Lack of semantic knowledge prevents meaningful generalization

  • Generalized sentences are limited to high-frequency terms

Evaluation

  • The model was trained on a subset of CHILDES

  • It was used to simulate verb island phenomenon, optional infinitive in English, subject omission

<p>MOSAIC (Model Of Syntax Acquisition In Children</p><ul><li><p>Learns from raw text, and produces utterances similar to what children produce using a discrimination network</p></li></ul><p></p><p><strong>Underlying mechanisms</strong></p><ul><li><p>Learning: expand the network based on input data</p></li><li><p>production: traverse the network and output contents of the nodes</p></li></ul><p><strong>Generalization</strong></p><ul><li><p>Similarity links allow limited generalization abilities</p></li><li><p>Lack of semantic knowledge prevents meaningful generalization</p></li><li><p>Generalized sentences are limited to high-frequency terms</p></li></ul><p><strong>Evaluation</strong></p><ul><li><p>The model was trained on a subset of CHILDES</p></li><li><p>It was used to simulate verb island phenomenon, optional infinitive in English, subject omission</p></li></ul><p></p>
20
New cards
<p>Describe 2 examples for Treating LLM’s as Human Subjects</p>

Describe 2 examples for Treating LLM’s as Human Subjects

Systematically manipulate aspects of input and monitor model output

<p>Systematically manipulate aspects of input and monitor model output</p>
21
New cards
<p>Describe the process of <span>Treating LLM’s as Human Subjects</span></p>

Describe the process of Treating LLM’s as Human Subjects

  • Use a small-scale, controlled test set

    • ... often borrowed from a published psycholinguistic study

  • Subject an LLM to these tests

  • Make conclusions about

    • the nature of linguistic knowledge learned by the LLM

    • the processing mechanisms employed by the LLM

  • Compare their performance to that of human subjects

22
New cards

Open questions

  • How various aspects of language acquisition interact with each other?

    • Various learning procedures are most probably interleaved (e.g.,word leaning and syntax acquisition)

    • Most of the existing models of language acquisition focus on one aspect, and simplify the problem

  • How to evaluate the models on realistic data?

    • Large collections of child-directed utterances/speech are available, but no such collection of semantic input

    • A wide-spread evaluation approach is lacking in the community