LING 1010 Language and Mind: Artificial Intelligence and Human Language
Artificial Intelligence and Human Language
Chatbot: A computer program simulating human conversation via text responses to user input.
Historical examples: ELIZA ($1960$s).
Recent examples: ChatGPT, Bard.
Recent chatbots produce novel, grammatical, and relevant sentences.
Unlike traditional grammars (symbolic rules), these chatbots are based on Large Language Models (LLMs).
Cognitive Science and AI Historical Context
Cognitive Science: Interdisciplinary field scientifically investigating information processing in the human brain (perception, reasoning, language, memory). Linguistics is the branch focusing on language.
Classical Computational Theory of Mind:
Human mind is like a digital computer, performing rule-governed computations on symbolic representations.
Influenced by logic and early computer science.
Connectionism:
Human mind is a product of the human brain.
Models inspired by neural wiring; information is distributed in neural networks without explicit symbols or rules.
Influenced by neuroscience.
Artificial Intelligence (AI): An engineering project to build intelligent computers/machines.
Intelligence typically involves reasoning, problem-solving, language, decision-making.
Natural Language Processing (NLP) is the AI branch dealing with language.
AI aims for efficient solutions, not necessarily mimicking human processes.
Large Language Models: Mechanics and Scale
Language Model: Any NLP program predicting the next word from a sequence of words.
Modern chatbots use LLMs with significant computational power.
LLMs (e.g., GPT-$4$, LLaMa, PaLM-$2$) run on Artificial Neural Networks (ANNs).
ANNs are not symbolically programmed but trained on vast datasets, adjusting numerical weights (parameters) for better outputs.
Scale of LLMs (e.g., GPT-$4$):
1.76 trillion parameters.
Trained on 13 trillion word tokens.
ANN has 120 layers of hidden nodes.
Training required 10,000+ advanced semiconductor chips.
LLMs and Human Language Understanding
Experts cannot interpret the inner workings of LLMs and lack techniques to understand their knowledge or reasoning. LLMs are often considered "black boxes."
Benchmarking (e.g., BLiMP): Used to test LLM linguistic abilities, often based on surprisal, showing high similarity to human language on tests.
It is unclear if LLMs follow symbolic grammatical rules or ascribe meaning to generated text. Some argue no actual language understanding occurs.
LLMs and Human Language Acquisition
LLMs do not learn like humans; they are trained on datasets too large for any human lifetime (e.g., children acquire language by age 6 with tens of millions of word tokens).
LLMs generally fail to reach human-level accuracy when restricted to human-scale data.
Human advantages in language acquisition include sensorimotor stimuli, inter-agent interaction, environmental interaction, and prosody, which LLMs typically lack.
LLMs serve as powerful tools for studying human language acquisition, with research involving realistic datasets like the BabyLM Challenge.
Concerns and Limitations of LLMs
Debate: Are LLMs creative or "stochastic parrots" (haphazardly stitching linguistic forms without meaning)?
Risk of LLMs parroting biases and prejudices present in their vast training data.
"No reliable techniques for steering the behavior of LLMs" raises concerns about controlling output.