1/14
Classical vs. nonclassical views of cognitive architecture
Name | Mastery | Learn | Test | Matching | Spaced | Call with Kai |
|---|
No analytics yet
Send a link to your students to track their progress
LOT and Cognitive Architecture
The propositional content of a thought is the meaning of a LOT sentence.
The propositional attitude is the subjectâs âcomputational relationâ to the LOT sentence, aka. the causal or functional role the sentence is playing in the subjectâs mental economy
Ex. S thinks that P if and only if S is cognitively related to a sentence (in Sâs internal LOT) that means P
Ex. Amy believes that grass is green if and only if Amy stands in the belief-relation to a LOT sentence that means grass is green.
MotivationÂ
Cognition has a language-like structureÂ
Rules for concatenating symbols is recursive â increasingly complex sentencesÂ
The meaning of a complex LOT sentence is a systematic function of the semantic values of the constituent elements and their syntactic organization (compositionality)
ANNs - HistoryÂ
McCulloch and Pitts (1949)Â
Simple binary threshold modelÂ
Suggested what neurons do + how synapses affect neural activityÂ
Could compute basic logical operations (proof neural nets can perform Turing-compatible tasks, ex, AND/OR/NOT)Â
Did not answer how neural networks evolve + learn
Rosenblatt (1958)
2 layer net called âperceptronâ (no hidden layer), just input + outputÂ
âNeurons that wire together, fire together!â
Gave suggestions for training two-layered networks using error correction techniquesÂ
Suggested an answer to how neural networks evolve + learn!Â
Minsky and Papert (1969)
Showed Rosenblattâs perceptron couldn't compute XORÂ
Showed its not obvious on how to train multi-layered networks
Connectionism
Modern look at computation â system to execute certain functions!Â
Many simple units/nodes that interact via connections with varying âweightsâ (excitatory/inhibitory)
Pattern of activation across nodes. Can be categorized/used for perceptual discrimination tasksÂ
Engages in representational activity!
Nodes are interconnected across adjacent layers
Memory distributed in connection weights (no central controller), rather than a local representation
Short term memory stored in unitâs changing state of activationÂ
Long term memory stored in strength of connection weights between unitsÂ
The only activity in the system is the changing activation level of units, in response to cumulative signal from connected unitsÂ
Sending signal: Simple function of activation level (representable numerically)
Receiving signal: Complex function of signal sent + the excitatory/inhibitory weight of the connection w/ the unit that sent signalÂ
Many more connections in network than units
Pattern of weighted connection among units determines the character of a given network (ex. particular pattern of activation given a certain input)
Why connectionism?
Provides a model of cognition that is closer to the biological structure of the brain (neural networks + parallel processing)
Offers superior learning capabilities from environmental data (learns from environment + graceful degradation)
Good at ambiguous tasks/handling patterns! Ex. filling in gaps, finding similarities, recognizing repeated instancesÂ
Patterns that networks are trained to respond are drawn from massive amounts of info, involving large number of weakly correlated variablesÂ
Machine Learning Training
Networks are trained to execute tasks (mapping functions) via training sets (exposure to examples)Â
Examples are precategorized/pre-labelled â network gets better over time after using learning algorithmÂ
Error propogation:
Compare actual output with desired output; adjust connections based on feedback to reduce discrepancy. Gradual error reduction over time!
Learning algorithm and Error Propagation
Training technique, automated procedure for adjusting connection weights/activation thresholds until desired results are reached!Â
Gradually nudge connection weights towards target response (ex. Back propagation of error)Â
Once sufficient at training sets, will adjust and perform with novel dataÂ
Advantages of Error Propagation/Supervised Learning
Graceful Degradation: Damage to node = slight decrease in performance, not collapse of whole systemÂ
Fault Tolerance: System can continue to operate in the event of failing component(s)Â
Sensitive to pattern differences undetectable by humansÂ
Resilient to noise/ambiguity/damageÂ
Spontaneous generalization to novel cases
Concerns of Error Propagation/Supervised Learning
Predicts forms of neural connections not found in mammalian brainsÂ
Problems with scaling up to more complex tasksÂ
Catastrophic forgetting: System matures up to a point - when exposed to novel data the system dramatically declines + does not recover
Require huge quantities of training data and repetition to learn very simple tasks; no one-shot learning (as seen in humans)Â
Unobvious how to train multilayer networks â more hidden layers, hard to distribute blame for error signalÂ
Requires pre-labelled + pre-categorized training dataÂ
ANN learning how to classify data on its own. All answers are provided to the ANNÂ
Can cause discrepancies between desired output + actual outputÂ
If a system already possesses target mapping (has the answer through the data set), is it really learning?
Connectionism was presumed to be an alternative to hard-coded knowledge!
Respects in which ANNs and LOT differ
LOT: Symbolic, rule based system (top-down)
Knowledge represented in a propositional, language format (similar to human thought)Â
Good at structured tasks
Step-by-step reasoning (sequential processing)Â
Explicit programming of rules/structured dataset â learns from fewer examples
Semantic constituents â combinatorial structure allows different organizations to carry contentÂ
Ex. someone who can think âJohn loves Maryâ can also think âMary loves JohnâÂ
ANNs: Connectionist, learning based systems (Bottom-up)Â
Knowledge represented through numerical weights distributed across layers of nodesÂ
Knowledge distributed through the network!
Parallel processing, good for unstructured data (pattern recognition)Â
Require massive datasetsÂ
Lack combinatorial structureÂ
The argument from systematicity for the claim that the human mind has a LOT structure
Human mind is combinatorial! Key feature:Â
Semantic constituents + structure of propositional contents are isomorphic to the syntactic constituents + organization of the representational vehicles that carry its connects
Hence why someone can think John loves Mary and also Mary loves JohnÂ
ANNs lack combinatorial structureÂ
Fodorâs argument against systemacity in ANNsÂ
If human thought is systematic, then it must have vehicles whose structure maps on to the structure of their contents.
If ANNs are good models of human thought, then human thought cannot have vehicles whose structure maps on to the structure of their contents.
Human thought is systematic.
Human thought must have vehicles whose structure maps on to the structure of their contents.
Artificial neural networks are not good models of human thought.
ANN advocateâs reply to Fodorâs argument for systematicity
Radical response: Deny that human thought is systematic
ANN advocateâs argument against systemacity in humans
If human thought is systematic, then it must have vehicles whose structure maps on to the structure of their contents.
If ANNs are good models of human thought, then human thought cannot have vehicles whose structure maps on to the structure of their contents.
ANNs are good models of human thought.
Human thought cannot have vehicles whose structure maps on to the structure of their contents.
Human thought is not systematic
ANN theorists wishing to explain propositionally structured thought face a dilemma:
either an ANN can be trained to display systematicity or it cannot.Â
If it cannot, it is hopeless as a general account of human cognition.Â
If it can, it simply vindicates the classical approach.
Theoretical morals Smith has about AI and its failings (general)
Recap of Smithâs Project: Arguing for the distinction between 2 kinds of intelligence
The sort we exemplifyÂ
Sees this as âgenuine intelligenceâ â manifests the capacity for judgment as opposed to just reckoning (calculative reasoning) like computers
The sort modern computers exemplifyÂ
AI systems need to deal with reality as it actually is, not the way we think it is!
We represent the world through thoughts/language â> but the world is not inherently split into neat objects, as we may perceive it. Not complete representations of reality!
He finds ANNs more compatible with his view of reality (can get a more objective understanding of world)
1st Wave AI Failures
NeurologicalÂ
Architecture of digital computer is much different from the human brainÂ
Brain operations are much slower!
PerceptualÂ
Classism rests on the false assumption that perception = recovering info about objects from retinal stimulationÂ
EpistemiologicalÂ
AI can know a label but not the referent in the real word (ex. Symbol dog is a node in a database, not an object with weight/fur/life)Â
Inability to generate genuine knowledge or understand the world, despite being able to manipulate symbols and follow rules
OntologicalÂ
We exist in a ânon-conceptualâ world
1st wave AI thinks that the world comes in neat, ontologically discrete objects. This misconception explains its ultimate inadequacyÂ
We cannot assume an intelligent system that we build will parse reality in the same way we do
If the system is intelligence, it should make its own sense of things, including constructing its own conceptual schemes
Theoretical morals Smith draws from 1st wave AI
The good: Representational Mandate
Proper functioning of world-directed system must be governed by normative criteria applying to its mechanical operations
Operations framed in terms of situations + states of affairs of the world that the system is representing/reasoning about, which situations and states of affair will not in be in effective (causal) reach
The bad: Deepest failing = AI took objects for grantedÂ
Rested on naive conception of the world outside the mind
Naive to assume that the world comes chopped into neat, ontologically discrete objects
Misconception of world in GOFAI: interpret/filter the world though abstractions or idealizations
Highlights some aspects of what is represented
Minimizes/distorts others, and ignores/abstracts in-the-world detail
Theoretical morals Smith draws from 2nd wave AI
The good: Yields valuable philosophical insightsÂ
Suggest rich + ineffable web of statistical relatedness, weaves the world together into an integrated âsubconceptualâ wholeÂ
However, reality itself surpasses an intelligent beingâs capacity to conceptualize verbally articulate itÂ
Intelligent systems must come to grip with world on conceptual level
The bad: Similar error as GOFAI, fails to explain how a system comes to understand what it is representing/talking aboutÂ
To achieve understanding (and genuine intelligence), Smith suggests the system would need to be capable of assuming a stance of deference to the worldÂ
Must have its own stance, not just act in accord to human deference
To do this, it will have to know:Â
That there is a worldÂ
That its representations are about that worldÂ
That it + its representations must defer to the world that they represent