2-2-Patterns Neuropsychological Data Agnosia Patient GS

2-2-Patterns Neuropsychological Data Agnosia Patient GS

Neuropsychological Data • Agnosia • Term coined by Sigmund Freud 2-2-patterns • From the Greek word for “lack of knowledge” • The inability to recognize objects when using a given sense (e.g. vision), even though that sense is basically intact (Nolte, 1999) Agnosia Patient GS • Sensory abilities intact • Language normal • Usually involves damage to the occipito-parietal • Unable to name objects pathway Agnosia Prosopagnosia • Apperceptive • Specific inability to recognize faces – Object recognition failure due to perceptual processing – Difficulty recognizing pictures w/deleted segments • Are faces and other objects in the world – Unable to utilize top-down information for pattern recognition represented in fundamentally different • Associative – Perceptual processing intact but subject cannot use information ways in memory? to recognize objects – Can draw objects but not say what they are • Does face-memory depend on – Language otherwise intact fundamentally different brain systems? – Often don’t know other things about object (how it’s used, etc.) 1 Are Faces Special? Are Faces Special? • Subjects presented with a face and asked to represent a face-part • Houses: similar performance for parts & wholes • Subjects presented with a house and asked to • Faces: whole-object advantage represent a house-part Are Faces Special? Models of Pattern Recognition • Template Models • Feature Models • Prototype Models • Neural Network Models • Objects represented in parts and holistically • Faces represented holistically Word Superiority Effect IAC Model WORK K C/K C/K 2 Connectionist (PDP) Models Neural Network Models • Information evaluated in • Nodes parallel and distributed • Connections throughout the network • Neurally inspired • Activation Rules – Neural Networks • Activation • Motivated by problems • Output function w/classical architecture • Learning rule Problems w/Classical Slow Architecture •Slow • 100-step Constraint – Active neuron spikes • Brittle approximately once every 1-3 ms •Inflexible – People can perform cognitive tasks in 100-300 ms – Cognitive tasks must be performed w/100 serial operations – Or, cognitive tasks performed in parallel Brittle Inflexible • 1 damaged symbol • Digital halts computation • Hard Constraints • 1 missing line of code • No Soft Constraints could be fatal • In contrast, effects of brain damage, local – BD often hard to notice 3 Connectionism, NNs Perceptron • Hardware Matters • Neurally-inspired cognitive modeling Biological Nets Connectionist Nets Basics •1010 -1011 neurons • Neuron basic • Set of processing •105 interconnections processing unit units per neuron • Highly interconnected • Input Function • Excitatory & Inhibitory • Excitatory & Inhibitory • State of activation • Learning done by • Learning involves changing strength of • Output function for modifying synapses connections each unit More Basics The Environment • Pattern of • Specifying the environment specifies the sorts of Connectivity problems a particular network addresses • Propagation Rule – Specify Inputs • Activation Rule –Specify Outputs • Learning Rule • Networks used to model many things –HebbRule – Stock Market Predictions – Widrow-Hoff Rule – Diagnose Heart Attacks – Generalized Delta – Model Cognition Rule (back-prop) • Pattern Recognition, Categorization, Memory, Learning 4 Appeal of Neural Nets Buzzwords • Parallel • Spontaneous Generalization – Not subject to 100- • Graceful Degradation step constraint • Distributed • Mutual Constraint Satisfaction Representations • Capacity for Learning and Self- – Less Brittle Organization • Graded Rules • Biologically Plausible (?) – More Flexible Example Possibilities •3-layer • Input • Output • Input Layer: Hidden • Home Team • Home Team Layer, fully – Quarterback Rating –Points interconnected – Rushing Yards •Away Team • Hidden Layer: Output – Receiving Yards –Points Layer, fully – Field Goals Made interconnected – Field Goals Attempted • Train with BackProp •Away Team Features Tri-Level Hypothesis • Incorporates • Not symbolic information about – Yes, sub-symbolic context • Yes, information processing • Doesn’t assume variables are independent “…dissimilarities do not imply that brains are not • Captures higher-order computers, but only that brains are not serial, digital statistical regularities computers.” –Churchland, Koch, Sejnowski 5 Computational Level Algorithmic Level • What problem is the network solving? • How do they do those crazy things they • Use of formal methods to do? – Determine limits of networks • Neural networks blur the structure/process – Determine limits of learning rules distinction • Architecture/Program Implementation Level IAC Model • Biological Plausibility – Only capture computationally relevant properties (cognitive modelling) – Closely model particular brain systems (computational neuroscience) Words versus Letters Word Superiority Effect • WORD and WORK both begin to get activated because of WOR • Activations feedback and activate K and D nodes • But only K gets bottom-up activation • D gets top-down activation that gets inhibited • Letters presented alone don’t get top-down activation from word level! 6 Template Theories of Pattern IAC Model Evaluation Template Theories of Pattern Recognition • Illustrates important principles • Template theories: a • Doesn’t hold up to 21st century standards miniature copy or template of each known pattern is stored – Scope too narrow in long-term memory – No learning mechanism – straightforward template theory – Too many feature units – normalized template theory • Problems – not adaptable – impose large storage requirements Feature Theories of Pattern Pattern Recognition Recognition • Pattern Recognition • Feature theories: patterns requires Feature Analysis consist of a set of specific – Break down pattern into features or attributes primitives • Advantages: – Recognize combination of – elementary features can combine to form multiple features objects • Integration of bottom-up – Features can be used to describe novel objects and top-down processing • Problems: – context effects in perception – recognition can take place when features are occluded 7.

View Full Text

Details

  • File Type
    pdf
  • Upload Time
    -
  • Content Languages
    English
  • Upload User
    Anonymous/Not logged-in
  • File Pages
    7 Page
  • File Size
    -

Download

Channel Download Status
Express Download Enable

Copyright

We respect the copyrights and intellectual property rights of all users. All uploaded documents are either original works of the uploader or authorized works of the rightful owners.

  • Not to be reproduced or distributed without explicit permission.
  • Not used for commercial purposes outside of approved use cases.
  • Not used to infringe on the rights of the original creators.
  • If you believe any content infringes your copyright, please contact us immediately.

Support

For help with questions, suggestions, or problems, please contact us