Title: fetzerreply-rpi.ppt
1fetzerreply-rpi.ppt
- version 20110906-2
- for RPI
- based on fetzer-reply.ppt-20100524 for NACAP_at_CMU
2How Cognition Could Be ComputingSemiotic
Systems, Computers, the Mind
- William J. Rapaport
- Department of Computer Science Engineering,
- Department of Philosophy, Department of
Linguistics, - and Center for Cognitive Science
- rapaport_at_buffalo.edu
- http//www.cse.buffalo.edu/rapaport
3Summary
- Computationalism cognition is computable.
- Mental processes can be the result of algorithmic
procedures - that can be affected by emotions/attitudes/indivi
dual histories. - Computers that implement these (cognitive)
procedures really exhibit those mental
processes. - They are semiotic ( sign-using) systems.
- They really think.
- Computers can possess minds.
- Syntactic semantics explains how all this is
possible.
4I. What Is Computationalism?
- What is AI?
- Not artificial Its a computational theory
- Not about intelligence Its about cognition
- Better name
- Computational cognition
- cf. computational linguistics
computational statistics computational
geometry, etc.
5What Is Computationalism?
- Computationalism ? cognition is computation
- Hobbes, McCulloch/Pitts, Putnam, Fodor, Pylyshyn,
- interesting, worth exploring, possibly true
- BUT
- Not what computational usually means!
- What should computationalism be?
- Must preserve crucial insight
- cognition is explainable via mathematical theory
of computation - First, some definitions
6Preliminary Definitions
- Cognition
- whatever cognitive scientists study,
including - believing
- consciousness
- emotion
- language
- learning
- memory
- perception
- planning
- problem solving
- reasoning
- representation
- including categorization, concepts, mental
imagery, etc. - sensation
- thought, etc.
7Preliminary Definitions
- Algorithm (informal notion)
- A is an algorithm for executor E to accomplish
goal G informally - A is a procedure (finite set/seq of
statements/rules/instructions) such that - Each statement/rule/instruction S is such that
- S is composed of finite of symbols/marks from
finite alphabet - S is unambiguous for Ei.e.
- E knows how to execute S
- E can execute S
- S can be executed in finite time
- after executing S, E knows what to do next
- A halts ( takes finite time)
- A halts with G accomplished
8Preliminary Definitions
- Effective
- Church left undefined
- Rosser
- each step is precisely determined
- method produces an answer
- in finite of steps
- Markov
- process produces an answer
- Kleene
- effective procedure algorithm
- Knuth
- all operations can be done exactly, in finite time
9Preliminary Definitions
- Algorithm (formal notion)
- A is an algorithm formallyA is (logically
equivalent to) a Turing machine - Church-Turing Thesis
- algorithminformal algorithmformal
10Preliminary Definitions
- Computable
- task/goal/field of study G is computable iff
- ? algorithm(s)formal for G
11The Proper Treatment of Computationalism
- Computationalism ? Cognition is computation
12The Proper Treatment of Computationalism
- Computationalism Cognition is computable
- i.e., ? algorithm(s) that compute cognitive
functions - Basic research question of computational
cognitive science - How much of cognition is computable?
- Working assumption of computational cognitive
science - All cognition is computable
13Proper Treatment of Computationalism
- Implementational implication(multiple
realization) - If cognition is computable, then
- anything that implements cognitive
computationswould be cognitive (would really
think) - even if humans dont do it that way!
- Piccinini
- neural spike trains are not representable as
digit strings - ? not computational
- BUT
- ? functions whose O/P they produce are not
computable - ? human cognition is computable but not computed
14Proper Treatment of Computationalism
- 2 Views of Cognitive Science
- How do humans cognize? (part of CogSci)
- Might not be computationally.
- BUT Can abstract away from the specifically
human to a more general issue - How is cognition possible? (also part of
CogSci) - Might be computable.
15The Proper Treatment of Computationalism
- 2 Views of Computationalism
- Cognition is computation
- strong / narrow / nearsighted view
- the mind or brain is a computer
- how the M/B does what it does is by computing
- vs. the proper treatment
- Cognition is computable
- weak / wide / farsighted view
- what the mind or brain does can be described in
computational terms - how it does it is a matter for neuroscience to
determine
16Proper Treatment of Computationalism
- Turing on his test
- the use of words and general educated opinion
will alter so much that one will be able to
speak of machines thinking without expecting to
be contradicted. - general educated opinion
- changes when we abstract generalize
- the use of words
- changes when reference shifts from words
initial / narrow application to more
abstract / general phenomenon - cf. fly, compute, algorithm
- ditto for cognition / think
17II. Syntactic Semanticsas a theory underlying
computationalism
- Cognition is internal
- Cognitive agents have direct access only to
internal representatives of external objects - Semantics is syntactic
- ? Words, meanings, semantic relations between
them are all syntactic items - Understanding is recursive
- Recursive Case
- We understand one thing in terms of another that
must already be understood - Base Case
- We understand something in terms of
itself(syntactic understanding)
18Syntactic Semantics
- InternalismCognitive agents have direct access
only to internal representatives of external
objects - A cognitive agent understands the world by
pushing the world into the mind (Jackendoff
2002) - ? Both words their meanings (including
external objects) are represented
internally in a single LOT - Humans biological neural network
- Adrian 1926,1928 nervous systems transduce
different physical stimuli into a common internal
medium (cf. Piccinini) - Computers
- artificial neural network
- symbolic knowledge-representation reasoning
system
19Syntactic Semantics Internalism
- Hume argument from double vision
- Kant noumena vs. phenomena
- Ayer argument from illusion
- Fodor anti-Putnam methodological solipsism
- G.Segal anti-Burge individualism
- Pylyshyn
- output of sensory transducers is the only
contact the cognitive system ever has with the
environment - Changizi argument from time delay in perception
20Syntactic Semantics
- (Internalism ?) Syntacticism? Words, meanings,
semantic relations between them are all
syntactic - syntax study of relations among members of
a single set - set of signs / marks / neurons /
- semantics study of relations between members of
two sets - set of signs / marks / neurons /
- set of (external) meanings / (with its own
syntax!) - Pushing meanings into same set as symbols for
them allows semantics to be done syntactically - turns semantic relations between 2 sets (internal
signs, external meanings) into syntactic
relations among the marks of a single (internal)
LOT - e.g. truth tables formal semantics are both
syntactic - e.g. neurons representing both signs external
meanings - Symbol-manipulating computers can do semantics
by doing syntax
21Syntactic Semantics Syntacticism
- Syntactic semantics underlies the Semantic
Web - syntactic info on web pages gains meaning
from - syntactic info (metadata encoded in RDF) in
HTML source files - metadata annotates webpage data
- metadata provides semantic interpretation of
webpage data
22Syntactic Semantics
- Understanding is recursive
- Recursive cases
- We understand a syntactic domain (SYN-1)
indirectly by interpreting it in terms of a
semantic domain (SEM-1) - but SEM-1 must be antecedently understood
- SEM-1 can be understood by considering it as a
syntactic domain SYN-2 interpreted in terms of
yet another semantic domain - which also must be antecedently understood, etc.
- Base case
- A domain that is understood directly (i.e., not
antecedently) - in terms of itself
- i.e., syntactically
- perhaps holistically
23Syntactic Semantics Recursiveness
- Syntactic understanding
- the meaning of an internal state (which may or
may not be linked to an external state of
affairs) for the system itself is most
naturally defined in terms of that states
relations to its other states. - Edelman, Shimon (2008), On the Nature of Minds,
or Truth and Consequences, JETAI 20 181-196
quote on pp.188-189. - I.e., syntactically
24III. Rapaports Thesis
- Syntax suffices for semantic cognition
- cognition is computable
- ? computers are capable of thinking
- James H. Fetzers Thesis
- It doesnt,
- it isnt,
- they arent
25The Nature of Signs
Something S
Ground
Causation
z Interpretant (wrt a
context) x Somebody Something
Figure 2. The Nature of Signs
26Questions about Semiotic Systems
- What is the causation relation between sign-user
z and sign S? - Which causes which?
- What is the grounding relation between sign S
and thing x that S stands for? - If S is grounded by that (x) which it stands
for,does is grounded by stands for? - If sign S stands for thing x for sign-user
z,then do we need a different diagram?
27- Sign-user z causes sign S
- Thing x grounds sign S
- Sign S stands for thing x for sign-user z
28Input-Output Systems
Input i
No Grounding
Causation
c (wrt a context)
o Computer
Output
Figure 4. An Input-Output System
29Input-Output Systems
- Differences
- Sign-user z is now computer c
- Sign S is now input i
- Thing x is now output o
- No grounding relation between i o
- BUT
- The marks by means of which computers operate
include more than merely the (external) input - Can include internally stored marks
- What the marks stand for is not necessarily the
output - What does it mean for sign S to stand for thing x
yet not be grounded by x?
30Fetzers Thesis
- Computers differ from cognitive agents in 3 ways
- statically (symbolically)
- dynamically (algorithmically)
- affectively (emotionally)
- Simulation is not the real thing
31Fetzers Static Difference
ARGUMENT 1 Computers are mark-manipulating
systems, minds are not. Premise 1
Computers manipulate marks on the basis of their
size, shapes, and relative
locations. Premise 2 (a) These shapes,
sizes, and relative locations exert causal
influence upon
computers but (b) do not stand
for anything for those systems. Premise 3
Minds operate by utilizing signs that stand for
other things in some
respect or other for them as sign-using (or
semiotic) systems.
__________________________________________________
________________ Conclusion 1 Computers are
not semiotic (or sign-using) systems.
______________________________________________
_____________________ Conclusion 2
Computers are not the possessors of minds.
Figure 9.
The Static Difference
32The Static Difference
- Static Premise 1
- Is computer manipulation of symbols independent
of meaning? - depends on what meaning means
- Computational symbol-manipulation is independent
of external, 3rd-person meaning imposed on the
symbols - But not independent of internal, 1st-person
meaning - arises from syntactic relations among internal
symbols - intrinsic meaning
33The Static Difference
- Static Premise 2b
- The symbols that computers manipulate do not
stand for anything for those computers. - But
- Fetzers locution allows for the possibility that
symbols could stand for something for the
computer - Insofar as they could, such machines might be
capable of thinking - He should have said could not stand for
anything - But then hed be wrong -)
34Fetzer - Computers Are Not Semiotic Systems
- But
- Semiotic systems interpret signs
- An algorithms O/P is an interpretation of its
I/P - Algorithms ground the I/P-O/P relation
- Computers are algorithm machines
- Computers are semiotic systems
35Fetzer - Computers Are Not Semiotic Systems
- In a semiotic system (e.g., a mind)
- something (S) is a sign of something (x) for
somebody (z) - x grounds sign S
- x is an interpretant w.r.t. a context to
sign-user z - S is in a causation relation with z
36Fetzer - Computers Are Not Semiotic Systems
- In a computer (I/O) system
- input i (playing role of sign S) is in a
causation relation with computer c (playing
role of sign-user z) - output o (playing role of thing x) is in an
interpretant relation with computer c - BUT No grounding relation between i o
37Fetzer - Computers Are Not Semiotic Systems
- ? Computers only have causal relationships,
no mediation between I/P O/P (?!) - But semiotic systems require such mediation
- Peirceinterpretant is mediately determined by
the sign - interpretant is really the sign-users mental
concept of the thing x (!!) - ? Computers are not semiotic systems
- But minds are.
- ? Minds are not computers computers cant be
minds.
38Incardona - Computers Are Semiotic Systems!
- Something is a semiotic system iffit carries out
a process that mediates between a sign its
interpretant - Semiotic systems interpret signs
- Algorithms describe processes that mediate
between I/Ps O/Ps - An algorithms O/P is an interpretation of its
I/P - Algorithms ground the I/O relation
- Computers are algorithm machines.
- ? Computers are semiotic systems
39The Static Difference
- Argument that computers are semiotic systems from
embedding in the world - Fetzers (counter?)example
- A red light at an intersection stands for
applying the brakes and coming to a complete
halt, only proceeding when the light turns green,
for those who know the rules of the road. - Can such a red light stand for applying the
brakes, etc., for a computer? - It could, if the computer knows the rules of the
road - But a computer can know those rules
- if it has those rules stored in a knowledge base
- and if it uses those rules to drive a vehicle
- cf. Stanley the VW (2005 DARPA Grand Challenge)
- Parisien Thagard 2008, Robosemantics
How Stanley Represents the World, Minds
Machines
40The Static Difference
- Does a calculator that computes GCDs understand
what its doing? - Fetzer Rapaport No
- Could a computer that computes GCDs understand
what its doing? - Fetzer No
- Rapaport Goldfain Yes, it could
- as long as it had enough background / contextual
/ supporting information - a computer with a full-blown theory of math at
the level of an algebra student learning GCDs
could understand GCDs as well as the student
41The Static Difference
- Goldfain - Computers could be semiotic systems
- G1 The natural s that a cognitive agent refers
to are denoted by a sequence of unique
marks exemplifying a finite initial
segment of the natural- structure. - G2 Such a finite initial segment can be
generated by a computational cognitive
agent (a computer) via perception action in the
world during an act of counting (e.g.,
using Lisps gensym) - they have a history of how they became marks
that signify something for the agent (the
computer). - G3 These marks (e.g., b4532, b182, b9000) have
no meaning for a human user who lacks
access to their ordering. - G4 Such private marks (numerons) are
associable with publicly meaningful
marks (numerlogs) - e.g., b4532 denotes the same number as 1,
b182 denotes the same number as 2, etc. - G5 A computational cognitive agent (a computer)
can do math solely on the basis of its
numerons. - C1 ? These marks stand for something for the
computer (the agent). - C2 we can check the math because of G4.
42The Static Difference
- Static Premise 1
- Computers do manipulate marks on the basis
of size, shape, location - but also on the basis of
- relations of those symbols to other symbols
- i.e., on the basis of their syntax
- which Fetzer can safely add to his theory
- this processing is not independent of their
meaning - (by SS II)
- but is independent of (external) reference
- In this way, such symbols can stand for something
for the computer - Computers are indeed string-processing systems
- But meaning can arise from (appropriate)
combinations of strings
43Summary No Static Differences
- Both computers minds manipulate marks
- The marks can stand for something for both
computers minds - Computers (and minds) are semiotic systems
- Computers can possess minds
44Fetzers Dynamic Difference
ARGUMENT 2 Computers are governed by
algorithms, but minds are not. Premise 1
Computers are governed by programs, which
are causal models of algorithms. Premise 2
Algorithms are effective decision procedures for
arriving at definite
solutions to problems in a finite number of
steps. Premise 3 Most human thought
processes, including dreams, daydreams, and
ordinary thinking, are not procedures for
arriving at solutions to problems in a finite
number of steps.
__________________________________________________
____________________ Conclusion 1 Most human
thought processes are not governed by programs
as causal models of algorithms.
__________________________________________________
_____________________ Conclusion 2 Minds
are not computers.
Figure 10. The Dynamic
Difference
45The Dynamic Difference A Red Herring
- Fetzer
- If thinking is computing computing is
thinking - if computing is algorithmic
- then thinking is algorithmic
- but it isnt
- 2nd conjunct is irrelevant and false
- A computer executing a non-cognitive program
(e.g., an operating system) is computing but not
thinking
46The Dynamic Difference
- Premises 1 2
- Def of algorithm is OK
- But algorithms may be the wrong entity
- may need a more general notion of procedure
(Shapiro) - like an algorithm, but
- need not halt
- need not yield correct output
47The Dynamic Difference
- Premise 3 Most human thinking is not
algorithmic - Dreams are not algorithms
- Ordinary stream-of-consciousness thinking is not
algorithmic - BUT
- Some human thought processes may indeed not be
algorithms - But thats not the real issue, which is
- Could there be algorithms/procedures that produce
these(or other mental states or processes)? - If dreams are our interpretations of random
neuron firings during sleep, as if they were
due to external causes - then if non-dream neuron-firings are
computable ( theres every reason to
think they are) then so are dreams - Stream of consciousness might be computable
- e.g., via spreading activation in a semantic
network
48The Dynamic Difference
- Whether a mental state/process is computable is
at least an empirical question - Must avoid the Hubert Dreyfus fallacy
- one philosophers idea of a non-computable
processis another computer scientists research
project - what no one has yet written a program for is not
thereby necessarily non-computable - In fact Mueller, Erik T. (1990), Daydreaming
in Humans Machines A Computer Model of the
Stream of Thought (Ablex) - Cf. Edelman, Shimon (2008), Computing the Mind
(Oxford) - ? burden of proof is on Fetzer!
49The Dynamic Difference
- Dynamic Conclusion 2
- Are minds computers?
- Maybe, maybe not
- I prefer to say (with Shimon Edelman, et al.)
- The (human) mind is a virtual machine,computation
ally implemented (in the nervous system)
50Summary No Dynamic Difference
- All (human) thought processes are/might be
describable by algorithms/procedures - computationalism properly treated
51Fetzers Affective Difference
ARGUMENT 3 Mental thought transitions are
affected by emotions, attitudes, and
histories, but computers are not. Premise 1
Computers are governed by programs, which are
causal models of algorithms. Premise 2
Algorithms are effective decisions, which are
not affected by emotions, attitudes, or
histories. Premise 3 Mental thought
transitions are affected by values of variables
that do not affect computers.
___________________________________________
__________________________ Conclusion 1
The processes controlling mental thought
transitions are fundamentally different than
those that control computer procedures.
_________________________________________________
____________________ Conclusion 2 Minds are
not computers.
Figure 11. The Affective Difference
52The Affective Difference
- Fetzers definitions
- intension of expression E def
- conditions that need to be satisfied for
something to be an E - extension of E def
- class of all things that satisfy Es intension
- denotation of E for agent A def
- subset of Es extension that A comes into contact
with - connotation of E for A def
- As attitudes emotions in response to As
interactions with Es denotation for A - Somewhat non-standard, but useful
- e.g. meaning of E for A ? Es denotation
connotation for A
53Contra Affective Premises 2 3
- Programs can be based on (idiosyncratic)emotions,
attitudes, histories - Rapaport-Ehrlich contextual vocabulary
acquisition program - Learns a meaning for an unfamiliar word from
- the words textual context
- integrated with the readers idiosyncratic
- denotations, connotations,
- emotions, attitudes, histories,
- prior beliefs
- Sloman, Picard, Thagard
- Developing computational theories of affect,
emotion, etc. - Emotions, attitudes, histories can affect
computers that model them.
54Summary No Affective Differences
- Processes controlling mental thought transitions
are not fundamentally different from those
controlling algorithms/procedures. - Algorithms can take emotions/attitudes/histories
into account. - Both computers minds can be affected by
emotions/attitudes/histories
55The Matter of Simulation
ARGUMENT 4 Digital machines can nevertheless
simulate thought processes and other
forms of human behavior. Premise 1 Computer
programmers and those who design the systems that
they control can increase their performance
capabilities, making them better and better
simulations. Premise 2 Their performance
capabilities may be closer and closer
approximations to the performance capabilities of
human beings without turning them into
thinking things. Premise 3 Indeed, the
static, dynamic, and affective differences that
distinguish computer performance from human
performance preclude them from being thinking
things. ________________________________________
______________________________________
Conclusion Although the performance
capabilities of digital machines can
become better and better approximations of human
behavior, they are still not thinking
things.
Figure 15. The Matter of Simulation
56Argument from Simulation
- AgreedA computer that simulates some process
P is not necessarily really doing P - But what is really doing P vs. simulating P?
- What is the scope of a simulation?
- Computer simulations of hurricanes dont get
real people really wet - Real people are outside the scope of the
simulation - BUT a computer simulation of a hurricane could
get simulated people simulatedly wet
- Computer simulation of the daily operations of a
bank is not thereby the daily operations of a
(real) bank - BUT I can do my banking online
- Simulations can be used as if they were real
57Argument from Simulation
- Some simulations of Xs are real Xs
- scale model of a scale model of X is a scale
model of X - Xeroxed/faxed/PDF copies of documents are those
documents - A computer that simulates an informational
process is thereby actually doing that
informational process - Because a computer simulation of information is
information
58Argument from Simulation
- Computer simulation of a picture is a picture
- digital photography
- Computer simulation of language is language
- computers really do parse sentences (Woods)
- IBMs Watson really answers questions
- Computer simulation of math is math
- A simulation of a computation and the
computation itself are equivalent try to
simulate the addition of 2 and 3, and the result
will be just as good as if you actually carried
out the additionthat is the nature of numbers
(Edelman) - Computer simulation of reasoning is reasoning
- automated theorem proving, computational logic,
59Argument from Simulation
- Computer simulation of cognition is cognition
- if the mind is a computational entity, a
simulation of the relevant computations would
constitute its fully functional replica
(Edelman) - cf. implementational implication
60Argument from Simulation
- A simulation of a computation and the
computation itself are equivalent try to
simulate the addition of 2 and 3, and the result
will be just as good as if you actually carried
out the additionthat is the nature of numbers.
Therefore, if the mind is a computational entity,
a simulation of the relevant computations would
constitute its fully functional replica. - Shimon Edelman (2008), Computing the Mind
61Summary Simulation Can Be(come) the Real Thing
- Close approximation to human thought processes
can turn computers into thinking things - only asymptotically?
- actually?
- cf. Turing on general educated opinion
the use of words
62Summary
- Computers are semiotic (sign-using) systems.
- Computationalismproperly treated cognition is
computable - not necessarily computational.
- Any non-computable residue will be negligible
- Mental processes are describable (?governable) by
algorithmic procedures - that can be affected by emotions/attitudes/indivi
dual histories. - Computers that implement these cognitive
procedures really exhibit those cognitive
behaviors. - They really think.
- Computers can possess minds.
- Syntactic semantics explains how all this is
possible.
63Any non-computable residue will be negligible
- On negligible differences
- cf. music on CDs with music on vinyl
- discrete/digital vs. continuous/analog
- Does it matter whether a cognitive computer
really thinks? - on the meaning of really
- human-specific thinking
- vs. abstract/general notion of thinking
- on does it matter
- an android will need to behave ethically
- to be treated ethically