In this paper, I attempt to give a brief account of
the current debate in Artificial Intelligence between the so-called
classical computationalist view and the more recent trend of
connectionist views concerning what to count as an appropriate
explanation of intelligent behaviour. In particular, I will focus on the
question of whether connectionist models of cognition constitute a
genuine alternative to more traditional views of mental processing, or
if, as some have argued, they have nothing to offer but a story about
the implementation of symbolic structures at the level of neural
functioning. The conclusion I wish to support is that connectionism does
offer an alternative account of cognition, although the debate as to
which is the best approach - the classical or the connectionist - is far
from settled.
It is an article of faith among Artificial Intelligence practitioners
that cognition, or intelligent behaviour, is best understood as a
complex of information processing transformations carried over
representations. However, and despite significant progress in devising a
computational view of the mind, the past two decades have seen the field
divide in two views on the nature of information processing, and it is
fair to say there is no shared agreement on what to count as the proper
theoretical approach to cognition. Accordingly, on one side of the fence
there are those who maintain the traditional view in AI that intelligent
behaviour can only be accounted for by appeal to syntactically
structured symbols that are manipulated according to certain formal
rules. On the other, there are those who defend that cognition is better
understood in terms of sub-symbolic patterns of activation in a group of
highly interconnected units. As can be expected, discord over who holds
the key to the enigmas of cognition runs deep, with the symbolic AI
people accusing connectionist models of having nothing new to offer to
the central issues of intelligence, and the connectionist camp arguing
that they do.
Now, before moving further, it is important to notice that both the
classic view and the connectionist share the goal of understanding and
modeling (that is, implementing) cognitive processes in artificial
devices. Also, it worth emphasizing that both traditional and
connectionist models have inputs and outputs and these inputs and
outputs can be given (in the right circumstances) intentional
interpretations. What makes them different is that, unlike the
traditional computational model, connectionism claims that the output is
not produced as a result of the rule governed manipulation of
functionally discrete symbols. That is, proponents of connectionism
argue that the whole system can be interpreted as representing
some state of affairs without it having to be supposed that this has to
be achieved by the manipulation of token symbols or representations. The
accusation that connectionism routinely meets is that its view of
cognition fails to capture the essential properties of (human)
information processing, and therefore cannot serve has an appropriate
account of intelligent behaviour. According to the defenders of GOFAI
(Good Old Fashioned Artificial Intelligence) the most that connectionism
can offer is an account of the implementation details of traditional
models at a very low level, but as far as cognition is concerned,
connectionism remains an irrelevant appendix.
In order to evaluate whether or not the pretensions of connectionism
to establish itself as a competing model of cognition are valid, we need
to have a clear idea of what both views are claiming and what exactly
are the distinguishing features of each model. Let us start by
considering the classical view. What does it say about cognition?
Roughly put, it says that it is rule-based manipulation of syntactically
structured symbols. One of the important characteristics of traditional
models is that they operate though rule-governed transformations of
discrete functional elements (the symbols), and have access only to the
form of the symbols (their syntax), not their meaning (that is, their
semantics). This means that, given a task (meaning, a problem to solve)
a symbolic system will proceed by a step-by-step transformation of the
original representation until it reaches its end-goal (solves the
problem it was given.)
Why take this to be a good model of cognition? According to Fodor and
Pylyshyn the advantage of classic AI, and also the reason why it stands
without qualified competitors, is its ability to accommodate the folk
psychological view that mental processes involve real, discrete,
causally significant, syntactically structured mental entities. Note
that what is important in these authors argument is not, strictly
speaking, the idea that GOFAI saves propositional attitudes from greedy
reductionists (e.g., those who would dispense with them in favour of
talk in terms of C-fibers and P-3000 waves), but their claim that a
model of cognition that fails to capture certain features of our common
sense conception of the mind simply cannot be a model of (human)
cognitive capacities.
Thus, what traditional models can do, and connectionist ones can't,
is to emulate the two most important features of our reasoning
processes. To be clear, these two features are a compositional syntax
and a combinatorial semantics. That is, according to Fodor, thoughts
have combinatorial structure which allows atomic units to be combined in
more complex structures to create new thoughts (e.g., if you can think
'honest' and you can think 'friend', then you can also think 'honest
friend'.) By compositional structure, Fodor means that cognitive
processes are sensitive to the configuration of the thoughts they
operate on. That is, the computational processes that drive cognition
are syntactically governed transformations defined exclusively over the
structure and place occupied by the symbols recruited in current tasks.
In this sense, a classical model of intelligent behaviour (basically)
consists in a store of information (data structures) and a set of search
algorithms (rules of inference) that allow it to arrive at a
satisfactory solution based only on transformations guided by the shape
of the symbols available to it. As classicists claim, this basic picture
of a rule-governed system provides the formal architecture of all
cognition.
In opposition to this view, connectionists have long argued that
symbolic systems have only limited success in accounting for all types
of cognition, are unable to deal with incomplete or partial information,
and provide a rather implausible picture of our brains due the
relatively slow speed at which they perform the tasks they are assigned.
What is more, as Smolensky has pointed out (1988), there is the danger
that traditional models may be illegitimately projecting certain
elements of our conscious, linguistically and culturally informed
thought processes onto the brain. In other words, it may well be the
case that the Language of Thought hypothesis, so dear to Fodor, is
nothing more than a misguided view of what our cognitive processes are
like, and thus one that is more likely to delay progress in
psychological research than be of help to it.
But, then, how do connectionist propose to account for cognitive
performance? While friends of connectionism come with many different
flags (some being more extremist about the implications of connectionism
(Churchland 1988), some more cautious (Clark, 1989)), they all share the
idea that representation of information in a system is the emergence of
stable patterns of activation in a network of simple components. Roughly
put, a connectionist network consists in a set of units connected to
each other, each of which has a specific level of activation at any
given time. The connections between units allow the activation level of
one unit to influence another's activation level, and this can be done
either through excitatory or inhibitory input. That is, the activation
level of a given unit may either increase or decrease the level of
activation of those units to which it is linked. Moreover, it is a
fundamental characteristic of connectionist models that the connections
between units can be modified; that is, how the activation of one unit
bears on the activation of another is not a matter settle once and for
all, and the network can learn new 'routes' of activation.
This said, how exactly do connectionist models represent, and how do
they work? As said before, representation in a connectionist network is
a function of the whole system. That is, typically a connectionist
representation is an emergent global state into which the network
stabilizes in response to its initial input. The way these global states
are achieved is through the combined activity of multiple units working
in parallel to achieve a stable and coherent pattern of activation.
This, in turn, is made possible due to local rules for the individual
activation of units and rules for changes in the strength ('weight') of
the connections between units. However, it is important to notice that
these rules, contrary to what happens in traditional models, are not
hardwired in the system. They are embedded in the connections between
units and, as mentioned before, can be modified or adjusted. Put in
other words, the representational capacities of connectionist networks
depend on the plasticity of the connections between units, and the rules
that operate to transform a current state of activation into a
subsequent one are mathematical rules for updating the activation of a
unit or change the strength of a connection. Accordingly, 'rules' in a
connectionist system should be understood as 'vectors of activation',
and not as 'if-then' clauses like in classical models. Also, it is worth
emphasizing that representation in a connectionist network is
distributed across activated units, and not a function of a discrete
symbol being tokened within the system as in traditional models.
What, then, are the advantages of connectionist models? To begin
with, connectionist networks have been able to model cognitive
capacities so far deemed intractable by classical models. Examples
include visual pattern recognition, the production of speech, and
decision-making in real-time. Other advantages include flexibility in
processing, the ability to learn, and the capacity to perform on the
basis of partial or incomplete information (that is, connectionist
networks exhibit what is otherwise known as graceful degradation). On
the side of disadvantages, connectionist models are often said to be
unable to capture the rational, or inferential character of thought. As
seen before, this amounts to the accusation that connectionist networks
fail to capture the compositionality of thought, and thus cannot gives
us an account of the essential architecture of cognition.
Now, as seen before, Fodor and Pylyshyn have often championed the
thesis that connectionism can at most provide a low-level implementation
of symbolic structures and processes, but as regards the principles of
(human) information processing they have nothing new to offer. In fact,
if we were to take connectionism as an explanatory model of cognition we
would be heading for a dead-end, since cognition as the connectionists
portray it simply makes no sense. In other words, according to Fodor and
Pylyshyn, either a model of cognition entails rules of composition, or
it isn't a model of cognition at all. Connectionist models operate under
no such rules unless they are simulated in symbolic systems. But this,
as these authors see it, can only be proof that connectionism is merely
a theory of implementation of symbolic structures, and not a model of
cognition on its own right.
Is this true? That is, Are connectionist networks mere implementation
of symbolic AI models? Paul Smolensky has on a number of occasions
argued against this accusation, and tried to show that the structural
and systematic aspects of reasoning can arise out of processes that, in
fact, do not have such a structure. Put otherwise, Smolensky thesis is
that compositionality can be accounted for in connectionist models by
looking at the holistic patterns of activation in the whole network.
However, as Smolensky also notes, while connectionist networks involve a
notion of compositionality, this notion is significantly distinct from
the one offered by traditional models. To be more specific, the
connectionist understanding of compositionality rests on the use of
distributed representation across patterns of activation, and it is the
sum of unit activations that properly constitute the composit elements
of representation. Plus, as Smolensky also notes, connectionist models
are sensitive to the structure of the inputs manipulated, the difference
being that connectionist information processing operate via
statistical-sensitive processes, and not via processes sensitive to the
syntax of discrete symbols composed in a Language of Thought. For the
connectionist, there is no language of thought at the level of mental
structure and processing of representation.
As can be seen, there are significant differences in the story each
side has to tell about the symbolic level of cognition. Classicists
insist that knowledge is encoded in discrete symbolic structures, and
that thought proceeds through rules of inference. What is more,
according to the classicist model, the level of semantic interpretation
pertains to the level of symbols being manipulated according to the
rules that define the system. In this sense, representation in a classic
model is a discrete function of symbols being tokened. For
Connectionists, on the contrary, representations are highly distributed
patterns of activation in a network of individuals units, and there is a
radical separation between the level at which formal rules of
transformation are applied and the level at which semantic
interpretation can be ascribed. That is, in a connectionist model the
semantic interpretable level pertains to whole patterns of activation,
and the level of rule-manipulation to the sub-symbolic domain of
individual activation of single units.
Now, as I see it, these differences in mode of representation and
information processing (or rule-governed manipulations) are enough to
set the two models as competing theories of cognition. In other words,
connectionism is not merely a theory of how a language of thought could
be implemented in physical structures. It offers an alternative account
of the fundamental principles of intelligent behaviour, as well as a
different view of the architecture of cognition. Moreover, it should be
noticed that the mode of explanation in connectionist models is
significantly different from that offered by traditional models. These
remain fairly close to folk psychological explanations of behaviour, in
that they cite the tokening of beliefs and desires (symbols) to explain
the causal structure of thinking processes. Connectionist models, on the
other hand, appeal to statistical rules of inference and inductive
processes to account for the pattern transitions (that is, sequence of
representations) that constitute the order and coherence of a system's
behaviour.
In brief, then, I consider it to be a mistake to say that the most
that connectionism can offer is a story about the implementation level
of symbolic systems. This, however, does not mean that the debate
between classicism and connectionism has been solve. Both models of
cognition face difficulties and, so far, none can claim to be in a
better position than the other to solve all the problem in AI. Still, it
seems to be undeniable that connectionism provides an alternative view
of our processes of thinking, and should therefore bee evaluated for its
merits and disadvantages in accounting for these, instead of judged for
its ability or inability of showing how our brains could be speaking the
language of thought.
This paper was written with the support of Ministerio da Ciência e
Tecnologia (Grant BM/10514/97). I would also like to thank Dr. Ian
Ravenscroft for comments and suggestions on earlier drafts.
isabel.gois@kcl.ac.uk
References
BETCHEL, W.: (1990) 'Connectionism and the Philosophy of Mind: an
overview' in Lycan (ed.) Mind and Cognition; Oxford; Blackwell,
pp.252-276.
BETCHEL, W. and ABRAHAMSEN, A.: (1991) Connectionism and the Mind;
Oxford; Blackwell.
BODEN, M.: (1991) 'Horses of a Different Colour?' in Ramsey, Stich and
Rumelhart (eds.) Philosophy and Connectionist Theory; Hillsdale;
Lawrence Earlbeum Associates; pp.3-20.
CHURCHLAND, P.: (1988) Matter and Consciousness (revised edition); Ch.
7; Cambridge-Mass.; The MIT Press; pp.123- 166.
CHURCHLAND, P. M. and SEJNOWSKI, T.: (1990) 'Neural Representation and
Neural Computation' in Lycan (ed.) Mind and Cognition; Oxford;
Blackwell; pp.224-252.
CLARK, A.: (1989) 'Beyond Eliminativism' in Mind and Language; vol.4,
Nº4, pp. 251-279.
DAVIS, M.: (1991) 'Concepts, Connectionism, and the Language of Thought'
in Ramsey, Stich and Rumelhart (eds.) Philosophy and Connectionist
Theory; Hillsdale; Lawrence Earlbaum Associates; pp. 229-258.
DENNETT, D.: (1984) 'Cognitive Wheels: the frame problem in AI' in
Hookway (ed) Minds, Machines and Evolution; Cambridge; Cambridge
University Press; pp.190-213.
FODOR, J.: (1983) The Modularity of Mind; Cambridge-Mass.; The MIT
Press.
FODOR and PYLYSHYN: (1988) 'Connectionism and cognitive architecture: a
critical analysis'; Cognition 28, pp. 3-71.
FODOR (1990) ' Why There Still Has To Be A Language of Thought' in Lycan
(ed) Mind and Cognition; Oxford; Blackwell; pp. 282-300.
RAMSEY, STICH and RUMELHART: (1991) 'Connectionism, Eliminativism and
the Future of Folk Psychology' in Greenwood (ed) The Future of Folk
Psychology; Cambridge; Cambridge University Press; pp. 93-119.
SMOLENSKY, P.: (1987) 'The constituent structure of connectionist mental
states: a reply to Fodor and Pylyshyn' in Southern Journal of Philosophy
26 (supplement) pp. 137-161.
SMOLENSKY, P.: (1988) 'On the Proper Treatment of Connectionism' in
Behavioural and Brain Sciences 11, pp.1-23.
SMOLENSKY, P.: (1990) 'Connectionism and the foundations of AI' in
Partridge and Wilks (eds.) The Foundations of Artificial Intelligence;
Cambridge; Cambridge University Press; pp-306-326.
STERNELY, K. (1990) 'Connectionism' in The Representational Theory of
the Mind; Ch 8; Oxford, Blackwell.