Michael G. Dyer (2001) Computationalism, Neural Networks and Minds, Analog or Otherwise. Psycoloquy: 12(042) Symbolism Connectionism (9)

Volume: 12 (next, prev) Issue: 042 (next, prev) Article: 9 (next prev first) Alternate versions: ASCII Summary
Topic:
Article:
PSYCOLOQUY (ISSN 1055-0143) is sponsored by the American Psychological Association (APA).
Psycoloquy 12(042): Computationalism, Neural Networks and Minds, Analog or Otherwise

COMPUTATIONALISM, NEURAL NETWORKS AND MINDS, ANALOG OR OTHERWISE
Commentary on Harnad on Symbolism-Connectionism

Michael G. Dyer
Computer Science Department
3532 Boelter Hall
UCLA
Los Angeles CA 90024, USA

Dyer@cs.ucla.edu

Abstract

Rebuttals are given to Harnad's (2001) argument that only analog systems (by virtue of their physicality) might be capable of supporting Mind. These rebuttals are based on the assumption that Mind arises due to how matter is organized (as opposed to its specific physicality). Mind could arise in non-analog systems by means of simulating Mind's organizational properties and dynamics at an appropriate level. Many observable behaviors of Mind (e.g., language comprehension, creativity, learning, common-sense reasoning) have already been demonstrated (albeit at primitive stages) in non-analog (i.e. computational) systems, at both symbolic and artificial neural network levels of organization.

    REPRINT OF: Dyer, G. (1993) Computationalism, neural networks and
    minds, analog or otherwise. Think 2: 12-78 (Special Issue on
    "Connectionism versus Symbolism" D.M.W. Powers & P.A. Flach, eds.).
    http://cwis.kub.nl/~fdl/research/ti/docs/think/2-1/index.stm

1. A working hypothesis of computationalism is that Mind arises, not from the intrinsic nature of the causal properties of particular forms of matter, but from the organization of matter. If this hypothesis is correct, then a wide range of physical systems (e.g. optical, chemical, various hybrids, etc.) should support Mind, especially computers, since they have the capability to create/manipulate organizations of bits of arbitrarily complexity and dynamics. In any particular computer, these bit patterns are quite physical, but their particular physicality is considered irrelevant (since they could be replaced by other physical substrata).

2. When an organizational correspondence is set up between patterns in a computer and patterns in some other physical system, we tend to call the computer patterns 'symbols'. The correspondence, however, is usually only to some level of organization. In traditional Artificial Intelligence (AI), a small number of symbols may correspond, for instance, to a entire proposition. In Connectionist Modeling (CM), a symbol more commonly will correspond to a single neuron (or perhaps just a single chunk of a neurotransmitter within a neuron). Thus, the major issues that distinguish AI from CM are more involved with what appropriate levels of granularity capture essential organizational dynamics, rather than with any (purported) abandonment of computationalism within the CM paradigm (Dyer, 1991). To my knowledge, both paradigms are strongly committed to computationalism.

3. In analog systems, however, physicality is central to organizational dynamics. For instance, to find the minimal energy state of water flowing downhill, we simply set up the terrain in a gravitational field, add water, and then let nature 'be itself'. But is there some extra capability that an analog system (A1) has over its organizational counterpart (C1) on the computer?

5. Clearly, A1 remains physically distinct from C1 -- e.g. simulating the organization of water molecules in a computer will never make the computer physically wet. But what about the organizational capabilities of A1 with respect to C1? Is there some organizational behavior that A1 is capable of but C1 is not? The answer to this question depends on the level at which the organizational correspondence has been established.

6. If nature is fundamentally discrete (as is the current view of quantum physics), then each symbol could conceivably correspond to some smallest (indivisible) unit of matter-energy or space-time. Thus, C1 modeled at a quantum level would have all the organizational properties of A1 (still without exhibiting any of A1's physicality). Hopefully, this extremely detailed level of organization is not needed to exhibit Mind.

7. If Minds do not arise solely from the organization of matter (but require specific forms of physicality) then both Harnad (2001) and Searle (1980) are right -- no computer could ever have Mind just by virtue of its organization. But are there any persuasive arguments for needing some particular physical substratum?

8. Searle's 'Chinese Room' argument is unpersuasive because there should be no expectation that Searle, in acting as an interpreter (whether at AI, CM or more detailed levels of organization), would understand Chinese. When we implement a natural language understanding (NLP) system 'on top of', say, a Lisp (or Prolog) interpreter, we do not expect that interpreter to understand what the NLP system understands. Thus, Searle's lack of Chinese understanding should come as no surprise (Dyer, 1990a,b).

9. Harnad's 'Transducer' argument is that physical transducers are required for Mind (with analog ones apparently now being Harnad's best candidates). Harnad's argument suffers from the 'Out of Sight, Out of Mind' problem. That is, if we build a Mind-like system (for instance, able to read and understand Harnad's position paper and this commentary) and disconnect it's eyes (and any other sensors/effectors), the system (according to Harnad) would lose its Mind. Harnad's argument also suffers from the 'Virtual Reality' rebuttal, in which we hook up a Mind-like system M to a Virtual Reality system. M is grounded in a sensory reality, but since that entire reality is computer generated, no physical transducers (only simulated ones) are needed (Dyer, 1990a,b).

10. Where does this leave us? Without definitive arguments for the need for special forms of physicality, we are left with both sides essentially arguing over the definition of Mind. The Computationalists define Mind in terms of Mind-like behavior, resulting from the organization of matter at some level of granularity (usually enough to pass either the TT or TTT). The Physicalists simply define Mind as requiring some extra (as yet unexplained) physicality (analog or otherwise). But until some convincing pro-physicality arguments come along, our best strategy should be to judge potential minds in terms of their Mind-like capabilities and behaviors, not their physical substrata.

REFERENCES

Dyer, M. G. (1990a) `Intentionality and Computationalism: Minds, Machines, Searle and Harnad' In: Journal of Experimental and Theoretical Artificial Intelligence 2.4. pp. 303--319.

Dyer, M. G. (1990b) `Finding Lost Minds' (Author's reply to S. Harnad's "Lost in the Hermeneutic Hall of Mirrors"). In: Journal of Experimental and Theoretical Artificial Intelligence, Vol. 2, No. 4.

Dyer, M. G. (1991)`Connectionism versus Symbolism in High-Level Cognition'. In T. Horgan and J. Tienson (eds.). Connectionism and the Philosophy of Mind. Kluwer Academic Publishers, Boston MA. pp. 382--416.

Harnad, S. (2001) Grounding symbols in the analog world with neural nets -- A hybrid model. PSYCOLOQUY 12(034) http://www.cogsci.soton.ac.uk/cgi/psyc/newpsy?12.034

Searle, J. R. (1980) "Minds, brains and programs." Behavioral and Brain Sciences 3: 417-424. http://www.cogsci.soton.ac.uk/bbs/Archive/bbs.searle2.html http://www.bbsonline.org/documents/a/00/00/04/84/index.html


Volume: 12 (next, prev) Issue: 042 (next, prev) Article: 9 (next prev first) Alternate versions: ASCII Summary
Topic:
Article: