Presentation

Download Report

Transcript Presentation

FUSION OF
LANGUAGE AND THOUGHT
PROCESSES FOR CTS
Collaborative Technologies and
Systems Conference 2003, Orlando,
Florida
Leonid Perlovsky
Technical Advisor, AFRL
CTS AND INTELLIGENT AGENTS

Collaborative systems include multiple interacting
intelligent agents
– a human, machine, device or software code

Agents are
– significantly autonomous and goal-oriented, perform
various functions, and communicate with other agents
– equipped with sensors or collect data, receive
communications, extracts information
– use existing knowledge, integrate new information into
producing new knowledge, send communications
– embody the concept of life and intelligence

What are the required intelligent agent
technologies?
INTELLIGENT AGENT
TECHNOLOGIES

Interfaces and access
– man-machine and machine-machine interfaces
– knowledge and data access

Understanding of
– language, situations, and environment

Fusion
– knowledge and data from diverse sources and disciplines,

Decision making
– heterogeneous environment, with inaccurate data,
uncertain knowledge and intuitions
– information exchange, knowledge management

Abilities for thinking and language
– a mysterious territory
LANGUAGE AND THINKING
PAST

Artificial Intelligence, 1950s-1980s
– logical rules
– no principal difference between thinking and language
– failed, yet no replacements to logic when combining L. & T.

Linguistics
– Chomskyan linguistics, computational linguistics, cognitive
linguistics: relate words to words
– No relation to surrounding world

Closely related and intertwined evolution in
human mind
– science cannot tell us yet what is language without thinking
or v.v.
LANGUAGE AND THINKING
FUTURE

Thinking and understanding
– identify “concepts” of objects, relationships and situation
in sensory data
– relate concepts to needs and emotions
– relate concepts to behavior

Language is a part of mind
– involved in concepts, emotions, instincts
– closest to concepts
– words are combined into phrases like retinal signals are
combined into objects (?)

Where language and thoughts come
together?
– Concepts? Not logical rules and artificial intelligence
again?
PHYSICS AND MATHEMATICS OF MIND
RANGE OF CONCEPTS

Logic is sufficient to explain mind
– [Newell, “Artificial Intelligence”]

No new specific mathematical concepts are needed
– Mind is a collection of ad-hoc principles, [Minsky]

Specific mathematical constructs describe the
multiplicity of mind phenomena
– “first physical principles of mind”
– [Grossberg, Zadeh, Perlovsky,…]

Quantum computation
– [Hameroff, Penrose, Perlovsky,…]

New unknown yet physical phenomena
– [Josephson, Penrose]
GENETIC ARGUMENTS
FOR THE “FIRST PRINCIPLES”

There is about 30,000 genes in human genome
– relate concepts to needs

Only about 2% difference between human and apes

Say, 1% difference between human and ape minds
– Only about 300 proteins

It is likely, that few general principles of concept
learning are required to explain our ability to operate
with concepts
– If we count “a protein per concept”
– If we count combinations: 300300 ~ unlimited => all
languages could have been genetically h/w-ed (!?!)

Languages are not genetically hardwired
– Because they have to be flexible and adaptive
INFORMATION PROCESSING
AND UNDERSTANDING

Understanding the meaning of signals (visual,
acoustic, text)
– Identify objects in signals
• signals -> concepts; or words -> phrases
– Associate relevant objects
• objects -> scenes;

or phrases -> more general concepts
In this task the human mind is by far superior
qualitatively to existing mathematical methods
– Effort has been devoted toward incorporating “biological
lessons” into smart algorithms, yet success has been
limited
– Why is this so and how to overcome existing limitations?
REASONS FOR PAST LIMITATIONS

The basis of human intelligence is in combining
conceptual understanding with emotional evaluation
– became exceedingly well appreciated among psychologists and
neurobiologists during the last ten years
– human understanding without emotional involvement is basically
flawed [Damasio]

This new understanding has not been accepted by
mathematical and engineering community
– mathematical laws governing emotional involvement into thinking
process have not been well known
– there is a long-standing cultural belief that emotions are opposite
to thinking and intellectually inferior
• Socrates, Plato, Aristotle
• reiterated by founders of Artificial Intelligence [Newell]
FUNDAMENTAL MATHEMATICAL
PROBLEM

Combinatorial Complexity (CC):
– understanding involves evaluating a large number
of combinations
– words into sentences, pixels or samples into
objects, objects into scenes
– a general problem of all data association methods

CC was encountered for over 50 years
– statistical pattern recognition and neural networks:
CC of learning requirements
– rule-based systems, expert systems, and AI, in the
presence of variability: CC of rules
– model-based systems, utilizing adaptive models:
CC of computations (N and NP complete)
– Chomskyan linguistics: (1-1957) rule-based, (21981) model-based (rules and parameters)
CC AND TYPES OF LOGIC

CC is related to formal logic
– law of excluded third
• every logical statement is either true or false
– there could be no similarity/distance measure between
logical statements
– CC is Godel’s “incompletenes” in a finite system

Multivalued and fuzzy logic eliminated the “law of
excluded third”
– yet, they are based on the math. of formal logic
– e.g. fuzzy logic systems are either too fuzzy or too crisp

A similarity measure between logical statements was
not introduced
STRUCTURE OF MIND

Concepts
– Models of objects, their relations, and situations
– Evolved to satisfy instincts

Instincts
– Internal sensors (e.g. sugar level in blood)

Emotions
– Neural signals connecting instincts and concepts
• e.g. a hungry person sees food all around

Behavior
– Models of goals (desires) and muscle-movement

Hierarchy
– Concept-models and behavior-models are organized in a “loose”
hierarchy
SIMILARITY MEASURE AND
DYNAMIC LOGIC

A similarity measure
– based on a similarity between models and data (words)
• equivalent to emotional evaluative signals
– leads to dynamics (equations of motions) improving
concept-models by maximizing similarity (dynamic logic)

Dynamic Logic unifies formal and fuzzy logic
– initial “fuzzy model-concept” dynamically evolve into
“formal-logic or crisp model-concept”

Overcomes CC of model-based recognition
– fast algorithms
• low-polynomial M*N, instead of combinatorial MN
– associates pixels into objects (or words into phrase-models)
without CC
THINKING

Understanding and learning
– Model-concepts always have to be adapted to incoming
signals or data
• lighting, surrounding, new objects and situations

Instinct for knowledge and understanding
– Models-concepts are adapted and improved even when there
is no concrete “bodily” needs
– Increase similarity between models and world

Emotions related to knowledge instinct
– Satisfaction or dissatisfaction
• change in similarity between models and world
– Harmony or disharmony: aesthetic emotion

Behavior related to knowledge
– Adaptation and learning of concepts (behavior in the mind)
MODELING FIELD THEORY
basic two-layer hierarchy: from signals to objects

Signals and Concept-Models
– signals x(n), n = 1,…,N
– model-object Mm(Sm,n), parameters Sm, m = 1, …;

Goal: learn object-models and understand signals
– associate samples n with models m and find parameters Sm
– learn signal-contents of objects (and object properties)

Maximize similarity, between signals and models, L
– knowledge instinct
– Likelihood or mutual information, L = l({x}) =  l(x(n))
– l(x(n)) =  r(m) l(x(n) | Mm(Sm,n)) (M may depend on n)
– CC: L contains MN items: all associations of words and models
n
m
DYNAMIC LOGIC ALGORITHM (DLA)
(non-combinatorial solution)

Start with a set of signals and unknown object-models
– any parameter values Sm
– associate fuzzy object-model with its contents (signal composition)
– (1)
f(m|n) = r(m) l(n|m) /  r(m') l(n|m')
m'

Improve parameter estimation
– (2)
Sm = (1- a) Sm + a  f(m|n) [ll(n|m)/Mm]*[Mm/Sm]
n
• (a determines speed of convergence)
– learn signal-contents of objects

Continue iterations (1)-(2). Theorem: MFT is converging
- similarity increases on each iteration
- aesthetic emotion is positive during learning

Each concept-model is an agent
– semi-independent, interacting with other agents
• competing for evidence (among signals)
– learning its properties and recognizing its signals
LINGUISTICS
WORDS, CONCEPTS AND GOALS

Text is a (loose) hierarchy of concepts
– Word is a concept; it acquires meaning in a phrase
– Phrase-concept acquires meaning in a “paragraph”,…

Model-concepts (e.g. phrases made up of words)
– Simplistic “bag”-model
• a set or collection of words
– More complex models: word order and relationships
– Real-language models
• grammar (Chomsky, Pinker, Jackendoff, Rieger, Mehler…)

Goal-instinct
– A search engine: find conceptual similarity between a query and
text
• analyze both in terms of concepts
– Learn and identify model-concepts in texts (language instinct =
knowledge instinct)
MODELING FIELD THEORY
basic two-layer hierarchy: words and phrase-concepts

Words and Concept-Models
– words w(n), n = 1,…,N
– model-phrase Mm(Sm,n), parameters Sm, m = 1, …;
– Simplistic “bag”-model: Mm = Sm = {w}m

Goal: learn phrase-models
– associate words n with models m and find parameters Sm
– learn word-contents of phrases (and grammatical relationships)

Maximize similarity, between words and models, L
– language instinct = knowledge instinct
– Likelihood or mutual information, L = l({w}) =  l(w(n))
– l(w(n)) =  r(m) l(w(n) | Mm(Sm,n)) (M may depend on n)
– CC: L contains MN items: all associations of words and models
n
m
DYNAMIC LOGIC ALGORITHM (DLA)
(non-combinatorial solution)

Start with a large body of text and unknown phrase-models
– any parameter values Sm
– associate fuzzy phrase-model with its contents (words)
– (1)
f(m|n) = r(m) l(n|m) /  r(m') l(n|m')
m'

Improve parameter estimation
– (2)
Sm = (1- a) Sm + a  f(m|n) [ll(n|m)/Mm]*[Mm/Sm]
n
• (a determines speed of convergence)
– learn word-contents of phrases (and grammatical relationships)

Continue iterations (1)-(2). Theorem: MFT is converging
- similarity increases on each iteration
- aesthetic emotion is positive during learning

Each phrase-concept-model is an agent
– semi-independent, interacting with other agents
• competing for evidence (among words)
– learning its properties and recognizing its words
CATCH ME BY HAND

If I give you enough time you’ll be able to catch me
on the previous slide
– the “bag”-model is non-differentiable
– this is a principal moment, learning non-differentiable models requires
sorting through combinations
– lead to combinatorial complexity
– differentiable models can be defined with a little trick
INTEGRATED
LANGUAGE AND THINKING

Where language and thoughts come together?
– concept-models have linguistic and objective aspects
– a fuzzy concept m has sensory and linguistic model
• Mm = { Mmsensory,Mmlinguistic };
– language and thoughts are fused at fuzzy pre-conscious level
• before concepts are learned

Understanding language and sensory data
– baby learning: “look, this is a car”
– each linguistic model has an empty “slot” for objects and
situations in surrounding world
– each sensory or situational model has an empty “slot” for a word
or phrase
– language participates in thinking and v.v.

Two types of information help learning and
understanding each other
– help associating signals, words, models, and behavior
INNER LINGUISTIC FORM
HUMBOLDT, the 1830s

In the 1830s Humboldt discussed two types of
linguistic forms
– words’ outer linguistic form (sounds) – a formal designation
– and inner linguistic form (???) – creative, full of potential

This remained a mystery for rule-based AI,
structural linguistics, Chomskyan linguistics
– rule-based approaches using the mathematics of logic make
no difference between formal and creative

In MFT and DLA there is a difference
– static form of learned (converged) concept-models
– dynamic form of fuzzy concepts, with creative learning
potential and emotional content
WHY MIND AND EMOTIONS?

A lot about the mind can be explained: concepts, instincts,
emotions, conscious and unconscious, intuition, aesthetic ability,…
– but, isn’t it sufficient to solve mathematical equations or to code a
computer and execute the code?

A simple yet profound question
–
–
–
–
–

the answer is in history and practice of science
Newton laws do not contain all of the classical mechanics
Maxwell equations do not exhaust radars and radio-communication
a physical intuition about the system is needed
an intuition about MFT was derived from biological, linguistic,
cognitive, neuro-physiological, and psychological insights into
human mind
Practical engineering applications of DLA require
biological, linguistic, cognitive, neuro-physiological,
and psychological insights in addition to mathematics
MFT THEORY OF MIND

MFT dynamics: elementary thought process
– a large number of model-concepts compete for incoming signals
– uncertainty in models corresponds to uncertainty in associations f(m|n)
– eventually, one model (m') wins a competition for a subset {n'} of input
signals w(n), when parameter values match object properties, and f(m'|n)
values become close to 1 for n{n'} and 0 for n{n'}
– upon convergence, the entire set of input signals {n} is divided into subsets,
each associated with one model-object
– fuzzy a priori concepts (unconscious) become crisp concepts (conscious)
• dynamic logic

Elementary thought process, consciousness and unconscious
– Aristotle: in thinking, an a priori form-as-potentiality (fuzzy model) meets
matter (signals) and becomes a form-as-actuality (a concept)
– Jung: conscious concepts are developed by mind based on inherited
structures of mind, archetypes, inaccessible to consciousness
– Grossberg: models attaining a resonant state (winning the competition for
signals and becoming crisp in this process) reach consciousness
MFT THEORY OF MIND
-understanding
Incoming signals, {x,w} are associated with model-concepts (m)
– creating phenomena (of the MFT-mind), which are understood as objects, situations,
phrases,…
– in other words signal subsets acquire meaning (e.g., a subset of retinal signals
acquires a meaning of a chair)

Several aspects of understanding and meaning
– concept-models are connected (by emotional signals) to instincts and to behavioral
models that can make use of them for satisfaction of bodily instincts
– an object is understood in the context of a more general situation in the next layer
consisting of more general concept-models (satisfaction of knowledge instinct)
• each recognized concept-model (phenomenon) sends (in neural terminology: activates) an
output signal
• a set of these signals comprises input signals for the next layer models, which ‘cognize’ more
general concept-models
• this process continues up and up the hierarchy towards the most general models: models of
universe (scientific theories), models of self (psychological concepts), models of meaning of
existence (philosophical concepts), models of a priori transcendent intelligent subject
(theological concepts)
– neural brain organization: individual modules, which form approximate hierarchies,
along with a number of “parallel” and “loop-like” pathways
SIGNS AND SYMBOLS
mathematical semiotics

Signs: stand for something else
– non-adaptive entities (mathematics, AI)
– brain signals insensitive to context (Pribram)

Symbols
– signs (mathematics, AI)
– psychological processes connecting conscious
and unconscious (Jung)
– brain signals sensitive to context (Pribram)
– processes of sign interpretation

Mathematics of symbol-processes
– relationships to thinking
– relationships to language
MFT THEORY OF SYMBOLS
-mathematical semiotics
Semiotics studies symbol-content of culture

Example: consider a written word "chair"
– It can be interpreted by mind to refer to something else: an entity in the
world, a specific chair, or the concept "chair" in the mind
– In this process, the mind, or an intelligent system is called an interpreter,
the written word is called a sign, the real-world chair is called a
designatum, and the concept in the interpreter's mind, the internal
representation of the results of interpretation is called an interpretant of the
sign
– The essence of a sign is that it can be interpreted by an interpreter to refer
to something else, a designatum

This is a simplified description of a thinking process, called semiosis
– its mechanism is given by the elementary thought process

Elementary thought process involving consciousness and unconscious,
concepts and emotions, is a dynamic symbol process
– a much more complicated entity than was originally envisioned by founders
of “symbolic AI”
MFT THEORY OF MIND
- aesthetic emotions and beauty 
Aesthetic emotions (not related to bodily satisfaction)
– Instincts for knowledge and language (learning concept-models)
– Emotions (satisfaction-dissatisfaction): harmony-disharmony
– Maximize similarity between models and world
• between our understanding of how things ought to be and how they actually are
in the surrounding world; Kant: aesthetic emotions

Beauty
– Harmony is an elementary aesthetic emotion; higher aesthetic emotions
• development of more complex “higher” models
– The highest forms of aesthetic emotion, beauty
• related to the most general and most important models
• models of the meaning of our existence, of our purposiveness or intentionality
• beautiful object stimulates improvement of the highest models of meaning
– Beautiful “reminds” us of our purposiveness
• Kant called beauty “aimless purposiveness”: not related to bodily purposes
• he was dissatisfied by not being able to give a positive definition
– knowledge instinct
• absence of positive definition remained a major source of confusion in
philosophical aesthetics till this very day
MFT THEORY OF MIND
- physical intuition 
Intuitive perception (imagination) of object-models and
their relationships with objects in the world
– involves fuzzy unconscious concept-models
– in process of being learned and adapted
• toward crisp and conscious models, a theory
– such models satisfy or dissatisfy the knowledge instinct before they
are accessible to consciousness, hence the complex emotional feel of
an intuition

Beauty of a physical theory discussed often by physicists
– related to satisfying our feeling of purpose in the world
– satisfying our need to improve the models of the meaning in our
understanding of the universe
REAL WORLD APPLICATIONS

Many applications have been developed
– Government
– Medical
– Commercial

Sensor signals processing and object recognition
– Variety of sensors

Internet search engines
– Based on text understanding

Financial market predictions
– Market crash on 9/11 predicted a week ahead
BACK UP
MFT
and Buddhism
MFT
vs. biology
Classical
methodology flowchart
MFT
flowchart
MFT
vs. inverse problems
MFT
predictions and testing
MFT
future directions
Publications
MFT AND BUDDHISM

Fundamental Buddhist notion of “Maya”
– the world of phenomena, “Maya”, is meaningless deception
– penetrates into the depths of perception and cognition
– phenomena are not identical to things-in-themselves

Fundamental Buddhist notion of “Emptiness”
– “consciousness of bodhisattva wonders at perception of emptiness
in any object” (Dalai Lama 1993)
– any object is first of all a phenomenon accessible to cognition
– value of any object for satisfying the “lower” bodily instincts is
much less than its value for satisfying higher needs, knowledge
instinct
– Bodhisattva’s consciousness is directed by knowledge instinct
– concentration on “emptiness” does not mean emotional emptiness,
but the opposite, the fullness with highest emotions related to the
knowledge instinct, beauty and spiritually sublime
MFT vs. BIOLOGY OF EYE
 Human eye is part of the brain
- Integrated sensor-processor system in both design and in operations
• multi-layer hierarchical system
• integrated adaptive optimized resource allocation
- Feedback from higher layers to lower layers is essential
• eye-brain neural pathway contains more feedback connections than feedforward
ones
 Adaptive, joint optimization of sensor-processor system network
- Based on a hierarchy with feedback among layers and modules
- What is the nature of this feedback?
 MFT hierarchy with feedback
- Every layer has 5 basic modules/elements:
(1) incoming signals (structured at lower layer, unstructured at the current layer)
(2) models: phenomenology (emissivity, geometry) and simulation codes
(3) similarity measure between signals and models
(4) adaptation mechanism
(5) outgoing signals (a structure: sign-concept)
CLASSICAL METHODOLGY
Result: Conceptual objects
Recognition
signals
MODELS/templates
•objects, sensors
•physical models
Sensors /
Effectors
Input: World/scene
MFT
basic two-layer hierarchy: signals and concepts
Result: Conceptual objects
Correspondence /
Similarity measures
signals
Attention / Action
Sim.signals
MODELS
•objects, sensors
•physical models
Sensors /
Effectors
signals
Input: World/scene
MFT VS. INVERSE SCATTERING

Inverse Scattering in Physics
– reconstruction of target properties by “propagating back” scattered fields
– usually complicated, ill-posed problems (exception: CATSCAN)

Biological systems (mind) solve this problem all the time
– by utilizing prior information (in feedback neural pathways)

Classical Tikhonov’s inversion cannot use knowledge
– regularization parameter () is a constant
– Morozov’s modification can utilize prior estimate of errors for 

Inverse Scattering using MFT
– can utilize any prior knowledge ( became an operator)
– utilizes prior knowledge adaptively ( depends on parameters)
MFT PREDICTIONS AND TESTING

General neural mechanisms of the elementary thought process
– confirmed by neural and psychological experiments
– includes neural mechanisms for bottom-up (sensory) signals, top-down
(“imagination”) model-signals, and the resonant matching between the two

Adaptive modeling abilities
– well studied: adaptive parameters are synaptic connections

Instinctual learning mechanisms
– studied in psychology and linguistics

Ongoing and future research will confirm, disprove, or suggest
modifications to
–
–
–
–
mechanisms of language and thinking integration
mechanisms of model parameterization and parameter adaptation
reduction of fuzziness during learning
similarity measure as a foundation of knowledge and language instincts
MFT FUTURE DIRECTIONS

Developing MFT models based on known linguistic
models

Differentiated forms of knowledge instinct
– highly differentiated emotions are involved in human
conversation and human thinking
– multiple measures of similarity, differentiated knowledge
instinct
• differentiated emotional concepts

Quantum Computing MFT devices
PUBLICATIONS
OXFORD UNIVERSITY PRESS
www.oup-usa.org