Intel Presentation

Download Report

Transcript Intel Presentation

Neural Networks:
Capabilities and Examples
L. Manevitz
Computer Science Department
HIACS Research Center
University of Haifa
L. Manevitz
U. Haifa
1
What Are Neural Networks?
What Are They Good for?
How Do We Use Them?
• Definitions and some history
• Basics
– Basic Algorithms
– Examples
• Recent Examples
• Future Directions
L. Manevitz
U. Haifa
2
Natural versus Artificial Neuron
• Natural Neuron
L. Manevitz
McCullough Pitts Neuron
U. Haifa
3
Definitions and History
• McCullough –Pitts Neuron
• Perceptron
• Adaline
• Linear Separability
• Multi-Level Neurons
• Neurons with Loops
L. Manevitz
U. Haifa
4
Sample Feed forward Network
(No loops)
•Weights
•Output
•Weights
•Weights
•Input
•Wji
•Vik
F(S wji xj
L. Manevitz
U. Haifa
5
Replacement of Threshold
Neurons with Sigmoid or
Differentiable Neurons
•Sigmoid
•Threshold
L. Manevitz
U. Haifa
6
Reason for Explosion of
Interest
• Two co-incident affects (around 1985 –
87)
– (Re-)discovery of mathematical tools and
algorithms for handling large networks
– Availability (hurray for Intel and company!) of
sufficient computing power to make
experiments practical.
L. Manevitz
U. Haifa
7
Some Properties of NNs
• Universal: Can represent and
accomplish any task.
• Uniform: “Programming” is changing
weights
• Automatic: Algorithms for Automatic
Programming; Learning
L. Manevitz
U. Haifa
8
Networks are Universal
• All logical functions represented by three level
(non-loop) network (McCullough-Pitts)
• All continuous (and more) functions represente
by three level feed-forward networks (Cybenko
al.)
• Networks can self organize (without teacher).
• Networks serve as associative memories
L. Manevitz
U. Haifa
9
Universality
• McCullough-Pitts: Adaptive Logic Gates;
can represent any logic function
• Cybenko: Any continuous function
representable by three-level NN.
L. Manevitz
U. Haifa
10
Networks can “LEARN” and
Generalize (Algorithms)
• One Neuron (Perceptron and Adaline) Very
popular in 1960s – early 70s
– Limited by representability (only linearly separable
• Feed forward networks (Back Propagation)
– Currently most popular network (1987 –now)
• Kohonen self-Organizing Network (1980s –
now)(loops)
• Attractor Networks (loops)
L. Manevitz
U. Haifa
11
Learnability (Automatic
Programming)
• One neuron: Perceptron and Adaline
algorithms (Rosenblatt and Widrow-Hoff)
(1960s –now)
Feed forward Networks: Backpropagation
(1987 – now)
Associative Memories and Looped
Networks (“Attractors”) (1990s – now)
L. Manevitz
U. Haifa
12
Generalizability
• Typically train a network on a sample set
of examples
• Use it on general class
• Training can be slow; but execution is
fast.
L. Manevitz
U. Haifa
13
Perceptron
•weights
w x
 threshold
 A in receptive field
kdkdkfjlll
w x
i i
i
i
   The letter A is in the receptive field.
•Pattern
Identification
•(Note: Neuron
is trained)
L. Manevitz
U. Haifa
14
Feed Forward Network
•weights
w x
i i
•weights
 threshold
 A in receptive field
kdkdkfjlll
L. Manevitz
U. Haifa
15
Classical Applications
(1986 – 1997)
• “Net Talk” : text to speech
• ZIPcodes: handwriting analysis
• Glovetalk: Sign Language to speech
•
Data and Picture Compression: “Bottleneck”
• Steering of Automobile (up to 55 m.p.h)
• Market Predictions
• Associative Memories
• Cognitive Modeling: (especially reading, …)
L. Manevitz U.(Finnish)
Haifa
• Phonetic Typewriter
16
Neural Network
• Once the architecture is fixed; the only
free parameters are the weights
• Thus Uniform Programming
• Potentially Automatic Programming
• Search for Learning Algorithms
L. Manevitz
U. Haifa
17
Programming: Just find the
weights!
• AUTOMATIC PROGRAMMING
• One Neuron: Perceptron or Adaline
• Multi-Level: Gradient Descent on
Continuous Neuron (Sigmoid instead of
step function).
L. Manevitz
U. Haifa
18
Prediction
•delay
•Input/Output
•NN
L. Manevitz
U. Haifa
•Compare
19
Training NN to Predict
L. Manevitz
U. Haifa
20
Finite Element Method
• Numerical Method for solving p.d.e.s
• Many user chosen parameters
• Replace user expertise with NNs.
L. Manevitz
U. Haifa
21
FEM Flow chart
L. Manevitz
U. Haifa
22
Problems and Methods
L. Manevitz
U. Haifa
23
Finite Element Method and
Neural Networks
• Place mesh on body
• Predict where to adapt mesh
L. Manevitz
U. Haifa
24
Placing Mesh on Body
(Manevitz, Givoli and Yousef)
• Need to place geometry on topology
• Method: Use Kohonen algorithm
• Idea: Identify neurons with FEM nodes
– Identify weights of nodes with geometric location
– Identify topology with adjaceny
– RESULT: Equi-probably placement
L. Manevitz
U. Haifa
25
Kohonen Placement for FEM
• Include slide from Malik’s work.
L. Manevitz
U. Haifa
26
Self-Organizing Network
•Weights from
input to
neurons
•Topology
between
neurons
L. Manevitz
U. Haifa
27
Self-Organizing Network
•Weights from input give “location”
to neuron
•Kohonen algorithm results in
“winner” neuron
•After training, close input patterns
have topologically close winners
•Results in Equiprobable Continuous
Mapping (without
teacher)
L. Manevitz U. Haifa
28
Placement of Mesh via Self
Organizing NNs
L. Manevitz
U. Haifa
29
Placement of Mesh via Self
Organizing NNs2
Iteration 0
Iteration 6000;
Quality =223
Iteration 500;
Quality =288
Iteration 12000;
Quality
= 208
L. Manevitz
U. Haifa
Iteration 2000;
Quality = 238
Iteration 30000;
Quality =20230
Comparison of NN and
PLTMG
PLTMG (249 nodes)
NN (225 nodes); Quality = 279
f ( x, y )  u xx  u yy where u ( x, y )  e
Node
Value
Error
Error
Pltmg
2.4 E-02
4.51 E-02
NN
7.5 E-03
9.09E-03
L. Manevitz
U. Haifa
 ( x 2)2
e
( y 2)2
31
FEM
Temporal
Adaptive
Meshes
L. Manevitz
U. Haifa
32
Prediction of Refinement of
Elements
• Method simulates time
• Current adaptive method uses gradient
• Can just MISS all the action.
• We use NNs to PREDICT the gradient.
• Under development with Manevitz, Givoli
and Bitar.
L. Manevitz
U. Haifa
33
Training NN to Predict2
L. Manevitz
U. Haifa
34
Refinement Predictors
•Need to choose features
•Need to identify kinds of elements
L. Manevitz
U. Haifa
35
Other Predictions?
• Stock Market (really!)
• Credit Card Fraud (Master Card, USA)
L. Manevitz
U. Haifa
36
Surfer’s Apprentice Program
• Manevitz and Yousef
• Make a “model” of user for retrieving
information from internet.
• Many issues: here focus on retrieval of
new pages similar to other pages of
interest to user. Note ONLY POSITIVE
DATA.
L. Manevitz
U. Haifa
37
L. Manevitz
U. Haifa
38
Bottleneck Network
•Train to Identity on Sample Data
•Should be identity only on similar data
•NOVELTY FILTER
L. Manevitz
U. Haifa
39
How well does it work?
• Tested on Standard Reuter’s Data Base.
• Used 25% for training
• Withholding information on
representation
• The best method for retrieval using only
positive training. (Better than SVM, etc.)
L. Manevitz
U. Haifa
40
How to help Intel? (Make
Billions? Reset NASDAQ)
• Branch prediction?
• (Note similarity to FEM refinement.)
• Perhaps can use to give predictor that is even
user or application dependent.
• (Note: Neural activity is, I am told, natural for
VLSI design and there have been several such
chips produced.)
L. Manevitz
U. Haifa
41
Other Different Directions
• Modify basic model to handle temporal
adaptivity. (Occurs in real neurons
according to latest biological information.)
• Apply to model human diseases, etc.
L. Manevitz
U. Haifa
42