Академический Документы
Профессиональный Документы
Культура Документы
Soft
AI
Machine
Learning
automatic
improvement
with
experience
Probability
(not possibility)
to do better
can explain their decisions
Soft Computing
In computer science, soft computing is the use of inexact
solutions to solve computationally hard tasks such as the
solution of NP-complete problems, for which there is no known
algorithm that can compute an exact solution in polynomial time
The essence of soft computing is that unlike the traditional, hard
computing, soft computing is aimed at an accommodation with
the pervasive imprecision of the real world.
Thus, the guiding principle of soft computing is to exploit the
tolerance for imprecision, uncertainty, and partial truth to
achieve tractability, robustness, low solution cost, and better
rapport with reality
- Lotfi Zadeh
Soft computing
Aims to surmount NP-complete problems.
Uses inexact methods to give useful but inexact answers to
intractable problems.
Represents a significant paradigm shift in the aims of
computing - a shift which reflects the human mind.
Tolerant to imprecision, uncertainty, partial truth, and
approximation.
Well suited for real world problems where ideal models are
not available.
Conventional
computing
Soft Computing
requires
a Soft
computing
is
tolerant
of
imprecision.
Often requires a lot of computation time. Can solve some real world problems in
reasonably less time.
Not suited for real world problems for Suitable for real world problems.
which ideal model is not present.
It requires full truth
Imprecise.
Copyright
Unit-1
Neural Networks-1(Introduction & Architecture)
Neuron, Nerve structure and synapse, Artificial Neuron and
rule,
Levels of Brain
Organization
The brain contains both large scale and small scale structures
and different functions take place at the higher and lower
levels.
There is a hierarchy of interwoven levels of organization:
1. Molecules and Ions
2. Synapses
3. Neuronal microcircuits
4. Dendritic trees
5. Neurons
6. Local circuits
7. Inter-regional circuits
8. Central nervous system
The ANNs studied in this module are crude approximations to
levels 5 and 6.
Brain Vs computer
Term
Brain
Speed
Execution time
milliseconds
Processing
Size and
complexity
Storage capacity
14
Computer
is
is
few
Contd
Tolerance
i) Fault tolerant
i) No
fault
ii) Store
and
tolerance
retrieve
ii) Information
information
corrupted
if
even
the
network
interconnectio
connections
ns fails
disconnected.
iii)Accept
iii)No
redundancies
redundancies
Control
mechanism
Depends
on
active chemicals
and
neuron
connections are
15
CPU
Control
mechanism
very simple
is
A BIOLOGICAL NEURON
The most basic element of the human brain is a specific type of
Structure of A BIOLOGICAL
NEURON
Cell body (Soma):The body of neuron cell contains the nucleus and
carries out biochemical transformation necessary to the life of
neurons.
Dendrite: Each neuron has fine, hair like tubular structures
(extensions) around it. They branch out into tree around the cell
body. They accept incoming signals and behave as an input channel.
Artificial Neuron
w0
w0
x1
w1 x1
w1
...
xn
wn
x1
x2
xn
21
wn x n
w1
w2
wn
w1 x1 ... winput
The weighted
signals are
n xn
w
summed.
Thus, 0 a
linear
combination of the input signals
is
z w0weight"
w1 x1 ... (or
wn xn
obtained.
A "free
bias)
, which does not
correspond to any input, is added
to this linear combination
y ( z ) and this
forms a weighted sum
.
A nonlinear activation function
is applied to the weighted sum.
Biological Neuron
Artificial Neuron
Cell
Neuron
Dendrites
Soma
Weights or
interconnections
Net input
Axon
Output
AN ARTIFICIAL NEURON
In Figure various inputs to the network are represented by
AN ARTIFICIAL NEURON
Component 2. Summation Function:
The inputs and corresponding weights are vectors which
AN ARTIFICIAL NEURON
Component 3. Transfer Function:
In the transfer function the summation can be compared
with some threshold to determine the neural output. If the
sum is greater than the threshold value, the processing
element generates a signal and if it is less than the
threshold, no signal (or some inhibitory signal) is
generated.
Both types of response are significant.
The threshold, or transfer function, is generally non-linear.
Component 4. Scaling and Limiting:
After the transfer function, the result can pass through
additional processes, which scale and limit.
This scaling simply multiplies a scale factor times the
transfer value and then adds an offset.
Limiting is the mechanism which insures that the scaled
result does not exceed an upper, or lower bound.
AN ARTIFICIAL NEURON
Component 5. Output Function (Competition):
Each processing element is allowed one output signal,
which it may give to hundreds of other neurons.
Some network topologies modify the transfer result to
incorporate competition among neighboring processing
elements.
First, competition determines which artificial neuron will
be active or provides an output. Second, competitive
inputs help to determine which processing element will
participate in the learning or adaptation process.
Component 6. Error Function and Back-Propagated
Value:
In most learning networks the difference between the
current output and the desired output is calculated as an
error which is then transformed by the error function to
match a particular network architecture.
AN ARTIFICIAL NEURON
This back-propagated value, after being scaled by
Activation functions
The activation function acts as a squashing function,
such that the output of a neuron in a neural network
is between certain values (usually 0 and 1, or -1 and
1).
To make work more efficient and for exact output,
some force or activation is given.
Like that, activation function is applied over the net
input to calculate the output of an ANN.
Information processing of processing element has two
major parts: input and output.
An integration function (f ) is associated with input of
processing element.
Several activation functions are there.
Refer written notes
Neural Networks
Neural
DEFINITION s OF NEURAL
NETWORKS
According to Nigrin (1993)
A neural network is a circuit composed of a very large
number of simple processing elements that are
neurally based. Each element operates only on local
information.
Furthermore each element operates asynchronously;
thus there is no overall system clock.
According to Zurada (1992):
Artificial neural systems, or neural networks, are
physical cellular systems which can acquire, store and
utilize experiential knowledge.
traffic control
Appraisal and valuation of property, etc.,
Betting on horse races, stock markets
Criminal sentencing
Complex physical and chemical process
Data mining, cleaning and validation
Direct mail advertisers
Echo patterns
Employee hiring
Expert consultants
Fraud detection
Hand writing and typewriting
Machinery controls
Medical diagnosis
Music composition
Photos and finger prints
Recipes and chemical formulation
Traffic flows
Voice prediction
Weather prediction
in analyzing the
Characteristics of ANN:
It is Neurally implemented mathematical model
There exists a Large number of processing elements called
neurons in an ANN.
Interconnections with weighted linkage hold informative
knowledge.
Input signals arrive at processing elements through
connections and connecting weights.
Processing elements are able to learn, recall and
generalize from the given data.
Computational power is determined by the collective
behavior of neurons.
ANN is a connection models, parallel distributed
processing models, self-organizing systems, neurocomputing systems and neuro morphic system.
LEARNING ALGORITHMS
An example application
Another application
A credit card company receives thousands of
Problem:
39
Approved or not
applications into
Yes (approved) and
No (not approved)
40
Machine learning is
Like human learning which comes from past experiences.
A computer does not have experiences.
A computer system learns from data, which represent
Accuracy
42
SUPERVISED LEARNING
Generally, a set of patterns is given where the class label of each
Training Data:
k attributes: A1, A2, Ak.
a class: Each example is labelled with a pre-defined
class
The information in the training data should be used to identify the
SUPERVISED LEARNING
In ANN, each input vector requires a
Generalizations of supervised
learning
There are several ways in which the standard
supervised learning problem can be generalized:
Semi-supervised learning : In this setting, the
desired output values are provided only for a subset
of the training data. The remaining data is
unlabeled.
Active learning : Instead of assuming that all of the
training examples are given at the start, active
learning algorithms interactively collect new
examples, typically by making queries to a human
user. Often, the queries are based on unlabeled
data, which is a scenario that combines semisupervised learning with active learning.
Algorithms of Supervised
Learning
Artificial neural network
Boosting (meta-algorithm)
Bayesian statistics
Case-based reasoning
Decision tree learning
Inductive logic programming
Naive bayes classifier
Nearest Neighbor Algorithm
Support vector machines
Random Forests
Ensembles of Classifier
Applications of Supervised
Learning
Bioinformatics
Database marketing
Handwriting recognition
Information retrieval
Object recognition incomputer vision
Optical character recognition
Spam detection
Pattern recognition
Speech recognition
Unsupervised learning
In unsupervised learning, there is no explicit
teacher, and the system forms clusters or natural
groupings" of the input patterns.
That is, there is no supervisor telling us what is right
or wrong; we simply observe some data and try to
describe it in an efficient way with our model.
Approaches to unsupervised learning include
clustering
(e.g.,
k-means,
mixture
models,
hierarchical clustering), and among neural network
models, the self-organizing map (SOM) and adaptive
resonance theory (ART) are commonly used
unsupervised learning algorithms
Unsupervised learning
Example:
tadpole learn to
swim by itself.
In ANN, during training process,
network receives input patterns
and organize it to form clusters.
It is observed that no feedback
is applied from environment to
inform what output should be or
whether they are correct.
The
network itself discover
patterns, regularities, features/
categories from the input data
and relations for the input data
over the output.
Exact clusters are formed by
discovering
similarities
&
dissimilarities so called as self
Reinforcement learning
It is a type of Machine Learning that allows
machines and software agents to automatically
determine the ideal behaviour within a specific
context, in order to maximize its performance.
The basic reinforcement learning model consists of:
A set of environment states
A set of actions
Rules of transitioning between states;
Rules that determine the scalar immediate reward of a
transition; and
Rules that describe what the agent observes.
Examples: Applications to game playing and robot
control
Reinforcement learning
Similar to supervised learning.
Learning based on critic information is called
Contd..
Recurrent Network
Contd..
Maxnet competitive interconnections having
fixed weights.
On-center-off-surround/lateralinhibiton
60
Contd
Processing element
output
can be directed back to the
nodes in the preceding layer,
forming
a
multilayer
recurrent network.
Processing element output
can be directed to processing
element itself or to other
processing element in the
same layer.
61
Architecture
Architecture
Perceptron
Basic model, formulation of learning of weights discussed in class