ARTIFICIAL
NUERAL
NETWORK
 The simplest definition of a neural network, more
properly referred to as an 'artificial' neural network
(ANN), is provided by the inventor of one of the first
neurocomputers, Dr. Robert Hecht-Nielsen. He defines
a neural network as: a computing system made up of a
number of simple, highly interconnected processing
elements, which process information by their dynamic
state response to external inputs.
 ANNs are processing devices (algorithms or actual
hardware) that are loosely modeled after the neuronal
structure of the mamalian cerebral cortex but on much
smaller scales. A large ANN might have hundreds or
thousands of processor units, whereas a mamalian
brain has billions of neurons with a corresponding
increase in magnitude of their overall interaction and
emergent behavior. Although ANN researchers are
generally not concerned with whether their networks
accurately resemble biological systems, some have. For
example, researchers have accurately simulated the
function of the retina and modeled the eye rather
well.

 In information technology, a neural network is a system of
programs and data structures that approximates the operation
of the human brain. A neural network usually involves a large
number of processors operating in parallel, each with its own
small sphere of knowledge and access to data in its local
memory. Typically, a neural network is initially "trained" or fed
large amounts of data and rules about data relationships (for
example, "A grandfather is older than a person's father"). A
program can then tell the network how to behave in response
to an external stimulus (for example, to input from a computer
user who is interacting with the network) or can initiate activity
on its own (within the limits of its access to the external world)
(Eric Davalo and Patrick Naim Assimov, 1984).
 Nonlinearity: A neuron is basically a nonlinear device.
Consequently, a neural network, made up of an interconnection
of neurons, is itself nonlinear. Moreover, the nonlinearity is of a
special kind in the sense that it is distributed throughout the
network.
 A popular paradigm of learning called supervised
learning involves the modification of the synaptic
weights of a neural network by applying a set of
training samples. Each sample consists of a unique
input signal and the corresponding desired
response (Fausett L., 1994). The network is
presented a sample picked at random from the set,
and the synaptic weights (free parameters) of the
network are modified so as to minimize the
difference between the desired response and the
actual response of the network produced by the
input signal in accordance with an appropriate
criterion. The training of the network is repeated
for many samples in the set until the network
reaches a steady state, where there are no further
significant changes in the synaptic weights. The
previously applied training samples may be
reapplied during the training session, usually in a
different order. Thus the network learns from the
samples by constructing an input output mapping
for the problem at hand.
 Neural networks have a built-in
capability to adapt their synaptic
weights to changes in the
surrounding environment. In
particular, a neural network
trained to operate in a specific
environment can be easily
retrained to deal with minor
changes in the operating
environmental conditions.
Moreover, when it is operating in
a nonstationary environment a
neural network can be designed
to change its synaptic weights i
 The massively parallel
nature of a neural network
makes it potentially fast for
the computation of certain
tasks. This same feature
makes a neural network
ideally suited for
implementation using very-
large-scale-integrated
(VLS1) technology.
Artificial nueral network slideshare

Artificial nueral network slideshare

  • 1.
  • 2.
     The simplestdefinition of a neural network, more properly referred to as an 'artificial' neural network (ANN), is provided by the inventor of one of the first neurocomputers, Dr. Robert Hecht-Nielsen. He defines a neural network as: a computing system made up of a number of simple, highly interconnected processing elements, which process information by their dynamic state response to external inputs.  ANNs are processing devices (algorithms or actual hardware) that are loosely modeled after the neuronal structure of the mamalian cerebral cortex but on much smaller scales. A large ANN might have hundreds or thousands of processor units, whereas a mamalian brain has billions of neurons with a corresponding increase in magnitude of their overall interaction and emergent behavior. Although ANN researchers are generally not concerned with whether their networks accurately resemble biological systems, some have. For example, researchers have accurately simulated the function of the retina and modeled the eye rather well. 
  • 3.
     In informationtechnology, a neural network is a system of programs and data structures that approximates the operation of the human brain. A neural network usually involves a large number of processors operating in parallel, each with its own small sphere of knowledge and access to data in its local memory. Typically, a neural network is initially "trained" or fed large amounts of data and rules about data relationships (for example, "A grandfather is older than a person's father"). A program can then tell the network how to behave in response to an external stimulus (for example, to input from a computer user who is interacting with the network) or can initiate activity on its own (within the limits of its access to the external world) (Eric Davalo and Patrick Naim Assimov, 1984).  Nonlinearity: A neuron is basically a nonlinear device. Consequently, a neural network, made up of an interconnection of neurons, is itself nonlinear. Moreover, the nonlinearity is of a special kind in the sense that it is distributed throughout the network.
  • 4.
     A popularparadigm of learning called supervised learning involves the modification of the synaptic weights of a neural network by applying a set of training samples. Each sample consists of a unique input signal and the corresponding desired response (Fausett L., 1994). The network is presented a sample picked at random from the set, and the synaptic weights (free parameters) of the network are modified so as to minimize the difference between the desired response and the actual response of the network produced by the input signal in accordance with an appropriate criterion. The training of the network is repeated for many samples in the set until the network reaches a steady state, where there are no further significant changes in the synaptic weights. The previously applied training samples may be reapplied during the training session, usually in a different order. Thus the network learns from the samples by constructing an input output mapping for the problem at hand.
  • 5.
     Neural networkshave a built-in capability to adapt their synaptic weights to changes in the surrounding environment. In particular, a neural network trained to operate in a specific environment can be easily retrained to deal with minor changes in the operating environmental conditions. Moreover, when it is operating in a nonstationary environment a neural network can be designed to change its synaptic weights i
  • 6.
     The massivelyparallel nature of a neural network makes it potentially fast for the computation of certain tasks. This same feature makes a neural network ideally suited for implementation using very- large-scale-integrated (VLS1) technology.