SlideShare a Scribd company logo
Mabrouka AL-Shebany et al. Int. Journal of Engineering Research and Applications
ISSN : 2248-9622, Vol. 4, Issue 2( Version 1), February 2014, pp.07-12

RESEARCH ARTICLE

www.ijera.com

OPEN ACCESS

Artificial Neural Network : A Brief Overview
Magdi Zakaria, Mabrouka AL-Shebany, Shahenda Sarhan
Sirte University SIRTE, LIBYA

Abstract
Artificial Neural Network (ANN) is an information processing paradigm that is inspired by the way biological
nervous systems, such as the brain, process information. The key element of this paradigm is the novel structure
of the information processing system. Neural networks, have remarkable ability to derive meaning from
complicated or imprecise data, can be used to extract patterns and detect trends that are too complex to be
noticed by either humans or other computer techniques. A trained neural network can be thought of as an
"expert" in the category of information it has been given to analyze. This expert can then be used to provide
projections given new situations of interest and answer "what if" questions. so in this paper we tried to introduce
a brief overview of ANN to help researchers in their way throw ANN.

I.

Introduction

An Artificial Neural networks are very
powerful brain-inspired computational models.
Which have been employed in various areas such as
computing, medicine, engineering, economics, and
many others.An artificial neural network is based on
the optimization theory. An Artificial Neural
Network is a computational model inspired in the
functioning of the human brain. It is composed by a

set of artificial neurons (known as processing units)
that are interconnected with other neuron these
neurons depend on weights of the neural network. As
The word network in Neural Network refers to the
interconnection between neurons present in various
layers of a system. These weights represent the
connections between the neurons which determine
the impact of one neuron on another[1].

Fig. 1. Natural neuron
The first artificial neuron was firstly
proposed in a formal model in 1943 by McCulloch
and Pitts. They proved that this model of neuron was
able to perform any computable function using a
finite number of artificial neurons and synaptic
weights adjustable.
The neurons in the input layer receive the
data and transfer them to neurons in the first hidden
layer through the weighted links. Data are
mathematically processed and transferred the result
to the neurons in the next layer. The network’s output
is provided by the neurons in the last layer. The j-th
www.ijera.com

neuron in a hidden layer processes the incoming data
(xi) by: (i) calculating the weighted sum and adding a
“bias” term (θj) according to:
𝑚
𝑛𝑒𝑡 𝑗 = 𝑖=1 𝑥 𝑖 ∗ 𝑤 𝑖𝑗 + 𝜃𝑗
𝑗 = (1, 2, 3, 𝑛)
[2]
neural networks are capable of adaptation to given
data; neural networks are capable of generalization
even when the input data set contains noise or
missing values (trained network has the capability of
correctly filling the value without affecting the
prediction); neural networks act as an universal
approximation for an arbitrary continuous function
with arbitrary precision.
7|P age
Mabrouka AL-Shebany et al. Int. Journal of Engineering Research and Applications
ISSN : 2248-9622, Vol. 4, Issue 2( Version 1), February 2014, pp.07-12
ANN can be defined based on the following three
characteristics:
1. The Architecture indicating the number of layers
and the no. of nodes in each of the layers.
2. The learning mechanism applied for updating the
weights of the connections.
3. The activation functions used in various layers
[3].

www.ijera.com

 Neural network layers:
1. Dimensional layer (Single Layer Perceptron) is a
function of N real variables of the form
𝑁

𝑓 𝑥1 , … . . , 𝑥 𝑁 = 𝑠𝑔𝑛(

𝑤 𝑖 𝑥 𝑖 − 𝜃)
𝑖=1

Here xi are real variables, (x1, . . . , xN) takes
values in some domain U ⊂RN, wiare real
parameters(weights of the neuron), _ is the threshold
of activation of the neuron, the function sgm(x) = 1
for x ≥ 0 and is equal to zero for x < 0.

Fig. 2. An artificial neuron
We also consider the smoothed variant of
the above neuron for which instead of the function
sgm we use the smooth monotonous increasing
function sgm which varies from zero to unity. In
particular we consider the neuron of the form
𝑁

𝑓 𝑥1 , … . . , 𝑥 𝑁 = 𝑠𝑔𝑛

𝑤 𝑖 𝑥 𝑖 − 𝜃 , 𝑠𝑔𝑚 𝑥
𝑖=1

=

1
1 + 𝑒 −𝑥

2. Multi Layer Perceptron (MLP) : Multi-layer
consisting of three consecutive layers: an input, a
hidden, and an output layer. Every system is basically
a three layered system, which are Input layer, Hidden
Layer and Output Layer. The input layer has input
neurons which transfer data via synapses to the
hidden layer, and similarly the hidden layer transfers
this data to the output layer via more synapses. The
synapses stores values called weights which helps
them to manipulate the input and output to various
layers.

Fig.3. General structure of a neural network with two hidden layer

II.

Learning in ANN

There are three major learning paradigms;
supervised learning, unsupervised learning and
reinforcement learning. Usually they can be
employed by any given type of artificial neural
network architecture. Each learning paradigm has
many training algorithms.
www.ijera.com



Supervised learning
Supervised learning is a machine learning
technique that sets parameters of an artificial neural
network from training data. The task of the learning
artificial neural network is to set the value of its
parameters for any valid input value after having seen
output value. The training data consist of pairs of
8|P age
Mabrouka AL-Shebany et al. Int. Journal of Engineering Research and Applications
ISSN : 2248-9622, Vol. 4, Issue 2( Version 1), February 2014, pp.07-12
input and desired output values that are traditionally
represented in data vectors. Supervised learning can
also be referred as classification, where we have a
wide range of classifiers, each with its strengths and
weaknesses.
Choosing a suitable classifier (Multilayer
perceptron, Support Vector Machines, k-nearest
neighbor algorithm, Gaussian mixture model,
Gaussian, naive Bayes, decision tree, radial basis
function classifiers,…) for a given problem is
however still more an art than a science. In order to
solve a given problem of supervised learning various
steps has to be considered. In the first step we have to
determine the type of training examples[6][14]. In the
second step we need to gather a training data set that
satisfactory describe a given problem. In the third
step we need to describe gathered training data set in
form understandable to a chosen artificial neural
network. In the fourth step we do the learning and
after the learning we can test theperformance of
learned artificial neural network with the test
(validation) data set. Test data set consist of data that
has not been introduced to artificial neural network
while learning.
2.2 Unsupervised learning
Unsupervised learning is a machine learning
technique that sets parameters of an artificial neural
network based on given data and a cost function
which is to be minimized. Cost function can be any
function and it is determined by the task formulation.
Unsupervised learning is mostly used in applications
that fall within the domain of estimation problems
such as statistical modeling, compression, filtering,
blind source separation and clustering[14].In
unsupervised learning we seek to determine how the
data is organized. It differs from supervised learning
and reinforcement learning in that the artificial neural
network is given only unlabeled examples.
One common form of unsupervised learning
is clustering where we try to categorize data in
different clusters by their similarity. Among above
described artificial neural network models, the Selforganizing maps are the ones that the most
commonly use unsupervised learning algorithms.
2.3 Reinforcement learning
Reinforcement learning is a machine
learning technique that sets parameters of an artificial
neural network, where data is usually not given, but
generated by interactions with the environment.
Reinforcement learning is concerned with how an
artificial neural network ought to take actions in an
environment so as to maximize some notion of longterm reward. Reinforcement learning is frequently
used as a part of artificial neural network’s overall
learning algorithm.
www.ijera.com

www.ijera.com

After return function that needs to be
maximized is defined, reinforcement learning uses
several algorithms to find the policy which produces
the maximum return. Naive brute force algorithm in
first step calculates return function for each possible
policy and chooses the policy with the largest return.
Obvious weakness of this algorithm is in case of
extremely large or even infinite number of possible
policies[1][14]. This weakness can be overcome by
value function approaches or direct policy
estimation. Value function approaches attempt to find
a policy that maximizes the return by maintaining a
set of estimates of expected returns for one policy;
usually either the current or the optimal estimates.
These methods converge to the correct estimates for a
fixed policy and can also be used to find the optimal
policy.
Similar as value function approaches the
direct policy estimation can also find the optimal
policy. It can find it by searching it directly in policy
space what greatly increases the computational cost.
Reinforcement learning is particularly suited to
problems which include a long-term versus shortterm reward trade-off. It has been applied
successfully to various problems, including robot
control, telecommunications, and games such as
chess and other sequential decision
making tasks.

III.

Neural Network Algorithms:



Radial basis functions network (RBF)
A radial
basis
function
network is
an artificial neural network that uses radial basis
functions as activation functions. The output of the
network is a linear combination of radial basis
functions of the inputs and neuron parameters.


Auto correlation function
Autocorrelation is the cross-correlation of
a signal with itself. It is the similarity between
observations as a function of the time interval
between them. It is a mathematical tool for finding
repeating patterns, such as the presence of a periodic
signal obscured by noise, or identifying the missing
fundamental frequency in a signal implied by
its harmonic frequencies. It is often used in signal
processing for analyzing functions or series of values,
such as time domain signals.


Self-organizing map (SOM) neural network
A self-organizing map (SOM) is a type
of artificial neural network (ANN) that is trained
using unsupervised learning to produce a lowdimensional (typically two-dimensional), discretized
representation of the input space of the training
samples, and called a map. Self-organizing maps are
different from other artificial neural networks in the
9|P age
Mabrouka AL-Shebany et al. Int. Journal of Engineering Research and Applications
ISSN : 2248-9622, Vol. 4, Issue 2( Version 1), February 2014, pp.07-12
sense that they use a neighborhood function to
preserve the topological properties of the input space.
The ideas of the SOFM combined with the elastic net
algorithm to solve Euclidean problems like the

www.ijera.com

travelling salesman problem. A modified SOFM has
been used to solve broad classes of optimization
problems by freeing the technique from the Euclidean
plane [5].

Fig. 4. Architecture of a SOFM with nine neurons


Back propagation neural
Back propagation ANNs represented by
nonlinear networks. The back propagation (BP)
algorithmused for training multilayer networks by
means of error propagation via variational
calculations. It minimizes the sum of squared
approximation errors using a gradient descent
technique.When noisy training data are present, the
learned function can oscillate abruptly between data
points. This is clearly undesirable for function
approximation from noisy data [6].

IV.

Neural Network Applications:

Hop-field networks
Hop-field networks used to solve
optimization problems which are familiar to
theoperations researcher. although there is no
layerstructure to the architecture, and the weights are
constants and symmetric. Hop-field networks are a
fullyinterconnected system of N neurons. The
weights of the network store information about the
memories or stablestates of the network. Each neuron
has a state xi which is bounded between 0 and 1.
Neurons areupdated according to a differential
equation, and an energy function is minimized over
time because of the stable states of the network.
Hop-field and Tank showed that the weights
of a Hop-field network can be chosen so the process
of neurons minimizes the Hop-field energyfunction
and the optimization problem.
Each neuron i updates itself according to the
differential equation
dnet i
net i
=−
+
dt
τ

Image Indexing And Retrieval. SooBeom Park,
Jae Won Lee, Sang Kyoon Kim proposed a
content-based image classification method that
reflect the shape of an object based on pattern of
the texure feature[8].



Pattern Recognition : hierarchical neural
networks
improves
supervised
pattern
classification as it used in online back
propagation, improved records on MNIST, Latin
letters, Chinese characters, and traffic signs. Dan
Ciresan ,Ueli Meier, J¨urgenSchmidhuber
proposed a system to get data sets ranging from
handwritten digits (MNIST), handwritten
characters to 3D toys (NORB) and faces[9].



Space Exploration Robotic Vehicle And
Exploring The Land Of A Planet. It has the
capability to travel across the surface of a
landscape and other cosmic bodies. Artificial
neural networks have many advantages in space
applications due to its:
Generality
Performance
Adaptability
Low energy consumption

N

wij xj + Ii
j=1

xi=f(neti) Where f(.) is a sigmoidal output function
bounded by 0 and 1 and q is a constant [5].

www.ijera.com

Classification : such as classification of brain
tumors based on many sources of information.
KailashD.Kharat&PradyumnaP.Kulkarni&M.B.
Nagori proposed method that can do
classification of brain tumors by the analysis of
Magnetic Resonance (MR) images and Magnetic
Resonance Spectroscopy (MRS) data of the
patients with benign and malignant tumors to
determine the type of tumors [7].












10 | P a g e
Mabrouka AL-Shebany et al. Int. Journal of Engineering Research and Applications
ISSN : 2248-9622, Vol. 4, Issue 2( Version 1), February 2014, pp.07-12








Robustness & Fault Tolerance: Youssef Bassil
proposes a path-planning solution for
autonomous robotic planetary rover systems
based on artificial neural network (ANN). The
proposed ANN uses a mix of activation
functions including Sigmoid for the hidden
neurons and linear for the output neurons [10].
Where Madhusmita Swain, Sanjit Kumar Dash,
Sweta Dash and AyeskantaMohapatra Proposed
The Multi-Layer Feed Forward Neural network
which is able to classify the three different types
of IRIS plants of 150 instances with just few
errors for the other one[3].
Wavelet Feature Extraction : proposed for image
classification problem. jay Kumar Singh, Shamik
Tiwari, V.P. Shukla proposed feature extraction
and classification of multiclass images by using
Haar wavelet transform and back propagation
neural network [11].
Multi-class object recognition is a critical
capability for an intelligence robot to perceive
its environment. YuhuaZheng and Yan Meng
Proposed a model combined a number of
modular neural networks to recognize multiple
classes of objects for a robotic system. The
population of the modular neural networks
depends on the class number of the objects to be
recognized and each modular network only
focuses on learning one object class. For each
modular neural network, both the bottom-up
(sensory-driven) and top-down (expectationdriven) pathways are attached together, and a
supervised learning algorithm is applied to
update corresponding weights of both pathways.
Also there are two different training strategies
are evaluated: positive-only training and
positive-and-negative training [12]
Pattern Classification WawanSetiawan and
Wiweka proposed a classifier model with neural
network approach based on the method used
Expectations Maximum (EM)[13].

V.

Conclusion

The computing world has a lot to gain from
neural networks. Their ability to learn by example
makes them very flexible and powerful. Furthermore
there is no need to devise an algorithm in order to
perform a specific task; i.e. there is no need to
understand the internal mechanisms of that task.
They are also very well suited for real time systems
because of their fast response and computational
times which are due to their parallel architecture.

References
[1]

[2]

[3]

[4]

[5]

[6]

[7]

[8]

[9]

[10]

[11]

www.ijera.com

www.ijera.com

1S.V.
Kozyrev,”
Classification
by
ensembles of neural networks”, Steklov
Mathematical Institute, February 21, 2012.
Filippo Amato, Alberto López, Eladia María
Peña-Méndez, Petr Vaňhara, Aleš Hampl,
Josef Havel, ”Artificial neural networks in
medical diagnosis”, Journal of APPLIED
BIOMEDICINE, J Appl Biomed. 11: 47–58,
2013
Madhusmita Swain, Sanjit Kumar Dash,
Sweta Dash and Ayeskanta Mohapatra, “AN
APPROACH
FOR
IRIS
PLANT
CLASSIFICATION USING NEURAL
NETWORK”, International Journal on Soft
Computing ( IJSC ) Vol.3, No.1, February
2012
Jose´ L. Herrera-Aguilar, Herna´n Larralde,
Maximino Aldana,” Pattern Recognition in
Neural
Networks
with
Competing
Dynamics: Coexistence of Fixed-Point and
Cyclic Attractors”, PLOS ONE, 1 August
2012, Volume 7, Issue 8, e42348
Kate A. Smith, Jatinder N.D. Gupta,”Neural
networks in business: techniques and
applications for the operations researcher",
Computers & Operations Research 27
(2000) 1023}1044
P.SIBI,
S.ALLWYN
JONES,
P.SIDDARTH,”
ANALYSIS
OF
DIFFERENT ACTIVATION FUNCTIONS
USING BACK PROPAGATION NEURAL
NETWORKS”, Journal of Theoretical and
Applied Information Technology 31st
January 2013. Vol. 47 No.3
Kailash D. Kharat & Pradyumna P. Kulkarni
& M. B. Nagori ,”Brain Tumor
Classification Using Neural Network Based
Methods”,
International
Journal
of
Computer Science and Informatics ISSN
(PRINT): 2231 –5292, Vol-1, Iss-4, 2012.
SooBeom Park, Jae Won Lee, Sang Kyoon
Kim, “Content-based image classification
using
a
neural
network”,
Pattern
Recognition Letters 25 (2004) 287–300
Dan Ciresan, Ueli Meier, J¨urgen
Schmidhuber, “Multi-column Deep Neural
Networks for Image Classification” ,
Technical
Report
No.
IDSIA-04-12
February 2012.
Youssef Bassil, “ Neural Network Model for
Path-Planning Of Robotic Rover Systems”,
International Journal of Science and
Technology (IJST), E-ISSN: 2224-3577,
Vol. 2, No. 2, February, 2012
jay Kumar Singh,Shamik Tiwari, V.P.
Shukla, “ Wavelet based Multi Class image
11 | P a g e
Mabrouka AL-Shebany et al. Int. Journal of Engineering Research and Applications
ISSN : 2248-9622, Vol. 4, Issue 2( Version 1), February 2014, pp.07-12

[12]

[13]

[14]

www.ijera.com

classification using Neural Network” ,
International
Journal
of
Computer
Applications (0975 – 8887) Volume 37–
No.4, January 2012
YuhuaZheng and Yan Meng, ” Modular
Neural Networks for Multi-Class Object
Recognition”, 2011 IEEE International
Conference on Robotics and Automation
Shanghai International Conference Center
May 9-13, 2011, Shanghai, China
WawanSetiawan and Wiweka, ”Combining
of Image Classification With Probabilistic
Neural Network (PNN) Approaches Based
on Expectation Maximum (EM)”, IJCSI
International Journal of Computer Science
Issues, Vol. 9, Issue 4, No 3, July 2012
WWW. Wikipedia.or

www.ijera.com

12 | P a g e

More Related Content

PDF
APPLYING NEURAL NETWORKS FOR SUPERVISED LEARNING OF MEDICAL DATA
PDF
C42021115
PDF
Neural Network Classification and its Applications in Insurance Industry
PDF
Image Recognition With the Help of Auto-Associative Neural Network
PPTX
Unsupervised learning
PPTX
Introduction Of Artificial neural network
PDF
Nature Inspired Reasoning Applied in Semantic Web
PDF
X-TREPAN : A Multi Class Regression and Adapted Extraction of Comprehensible ...
APPLYING NEURAL NETWORKS FOR SUPERVISED LEARNING OF MEDICAL DATA
C42021115
Neural Network Classification and its Applications in Insurance Industry
Image Recognition With the Help of Auto-Associative Neural Network
Unsupervised learning
Introduction Of Artificial neural network
Nature Inspired Reasoning Applied in Semantic Web
X-TREPAN : A Multi Class Regression and Adapted Extraction of Comprehensible ...

What's hot (19)

PPTX
Neural Network Classification and its Applications in Insurance Industry
PPT
Supervised Learning
DOC
Question bank soft computing
PDF
Cq4201618622
PDF
ANALYSIS AND COMPARISON STUDY OF DATA MINING ALGORITHMS USING RAPIDMINER
PPTX
Neural network
PPTX
Artificial Neural Network
PDF
Artificial Neural Networks: Applications In Management
PDF
Machine learning and_neural_network_lecture_slide_ece_dku
PDF
X trepan an extended trepan for
PDF
Artificial Neural Network report
PPTX
02 Fundamental Concepts of ANN
PPT
Neural networks
PPTX
DRL Medical Imaging Literature Review
PDF
DESIGN AND IMPLEMENTATION OF BINARY NEURAL NETWORK LEARNING WITH FUZZY CLUSTE...
PPTX
Neural networks
PDF
Handwritten digits recognition report
PDF
Artificial neural networks and its application
PDF
Feed forward neural network for sine
Neural Network Classification and its Applications in Insurance Industry
Supervised Learning
Question bank soft computing
Cq4201618622
ANALYSIS AND COMPARISON STUDY OF DATA MINING ALGORITHMS USING RAPIDMINER
Neural network
Artificial Neural Network
Artificial Neural Networks: Applications In Management
Machine learning and_neural_network_lecture_slide_ece_dku
X trepan an extended trepan for
Artificial Neural Network report
02 Fundamental Concepts of ANN
Neural networks
DRL Medical Imaging Literature Review
DESIGN AND IMPLEMENTATION OF BINARY NEURAL NETWORK LEARNING WITH FUZZY CLUSTE...
Neural networks
Handwritten digits recognition report
Artificial neural networks and its application
Feed forward neural network for sine
Ad

Viewers also liked (20)

PDF
R4201124128
PDF
M42017584
PDF
U4201139143
PDF
Az4103318326
PDF
Ah4103196202
PDF
Ax4103307314
PDF
Hm3613211331
PPS
Busca dentro de ti.
PDF
Carolien Sertifikate
PDF
Standards Spotlight
DOCX
Aprendizage movel3
PPT
As figuras xeométricas
PDF
O Indizível Futuro do Brasil. Ou: Vamos construir um futuro para a nação?
PPT
Family Life
PPT
A Casa Dos Bicos
PPTX
Unidad ii Tecnología de la Informacion
DOCX
Universidad nacional de chimborazo drive trabajo colavorativo
PPSX
Para quintela
PPTX
Smart art iguales
PDF
1017_001
R4201124128
M42017584
U4201139143
Az4103318326
Ah4103196202
Ax4103307314
Hm3613211331
Busca dentro de ti.
Carolien Sertifikate
Standards Spotlight
Aprendizage movel3
As figuras xeométricas
O Indizível Futuro do Brasil. Ou: Vamos construir um futuro para a nação?
Family Life
A Casa Dos Bicos
Unidad ii Tecnología de la Informacion
Universidad nacional de chimborazo drive trabajo colavorativo
Para quintela
Smart art iguales
1017_001
Ad

Similar to B42010712 (20)

PDF
A Study of Social Media Data and Data Mining Techniques
PPT
Artificial-Neural-Networks.ppt
PPT
Neural-Networks.ppt
PDF
Neural networks are parallel computing devices.docx.pdf
DOCX
Neural networks of artificial intelligence
DOCX
ABSTRACT.docxiyhkkkkkkkkkkkkkkkkkkkkkkkkkkkk
PDF
ANN stands for Artificial Neural Network
PPT
artificial-neural-networks-rev.ppt
PPT
artificial-neural-networks-rev.ppt
PDF
ppt_artificial neural netwroK_module 3_KTU
PPTX
Acem neuralnetworks
PPT
Artificial Neural Network Learning Algorithm.ppt
PPTX
Artificial Neural Network - Basic Concepts.pptx
PPT
NeuralNetworksbasics for Deeplearning
PPT
Neural-Networks2Neural Network Classifier.ppt
PPT
Neural-Networks full covering AI networks.ppt
PPT
Neural-Networks, images and examples details
PDF
Deep Learning detailkesdSECA4002 doc.pdf
PPTX
Artificial neural network
PPT
Artificial Neural Network and Machine Learning
A Study of Social Media Data and Data Mining Techniques
Artificial-Neural-Networks.ppt
Neural-Networks.ppt
Neural networks are parallel computing devices.docx.pdf
Neural networks of artificial intelligence
ABSTRACT.docxiyhkkkkkkkkkkkkkkkkkkkkkkkkkkkk
ANN stands for Artificial Neural Network
artificial-neural-networks-rev.ppt
artificial-neural-networks-rev.ppt
ppt_artificial neural netwroK_module 3_KTU
Acem neuralnetworks
Artificial Neural Network Learning Algorithm.ppt
Artificial Neural Network - Basic Concepts.pptx
NeuralNetworksbasics for Deeplearning
Neural-Networks2Neural Network Classifier.ppt
Neural-Networks full covering AI networks.ppt
Neural-Networks, images and examples details
Deep Learning detailkesdSECA4002 doc.pdf
Artificial neural network
Artificial Neural Network and Machine Learning

Recently uploaded (20)

PPTX
Cloud computing and distributed systems.
PDF
Approach and Philosophy of On baking technology
PDF
Blue Purple Modern Animated Computer Science Presentation.pdf.pdf
PDF
cuic standard and advanced reporting.pdf
PDF
Shreyas Phanse Resume: Experienced Backend Engineer | Java • Spring Boot • Ka...
PDF
Mobile App Security Testing_ A Comprehensive Guide.pdf
PDF
CIFDAQ's Market Insight: SEC Turns Pro Crypto
PDF
Build a system with the filesystem maintained by OSTree @ COSCUP 2025
PDF
Reach Out and Touch Someone: Haptics and Empathic Computing
PDF
Machine learning based COVID-19 study performance prediction
PDF
NewMind AI Weekly Chronicles - August'25 Week I
PPTX
MYSQL Presentation for SQL database connectivity
PDF
KodekX | Application Modernization Development
PPTX
Understanding_Digital_Forensics_Presentation.pptx
PPTX
20250228 LYD VKU AI Blended-Learning.pptx
PDF
TokAI - TikTok AI Agent : The First AI Application That Analyzes 10,000+ Vira...
PDF
Encapsulation theory and applications.pdf
PDF
Bridging biosciences and deep learning for revolutionary discoveries: a compr...
PDF
Electronic commerce courselecture one. Pdf
PDF
How UI/UX Design Impacts User Retention in Mobile Apps.pdf
Cloud computing and distributed systems.
Approach and Philosophy of On baking technology
Blue Purple Modern Animated Computer Science Presentation.pdf.pdf
cuic standard and advanced reporting.pdf
Shreyas Phanse Resume: Experienced Backend Engineer | Java • Spring Boot • Ka...
Mobile App Security Testing_ A Comprehensive Guide.pdf
CIFDAQ's Market Insight: SEC Turns Pro Crypto
Build a system with the filesystem maintained by OSTree @ COSCUP 2025
Reach Out and Touch Someone: Haptics and Empathic Computing
Machine learning based COVID-19 study performance prediction
NewMind AI Weekly Chronicles - August'25 Week I
MYSQL Presentation for SQL database connectivity
KodekX | Application Modernization Development
Understanding_Digital_Forensics_Presentation.pptx
20250228 LYD VKU AI Blended-Learning.pptx
TokAI - TikTok AI Agent : The First AI Application That Analyzes 10,000+ Vira...
Encapsulation theory and applications.pdf
Bridging biosciences and deep learning for revolutionary discoveries: a compr...
Electronic commerce courselecture one. Pdf
How UI/UX Design Impacts User Retention in Mobile Apps.pdf

B42010712

  • 1. Mabrouka AL-Shebany et al. Int. Journal of Engineering Research and Applications ISSN : 2248-9622, Vol. 4, Issue 2( Version 1), February 2014, pp.07-12 RESEARCH ARTICLE www.ijera.com OPEN ACCESS Artificial Neural Network : A Brief Overview Magdi Zakaria, Mabrouka AL-Shebany, Shahenda Sarhan Sirte University SIRTE, LIBYA Abstract Artificial Neural Network (ANN) is an information processing paradigm that is inspired by the way biological nervous systems, such as the brain, process information. The key element of this paradigm is the novel structure of the information processing system. Neural networks, have remarkable ability to derive meaning from complicated or imprecise data, can be used to extract patterns and detect trends that are too complex to be noticed by either humans or other computer techniques. A trained neural network can be thought of as an "expert" in the category of information it has been given to analyze. This expert can then be used to provide projections given new situations of interest and answer "what if" questions. so in this paper we tried to introduce a brief overview of ANN to help researchers in their way throw ANN. I. Introduction An Artificial Neural networks are very powerful brain-inspired computational models. Which have been employed in various areas such as computing, medicine, engineering, economics, and many others.An artificial neural network is based on the optimization theory. An Artificial Neural Network is a computational model inspired in the functioning of the human brain. It is composed by a set of artificial neurons (known as processing units) that are interconnected with other neuron these neurons depend on weights of the neural network. As The word network in Neural Network refers to the interconnection between neurons present in various layers of a system. These weights represent the connections between the neurons which determine the impact of one neuron on another[1]. Fig. 1. Natural neuron The first artificial neuron was firstly proposed in a formal model in 1943 by McCulloch and Pitts. They proved that this model of neuron was able to perform any computable function using a finite number of artificial neurons and synaptic weights adjustable. The neurons in the input layer receive the data and transfer them to neurons in the first hidden layer through the weighted links. Data are mathematically processed and transferred the result to the neurons in the next layer. The network’s output is provided by the neurons in the last layer. The j-th www.ijera.com neuron in a hidden layer processes the incoming data (xi) by: (i) calculating the weighted sum and adding a “bias” term (θj) according to: 𝑚 𝑛𝑒𝑡 𝑗 = 𝑖=1 𝑥 𝑖 ∗ 𝑤 𝑖𝑗 + 𝜃𝑗 𝑗 = (1, 2, 3, 𝑛) [2] neural networks are capable of adaptation to given data; neural networks are capable of generalization even when the input data set contains noise or missing values (trained network has the capability of correctly filling the value without affecting the prediction); neural networks act as an universal approximation for an arbitrary continuous function with arbitrary precision. 7|P age
  • 2. Mabrouka AL-Shebany et al. Int. Journal of Engineering Research and Applications ISSN : 2248-9622, Vol. 4, Issue 2( Version 1), February 2014, pp.07-12 ANN can be defined based on the following three characteristics: 1. The Architecture indicating the number of layers and the no. of nodes in each of the layers. 2. The learning mechanism applied for updating the weights of the connections. 3. The activation functions used in various layers [3]. www.ijera.com  Neural network layers: 1. Dimensional layer (Single Layer Perceptron) is a function of N real variables of the form 𝑁 𝑓 𝑥1 , … . . , 𝑥 𝑁 = 𝑠𝑔𝑛( 𝑤 𝑖 𝑥 𝑖 − 𝜃) 𝑖=1 Here xi are real variables, (x1, . . . , xN) takes values in some domain U ⊂RN, wiare real parameters(weights of the neuron), _ is the threshold of activation of the neuron, the function sgm(x) = 1 for x ≥ 0 and is equal to zero for x < 0. Fig. 2. An artificial neuron We also consider the smoothed variant of the above neuron for which instead of the function sgm we use the smooth monotonous increasing function sgm which varies from zero to unity. In particular we consider the neuron of the form 𝑁 𝑓 𝑥1 , … . . , 𝑥 𝑁 = 𝑠𝑔𝑛 𝑤 𝑖 𝑥 𝑖 − 𝜃 , 𝑠𝑔𝑚 𝑥 𝑖=1 = 1 1 + 𝑒 −𝑥 2. Multi Layer Perceptron (MLP) : Multi-layer consisting of three consecutive layers: an input, a hidden, and an output layer. Every system is basically a three layered system, which are Input layer, Hidden Layer and Output Layer. The input layer has input neurons which transfer data via synapses to the hidden layer, and similarly the hidden layer transfers this data to the output layer via more synapses. The synapses stores values called weights which helps them to manipulate the input and output to various layers. Fig.3. General structure of a neural network with two hidden layer II. Learning in ANN There are three major learning paradigms; supervised learning, unsupervised learning and reinforcement learning. Usually they can be employed by any given type of artificial neural network architecture. Each learning paradigm has many training algorithms. www.ijera.com  Supervised learning Supervised learning is a machine learning technique that sets parameters of an artificial neural network from training data. The task of the learning artificial neural network is to set the value of its parameters for any valid input value after having seen output value. The training data consist of pairs of 8|P age
  • 3. Mabrouka AL-Shebany et al. Int. Journal of Engineering Research and Applications ISSN : 2248-9622, Vol. 4, Issue 2( Version 1), February 2014, pp.07-12 input and desired output values that are traditionally represented in data vectors. Supervised learning can also be referred as classification, where we have a wide range of classifiers, each with its strengths and weaknesses. Choosing a suitable classifier (Multilayer perceptron, Support Vector Machines, k-nearest neighbor algorithm, Gaussian mixture model, Gaussian, naive Bayes, decision tree, radial basis function classifiers,…) for a given problem is however still more an art than a science. In order to solve a given problem of supervised learning various steps has to be considered. In the first step we have to determine the type of training examples[6][14]. In the second step we need to gather a training data set that satisfactory describe a given problem. In the third step we need to describe gathered training data set in form understandable to a chosen artificial neural network. In the fourth step we do the learning and after the learning we can test theperformance of learned artificial neural network with the test (validation) data set. Test data set consist of data that has not been introduced to artificial neural network while learning. 2.2 Unsupervised learning Unsupervised learning is a machine learning technique that sets parameters of an artificial neural network based on given data and a cost function which is to be minimized. Cost function can be any function and it is determined by the task formulation. Unsupervised learning is mostly used in applications that fall within the domain of estimation problems such as statistical modeling, compression, filtering, blind source separation and clustering[14].In unsupervised learning we seek to determine how the data is organized. It differs from supervised learning and reinforcement learning in that the artificial neural network is given only unlabeled examples. One common form of unsupervised learning is clustering where we try to categorize data in different clusters by their similarity. Among above described artificial neural network models, the Selforganizing maps are the ones that the most commonly use unsupervised learning algorithms. 2.3 Reinforcement learning Reinforcement learning is a machine learning technique that sets parameters of an artificial neural network, where data is usually not given, but generated by interactions with the environment. Reinforcement learning is concerned with how an artificial neural network ought to take actions in an environment so as to maximize some notion of longterm reward. Reinforcement learning is frequently used as a part of artificial neural network’s overall learning algorithm. www.ijera.com www.ijera.com After return function that needs to be maximized is defined, reinforcement learning uses several algorithms to find the policy which produces the maximum return. Naive brute force algorithm in first step calculates return function for each possible policy and chooses the policy with the largest return. Obvious weakness of this algorithm is in case of extremely large or even infinite number of possible policies[1][14]. This weakness can be overcome by value function approaches or direct policy estimation. Value function approaches attempt to find a policy that maximizes the return by maintaining a set of estimates of expected returns for one policy; usually either the current or the optimal estimates. These methods converge to the correct estimates for a fixed policy and can also be used to find the optimal policy. Similar as value function approaches the direct policy estimation can also find the optimal policy. It can find it by searching it directly in policy space what greatly increases the computational cost. Reinforcement learning is particularly suited to problems which include a long-term versus shortterm reward trade-off. It has been applied successfully to various problems, including robot control, telecommunications, and games such as chess and other sequential decision making tasks. III. Neural Network Algorithms:  Radial basis functions network (RBF) A radial basis function network is an artificial neural network that uses radial basis functions as activation functions. The output of the network is a linear combination of radial basis functions of the inputs and neuron parameters.  Auto correlation function Autocorrelation is the cross-correlation of a signal with itself. It is the similarity between observations as a function of the time interval between them. It is a mathematical tool for finding repeating patterns, such as the presence of a periodic signal obscured by noise, or identifying the missing fundamental frequency in a signal implied by its harmonic frequencies. It is often used in signal processing for analyzing functions or series of values, such as time domain signals.  Self-organizing map (SOM) neural network A self-organizing map (SOM) is a type of artificial neural network (ANN) that is trained using unsupervised learning to produce a lowdimensional (typically two-dimensional), discretized representation of the input space of the training samples, and called a map. Self-organizing maps are different from other artificial neural networks in the 9|P age
  • 4. Mabrouka AL-Shebany et al. Int. Journal of Engineering Research and Applications ISSN : 2248-9622, Vol. 4, Issue 2( Version 1), February 2014, pp.07-12 sense that they use a neighborhood function to preserve the topological properties of the input space. The ideas of the SOFM combined with the elastic net algorithm to solve Euclidean problems like the www.ijera.com travelling salesman problem. A modified SOFM has been used to solve broad classes of optimization problems by freeing the technique from the Euclidean plane [5]. Fig. 4. Architecture of a SOFM with nine neurons  Back propagation neural Back propagation ANNs represented by nonlinear networks. The back propagation (BP) algorithmused for training multilayer networks by means of error propagation via variational calculations. It minimizes the sum of squared approximation errors using a gradient descent technique.When noisy training data are present, the learned function can oscillate abruptly between data points. This is clearly undesirable for function approximation from noisy data [6]. IV. Neural Network Applications: Hop-field networks Hop-field networks used to solve optimization problems which are familiar to theoperations researcher. although there is no layerstructure to the architecture, and the weights are constants and symmetric. Hop-field networks are a fullyinterconnected system of N neurons. The weights of the network store information about the memories or stablestates of the network. Each neuron has a state xi which is bounded between 0 and 1. Neurons areupdated according to a differential equation, and an energy function is minimized over time because of the stable states of the network. Hop-field and Tank showed that the weights of a Hop-field network can be chosen so the process of neurons minimizes the Hop-field energyfunction and the optimization problem. Each neuron i updates itself according to the differential equation dnet i net i =− + dt τ Image Indexing And Retrieval. SooBeom Park, Jae Won Lee, Sang Kyoon Kim proposed a content-based image classification method that reflect the shape of an object based on pattern of the texure feature[8].  Pattern Recognition : hierarchical neural networks improves supervised pattern classification as it used in online back propagation, improved records on MNIST, Latin letters, Chinese characters, and traffic signs. Dan Ciresan ,Ueli Meier, J¨urgenSchmidhuber proposed a system to get data sets ranging from handwritten digits (MNIST), handwritten characters to 3D toys (NORB) and faces[9].  Space Exploration Robotic Vehicle And Exploring The Land Of A Planet. It has the capability to travel across the surface of a landscape and other cosmic bodies. Artificial neural networks have many advantages in space applications due to its: Generality Performance Adaptability Low energy consumption N wij xj + Ii j=1 xi=f(neti) Where f(.) is a sigmoidal output function bounded by 0 and 1 and q is a constant [5]. www.ijera.com Classification : such as classification of brain tumors based on many sources of information. KailashD.Kharat&PradyumnaP.Kulkarni&M.B. Nagori proposed method that can do classification of brain tumors by the analysis of Magnetic Resonance (MR) images and Magnetic Resonance Spectroscopy (MRS) data of the patients with benign and malignant tumors to determine the type of tumors [7].        10 | P a g e
  • 5. Mabrouka AL-Shebany et al. Int. Journal of Engineering Research and Applications ISSN : 2248-9622, Vol. 4, Issue 2( Version 1), February 2014, pp.07-12     Robustness & Fault Tolerance: Youssef Bassil proposes a path-planning solution for autonomous robotic planetary rover systems based on artificial neural network (ANN). The proposed ANN uses a mix of activation functions including Sigmoid for the hidden neurons and linear for the output neurons [10]. Where Madhusmita Swain, Sanjit Kumar Dash, Sweta Dash and AyeskantaMohapatra Proposed The Multi-Layer Feed Forward Neural network which is able to classify the three different types of IRIS plants of 150 instances with just few errors for the other one[3]. Wavelet Feature Extraction : proposed for image classification problem. jay Kumar Singh, Shamik Tiwari, V.P. Shukla proposed feature extraction and classification of multiclass images by using Haar wavelet transform and back propagation neural network [11]. Multi-class object recognition is a critical capability for an intelligence robot to perceive its environment. YuhuaZheng and Yan Meng Proposed a model combined a number of modular neural networks to recognize multiple classes of objects for a robotic system. The population of the modular neural networks depends on the class number of the objects to be recognized and each modular network only focuses on learning one object class. For each modular neural network, both the bottom-up (sensory-driven) and top-down (expectationdriven) pathways are attached together, and a supervised learning algorithm is applied to update corresponding weights of both pathways. Also there are two different training strategies are evaluated: positive-only training and positive-and-negative training [12] Pattern Classification WawanSetiawan and Wiweka proposed a classifier model with neural network approach based on the method used Expectations Maximum (EM)[13]. V. Conclusion The computing world has a lot to gain from neural networks. Their ability to learn by example makes them very flexible and powerful. Furthermore there is no need to devise an algorithm in order to perform a specific task; i.e. there is no need to understand the internal mechanisms of that task. They are also very well suited for real time systems because of their fast response and computational times which are due to their parallel architecture. References [1] [2] [3] [4] [5] [6] [7] [8] [9] [10] [11] www.ijera.com www.ijera.com 1S.V. Kozyrev,” Classification by ensembles of neural networks”, Steklov Mathematical Institute, February 21, 2012. Filippo Amato, Alberto López, Eladia María Peña-Méndez, Petr Vaňhara, Aleš Hampl, Josef Havel, ”Artificial neural networks in medical diagnosis”, Journal of APPLIED BIOMEDICINE, J Appl Biomed. 11: 47–58, 2013 Madhusmita Swain, Sanjit Kumar Dash, Sweta Dash and Ayeskanta Mohapatra, “AN APPROACH FOR IRIS PLANT CLASSIFICATION USING NEURAL NETWORK”, International Journal on Soft Computing ( IJSC ) Vol.3, No.1, February 2012 Jose´ L. Herrera-Aguilar, Herna´n Larralde, Maximino Aldana,” Pattern Recognition in Neural Networks with Competing Dynamics: Coexistence of Fixed-Point and Cyclic Attractors”, PLOS ONE, 1 August 2012, Volume 7, Issue 8, e42348 Kate A. Smith, Jatinder N.D. Gupta,”Neural networks in business: techniques and applications for the operations researcher", Computers & Operations Research 27 (2000) 1023}1044 P.SIBI, S.ALLWYN JONES, P.SIDDARTH,” ANALYSIS OF DIFFERENT ACTIVATION FUNCTIONS USING BACK PROPAGATION NEURAL NETWORKS”, Journal of Theoretical and Applied Information Technology 31st January 2013. Vol. 47 No.3 Kailash D. Kharat & Pradyumna P. Kulkarni & M. B. Nagori ,”Brain Tumor Classification Using Neural Network Based Methods”, International Journal of Computer Science and Informatics ISSN (PRINT): 2231 –5292, Vol-1, Iss-4, 2012. SooBeom Park, Jae Won Lee, Sang Kyoon Kim, “Content-based image classification using a neural network”, Pattern Recognition Letters 25 (2004) 287–300 Dan Ciresan, Ueli Meier, J¨urgen Schmidhuber, “Multi-column Deep Neural Networks for Image Classification” , Technical Report No. IDSIA-04-12 February 2012. Youssef Bassil, “ Neural Network Model for Path-Planning Of Robotic Rover Systems”, International Journal of Science and Technology (IJST), E-ISSN: 2224-3577, Vol. 2, No. 2, February, 2012 jay Kumar Singh,Shamik Tiwari, V.P. Shukla, “ Wavelet based Multi Class image 11 | P a g e
  • 6. Mabrouka AL-Shebany et al. Int. Journal of Engineering Research and Applications ISSN : 2248-9622, Vol. 4, Issue 2( Version 1), February 2014, pp.07-12 [12] [13] [14] www.ijera.com classification using Neural Network” , International Journal of Computer Applications (0975 – 8887) Volume 37– No.4, January 2012 YuhuaZheng and Yan Meng, ” Modular Neural Networks for Multi-Class Object Recognition”, 2011 IEEE International Conference on Robotics and Automation Shanghai International Conference Center May 9-13, 2011, Shanghai, China WawanSetiawan and Wiweka, ”Combining of Image Classification With Probabilistic Neural Network (PNN) Approaches Based on Expectation Maximum (EM)”, IJCSI International Journal of Computer Science Issues, Vol. 9, Issue 4, No 3, July 2012 WWW. Wikipedia.or www.ijera.com 12 | P a g e