SlideShare a Scribd company logo
Network Design & Training
Network Design &  Training Issues   Design: Architecture of network Structure of artificial neurons Learning rules  Training: Ensuring optimum training Learning parameters Data preparation and more ....
Network Design
Network Design Architecture of the network:  How many nodes? Determines number of network weights How many layers?  How many nodes per layer? Input Layer  Hidden Layer  Output Layer Automated methods:  augmentation (cascade correlation) weight pruning and elimination
Network Design Architecture of the network:  Connectivity? Concept of model or  hypothesis  space Constraining the number of hypotheses: selective connectivity shared weights recursive connections
Network Design Structure of artificial neuron nodes Choice of input integration: summed, squared and summed multiplied Choice of activation (transfer) function: sigmoid (logistic) hyperbolic tangent Gaussian linear soft-max
Network Design Selecting a Learning Rule  Generalized delta rule  (steepest descent) Momentum descent Advanced weight space search techniques Global Error function can also vary - normal  - quadratic  - cubic
Network Training
Network Training How do you ensure that a network has been well trained? Objective:  To achieve good generalization   accuracy on new examples/cases  Establish a maximum acceptable error rate  Train the network using a  validation test set  to tune it Validate the trained network against a separate test set which is usually referred to as a  production test set
Network Training Approach #1:  Large Sample When the amount of available data is large ... Available  Examples Training Set Production Set 70% 30% Used to develop one ANN model Compute Test error Divide randomly Generalization error = test error Test Set
Network Training Approach #2:  Cross-validation When the amount of available data is small ... Available Examples Training Set Pro. Set 10% 90% Repeat 10 times Used to develop 10 different ANN models Accumulate test errors Generalization error determined by mean test error and stddev Test Set
Network Training How do you select between two ANN designs ?  A statistical test of hypothesis is required to ensure that a significant difference exists between the error rates of two ANN models If  Large Sample  method has been used then apply  McNemar’s test* If  Cross-validation  then use a  paired  t  test  for difference of two proportions *We assume a classification problem, if this is function  approximation then use paired  t  test for difference of  means
Network Training Mastering ANN Parameters Typical   Range learning rate -  0.1  0.01 - 0.99 momentum -  0.8  0.1 - 0.9 weight-cost -  0.1  0.001 - 0.5 Fine tuning :  -   adjust individual parameters at each node and/or connection weight automatic adjustment during training
Network Training Network weight initialization Random initial values  +/- some range Smaller weight values for nodes with many incoming connections Rule of thumb:  initial weight range should be approximately coming into a node
Network Training Typical Problems During Training E # iter E # iter E # iter Would like: But sometimes: Steady, rapid decline in total error Seldom a local minimum -  reduce learning or momentum parameter Reduce learning parms. - may indicate data is not learnable
Data Preparation
Data Preparation Garbage in  Garbage out  The quality of results relates directly to quality of the data 50%-70% of ANN development time will be spent on data preparation The three steps of data preparation: Consolidation and Cleaning Selection and Preprocessing Transformation and Encoding
Data Preparation Data Types and ANNs Three basic data types: nominal  discrete symbolic ( A, yes, small ) ordinal  discrete numeric (-5, 3, 24) continuous  numeric (0.23, -45.2, 500.43)  bp ANNs accept only continuous numeric values  (typically 0 - 1 range)
Data Preparation Consolidation and Cleaning Determine appropriate input attributes  Consolidate data into working database Eliminate or estimate missing values Remove  outliers  (obvious exceptions) Determine prior probabilities of categories and deal with  volume bias
Data Preparation Selection and Preprocessing Select examples  random sampling Consider number of training examples?  Reduce attribute dimensionality remove redundant and/or correlating attributes combine attributes (sum, multiply, difference) Reduce attribute value ranges group symbolic discrete values quantize continuous numeric values
Data Preparation Transformation and Encoding Discrete symbolic or numeric values Transform to discrete numeric values Encode the value  4  as follows: one-of-N code ( 0 1 0 0 0 ) - five inputs thermometer code (  1 1 1 1 0 ) - five inputs real value ( 0.4 )*  - one input Consider relationship between values ( single, married, divorce )  vs.  ( youth, adult, senior ) * Target values should be 0.1 - 0.9 , not 0.0 - 1.0 range
Data Preparation Transformation and Encoding Continuous numeric values De-correlate example attributes via normalization of values: Euclidean:  n = x/sqrt(sum of all x^2) Percentage:  n =  x/(sum of all x) Variance based:  n = (x - (mean of all x))/variance Scale values  using a linear transform if data is uniformly distributed or use non-linear (log, power) if skewed distribution
Data Preparation Transformation and Encoding Continuous numeric values Encode the value  1.6   as: Single real-valued number ( 0.16 )* -  OK! Bits of a binary number ( 010000 ) -  BAD! one-of-N quantized intervals ( 0 1 0 0 0 )  -  NOT GREAT! - discontinuities distributed (fuzzy) overlapping intervals  (  0.3 0.8 0.1 0.0 0.0 ) -  BEST! * Target values should be 0.1 - 0.9 , not 0.0 - 1.0 range
TUTORIAL #5 Develop and train a BP network on real-world data
Post-Training Analysis
Post-Training Analysis Examining the neural net model: Visualizing the constructed model Detailed network analysis Sensitivity analysis of input attributes: Analytical techniques  Attribute elimination
Post-Training Analysis Visualizing the Constructed Model Graphical tools can be used to display output response as selected input variables are changed Response Size Temp
Post-Training Analysis Detailed network analysis Hidden nodes form internal representation Manual analysis of weight values often difficult - graphics very helpful Conversion to equation, executable code Automated ANN to symbolic logic conversion is a hot area of research
Post-Training Analysis Sensitivity analysis of input attributes Analytical techniques  factor analysis network weight analysis Feature (attribute) elimination forward feature elimination backward feature elimination
The ANN Application Development Process Guidelines for using neural networks 1.  Try the best existing method first 2.  Get a  big   training set 3.  Try a net without hidden units 4.  Use a sensible coding for input variables 5.  Consider methods of constraining network 6.  Use a test set to prevent over-training 7.  Determine confidence in generalization through cross-validation
Example Applications Pattern Recognition  (reading zip codes) Signal Filtering  (reduction of radio noise) Data Segmentation  (detection of seismic onsets) Data Compression  (TV image transmission) Database Mining  (marketing, finance analysis) Adaptive Control  (vehicle guidance)
Pros and Cons of Back-Prop
Pros and Cons  of Back-Prop   Cons: Local minimum - but not generally a concern Seems biologically implausible Space and time complexity:   lengthy training times It’s a black box!  I can’t see how it’s making decisions? Best suited for supervised learning Works poorly on dense data with few input variables
Pros and Cons  of Back-Prop   Pros: Proven training method for multi-layer nets Able to learn any arbitrary function ( XOR ) Most useful for non-linear mappings Works well with noisy data Generalizes well given sufficient examples Rapid recognition speed Has inspired many new learning algorithms
Other Networks and  Advanced Issues
Other Networks and  Advanced Issues Variations in feed-forward architecture jump connections to output nodes hidden nodes that vary in structure Recurrent networks with feedback connections Probabilistic networks General Regression networks Unsupervised self-organizing networks
THE END Thanks for your participation!
 

More Related Content

PPTX
Artificial neural network
PPTX
ANN load forecasting
PPTX
Neural Network Research Projects Topics
PDF
Neural Computing
PPT
Artificial Neural Networks - ANN
PPTX
Neural network
PPTX
Neural network
PPTX
Artificial Neural Network(Artificial intelligence)
Artificial neural network
ANN load forecasting
Neural Network Research Projects Topics
Neural Computing
Artificial Neural Networks - ANN
Neural network
Neural network
Artificial Neural Network(Artificial intelligence)

What's hot (20)

PPTX
Neural networks
PPTX
neural network
PDF
Artificial neural network for machine learning
PPTX
Artificial neural network
PPT
artificial neural network
PDF
Neural network
PDF
Lecture 11 neural network principles
PPTX
Artificial Neural Network
PPTX
Forecasting of Sales using Neural network techniques
PPTX
Neural Network Classification and its Applications in Insurance Industry
PPTX
Artificial Neural Network
PPTX
Artifical Neural Network and its applications
PPT
Artificial neural network
PPT
Artificial neural network
PDF
Neural networks introduction
PPT
Artificial Intelligence: Artificial Neural Networks
PPT
Neural network final NWU 4.3 Graphics Course
PDF
Artificial Neural Network Paper Presentation
PPT
Artificial neural networks in hydrology
PPTX
Introduction Of Artificial neural network
Neural networks
neural network
Artificial neural network for machine learning
Artificial neural network
artificial neural network
Neural network
Lecture 11 neural network principles
Artificial Neural Network
Forecasting of Sales using Neural network techniques
Neural Network Classification and its Applications in Insurance Industry
Artificial Neural Network
Artifical Neural Network and its applications
Artificial neural network
Artificial neural network
Neural networks introduction
Artificial Intelligence: Artificial Neural Networks
Neural network final NWU 4.3 Graphics Course
Artificial Neural Network Paper Presentation
Artificial neural networks in hydrology
Introduction Of Artificial neural network
Ad

Viewers also liked (18)

PDF
Analysis and applications of artificial neural networks
PPT
Sublimation vs Digital Printing By Sukhvir Sabharwal
PPTX
How does rotary heat machine work on fabric
PPT
Micromachining Technology Seminar Presentation
PPT
Cryptography
PPT
Smartplug ppt
PPT
Cryptography and E-Commerce
PPTX
Lessons from Software for Synthetic Biology
PPTX
Thesis presentation
PPTX
Virtual manufacturing
PPTX
Laser Assisted Micro Machining (lamm)
PDF
IBA Admission - Mystery Revealed (Infographics)
PPTX
Artificial intelligence NEURAL NETWORKS
PPTX
Cryptography
PPTX
Cryptography.ppt
PPTX
Cryptography
PDF
Artificial neural networks
PPTX
Neural network & its applications
Analysis and applications of artificial neural networks
Sublimation vs Digital Printing By Sukhvir Sabharwal
How does rotary heat machine work on fabric
Micromachining Technology Seminar Presentation
Cryptography
Smartplug ppt
Cryptography and E-Commerce
Lessons from Software for Synthetic Biology
Thesis presentation
Virtual manufacturing
Laser Assisted Micro Machining (lamm)
IBA Admission - Mystery Revealed (Infographics)
Artificial intelligence NEURAL NETWORKS
Cryptography
Cryptography.ppt
Cryptography
Artificial neural networks
Neural network & its applications
Ad

Similar to NEURAL Network Design Training (20)

PPTX
Nimrita deep learning
PDF
The Power of Auto ML and How Does it Work
PDF
Intro to Machine Learning by Microsoft Ventures
PDF
Heuristic design of experiments w meta gradient search
PPTX
A Framework for Scene Recognition Using Convolutional Neural Network as Featu...
PPTX
Metabolomic Data Analysis Workshop and Tutorials (2014)
PDF
Automated Testing and Safety Analysis of Deep Neural Networks
PPTX
Comparative Study of Machine Learning Algorithms for Sentiment Analysis with ...
PDF
Getting started with Machine Learning
PPT
backpropagation in neural networks
PDF
AI and Deep Learning
PDF
featurers_Machinelearning___________.pdf
PDF
EssentialsOfMachineLearning.pdf
PPTX
Deeplearning for Computer Vision PPT with
PDF
Integrating Artificial Intelligence with IoT
PDF
Introduction to Machine Learning with SciKit-Learn
PDF
Predicting Moscow Real Estate Prices with Azure Machine Learning
PDF
Predicting Moscow Real Estate Prices with Azure Machine Learning
PDF
Predicting Moscow Real Estate Prices with Azure Machine Learning
PPTX
Sachpazis: Demystifying Neural Networks: A Comprehensive Guide
Nimrita deep learning
The Power of Auto ML and How Does it Work
Intro to Machine Learning by Microsoft Ventures
Heuristic design of experiments w meta gradient search
A Framework for Scene Recognition Using Convolutional Neural Network as Featu...
Metabolomic Data Analysis Workshop and Tutorials (2014)
Automated Testing and Safety Analysis of Deep Neural Networks
Comparative Study of Machine Learning Algorithms for Sentiment Analysis with ...
Getting started with Machine Learning
backpropagation in neural networks
AI and Deep Learning
featurers_Machinelearning___________.pdf
EssentialsOfMachineLearning.pdf
Deeplearning for Computer Vision PPT with
Integrating Artificial Intelligence with IoT
Introduction to Machine Learning with SciKit-Learn
Predicting Moscow Real Estate Prices with Azure Machine Learning
Predicting Moscow Real Estate Prices with Azure Machine Learning
Predicting Moscow Real Estate Prices with Azure Machine Learning
Sachpazis: Demystifying Neural Networks: A Comprehensive Guide

More from ESCOM (20)

PDF
redes neuronales tipo Som
DOC
redes neuronales Som
PDF
redes neuronales Som Slides
PDF
red neuronal Som Net
PDF
Self Organinising neural networks
DOC
redes neuronales Kohonen
DOC
Teoria Resonancia Adaptativa
DOC
ejemplo red neuronal Art1
DOC
redes neuronales tipo Art3
DOC
Art2
DOC
Redes neuronales tipo Art
DOC
Neocognitron
PPT
Neocognitron
PPT
Neocognitron
PPT
Fukushima Cognitron
PPT
Counterpropagation NETWORK
PPT
Counterpropagation NETWORK
PPT
Counterpropagation
PPT
Teoría de Resonancia Adaptativa Art2 ARTMAP
PPT
Teoría de Resonancia Adaptativa ART1
redes neuronales tipo Som
redes neuronales Som
redes neuronales Som Slides
red neuronal Som Net
Self Organinising neural networks
redes neuronales Kohonen
Teoria Resonancia Adaptativa
ejemplo red neuronal Art1
redes neuronales tipo Art3
Art2
Redes neuronales tipo Art
Neocognitron
Neocognitron
Neocognitron
Fukushima Cognitron
Counterpropagation NETWORK
Counterpropagation NETWORK
Counterpropagation
Teoría de Resonancia Adaptativa Art2 ARTMAP
Teoría de Resonancia Adaptativa ART1

Recently uploaded (20)

PDF
102 student loan defaulters named and shamed – Is someone you know on the list?
PDF
Supply Chain Operations Speaking Notes -ICLT Program
PDF
The Lost Whites of Pakistan by Jahanzaib Mughal.pdf
PPTX
Lesson notes of climatology university.
PDF
Complications of Minimal Access Surgery at WLH
PPTX
GDM (1) (1).pptx small presentation for students
PPTX
Renaissance Architecture: A Journey from Faith to Humanism
PPTX
school management -TNTEU- B.Ed., Semester II Unit 1.pptx
PPTX
master seminar digital applications in india
PDF
VCE English Exam - Section C Student Revision Booklet
PPTX
Cell Types and Its function , kingdom of life
PDF
Abdominal Access Techniques with Prof. Dr. R K Mishra
PDF
TR - Agricultural Crops Production NC III.pdf
PPTX
Introduction_to_Human_Anatomy_and_Physiology_for_B.Pharm.pptx
PDF
FourierSeries-QuestionsWithAnswers(Part-A).pdf
PPTX
human mycosis Human fungal infections are called human mycosis..pptx
PDF
01-Introduction-to-Information-Management.pdf
PDF
STATICS OF THE RIGID BODIES Hibbelers.pdf
PDF
Insiders guide to clinical Medicine.pdf
PPTX
Final Presentation General Medicine 03-08-2024.pptx
102 student loan defaulters named and shamed – Is someone you know on the list?
Supply Chain Operations Speaking Notes -ICLT Program
The Lost Whites of Pakistan by Jahanzaib Mughal.pdf
Lesson notes of climatology university.
Complications of Minimal Access Surgery at WLH
GDM (1) (1).pptx small presentation for students
Renaissance Architecture: A Journey from Faith to Humanism
school management -TNTEU- B.Ed., Semester II Unit 1.pptx
master seminar digital applications in india
VCE English Exam - Section C Student Revision Booklet
Cell Types and Its function , kingdom of life
Abdominal Access Techniques with Prof. Dr. R K Mishra
TR - Agricultural Crops Production NC III.pdf
Introduction_to_Human_Anatomy_and_Physiology_for_B.Pharm.pptx
FourierSeries-QuestionsWithAnswers(Part-A).pdf
human mycosis Human fungal infections are called human mycosis..pptx
01-Introduction-to-Information-Management.pdf
STATICS OF THE RIGID BODIES Hibbelers.pdf
Insiders guide to clinical Medicine.pdf
Final Presentation General Medicine 03-08-2024.pptx

NEURAL Network Design Training

  • 1. Network Design & Training
  • 2. Network Design & Training Issues Design: Architecture of network Structure of artificial neurons Learning rules Training: Ensuring optimum training Learning parameters Data preparation and more ....
  • 4. Network Design Architecture of the network: How many nodes? Determines number of network weights How many layers? How many nodes per layer? Input Layer Hidden Layer Output Layer Automated methods: augmentation (cascade correlation) weight pruning and elimination
  • 5. Network Design Architecture of the network: Connectivity? Concept of model or hypothesis space Constraining the number of hypotheses: selective connectivity shared weights recursive connections
  • 6. Network Design Structure of artificial neuron nodes Choice of input integration: summed, squared and summed multiplied Choice of activation (transfer) function: sigmoid (logistic) hyperbolic tangent Gaussian linear soft-max
  • 7. Network Design Selecting a Learning Rule Generalized delta rule (steepest descent) Momentum descent Advanced weight space search techniques Global Error function can also vary - normal - quadratic - cubic
  • 9. Network Training How do you ensure that a network has been well trained? Objective: To achieve good generalization accuracy on new examples/cases Establish a maximum acceptable error rate Train the network using a validation test set to tune it Validate the trained network against a separate test set which is usually referred to as a production test set
  • 10. Network Training Approach #1: Large Sample When the amount of available data is large ... Available Examples Training Set Production Set 70% 30% Used to develop one ANN model Compute Test error Divide randomly Generalization error = test error Test Set
  • 11. Network Training Approach #2: Cross-validation When the amount of available data is small ... Available Examples Training Set Pro. Set 10% 90% Repeat 10 times Used to develop 10 different ANN models Accumulate test errors Generalization error determined by mean test error and stddev Test Set
  • 12. Network Training How do you select between two ANN designs ? A statistical test of hypothesis is required to ensure that a significant difference exists between the error rates of two ANN models If Large Sample method has been used then apply McNemar’s test* If Cross-validation then use a paired t test for difference of two proportions *We assume a classification problem, if this is function approximation then use paired t test for difference of means
  • 13. Network Training Mastering ANN Parameters Typical Range learning rate - 0.1 0.01 - 0.99 momentum - 0.8 0.1 - 0.9 weight-cost - 0.1 0.001 - 0.5 Fine tuning : - adjust individual parameters at each node and/or connection weight automatic adjustment during training
  • 14. Network Training Network weight initialization Random initial values +/- some range Smaller weight values for nodes with many incoming connections Rule of thumb: initial weight range should be approximately coming into a node
  • 15. Network Training Typical Problems During Training E # iter E # iter E # iter Would like: But sometimes: Steady, rapid decline in total error Seldom a local minimum - reduce learning or momentum parameter Reduce learning parms. - may indicate data is not learnable
  • 17. Data Preparation Garbage in Garbage out The quality of results relates directly to quality of the data 50%-70% of ANN development time will be spent on data preparation The three steps of data preparation: Consolidation and Cleaning Selection and Preprocessing Transformation and Encoding
  • 18. Data Preparation Data Types and ANNs Three basic data types: nominal discrete symbolic ( A, yes, small ) ordinal discrete numeric (-5, 3, 24) continuous numeric (0.23, -45.2, 500.43) bp ANNs accept only continuous numeric values (typically 0 - 1 range)
  • 19. Data Preparation Consolidation and Cleaning Determine appropriate input attributes Consolidate data into working database Eliminate or estimate missing values Remove outliers (obvious exceptions) Determine prior probabilities of categories and deal with volume bias
  • 20. Data Preparation Selection and Preprocessing Select examples random sampling Consider number of training examples? Reduce attribute dimensionality remove redundant and/or correlating attributes combine attributes (sum, multiply, difference) Reduce attribute value ranges group symbolic discrete values quantize continuous numeric values
  • 21. Data Preparation Transformation and Encoding Discrete symbolic or numeric values Transform to discrete numeric values Encode the value 4 as follows: one-of-N code ( 0 1 0 0 0 ) - five inputs thermometer code ( 1 1 1 1 0 ) - five inputs real value ( 0.4 )* - one input Consider relationship between values ( single, married, divorce ) vs. ( youth, adult, senior ) * Target values should be 0.1 - 0.9 , not 0.0 - 1.0 range
  • 22. Data Preparation Transformation and Encoding Continuous numeric values De-correlate example attributes via normalization of values: Euclidean: n = x/sqrt(sum of all x^2) Percentage: n = x/(sum of all x) Variance based: n = (x - (mean of all x))/variance Scale values using a linear transform if data is uniformly distributed or use non-linear (log, power) if skewed distribution
  • 23. Data Preparation Transformation and Encoding Continuous numeric values Encode the value 1.6 as: Single real-valued number ( 0.16 )* - OK! Bits of a binary number ( 010000 ) - BAD! one-of-N quantized intervals ( 0 1 0 0 0 ) - NOT GREAT! - discontinuities distributed (fuzzy) overlapping intervals ( 0.3 0.8 0.1 0.0 0.0 ) - BEST! * Target values should be 0.1 - 0.9 , not 0.0 - 1.0 range
  • 24. TUTORIAL #5 Develop and train a BP network on real-world data
  • 26. Post-Training Analysis Examining the neural net model: Visualizing the constructed model Detailed network analysis Sensitivity analysis of input attributes: Analytical techniques Attribute elimination
  • 27. Post-Training Analysis Visualizing the Constructed Model Graphical tools can be used to display output response as selected input variables are changed Response Size Temp
  • 28. Post-Training Analysis Detailed network analysis Hidden nodes form internal representation Manual analysis of weight values often difficult - graphics very helpful Conversion to equation, executable code Automated ANN to symbolic logic conversion is a hot area of research
  • 29. Post-Training Analysis Sensitivity analysis of input attributes Analytical techniques factor analysis network weight analysis Feature (attribute) elimination forward feature elimination backward feature elimination
  • 30. The ANN Application Development Process Guidelines for using neural networks 1. Try the best existing method first 2. Get a big training set 3. Try a net without hidden units 4. Use a sensible coding for input variables 5. Consider methods of constraining network 6. Use a test set to prevent over-training 7. Determine confidence in generalization through cross-validation
  • 31. Example Applications Pattern Recognition (reading zip codes) Signal Filtering (reduction of radio noise) Data Segmentation (detection of seismic onsets) Data Compression (TV image transmission) Database Mining (marketing, finance analysis) Adaptive Control (vehicle guidance)
  • 32. Pros and Cons of Back-Prop
  • 33. Pros and Cons of Back-Prop Cons: Local minimum - but not generally a concern Seems biologically implausible Space and time complexity: lengthy training times It’s a black box! I can’t see how it’s making decisions? Best suited for supervised learning Works poorly on dense data with few input variables
  • 34. Pros and Cons of Back-Prop Pros: Proven training method for multi-layer nets Able to learn any arbitrary function ( XOR ) Most useful for non-linear mappings Works well with noisy data Generalizes well given sufficient examples Rapid recognition speed Has inspired many new learning algorithms
  • 35. Other Networks and Advanced Issues
  • 36. Other Networks and Advanced Issues Variations in feed-forward architecture jump connections to output nodes hidden nodes that vary in structure Recurrent networks with feedback connections Probabilistic networks General Regression networks Unsupervised self-organizing networks
  • 37. THE END Thanks for your participation!
  • 38.