Artificial Neural network model & hidden layers in multilayer artificial neural networks
Outline • Introduction • Single Layer • Multi Layer • Learning • Backpropogation Algorithm
Introduction • Artificial Neural Network are computational models inspired by biological neural network used for processing large number of inputs which are mostly unknown.
What is Problem? •Our world is full of data. After collection and organization, data, if we are lucky, becomes information. In today's interconnected world, information exists in electronic form that can be stored and transmitted instantly. Challenge is to understand, integrate, and apply information to generate useful knowledge.
What is Solution? • Interpretation requires data acquisition, cleaning (preparing the data for analysis), analysis, and presentation in a way that permits knowledgeable decision making and action. Key is to extract information about data from relationships buried within the data itself.
Why Human System? •Human brain is most powerful pattern recognition engine ever invented.
Biological Neuron
Artificial Neuron
Artificial Neuron
History • 1943: McCulloch–Pitts “neuron” – Started the field • 1962: Rosenblatt’s perceptron – Learned its own weight values; convergence proof • 1969: Minsky & Papert book on perceptrons – Proved limitations of single-layer perceptron networks • 1982: Hopfield and convergence in symmetric networks – Introduced energy-function concept • 1986: Backpropagation of errors – Method for training multilayer networks • Present: Probabilistic interpretations, Bayesian and spiking networks
Single Layer • Input Layer • Output Layer
Multi Layer • Input Layer • Hidden Layer(s) • Output Layer
Multi Layer
• Instead of programming computational system to do specific tasks, teach system how to perform task • To do this, generate Artificial Intelligence System- AI • AI systems must be adaptive – able to learn from data on a continuous basis
ANN Architecture a. Single Layer Feed Forward
ANN Architecture Multilayer Feed Forward
ANN Architecture Recurrent Network
ANN Architecture a. Single Layer Feed Forward
ANN Architecture Multilayer Feed Forward
ANN Architecture Recurrent Network
Clasification of learning strategies
Supervised • Process of using desired output for training the NN • It employees a teacher to assist the network by telling the network what the desired response to a given input • Weights are modified according to the required output • Not practicable in all cases
Understanding Supervised and Unsupervised Learning A BA B A B
Two possible Solutions… A B A B A B A B A B A B
Supervised Learning • It is based on a labeled training set. • The class of each piece of data in training set is known. • Class labels are pre-determined and provided in the training phase. A B A B A B ε Class λ Class λ Class λ Class ε Class ε Class
26 A single perceptron can be used to represent many boolean functions. • AND function : x1 x2 output 0 0 -1 0 1 -1 1 0 -1 1 1 1 <Training examples> Decision hyperplane : w0 + w1 x1 + w2 x2 = 0 -0.8 + 0.5 x1 + 0.5 x2 = 0 x1 x2 wixi output 0 0 -0.8 -1 0 1 -0.3 -1 1 0 -0.3 -1 1 1 0.2 1 <Test Results> - - - + x1 x2 -0.8 + 0.5 x1 + 0.5 x2 = 0 - - - + x1 x2 -0.8 + 0.5 x1 + 0.5 x2 = 0
27 OR function • The two-input perceptron can implement the OR function when we set the weights: w0 = -0.3, w1 = w2 = 0.5 x1 x2 o utput 0 0 - 1 0 1 1 1 0 1 1 1 1 <Training e x ample s > Decision hyperplane : w0 + w1 x1 + w2 x2 = 0 -0.3 + 0.5 x1 + 0.5 x2 = 0 x1 x2 Σwixi output 0 0 -0.3 - 1 0 1 0.2 - 1 1 0 0.2 - 1 1 1 0.7 1 <Te s t Re s ults > - + + + x1 x2 -0.3 + 0.5 x1 + 0.5 x2 = 0 - + + + x1 x2 -0.3 + 0.5 x1 + 0.5 x2 = 0
Unsupervised • No teacher Required • Similar to the students learning on their own • Adaption rules • Adaption rule generate error signals
Reinforced • A teacher is assumed to b present but right answer is not given to the network • Network is given an indication whether output is right or wrong • Network use this indication to improve performance
Types of Neuron activation function 1. Heaviside 1 if s>=T F(s)= 0 if s<T
2. Linear f(s)=s
3. Piecewise Linear 1 if s>T F(s)= s if –T<s<T -1 if s<-T
Hardlimiter Unipolar Sigmoidal Bipolar Sigmoidal Unipolar Multimodal Radial Bessel’s Function
Backpropogation Algorithm
36 NEURAL NETWORK APPLICATION DEVELOPMENT The development process for an ANN application has eight steps. • Step 1: (Data collection). • Step 2: (Training and testing data separation For a moderately sized data set, 80% of the data are randomly selected for training, 10% for testing, and 10% secondary testing. • Step 3: (Network architecture) Important considerations are the exact number of perceptrons and the number of layers.
37 • Step 4: (Parameter tuning and weight initialization) • Step 5: (Data transformation) Transforms the application data into the type and format required by the ANN. • Step 6: (Training)
38 • Step 7: (Testing) – The testing examines the performance of the network using the derived weights by measuring the ability of the network to classify the testing data correctly. – Black-box testing (comparing test results to historical results) is the primary approach for verifying that inputs produce the appropriate outputs. • Step 8: (Implementation) Now a stable set of weights are obtained.
Thank You for your Attention..!! . .
We welcome any Questions….

Artificial neural network model & hidden layers in multilayer artificial neural networks

  • 1.
    Artificial Neural network model& hidden layers in multilayer artificial neural networks
  • 2.
    Outline • Introduction • SingleLayer • Multi Layer • Learning • Backpropogation Algorithm
  • 3.
    Introduction • Artificial NeuralNetwork are computational models inspired by biological neural network used for processing large number of inputs which are mostly unknown.
  • 4.
    What is Problem? •Ourworld is full of data. After collection and organization, data, if we are lucky, becomes information. In today's interconnected world, information exists in electronic form that can be stored and transmitted instantly. Challenge is to understand, integrate, and apply information to generate useful knowledge.
  • 5.
    What is Solution? •Interpretation requires data acquisition, cleaning (preparing the data for analysis), analysis, and presentation in a way that permits knowledgeable decision making and action. Key is to extract information about data from relationships buried within the data itself.
  • 6.
    Why Human System? •Humanbrain is most powerful pattern recognition engine ever invented.
  • 7.
  • 8.
  • 9.
  • 10.
    History • 1943: McCulloch–Pitts“neuron” – Started the field • 1962: Rosenblatt’s perceptron – Learned its own weight values; convergence proof • 1969: Minsky & Papert book on perceptrons – Proved limitations of single-layer perceptron networks • 1982: Hopfield and convergence in symmetric networks – Introduced energy-function concept • 1986: Backpropagation of errors – Method for training multilayer networks • Present: Probabilistic interpretations, Bayesian and spiking networks
  • 11.
    Single Layer • InputLayer • Output Layer
  • 12.
    Multi Layer • InputLayer • Hidden Layer(s) • Output Layer
  • 13.
  • 14.
    • Instead ofprogramming computational system to do specific tasks, teach system how to perform task • To do this, generate Artificial Intelligence System- AI • AI systems must be adaptive – able to learn from data on a continuous basis
  • 15.
    ANN Architecture a. SingleLayer Feed Forward
  • 16.
  • 17.
  • 18.
    ANN Architecture a. SingleLayer Feed Forward
  • 19.
  • 20.
  • 21.
  • 22.
    Supervised • Process ofusing desired output for training the NN • It employees a teacher to assist the network by telling the network what the desired response to a given input • Weights are modified according to the required output • Not practicable in all cases
  • 23.
  • 24.
  • 25.
    Supervised Learning • Itis based on a labeled training set. • The class of each piece of data in training set is known. • Class labels are pre-determined and provided in the training phase. A B A B A B ε Class λ Class λ Class λ Class ε Class ε Class
  • 26.
    26 A single perceptroncan be used to represent many boolean functions. • AND function : x1 x2 output 0 0 -1 0 1 -1 1 0 -1 1 1 1 <Training examples> Decision hyperplane : w0 + w1 x1 + w2 x2 = 0 -0.8 + 0.5 x1 + 0.5 x2 = 0 x1 x2 wixi output 0 0 -0.8 -1 0 1 -0.3 -1 1 0 -0.3 -1 1 1 0.2 1 <Test Results> - - - + x1 x2 -0.8 + 0.5 x1 + 0.5 x2 = 0 - - - + x1 x2 -0.8 + 0.5 x1 + 0.5 x2 = 0
  • 27.
    27 OR function • Thetwo-input perceptron can implement the OR function when we set the weights: w0 = -0.3, w1 = w2 = 0.5 x1 x2 o utput 0 0 - 1 0 1 1 1 0 1 1 1 1 <Training e x ample s > Decision hyperplane : w0 + w1 x1 + w2 x2 = 0 -0.3 + 0.5 x1 + 0.5 x2 = 0 x1 x2 Σwixi output 0 0 -0.3 - 1 0 1 0.2 - 1 1 0 0.2 - 1 1 1 0.7 1 <Te s t Re s ults > - + + + x1 x2 -0.3 + 0.5 x1 + 0.5 x2 = 0 - + + + x1 x2 -0.3 + 0.5 x1 + 0.5 x2 = 0
  • 28.
    Unsupervised • No teacherRequired • Similar to the students learning on their own • Adaption rules • Adaption rule generate error signals
  • 29.
    Reinforced • A teacheris assumed to b present but right answer is not given to the network • Network is given an indication whether output is right or wrong • Network use this indication to improve performance
  • 30.
    Types of Neuronactivation function 1. Heaviside 1 if s>=T F(s)= 0 if s<T
  • 31.
  • 32.
    3. Piecewise Linear 1if s>T F(s)= s if –T<s<T -1 if s<-T
  • 33.
    Hardlimiter Unipolar Sigmoidal Bipolar Sigmoidal UnipolarMultimodal Radial Bessel’s Function
  • 34.
  • 36.
    36 NEURAL NETWORK APPLICATION DEVELOPMENT Thedevelopment process for an ANN application has eight steps. • Step 1: (Data collection). • Step 2: (Training and testing data separation For a moderately sized data set, 80% of the data are randomly selected for training, 10% for testing, and 10% secondary testing. • Step 3: (Network architecture) Important considerations are the exact number of perceptrons and the number of layers.
  • 37.
    37 • Step 4:(Parameter tuning and weight initialization) • Step 5: (Data transformation) Transforms the application data into the type and format required by the ANN. • Step 6: (Training)
  • 38.
    38 • Step 7:(Testing) – The testing examines the performance of the network using the derived weights by measuring the ability of the network to classify the testing data correctly. – Black-box testing (comparing test results to historical results) is the primary approach for verifying that inputs produce the appropriate outputs. • Step 8: (Implementation) Now a stable set of weights are obtained.
  • 39.
    Thank You foryour Attention..!! . .
  • 40.
    We welcome anyQuestions….