CU7009 NEURAL NETWORKS AND APPLICATIONS SYLLABUS FOR ME ECE 1ST SEMESTER - Anna University Multiple Choice Questions

CU7009 NEURAL NETWORKS AND APPLICATIONS SYLLABUS FOR ME ECE 1ST SEMESTER

ANNA UNIVERSITY, CHENNAI
REGULATIONS - 2013
CU7009 NEURAL NETWORKS AND APPLICATIONS SYLLABUS
ME 1ST SEM ELECTRONICS AND COMMUNICATION ENGINEERING SYLLABUS
CU7009 NEURAL NETWORKS AND APPLICATIONS SYLLABUS
CU7009 NEURAL NETWORKS AND APPLICATIONS SYLLABUS
UNIT I BASIC LEARNING ALGORITHMS
Biological Neuron – Artificial Neural Model - Types of activation functions – Architecture:Feedforward and Feedback – Learning Process: Error Correction Learning –Memory Based Learning – Hebbian Learning – Competitive Learning - Boltzman Learning – Supervised and Unsupervised Learning – Learning Tasks: Pattern Space – Weight Space – Pattern Association – Pattern Recognition – Function Approximation – Control – Filtering - Beamforming – Memory – Adaptation - Statistical Learning Theory – Single Layer Perceptron – Perceptron Learning Algorithm – Perceptron Convergence Theorem – Least Mean Square Learning Algorithm – Multilayer Perceptron – Back Propagation Algorithm – XOR problem – Limitations of Back Propagation Algorithm.

UNIT II RADIAL-BASIS FUNCTION NETWORKS AND SUPPORT VECTOR MACHINES RADIAL BASIS FUNCTION NETWORKS
Cover’s Theorem on the Separability of Patterns - Exact Interpolator – Regularization Theory – Generalized Radial Basis Function Networks - Learning in Radial Basis Function Networks Applications: XOR Problem – Image Classification.

SUPPORT VECTOR MACHINES
Optimal Hyperplane for Linearly Separable Patterns and Nonseparable Patterns – Support Vector Machine for Pattern Recognition – XOR Problem - -insensitive Loss Function – Support Vector Machines for Nonlinear Regression

UNIT III COMMITTEE MACHINES
Ensemble Averaging - Boosting – Associative Gaussian Mixture Model – Hierarchical Mixture of Experts Model(HME) – Model Selection using a Standard Decision Tree – A Priori and Postpriori Probabilities – Maximum Likelihood Estimation – Learning Strategies for the HME Model - EMAlgorithm – Applications of EM Algorithm to HME Model

NEURODYNAMICS SYSTEMS
Dynamical Systems – Attractors and Stability – Non-linear Dynamical Systems- Lyapunov Stability – Neurodynamical Systems – The Cohen-Grossberg Ttheorem.

UNIT IV ATTRACTOR NEURAL NETWORKS:
Associative Learning – Attractor Neural Network Associative Memory – Linear Associative Memory – Hopfield Network – Content Addressable Memory – Strange Attractors and Chaos- Error Performance of Hopfield Networks - Applications of Hopfield Networks – Simulated Annealing – Boltzmann Machine – Bidirectional Associative Memory – BAM Stability Analysis – Error Correction in BAMs - Memory Annihilation of Structured Maps in BAMS – Continuous BAMs – Adaptive BAMs – Applications

ADAPTIVE RESONANCE THEORY:
Noise-Saturation Dilemma - Solving Noise-Saturation Dilemma – Recurrent On-center – Off
surround Networks – Building Blocks of Adaptive Resonance – Substrate of Resonance Structural Details of Resonance Model – Adaptive Resonance Theory – Applications

UNIT V SELF ORGANISING MAPS:
Self-organizing Map – Maximal Eigenvector Filtering – Sanger’s Rule – Generalized Learning Law – Competitive Learning - Vector Quantization – Mexican Hat Networks - Self-organizing Feature Maps – Applications

PULSED NEURON MODELS:
Spiking Neuron Model – Integrate-and-Fire Neurons – Conductance Based Models – Computing with Spiking Neurons.

TOTAL: 45 PERIODS

REFERENCES:

1. Satish Kumar, “Neural Networks: A Classroom Approach”, Tata McGraw-Hill Publishing Company Limited, New Delhi, 2004.
2. Simon Haykin, “Neural Networks: A Comprehensive Foundation”, 2ed., Addison Wesley Longman (Singapore) Private Limited, Delhi, 2001.
3. Martin T.Hagan, Howard B. Demuth, and Mark Beale, “Neural Network Design”, Thomson Learning, New Delhi, 2003.
4. James A. Freeman and David M. Skapura, “Neural Networks Algorithms, Applications, and Programming Techniques, Pearson Education (Singapore) Private Limited, Delhi, 2003.

No comments:

Post a Comment