Dl Question Bank
Dl Question Bank
REFERENCES:
1. Aurélien Géron, “Hands-On Machine Learning with Scikit-Learn and TensorFlow”, Oreilly, 2018.
2. Josh Patterson, Adam Gibson, “Deep Learning: A Practitioner’s Approach”, O’Reilly Media, 2017.
CO1 2 2 1 - 1 - 1 - - 1 2 1 1 1 1
CO2 2 2 1 - 1 - 1 - - 1 2 1 1 1 1
CO3 3 2 2 2 1 - 1 - - 1 2 1 1 1 1
CO4 3 2 2 2 1 - 1 - - 1 2 1 1 1 1
CO5 3 2 2 2 1 - 1 - - 1 2 1 1 1 1
UNIT I INTRODUCTION
S.No Knowledge level Course Content Course Outcomes
1 R, U & Ap Neural Networks
2 U & Ap Application Scope of Neural Networks
3 An & Ap Artificial Neural Network
4 An & Ap ANN - An Introduction
5 An & Ap Evolution of Neural Networks CO1
6 U & Ap Basic Models of Artificial Neural Network
relationship between input features and a continuous target variable by fitting a linear equation to
the data.Types:Simple ,polynomial,multiple regression
32 What is meant by classification tasks in supervised learning(DEC/2022)
In supervised learning, classification tasks involve assigning input data to discrete
classes or categories based on the learned patterns. Examples include image classification, BL1
sentiment analysis, and spam detection.
Task: naive bayes,decision tree,Random forest,logistic regression
33 What is linear regression in supervised learning(DEC/2022)
Linear regression is a supervised learning technique used to model the relationship
between a dependent variable (also called the target or output) and one or more independent BL2
variables (predictors or input features) by fitting a linear equation to the observed data. It is
primarily used for regression tasks, where the goal is to predict a continuous numerical value.
34 What is decision tree in supervised learning(April/2021)
A decision tree is a supervised learning algorithm used for both classification and BL1
regression tasks. It builds a tree-like model of decisions and their possible consequences.
35 What is neural network in supervised learning
In supervised learning, a neural network is a type of model composed of
BL1
interconnected neurons or nodes that can be used for various tasks, including both regression
and classification. It can capture complex patterns in data through its layered structure.
UNIT I/ PART- B
1. Describe the concept of transfer learning in neural networks, and provide a detailed example
BL2
demonstrating its application in a specific domain.
2. Describe the concept of recurrent neural networks (RNNs) (DEC/2022) BL1
3. Provide a real-world example where CNNs have demonstrated superior performance compared to
traditional neural networks or other machine learning algorithms. BL5
4. Trace the evolution of neural networks from the perceptron model to modern deep learning
BL4
architectures. (JAN/2022)
5. Describe the structure and functioning of a Radial Basis Function Network (RBFN) BL2
6. Define and explain the significance of activation functions in neural networks. (JAN/2021) BL1
7. Describe the concept of weight initialization in neural networks(DEC/2022) BL1
8. Explain the purpose and functioning of dropout regularization in neural networks BL2
9. discuss the impact of different initialization techniques on training convergence and model BL4
performance. Provide examples to illustrate. (JAN/2021)
10 Provide examples of different activation functions and discuss scenarios where specific functions BL5
are more suitable. (JAN/2022)
UNIT I/ PART- C
1. Choose a specific industry or field (e.g., finance, healthcare, and robotics) and elaborate on three
distinct applications of Neural Networks within it, highlighting the unique advantages they bring to BL6
each application.
2. Explain the architecture and training process of Convolutional Neural Network (CNN) in
BL1
detail(JAN/2021)
3. Explain the architecture and training process of a Multilayer Perceptron (MLP) in detail, and BL6
provide a practical example where an MLP is well-suited for a specific task. (JAN/2022)
4. Elaborate on the architecture and training mechanisms of a Self-Organizing Map (SOM), and BL3
provide an example illustrating how SOMs can be applied in unsupervised learning tasks.
5. Explain the process of training a recurrent neural network (RNN) for sequential data analysis. BL2
(JAN/2021)
6. Explain the process of training a convolutional neural network (CNN) for image classification. BL3
Detail the architecture, including convolutional layers, pooling layers, and fully connected layers.
UNIT II
ASSOCIATIVE MEMORY AND UNSUPERVISED LEARNING NETWORKS
Training Algorithms for Pattern Association-Autoassociative Memory Network-Heteroassociative Memory
Network-Bidirectional Associative Memory (BAM)-Hopfield Networks-Iterative Autoassociative Memory
Networks-Temporal Associative Memory Network-Fixed Weight Competitive Nets-Kohonen Self-Organizing
Feature Maps-Learning Vector Quantization-Counter propagation Networks-Adaptive Resonance Theory Network.
UNIT II/PART A
1 Define the term "pattern association" in the context of neural networks BL1
In neural networks, pattern association refers to the ability to link and retrieve specific
patterns or features from a given input, allowing the network to recognize and recall familiar
patterns based on learned associations.
2 Give the difference between Hebbian learning and Perceptron learning in pattern BL3
association.
Hebbian learning is an unsupervised learning rule that strengthens connections between
neurons when they are simultaneously active, promoting associative memory formation.
Perceptron learning, on the other hand, is a supervised learning algorithm used for binary
classification tasks, adjusting weights based on misclassification errors.
3 How does an Auto associative Memory Network differ from a Heteroassociative Memory BL4
Network?
Autoassociative Memory Network retrieves complete patterns from partial or noisy
inputs, essentially completing missing information. Heteroassociative Memory Network
associates one pattern with another, enabling the retrieval of related patterns based on learned
associations.
4 What is Learning Vector Quantization (LVQ)? BL1
LVQ is a supervised learning algorithm where prototype vectors are iteratively adjusted
to better represent input data. It classifies input patterns based on the similarity between
prototypes and input vectors.
5 Explain the principle behind Bidirectional Associative Memory (BAM). BL2
BAM is a neural network architecture that reinforces connections between neurons
activated in pairs. It enables bidirectional pattern retrieval, allowing the network to recognize
patterns in both forward and reverse associations.
6 Define Autoassociative Memory Network. (JAN/2021) BL1
An Autoassociative Memory Network is a type of neural network that can reconstruct
complete patterns from partial or noisy inputs, effectively filling in missing information to
retrieve the original pattern.
7 Define Kohonen Self-Organizing Feature Maps. BL1
Kohonen Maps are unsupervised neural networks that map high-dimensional input data
onto a lower-dimensional grid while preserving the topological relationships between data
points. They're useful for tasks like clustering and dimensionality reduction.
8 Describe the functioning of Hopfield Networks. BL1
Hopfield Networks are recurrent neural networks where each neuron is connected to
every other neuron. They utilize iterative update rules to store and retrieve patterns, allowing
for content-addressable memory and associative recall.
9 What is the significance of Iterative Autoassociative Memory Networks? BL2
Iterative Autoassociative Memory Networks: These networks refine pattern
representations through iterative processing, enhancing the accuracy and robustness of pattern
retrieval, particularly from noisy or incomplete inputs.
10 How does Temporal Associative Memory Network differ from other associative BL3
memory models?
Unlike other associative memory models, Temporal Associative Memory
Network focuses on associating patterns over time. This allows for the sequential recall
of patterns, making it suitable for tasks involving temporal sequences.
11 Define Heteroassociative Memory Network. (April/2021) BL1
Heteroassociative Memory Network associates one pattern with another, enabling the
retrieval of related patterns based on the similarity of their associations. This is distinct from
autoassociative networks which retrieve similar patterns from partial inputs.
12 Explain the concept of Fixed Weight Competitive Nets. BL2
These neural networks have fixed connections between input neurons and output neurons.
Each output neuron competes to respond to a specific pattern, with the strongest response
representing the recognized pattern.
29 Provide an example scenario where Kohonen Self-Organizing Feature Maps have been BL3
successfully applied.
Kohonen Self-Organizing Feature Maps Example: In customer segmentation for
marketing, these maps group similar customer behaviors, enabling targeted marketing strategies
tailored to specific segments.
30 Explain Counterpropagation Networks. BL2
They have a two-stage training process: competitive selection of neurons in the first
stage, and error-correction in the second stage. This enables them to categorize input patterns
effectively.
31 Provide a practical example of an application where Adaptive Resonance Theory Network BL3
(ART) has shown significant advantage.
In medical diagnosis, ART networks adaptively adjust to variations in patient data,
allowing for accurate and reliable classification of diseases, even with variable symptoms and
test results.
32 Define Temporal Associative Memory Network. BL1
It's an associative memory model that associates patterns over time, making it ideal for
tasks like speech recognition, where understanding the temporal context is crucial.
33 Explain Fixed Weight Competitive Nets. BL2
These neural networks have fixed connections between input and output neurons, and
they're employed in scenarios like face recognition systems, where specific facial features must
be identified reliably.
34 Define Adaptive Resonance Theory Network (ART). BL2
Adaptive Resonance Theory Network (ART) Definition: ART networks
dynamically adjust their response thresholds and synaptic weights, making them excellent
for tasks where patterns may vary or evolve over time, such as adaptive learning systems.
35 Explain an application domain where Fixed Weight Competitive Nets are employed. BL4
In fingerprint recognition systems, they identify unique features by establishing fixed
connections, ensuring consistent and accurate recognition of individuals.
UNIT II/ PART- B
1 Analyze the key characteristics and limitations of Bidirectional Associative Memory (BAM) in BL4
pattern association.
2 Evaluate the effectiveness and limitations of Hopfield Networks in solving specific problem BL5
domains.
3 Explain the significance of Iterative Autoassociative Memory Networks in complex pattern BL2
recognition tasks.
4 Assess the suitability of Temporal Associative Memory Network in scenarios with dynamic data BL4
patterns.
5 Evaluate the strengths and weaknesses of Kohonen Self-Organizing Feature Maps in clustering BL5
and dimensionality reduction tasks.
6 Compare Learning Vector Quantization (LVQ) with other unsupervised learning algorithms, BL4
discussing scenarios where LVQ is most effective.
7 Evaluate the adaptability and learning capabilities of Adaptive Resonance Theory Network BL5
(ART) in dynamic environments. (JAN/2022)
8 Define Autoassociative Memory Network and provide an example of its application. BL1
9 What are the key characteristics of Bidirectional Associative Memory (BAM) and how is it BL2
utilized in pattern association?
10 Explain the concept of Iterative Autoassociative Memory Networks and provide a practical BL2
application example.
UNIT II/ PART- C
1 Compare and contrast Autoassociative Memory Network and Heteroassociative Memory BL4
Network, highlighting their respective applications and advantages.
2 Analyze the function and potential applications of Fixed Weight Competitive Nets in BL4
unsupervised learning.
11. Explain the rationale behind the Convolution Operation in CNNs and its role in feature BL2
extraction from images.
It scans a filter (also known as kernel) over the input to extract features. This operation
allows the network to recognize local patterns in the input, enabling it to learn hierarchical
representations of features.
12. Analyze the effectiveness of Convolutional Neural Networks in computer vision tasks BL4
compared to other machine learning approaches.
Effectiveness of CNNs in Computer Vision: CNNs outperform other machine learning
approaches in computer vision tasks because they can automatically learn and extract features
directly from raw pixel values, eliminating the need for manual feature engineering and enabling
better performance on complex visual recognition tasks.
13. What is the Extreme Learning Machine Model and in what scenarios is it particularly BL2
effective?
ELM is a machine learning algorithm where only the output weights are learned, while
the input weights are randomly generated. It is particularly effective in scenarios with large
datasets and high-dimensional input spaces, as it offers simplicity, speed, and good
generalization performance.
14. What is the significance of Datatypes in Third-Generation Neural Networks? BL1
Datatypes in Third-Generation Neural Networks determine how information is
represented and processed. For example, in Transformers, the use of float16 or bfloat16 can
significantly reduce memory requirements without sacrificing model accuracy.
15. Explain how different Data Types can affect the performance and memory requirements BL5
of Convolutional Neural Networks.
Using lower precision data types like float16 or int8 can reduce memory requirements and
improve inference speed in CNNs. However, it may also lead to a slight loss in model accuracy
due to reduced precision.
16. Compare and contrast Spiking Neural Networks with traditional artificial neural BL4
networks(JAN/2021)
Spiking Neural Networks model neuron communication with spikes, which is more bio-
realistic but can be computationally expensive. Traditional ANNs use continuous activation
functions. SNNs are efficient for event-based processing, while ANNs are more suitable for
traditional continuous data processing.
17. List Spiking Neural Networks’s advantages and limitations. BL1
Advantages: Event-driven processing, high energy efficiency, ability to
process asynchronous data, and biological plausibility.
Limitations: Higher computational requirements, lack of standardized frameworks, and
challenges in training.
18. Explain the key components of Convolutional Neural Networks, including Convolution BL2
Operation, Pooling, and variants of the basic Convolution Function.
Convolution Operation: It applies a filter over the input to extract features. This
enables the network to recognize local patterns.
Pooling: Reduces spatial dimensions, focusing on important features and aiding
in translation invariance.
Variants of Convolution Function: Different functions (e.g., dilated convolution)
modify the convolution operation for specific applications.
21. Describe the concept of Structured Outputs in Convolutional Neural Networks and their BL2
importance in specialized tasks.
Structured outputs provide detailed information about the spatial arrangement of objects
in tasks like object detection and segmentation. They are crucial for tasks where precise
localization and delineation of objects within an image are required.
22. Explain how different Datatypes can impact the performance of Third-Generation Neural BL2
Networks.
Different datatypes (e.g., float16, int8) can significantly affect the model's performance.
Choosing the right datatype is crucial for balancing accuracy and computational efficiency.
Floating Point has high precision values when compared to integer.
23. Define Datatypes in the context of Third-Generation Neural Networks. BL1
Datatypes in Third-Generation Neural Networks refer to the precision with which
numerical values are represented in the network's computations. They include options like
float32, float16, int8, etc., each with varying levels of precision and memory requirements.
24 Define Data Types in the context of Convolutional Neural Networks and explain their BL1
relevance.
Data types in CNNs refer to the precision of numerical values used in computations.
For example, using lower precision data types like float16 or int8 can reduce memory
requirements and improve inference speed without sacrificing much accuracy.
25 How do Datatypes affect the memory requirements of Third-Generation Neural BL2
Networks?
Using lower precision datatypes (e.g., float16) reduces memory requirements but may
result in some loss of model accuracy. Higher precision datatypes (e.g., float32) require more
memory but provide greater numerical precision.
26 Provide an example scenario where the choice of Datatypes is crucial for the effectiveness BL3
of a neural network.
In edge computing for IoT devices, choosing lower precision datatypes like int8 is crucial
for reducing memory usage and computational load, allowing the neural network to run
efficiently on resource-constrained devices.
27 Detail Efficient Convolution Algorithms and how they enhance the computational BL2
efficiency of CNNs.
These algorithms optimize the computation of convolutions, reducing the number of
operations required. Techniques like Winograd, FFT-based convolutions, and depthwise
separable convolutions significantly enhance the computational efficiency of CNNs.
28 Evaluate the diverse applications of Third-Generation Neural Networks in Computer BL3
Vision.
Third-Generation NNs, such as Transformers, have been applied in tasks like image
classification, object detection, and image captioning. They excel in tasks requiring an
understanding of contextual relationships within images.
29 Evaluate the diverse applications of Third-Generation Neural Networks in Image BL3
Generation.
Third-Generation NNs like Generative Adversarial Networks (GANs) have found
applications in generating realistic images for tasks like face synthesis, artistic style transfer, and
deepfake creation.
33 Analyze the neuroscientific principles incorporated into Spiking Neural Networks and BL4
their potential benefits in AI research.
SNNs replicate biological neurons' behavior with events based on timing and spikes.
They incorporate concepts like synaptic weight updates based on spike timings and refractory
periods, making them more biologically plausible and potentially useful in modeling complex
neural systems.
34 Provide examples of applications of Third-Generation Neural Networks in Computer BL2
Vision, Image Generation, and Image Compression.
Computer Vision: Transformers for image captioning, where understanding
image context is crucial.
Image Generation: GANs for creating photorealistic images in tasks like face synthesis.
Image Compression: VAEs for lossy compression, reducing storage requirements
while maintaining quality.
35 Evaluate the diverse applications of Third-Generation Neural Networks in Computer BL5
Vision, Image Generation, and Image Compression, providing specific examples.
Computer Vision: Transformers for fine-grained image classification, GANs for
super- resolution image reconstruction.
Image Generation: StyleGAN for generating high-quality art and realistic faces.
Image Compression: VAEs for efficient storage and transmission of medical
imaging data.
UNIT III/ PART- B
1. Define Spiking Neural Networks and explain their fundamental operation. BL1
2. Discuss the importance of Structured Outputs in specialized applications like object detection and BL2
segmentation using CNNs.
3. Define Data Types in the context of Convolutional Neural Networks and discuss their BL1
importance.
4. Define the Extreme Learning Machine Model and provide an example of its application. BL2
5. Analyze the variants of the basic Convolution Function in CNNs and discuss scenarios where BL4
specific variants are preferred.
6. Describe the core principles of Convolutional Neural Networks (CNNs) and their applications. BL2
(JAN/2021)
7. Evaluate the diverse applications of Third-Generation Neural Networks in Computer Vision, BL5
Image Generation, and Image Compression, providing specific examples.
8. Explain the Neuroscientific Basis behind Spiking Neural Networks and how it relates to BL3
biological neural systems.
9. Explain how different Data Types can affect the performance and memory requirements of BL4
Convolutional Neural Networks. (Dec/2022)
10. Examine the effectiveness of Convolutional Neural Networks in computer vision tasks compared BL5
to other machine learning approaches.
UNIT III/ PART- C
1. Analyze the neuroscientific principles incorporated into Spiking Neural Networks and their BL4
potential benefits in AI research.
BL1