practice MCQ
practice MCQ
practice MCQ
Answer: a) When the gradients in a deep neural network become extremely small
7. Which of the following is a common loss function used for binary classification in deep learning?
a) Mean Absolute Error (MAE)
b) Mean Squared Error (MSE)
c) Binary Cross-Entropy
d) Categorical Cross-Entropy
Answer: c) Binary Cross-Entropy
8. What is the purpose of dropout regularization in deep learning?
a) To reduce overfitting
b) To increase the model's capacity
c) To improve the training speed
d) To handle imbalanced datasets
Answer: a) To reduce overfitting
9. What is an epoch in deep learning?
a) The number of layers in a neural network
b) The number of training examples in a dataset
c) The number of times the entire dataset is passed through the neural network during training
d) The number of neurons in a layer
Answer: c) The number of times the entire dataset is passed through the neural network during training
10. What is the purpose of a convolutional layer in a convolutional neural network (CNN)?
a) To reduce the dimensionality of the input
b) To extract spatial features from the input
c) To classify the input data
d) To apply non-linear transformations to the input
Answer: b) To extract spatial features from the input
11. Which activation function is commonly used in the hidden layers of a deep neural network?
a) ReLU (Rectified Linear Unit)
b) Sigmoid
c) Tanh
d) Softmax
Answer: a) ReLU (Rectified Linear Unit)
12. What is the purpose of pooling layers in a convolutional neural network (CNN)?
a) To reduce the size of the input data
b) To perform spatial downsampling
c) To extract the most important features
d) All of the above
Answer: d) All of the above
13. Which of the following is a common optimization algorithm used in deep learning?
a) Gradient Descent
b) Stochastic Gradient Descent (SGD)
c) Adam
d) All of the above
Answer: d) All of the above
16. Which type of RNN architecture is used to address the vanishing gradient problem?
a) Long Short-Term Memory (LSTM)
b) Gated Recurrent Unit (GRU)
c) Simple RNN
d) Bidirectional RNN
Answer: a) Long Short-Term Memory (LSTM)
17. Which deep learning technique is used for generating new, realistic data samples?
a) Generative Adversarial Networks (GANs)
b) Convolutional Neural Networks (CNNs)
c) Reinforcement Learning
d) Transfer Learning
Answer: a) Generative Adversarial Networks (GANs)
18. What is the purpose of transfer learning in deep learning?
a) To reuse pre-trained models on new tasks
b) To transfer knowledge from one domain to another
c) To speed up the training process
d) All of the above
Answer: d) All of the above
20. Which deep learning technique is used for sequence-to-sequence tasks, such as machine translation?
a) Attention Mechanism
b) Convolutional Neural Networks (CNNs)
c) Transfer Learning
d) Autoencoders
Answer: a) Attention Mechanism
21. Which of the following is a common activation function used in the output layer for binary
classification?
a) Sigmoid
b) ReLU (Rectified Linear Unit)
c) Tanh
d) Softmax
Answer: a) Sigmoid
22. What is the purpose of batch normalization in deep learning?
a) To reduce internal covariate shift
b) To accelerate the training process
c) To improve the generalization of the model
d) All of the above
Answer: d) All of the above
23. Which deep learning technique is used for unsupervised feature learning?
a) Autoencoders
b) Convolutional Neural Networks (CNNs)
c) Recurrent Neural Networks (RNNs)
d) Reinforcement Learning
Answer: a) Autoencoders
25. Which of the following is a popular deep learning architecture for object detection?
a) YOLO (You Only Look Once)
b) LSTM (Long Short-Term Memory)
c) VGG (Visual Geometry Group)
d) GAN (Generative Adversarial Network)
Answer: a) YOLO (You Only Look Once)
27. Which deep learning technique is used for learning latent representations from unlabelled data?
a) Self-supervised learning
b) Reinforcement Learning
c) Transfer Learning
d) Unsupervised Learning
Answer: d) Unsupervised Learning
32. Which type of image segmentation assigns each pixel in the image a class label?
A. Object detection
B. Semantic segmentation
C. Instance segmentation
D. Supervised learning
Answer: B) Semantic segmentation
33. Which popular word embedding method captures the semantic meaning of words by
positioning similar words closely in vector space?
A. Bag of Words (BoW)
B. Term Frequency-Inverse Document Frequency (TF-IDF)
C. Word2Vec
D. One-Hot Encoding
Answer: C) Word2Vec
37. Which deep learning technique is used for learning from delayed rewards?
A. Reinforcement Learning
B. Supervised Learning
C. Unsupervised Learning
D. Transfer Learning
Answer: A) Reinforcement Learning
38. What are the general tasks that are performed with backpropagation algorithm?
A. pattern mapping
B. function approximation
C. prediction
D. all of the mentioned
Answer: D) all of the mentioned
39. Which of the following statements is not true about Kernel methods?
a) It works by embedding the input data to some high dimensional feature space
b) Embedding into feature space can be determined uniquely by specifying a kernel function that
computes the dot product between data points in the feature space
c) It defines only the linear mapping to the feature space
d) Expensive computations in the high dimensional feature space can be avoided by evaluating
the kernel function
Answer:c) It defines only the linear mapping to the feature space
41.What is the advantage of basis function over mutilayer feedforward neural networks?
a) training of basis function is faster than MLFFNN
b) training of basis function is slower than MLFFNN
c) storing in basis function is faster than MLFFNN
d) none of the mentioned
Answer:a) training of basis function is faster than MLFFNN
42. What consist of a basic counterpropagation network?
a) a feedforward network only
b) a feedforward network with hidden layer
c) two feedforward network with hidden layer
d) none of the mentioned
Answer:c) two feedforward network with hidden layer
43. How can false minima be reduced in case of error in recall in feedback neural networks?
a) by providing additional units
b) by using probabilistic update
c) can be either probabilistic update or using additional units
d) none of the mentioned
Answer: b) by using probabilistic update
44. What property should a feedback network have, to make it useful for storing information?
a) accretive behaviour
b) interpolative behaviour
c) both accretive and interpolative behaviour
d) none of the mentioned
Answer: a) accretive behaviour
45. What is the objective of a pattern storage task in a network?
a) to store a given set of patterns
b) to recall a give set of patterns
c) both to store and recall
d) none of the mentioned
Answer: c) both to store and recall
46. Which is a simplest pattern recognition task in a feedback network?
a) heteroassociation
b) autoassociation
c) can be hetero or autoassociation, depends on situation
d) none of the mentioned
Answer: b) autoassociation
47. What happens when number of available energy minima be less than number of patterns to
be stored?
a) pattern storage is not possible in that case
b) pattern storage can be easily done
c) pattern storage problem becomes hard problem for the network
d) none of the mentioned
Answer: c) pattern storage problem becomes hard problem for the network
48. What happens when number of available energy minima be more than number of patterns to
be stored?
a) no effect
b) pattern storage is not possible in that case
c) error in recall
d) none of the mentioned
Answer: c) error in recall
49. Why there is error in recall, when number of energy minima is more the required number of
patterns to be stored?
a) due to noise
b) due to additional false maxima
c) due to additional false minima
d) none of the mentioned
Answer: c) due to additional false minima
50. Generalization feature of a multilayer feedforward network depends on factors?
a) architectural details
b) learning rate parameter
c) training samples
d) all of the mentioned
Answer: a) architectural details
51. What is accretive behaviour?
a) not a type of pattern clustering task
b) for small noise variations pattern lying closet to the desired pattern is recalled.
c) for small noise variations noisy pattern having parameter adjusted according to noise variation
is recalled
d) none of the mentioned
Answer: b) for small noise variations pattern lying closet to the desired pattern is recalled.
52. What is Interpolative behaviour?
a) not a type of pattern clustering task
b) for small noise variations pattern lying closet to the desired pattern is recalled.
c) for small noise variations noisy pattern having parameter adjusted according to noise variation
is recalled
d) none of the mentioned
Answer: c) for small noise variations noisy pattern having parameter adjusted according to noise
variation is recalled
53. What is the feature that doesn’t belongs to pattern mapping in feeddorward neural networks?
a) recall is direct
b) delta rule learning
c) non linear processing units
d) two layers
Answer: d) two layers
54. What is the primary purpose of a Recurrent Neural Network (RNN)?
a) Image classification
b) Text generation
c) Reinforcement learning
d) Object detection
Answer: b) Text generation
55. Which layer type is typically used to capture sequential dependencies in an RNN?
a) Input layer
b) Hidden layer
c) Output layer
d) Activation layer
Answer: b) Hidden layer
60. Which layer type is commonly used to initialize the hidden state in an RNN?
a) Input layer
b) Hidden layer
c) Output layer
d) Activation layer
Answer: b) Hidden layer
64. Which layer type is commonly used in RNNs for sequence-to-sequence tasks?
a) Input layer
b) Hidden layer
c) Output layer
d) Attention layer
Answer: d) Attention layer
65. What is the purpose of the backpropagation through time (BPTT) algorithm in RNN
training?
a) To compute the gradients and update the network's parameters
b) To adjust the learning rate during training
c) To prevent overfitting by regularizing the model
d) None of the above
Answer: a) To compute the gradients and update the network's parameters
66. Which layer type is commonly used in RNNs to handle variable-length inputs?
a) Input layer
b) Hidden layer
c) Output layer
d) None of the above
Answer: a) Input layer
68. Which layer type is responsible for handling the output at each time step in an RNN?
a) Input layer
b) Hidden layer
c) Output layer
d) Activation layer
Answer: c) Output layer
69. What is the purpose of the teacher forcing technique in RNN training?
a) To adjust the learning rate during training
b) To propagate the gradients through time
c) To reduce the computational complexity of the network
d) None of the above
Answer: b) To propagate the gradients through time
70. Which layer type is commonly used in RNNs for language modeling tasks?
a) Input layer
b) Hidden layer
c) Output layer
d) None of the above
Answer: c) Output layer
73. What is the purpose of the forget gate in a Gated Recurrent Unit (GRU)?
a) To control the flow of information from the previous hidden state
b) To adjust the learning rate during training
c) To compute the gradients for backpropagation
d) None of the above
Answer: a) To control the flow of information from the previous hidden state
74. Which layer type is commonly used in RNNs for machine translation tasks?
a) Input layer
b) Hidden layer
c) Output layer
d) Attention layer
Answer: d) Attention layer
75. What is the purpose of the peephole connections in a Long Short-Term Memory (LSTM)
network?
a) To allow the cell state to influence the gating mechanisms
b) To adjust the learning rate during training
c) To introduce non-linearity to the network
d) None of the above
Answer: a) To allow the cell state to influence the gating mechanisms
76. Which layer type is responsible for handling variable-length outputs in an RNN?
a) Input layer
b) Hidden layer
c) Output layer
d) None of the above
Answer: c) Output layer
80. Which layer type is responsible for handling variable-length inputs and outputs in an RNN?
a) Input layer
b) Hidden layer
c) Output layer
d) None of the above
Answer: d) None of the above