practice_paper_2
practice_paper_2
practice_paper_2
12. What is the role of the loss function in training neural networks?
a) To compute the accuracy of the model
b) To determine the learning rate
c) To measure the difference between predicted and actual values
d) To initialize the weights of the network
Answer: c) To measure the difference between predicted and actual values
13. Which of the following is a common activation function used in neural networks
for binary classification tasks?
a) ReLU (Rectified Linear Activation)
b) Sigmoid
c) Tanh (Hyperbolic Tangent)
d) Softmax
Answer: b) Sigmoid
15. Which of the following is not a common optimizer used in training neural
networks?
a) Adam
b) Gradient Descent
c) RMSprop
d) Newton's Method
Answer: d) Newton's Method
20. In PCA, what does the term "explained variance" refer to?
a) The percentage of total variance explained by each principal component
b) The number of principal components in the dataset
c) The amount of noise added to the dataset
d) The number of features after dimensionality reduction
Answer: a) The percentage of total variance explained by each principal
component
Over-sampling Technique)
c) Feature scaling
d) L1 regularization
Answer: b) SMOTE (Synthetic Minority Over-sampling Technique)
27. Which of the following is a drawback of using the elbow method to determine the
optimal number of clusters in K-means clustering?
a) It is computationally expensive
b) It requires a large amount of data
c) It may not always provide a clear indication of the optimal number of
clusters
d) It is sensitive to outliers
Answer: c) It may not always provide a clear indication of the optimal number
of clusters
29. Which of the following is a disadvantage of using the RBF kernel in SVM?
a) It is not suitable for non-linearly separable data
b) It is computationally expensive
c) It requires the data to be linearly separable
d) It is not affected by the choice of hyperparameters
Answer: b) It is computationally expensive
31. Which activation function is commonly used in the output layer of a binary
classification neural network?
a) Sigmoid
b) ReLU
c) Tanh
d) Softmax
Answer: a) Sigmoid
33. What is the purpose of the learning rate in neural network optimization
algorithms?
a) To control the rate at which weights are updated during training
b) To determine the number of training epochs
c) To define the size of the input layer
d) To measure the accuracy of the model
Answer: a) To control the rate at which weights are updated during training
34. Which of the following statements about convolutional neural networks (CNNs) is
true?
a) CNNs are primarily used for sequence data
b) CNNs consist of fully connected layers only
c) CNNs automatically learn spatial hierarchies of features
d) CNNs are not suitable for image classification tasks
Answer: c) CNNs automatically learn spatial hierarchies of features
35. Which activation function is commonly used in the hidden layers of a neural
network?
a) Sigmoid
b) ReLU
c) Tanh
d) Softmax
Answer: b) ReLU
36. Which of the following is true about recurrent neural networks (RNNs)?
a) RNNs are only used for image recognition tasks
b) RNNs are not suitable for sequential data
c) RNNs have connections between neurons that form a directed cycle
d) RNNs do not suffer from the vanishing gradient problem
Answer: c) RNNs have connections between neurons that form a directed cycle
38. Which of the following is true about Long Short-Term Memory (LSTM) networks?
a) LSTMs do not suffer from the vanishing gradient problem
b) LSTMs have a simpler architecture compared to standard RNNs
c) LSTMs are not suitable for sequential data
d) LSTMs do not have memory cells
Answer: a) LSTMs do not suffer from the vanishing gradient problem
40. Which technique is commonly used to address the vanishing gradient problem in
neural networks?
a) Batch normalization
b) Dropout regularization
c) Gradient clipping
d) Learning rate scheduling
Answer: c) Gradient clipping
42. Which regularization technique penalizes large weights by adding their squared
magnitude to the loss function?
a) L1 regularization
b) L2 regularization
c) Elastic Net regularization
d) Ridge regression
Answer: b) L2 regularization
43. What is the purpose of the softmax function in the output layer of a multi-
class classification neural network?
a) To normalize the output probabilities
b) To introduce non-linearity
c) To compute the loss function
d) To regularize the model
Answer
44. Which optimization algorithm adjusts the learning rate during training based on
the performance of the model?
a) Gradient Descent
b) Stochastic Gradient Descent (SGD)
c) Adagrad
d) Adam
Answer: d) Adam
45. Which of the following is true about transfer learning in neural networks?
a) It involves training a model from scratch for a new task
b) It does not require pre-trained models
c) It is not applicable to image recognition tasks
d) It leverages knowledge gained from training on one task to improve
performance on another task
Answer: d) It leverages knowledge gained from training on one task to improve
performance on another task
47. Which of the following is a common technique used for handling missing data in
machine learning?
a) Dropping rows with missing values
b) Imputing missing values with the mean of the column
c) Imputing missing values with the median of the column
d) All of the above
Answer: d) All of the above
49. Which of the following is a common technique used for feature selection in
machine learning?
a) Principal Component Analysis (PCA)
b) Recursive Feature Elimination (RFE)
c) Regularization
d) All of the above
Answer: d) All of the above
53. Which of the following is a common technique used for handling categorical
variables in machine learning?
a) One-Hot Encoding
b) Label Encoding
c) Ordinal Encoding
d) All of the above
Answer: d) All of the above
56. Which of the following is a common technique used for text vectorization in
NLP?
a) Bag of Words (BoW)
b) Principal Component Analysis (PCA)
c) Support Vector Machines (SVM)
d) K-means clustering
Answer: a) Bag of Words (BoW)
57. What is the purpose of the IDF (Inverse Document Frequency) term in TF-IDF
vectorization?
a) To measure the frequency of a term in a document
b) To measure the importance of a term in a collection of documents
c) To normalize the term frequencies in a document
d) To compute the cosine similarity between documents
Answer: b) To measure the importance of a term in a collection of documents
65. Which
66. Which of the following is a common technique used for feature extraction in
image processing?
a) Principal Component Analysis (PCA)
b) Convolutional Neural Networks (CNNs)
c) Support Vector Machines (SVMs)
d) K-means clustering
Answer: b) Convolutional Neural Networks (CNNs)
69. Which of the following is true about transfer learning in machine learning?
a) It involves training a model from scratch for a new task
b) It does not require pre-trained models
c) It is not applicable to image recognition tasks
d) It leverages knowledge gained from training on one task to improve
performance on another task
Answer: d) It leverages knowledge gained from training on one task to improve
performance on another task
71. Which of the following is a common technique used for handling imbalanced
datasets?
a) SMOTE (Synthetic Minority Over-sampling Technique)
b) Downsampling the majority class
c) Upsampling the minority class
d) All of the above
Answer: d) All of the above