36. Which of the following best describes a ‘feedforward neural network’?
A) A network that processes input in a circular manner
B) A network where data flows in one direction from input to output
C) A network that learns from feedback signals
D) A network with recurrent connections
View AnswerB
37. What is ‘gradient descent’?
A) A technique to initialize weights
B) An algorithm for optimizing a function by iteratively moving towards the steepest descent
C) A method to increase model capacity
D) A type of activation function
View AnswerB
38. Which of the following represents a major disadvantage of deep learning models?
A) They can achieve high accuracy.
B) They often require large datasets for training.
C) They are interpretable.
D) They can be easily tuned.
View AnswerB
39. What does ‘batch size’ refer to in the context of neural network training?
A) The number of epochs in training
B) The number of samples processed before the model updates
C) The number of hidden layers in the model
D) The size of the training dataset
View AnswerB
40. In a neural network, what does ‘weight initialization’ refer to?
A) Setting all weights to zero
B) Randomly assigning values to the weights before training
C) Adjusting weights during backpropagation
D) Setting a fixed learning rate
View AnswerB
41. What is the purpose of an ’embedding layer’ in neural networks?
A) To convert categorical data into continuous vectors
B) To apply dropout
C) To reduce dimensionality
D) To initialize weights
View AnswerA
42. What is ‘transfer learning’?
A) Training a model on one dataset and applying it to another
B) Transferring weights between models
C) Changing the architecture of a neural network
D) A method for feature selection
View AnswerA
43. Which of the following is NOT a common type of neural network?
A) Feedforward Neural Network
B) Convolutional Neural Network
C) Recurrent Neural Network
D) Linear Regression Network
View AnswerD
44.What is ‘layer normalization’?
A) A technique to normalize input features
B) A method for normalizing the outputs of each layer
C) A method to reduce overfitting
D) A technique to initialize weights
View AnswerB
45. Which of the following activation functions is linear?
A) Sigmoid
B) ReLU
C) Tanh
D) Linear
View AnswerD
46. What is the primary function of a loss function in neural networks?
A) To calculate the accuracy of the model
B) To measure how well the model’s predictions match the actual data
C) To optimize the weights of the model
D) To initialize the model
View AnswerB
47. What does ‘regularization’ do in a neural network?
A) Increases model complexity
B) Decreases the learning rate
C) Helps prevent overfitting
D) Reduces the number of layers
View AnswerC
48. What is a common challenge when working with unbalanced datasets in neural networks?
A) Reduced training speed
B) Increased accuracy
C) Bias towards the majority class
D) Decreased model capacity
View AnswerC
49. Which of the following is an example of a recurrent neural network (RNN)?
A) LSTM (Long Short-Term Memory)
B) CNN (Convolutional Neural Network)
C) DNN (Deep Neural Network)
D) GAN (Generative Adversarial Network)
View AnswerA
50. What is ‘self-supervised learning’?
A) Learning without labeled data
B) Learning with only labeled data
C) Learning using unsupervised techniques
D) Learning through reinforcement
View AnswerA
51. Which of the following describes ‘convolution’ in a CNN?
A) A method of pooling data
B) A technique to detect features in the input
C) A way to increase data dimensions
D) A method for data augmentation
View AnswerB
52. What type of loss function would you use for regression tasks?
A) Cross-Entropy Loss
B) Mean Squared Error
C) Hinge Loss
D) Binary Cross-Entropy
View AnswerB
53. What is ‘sparsity’ in the context of neural networks?
A) A measure of model accuracy
B) A property where most weights are zero
C) A type of activation function
D) A training technique to reduce data
View AnswerB
54. Which type of neural network is best for generating new content such as images or music?
A) Convolutional Neural Network
B) Generative Adversarial Network
C) Recurrent Neural Network
D) Feedforward Neural Network
View AnswerB
55. What does ‘early stopping’ refer to in training neural networks?
A) Stopping training before the model learns
B) Monitoring validation performance and stopping training to avoid overfitting
C) Reducing the learning rate
D) Terminating the training process randomly
View AnswerB
56. What is a common method for handling missing data in neural networks?
A) Ignore missing data
B) Fill with zeros
C) Imputation
D) Increase batch size
View AnswerC
57. In the context of CNNs, what is a ‘kernel’?
A) A component that connects neurons
B) A small matrix used for convolution
C) A type of activation function
D) A method for regularization
View AnswerB
58. Which of the following best describes ‘data augmentation’?
A) Increasing the size of the training dataset by creating modified versions of data
B) Reducing the complexity of the model
C) Adding new features to the dataset
D) Removing outliers from the dataset
View AnswerA
59. What does the ‘vanishing gradient’ problem refer to?
A) Gradients becoming too large during training
B) Gradients becoming too small, hindering weight updates in deep networks
C) Gradients fluctuating during training
D) Gradients being ignored by the optimizer
View AnswerB
60. What is ‘batch normalization’ used for in neural networks?
A) To increase the training batch size
B) To normalize the output of each layer to improve training speed and stability
C) To reduce the number of hidden layers
D) To decrease model complexity
View AnswerB
61. What is ‘L2 regularization’ also known as?
A) Dropout
B) Weight decay
C) Early stopping
D) Batch normalization
View AnswerB
62. In neural networks, what is the purpose of the ‘output layer’?
A) To process the input features
B) To transform the activations of hidden layers
C) To provide the final predictions
D) To initialize the weights
View AnswerC
63. What type of neural network architecture is characterized by using skip connections?
A) Convolutional Neural Network
B) Recurrent Neural Network
C) Residual Network
D) Radial Basis Function Network
View AnswerC
64. Which of the following describes ‘fine-tuning’ in neural networks?
A) Adjusting model architecture
B) Modifying hyperparameters
C) Making small adjustments to a pre-trained model
D) Increasing batch size
View AnswerC
65. What is the role of ‘hidden layers’ in a neural network?
A) To directly input data
B) To map inputs to outputs through transformations
C) To apply regularization techniques
D) To produce the final output
View AnswerB
66. Which of the following optimization algorithms uses momentum?
A) Adam
B) Gradient Descent
C) RMSprop
D) AdaGrad
View AnswerA
67. Which activation function is commonly used for hidden layers in deep networks?
A) Sigmoid
B) Tanh
C) ReLU
D) Softmax
View AnswerC
68. In the context of machine learning, what does ‘hyperparameter tuning’ refer to?
A) Adjusting the architecture of the model
B) Modifying parameters that are set before training
C) Updating weights during training
D) Increasing the training dataset size
View AnswerB
69. What is the purpose of the ‘loss curve’ in neural network training?
A) To visualize the accuracy of the model
B) To monitor the training and validation loss over epochs
C) To represent the model complexity
D) To visualize the weights of the model
View AnswerB
70. Which type of data is most commonly used in natural language processing with RNNs?
A) Image data
B) Text data
C) Video data
D) Tabular data
View AnswerB
71. What does the term ‘class imbalance’ refer to in a dataset?
A) Equal representation of all classes
B) Unequal representation of classes in the training data
C) Only one class present in the data
D) The data having too many features
View AnswerB
72. Which of the following strategies is often used to handle class imbalance?
A) Increase batch size
B) Data augmentation
C) Undersampling the majority class
D) All of the above
View AnswerD
73. What is the primary use of ‘attention mechanisms’ in neural networks?
A) To simplify the model architecture
B) To focus on relevant parts of the input when making predictions
C) To improve the training speed
D) To increase model capacity
View AnswerB
74. What does ‘model ensembling’ refer to?
A) Training multiple models and combining their predictions
B) Increasing the size of a single model
C) Reducing the number of features in the model
D) A method for hyperparameter tuning
View AnswerA
75. Which of the following describes ‘one-hot encoding’?
A) Converting categorical variables into binary format
B) Normalizing input features
C) Creating a single feature from multiple features
D) Reducing the dimensionality of data
View AnswerA
76. What is the purpose of a ‘validation set’ in machine learning?
A) To train the model
B) To evaluate the model during training
C) To optimize the model parameters
D) To test the final model performance
View AnswerB
77. Which of the following is a feature of transfer learning?
A) Training a model from scratch
B) Adapting a pre-trained model to a new task
C) Using only labeled data
D) Creating a completely new architecture
View AnswerB
78. What is the primary benefit of using a pre-trained model?
A) Lower accuracy
B) Reduced training time
C) Increased complexity
D) More hyperparameter tuning
View AnswerB
79. What is the role of ‘gradient clipping’ in neural networks?
A) To increase gradients
B) To prevent exploding gradients
C) To optimize learning rate
D) To simplify model architecture
View AnswerB
80. Which of the following layers is typically used for dimensionality reduction in CNNs?
A) Dense layer
B) Pooling layer
C) Convolutional layer
D) Normalization layer
View AnswerB
81. What is ‘parameter sharing’ in CNNs?
A) Using different weights for different inputs
B) Using the same filter weights across different positions in the input
C) Sharing model weights between different models
D) Using multiple activation functions
View AnswerB
82. Which of the following neural network types is used to classify images?
A) Generative Adversarial Network
B) Recurrent Neural Network
C) Convolutional Neural Network
D) Linear Regression
View AnswerC
83. What is a common challenge faced when training very deep networks?
A) Too few parameters
B) Overfitting
C) Vanishing and exploding gradients
D) Lack of computational power
View AnswerC
84. Which of the following best describes a ‘feedforward layer’?
A) A layer where outputs are fed back into the same layer
B) A layer where inputs flow in only one direction
C) A layer that combines outputs from multiple layers
D) A layer that processes sequential data
View AnswerB
85. What does ‘early stopping’ help prevent during training?
A) Underfitting
B) Overfitting
C) Learning rate issues
D) Poor initialization
View AnswerB
86. In a neural network, which technique is used to optimize the model’s weights?
A) Forward propagation
B) Backpropagation
C) Dropout
D) Activation function
View AnswerB
87. Which of the following methods can be used to visualize how a model makes predictions?
A) Hyperparameter tuning
B) Feature importance analysis
C) Model ensembling
D) Data normalization
View AnswerB
88. Which of the following describes ‘ensemble learning’?
A) Using a single model for predictions
B) Combining predictions from multiple models to improve accuracy
C) Optimizing a model’s architecture
D) None of the above
View AnswerB
89. What does the ‘F1 score’ measure in classification tasks?
A) The average of precision and recall
B) The total number of correct predictions
C) The accuracy of the model
D) The speed of the model
View AnswerA
90. What is ‘class activation mapping’ used for?
A) To visualize class-specific information in a CNN
B) To improve training speed
C) To reduce overfitting
D) To optimize model architecture
View AnswerA
91. Which of the following is NOT a type of RNN?
A) LSTM
B) GRU
C) CNN
D) Vanilla RNN
View AnswerC
92. What is a ‘confusion matrix’?
A) A tool for data preprocessing
B) A summary of prediction results on a classification problem
C) A technique for hyperparameter tuning
D) A way to visualize the architecture of a neural network
View AnswerB
93. Which of the following describes ‘softmax regression’?
A) A method for dimensionality reduction
B) A type of activation function for the output layer of a multi-class model
C) A technique for reducing overfitting
D) A loss function for regression tasks
View AnswerB
94. What is ‘over-sampling’ in the context of imbalanced datasets?
A) Decreasing the size of the majority class
B) Increasing the size of the minority class by duplicating samples
C) Combining classes
D) Using a larger batch size
View AnswerB
95. Which of the following describes a ‘feedforward layer’?
A) A layer where inputs loop back to the same layer
B) A layer that processes inputs sequentially
C) A layer where data flows in one direction, from input to output
D) A layer that generates new data samples
View AnswerC
96. What is ‘time series forecasting’?
A) Predicting future data points based on historical data
B) Predicting categorical outputs
C) Classifying images
D) Reducing the dimensionality of data
View AnswerA
97. What is a potential drawback of using very deep neural networks?
A) They always overfit.
B) They may require extensive computational resources and training time.
C) They are always more accurate than shallow networks.
D) They are easier to interpret.
View AnswerB
98. Which technique is often used for sequential data prediction?
A) Convolutional Neural Network
B) K-Nearest Neighbors
C) Recurrent Neural Network
D) Support Vector Machine
View AnswerC
99. What does ‘hyperparameter’ tuning involve?
A) Adjusting weights during training
B) Modifying parameters that govern the training process
C) Evaluating model accuracy
D) Increasing the dataset size
View AnswerB
100. Which of the following describes ‘gradient boosting’?
A) A technique that focuses on minimizing error by combining weak learners
B) A way to visualize gradient flows
C) A method to adjust learning rates
D) A strategy to maximize model complexity
View AnswerA