The ability of SVMs to handle non-linear decision boundaries is achieved by transforming the input data into a higher-dimensional space using a ______.
- Classifier
- Dimensionality Reduction
- Ensemble
- Kernel
SVMs use a mathematical function called a kernel to transform data into a higher-dimensional space, enabling them to handle non-linear decision boundaries effectively.
The multi-armed bandit problem is a classic problem in which domain?
- Sequential Decision-Making Problems
- Natural Language Processing
- Computer Graphics
- Speech Recognition
The multi-armed bandit problem falls under the domain of Sequential Decision-Making Problems, specifically addressing scenarios where a decision must be made over time with limited resources.
In the context of RNNs, what problem does the introduction of gating mechanisms in LSTMs and GRUs aim to address?
- Vanishing and Exploding Gradients
- Overfitting and Data Loss
- Dimensionality Reduction and Compression
- Sequence Length Reduction and Truncation
The introduction of gating mechanisms in LSTMs and GRUs aims to address the problem of vanishing and exploding gradients, which occur during training due to the backpropagation of errors over long sequences. These mechanisms help RNNs capture long-range dependencies in data.
In K-means clustering, the algorithm iteratively updates the cluster centers until the within-cluster sum of squares is ________.
- Minimized
- Equal to 0
- Maximized
- Converged
In K-means clustering, the algorithm aims to minimize the within-cluster sum of squares (WCSS). This represents the total variance within clusters. As the algorithm iteratively updates the cluster centers, the goal is to minimize the WCSS, making "Minimized" the correct option.
In a scenario with noisy data, increasing the value of 'k' in the k-NN algorithm can help to ________ the noise.
- Amplify
- Eliminate
- Introduce
- Preserve
Increasing the value of 'k' in the k-NN algorithm can help eliminate noise. A higher 'k' value smooths the decision boundaries, reducing the impact of individual noisy data points.
Q-learning is a type of ________ learning algorithm that aims to find the best action to take given a current state.
- Reinforcement
- Supervised
- Unsupervised
- Semi-supervised
Q-learning is a type of reinforcement learning that focuses on finding the best action to take in a given state to maximize cumulative rewards.
When aiming to reduce both bias and variance, one might use techniques like ________ to regularize a model.
- Cross-Validation
- Data Augmentation
- Dropout
- L1 Regularization
L1 regularization is a technique used to reduce both bias and variance in a machine learning model. It does so by adding a penalty term to the model's loss function, which encourages the model to use fewer features, thus reducing complexity and variance. Dropout, Cross-Validation, and Data Augmentation are techniques but are not primarily used for regularization.
What does the "G" in GRU stand for when referring to a type of RNN?
- Gated
- Global
- Gradient
- Graph
The "G" in GRU stands for "Gated." GRU is a type of RNN that uses gating mechanisms to control information flow, making it capable of handling sequences efficiently.
One of the challenges in training deep RNNs is the ________ gradient problem, which affects the network's ability to learn long-range dependencies.
- Vanishing
- Exploding
- Overfitting
- Regularization
The vanishing gradient problem refers to the issue where gradients in deep RNNs become too small during training, making it challenging to capture long-range dependencies.
In the context of the bias-variance trade-off, which one is typically associated with complex models with many parameters?
- Balanced Bias-Variance
- High Bias
- High Variance
- Neither
High Variance is typically associated with complex models with many parameters. Complex models are more flexible and tend to fit the training data closely, resulting in high variance, which can lead to overfitting.