What is a binary classification problem?
- A problem with two algorithms
- A problem with two classes or outcomes
- A problem with two input variables
- A problem with two solutions
A binary classification problem involves predicting one of two possible classes or outcomes. It's a specific type of classification where the target variable has only two possible values.
What are the potential issues with using a large value of K in KNN, and how can they be mitigated?
- All of the above
- Complexity, can be mitigated by feature selection
- Overfitting, can be mitigated by reducing dimensions
- Underfitting, can be mitigated by reducing K
A large value of K in KNN may cause underfitting. It can be mitigated by reducing K or using techniques like cross-validation.
In a situation where the MAE is significantly lower than the RMSE, what might this tell you about the distribution of the errors in your model?
- Errors are normally distributed; no impact on model
- Errors are uniformly distributed; no large outliers
- Many large errors, few small outliers
- Many small errors, few large outliers
When the Mean Absolute Error (MAE) is significantly lower than the Root Mean Squared Error (RMSE), it likely indicates that the model has many small errors and a few large outliers. RMSE, being sensitive to larger errors, would be higher, while MAE would be less impacted by those larger errors. An analysis of the residuals can further elucidate the nature of these errors.
In the context of SVM, the support vectors are the data points that lie closest to the _________.
- Data center
- Hyperplane
- Kernel
- Margin
Support vectors are the data points that lie closest to the hyperplane and influence its position and orientation.
Can you explain what separates support vectors from other data points in SVM?
- Cluster membership
- Color
- Distance to hyperplane
- Size
Support vectors are the data points closest to the hyperplane, and they determine its position in SVM.
How do interpretability and explainability vary between AI, Machine Learning, and Deep Learning?
- AI and Deep Learning are equally interpretable, Machine Learning is least
- AI is least interpretable, Machine Learning and Deep Learning are equally interpretable
- AI is most interpretable, Machine Learning is moderate, Deep Learning is least
- Machine Learning is most interpretable, AI is moderate, Deep Learning is least
Generally, AI techniques can vary in interpretability, traditional Machine Learning models tend to be more interpretable, and Deep Learning models are often the least interpretable due to their complexity.
How does the choice of loss function affect the learning process in a Machine Learning model?
- It defines the optimization algorithm
- It determines the learning rate
- It measures how well the model's predictions match the true values
- It selects the type of regularization
The loss function measures the discrepancy between the predicted values and the actual values, guiding the optimization process. Different loss functions can emphasize different aspects of the error, influencing how the model learns.
How do the hyperparameters in Ridge and Lasso affect the bias-variance tradeoff?
- Increase bias, reduce variance
- Increase both bias and variance
- No effect
- Reduce bias, increase variance
The hyperparameters in Ridge and Lasso control the regularization strength. Increasing them increases bias but reduces variance, helping to prevent overfitting.
The _________ is a single summary value that illustrates the ability of a classification model to discriminate between positive and negative classes.
- AUC
- F1-Score
- Precision
- Recall
The AUC (Area Under the Curve) provides a single summary value representing the model's ability to differentiate between positive and negative classes, considering various thresholds.
How does the Kernel Trick transform the input data?
- Compresses data
- Maps data into higher-dimensional space
- Normalizes data
- Reduces dimensionality
The Kernel Trick maps the input data into a higher-dimensional space to enable linear separation in SVM.