You're given a dataset with several features, some of which are highly correlated. How would you handle this using dimensionality reduction techniques?
- Applying K-Means Clustering
- Applying L1 Regularization
- Applying Principal Component Analysis (PCA)
- Applying Random Forest
Principal Component Analysis (PCA) would be used to handle high correlation among features. It reduces dimensionality by creating new uncorrelated variables that capture the variance present in the original features.
The addition of _________ in the loss function is a common technique to regularize the model and prevent overfitting.
- bias
- learning rate
- regularization terms
- weights
Regularization terms (like L1 or L2 penalties) in the loss function constrain the model, reducing the risk of overfitting by preventing large weights.
You have trained an SVM but the decision boundary is not fitting well to the data. How could adjusting the hyperplane parameters help?
- Change the kernel's color
- Increase the size of the hyperplane
- Modify the regularization parameter 'C'
- Reduce the number of support vectors
Adjusting the regularization parameter 'C' controls the trade-off between margin maximization and error minimization, helping to fit the decision boundary better.
Discuss the difference between Euclidean distance and Manhattan distance metrics in the context of KNN.
- Euclidean is faster, Manhattan is more accurate
- Euclidean is for 3D, Manhattan for 2D
- Euclidean is for continuous data, Manhattan for categorical
- Euclidean uses squares, Manhattan uses absolutes
Euclidean distance is the square root of the sum of squared differences, while Manhattan distance is the sum of the absolute differences.
A dataset with very high between-class variance but low within-class variance is given. How would the LDA approach be beneficial here?
- LDA would be the same as PCA
- LDA would perform optimally due to the variance characteristics
- LDA would perform poorly
- LDA would require transformation of the dataset
LDA would "perform optimally" in this scenario, as high between-class variance and low within-class variance align perfectly with its objective of maximizing between-class variance and minimizing within-class variance.
When a Decision Tree is too complex and fits the training data too well, __________ techniques can be applied to simplify the model.
- Bagging
- Boosting
- Normalizing
- Pruning
When a Decision Tree is overfitting (too complex), pruning techniques can be applied to simplify the model. Pruning involves removing branches that have little predictive power, thereby reducing the complexity and the risk of overfitting.
What's the difference between simple linear regression and multiple linear regression?
- One is for classification, the other for regression
- One is linear, the other is nonlinear
- One uses one predictor, the other uses more than one
- One uses time series data, the other doesn't
Simple linear regression involves one independent variable to predict the dependent variable, while multiple linear regression uses two or more independent variables for prediction.
What is the main goal of Artificial Intelligence as a scientific discipline?
- To create complex algorithms
- To improve data storage
- To make money through technology
- To mimic human intelligence and decision-making
The primary goal of Artificial Intelligence is to develop systems that can mimic human intelligence and decision-making processes.
How can Cross-Validation help in hyperparameter tuning?
- By allowing repeated testing on the same validation set
- By improving model accuracy directly
- By providing robust performance estimates to select the best hyperparameters
- By reducing computation time
Cross-Validation enables hyperparameter tuning by providing a robust estimate of the model's performance across different data splits. This process helps to find hyperparameters that generalize well to unseen data, minimizing the risk of overfitting, and allowing a more informed selection of optimal hyperparameters.
How do Ridge and Lasso regularization techniques interact with Polynomial Regression to mitigate overfitting?
- By adding a penalty term to constrain coefficients
- By fitting low-degree polynomials
- By ignoring interaction terms
- By increasing the model's complexity
Ridge and Lasso regularization techniques mitigate overfitting in Polynomial Regression by adding a penalty term to the loss function. This constrains the coefficients, reducing the complexity of the model, and helps in avoiding overfitting.