You have a dataset with clusters of varying densities. How would you configure the Epsilon and MinPts in DBSCAN to handle this?

  • Increase Epsilon; Decrease MinPts
  • Increase both Epsilon and MinPts
  • Reduce both Epsilon and MinPts
  • Use a different clustering algorithm
DBSCAN's Epsilon and MinPts are global parameters that apply to all clusters. If clusters have varying densities, tuning these parameters to fit one density might not suit others, leading to misclustering. In such a scenario, a different clustering algorithm that can handle varying densities might be more appropriate.

What is the main difference between Ridge and Lasso regularization?

  • Both use L1 penalty
  • Both use L2 penalty
  • Ridge uses L1 penalty, Lasso uses L2 penalty
  • Ridge uses L2 penalty, Lasso uses L1 penalty
Ridge regularization uses an L2 penalty, which shrinks coefficients but keeps them non-zero, while Lasso uses an L1 penalty, leading to some coefficients being exactly zero.

In PCA, the Eigenvectors are also known as the ________ of the data.

  • components
  • directions
  • eigendata
  • principal directions
In PCA, the Eigenvectors, also known as the "principal directions," define the directions in which the data varies the most. They form the axes of the new feature space and capture the essential structure of the data.

What is the intercept in Simple Linear Regression, and how is it interpreted?

  • Maximum Value of Y
  • Minimum Value of X
  • Start of the Line on X-axis
  • Value of Y when X is Zero
The intercept in Simple Linear Regression is the value of the dependent variable (Y) when the independent variable (X) is zero. It represents where the regression line crosses the Y-axis.

Why might one prefer to use MAE over MSE in evaluating a regression model?

  • MAE considers the direction of errors
  • MAE gives more weight to larger errors
  • MAE is less sensitive to outliers
  • MAE is more computationally expensive
One might prefer to use Mean Absolute Error (MAE) over Mean Squared Error (MSE) because MAE is less sensitive to outliers. While MSE squares the differences and thus gives more weight to larger errors, MAE takes the absolute value of the differences, providing an equal weighting. This makes MAE more robust when there are outliers or when one doesn't want to overly penalize larger deviations from the true values.

What challenges might arise when using Hierarchical Clustering on very large datasets?

  • Computationally intensive and requires high memory
  • Less accurate and requires more hyperparameters
  • Less sensitive to distance metrics and more prone to noise
  • Prone to overfitting and less interpretable
Hierarchical Clustering can be computationally intensive and require a lot of memory, especially when dealing with very large datasets. The algorithm has to compute and store a distance matrix, which has a size of O(n^2), where n is the number of data points. This can lead to challenges in computational efficiency and memory usage, making it less suitable for large-scale applications.

Imagine a scenario where you want to assess the stability of a statistical estimator. How would Bootstrapping help in this context?

  • By fixing the bias in the estimator
  • By increasing the size of the dataset
  • By repeating the sampling process with replacement and calculating the variance
  • By repeating the sampling process without replacement
Bootstrapping assesses the stability of a statistical estimator by repeating the sampling process with replacement and calculating variance, standard error, or other statistics. By creating numerous "bootstrap samples," it allows insights into the estimator's distribution, thereby providing a measure of its stability and reliability.

Why might pruning be necessary in the construction of a Decision Tree?

  • Determine Leaf Nodes
  • Increase Complexity
  • Increase Size
  • Reduce Overfitting
Pruning is necessary to remove unnecessary branches, simplifying the model and reducing the risk of overfitting the training data.

Which field utilizes Machine Learning to recommend products or media to consumers based on their past behavior?

  • Autonomous Driving
  • Education
  • Healthcare
  • Recommender Systems
Recommender Systems use machine learning algorithms to suggest products, media, or content to users based on their past interactions and behavior, creating personalized experiences.

You built a regression model and it's yielding a very low R-Squared value. What could be the reason and how would you improve it?

  • Data noise; Apply data cleaning
  • Incorrect model; Change the model
  • Poorly fitted; Improve the model fit
  • Too many features; Reduce features
A low R-Squared value might indicate that the model doesn't fit the data well. This could be due to an incorrect choice of model, underfitting, or other issues. Improving the model fit by selecting an appropriate algorithm, feature engineering, or hyperparameter tuning can address this problem.