What method is commonly used to estimate the coefficients in Simple Linear Regression?

  • Maximum Likelihood Estimation
  • Minimizing the Sum of Absolute Errors
  • Minimizing the Sum of the Squares of the Residuals
  • Neural Networks Training
In Simple Linear Regression, the method used to estimate coefficients is by minimizing the sum of the squares of the residuals, known as the Ordinary Least Squares (OLS) method.

You've built a multiple linear regression model and found that two or more predictors are highly correlated. What problems might this cause, and how can you solve them?

  • High bias, Address by increasing the model complexity
  • High variance, Address by using Lasso regression
  • Overfitting, Address by removing correlated features or using Ridge regression
  • Underfitting, Address by adding more features
Multicollinearity, where predictors are highly correlated, can cause overfitting and unstable estimates. This can be addressed by removing correlated features or using Ridge regression, which penalizes large coefficients and reduces the impact of multicollinearity.

In Supervised Learning, _________ and ___________ are the two main types of problems.

  • Classification; Clustering
  • Classification; Regression
  • Regression; Clustering
  • Regression; Ensemble Learning
In Supervised Learning, the two main types of problems are Classification and Regression. Classification is about categorizing data into predefined classes, while Regression is predicting a continuous outcome.

You've trained a model with a small training set and a large testing set. What challenges might you encounter, and how could they be addressed?

  • Both Overfitting and Underfitting
  • Data is perfectly balanced
  • Overfitting
  • Underfitting
A small training set might lead to overfitting, where the model memorizes noise from the training data. Conversely, it might also lead to underfitting if the model fails to capture the underlying pattern. Cross-validation, bootstrapping, or augmenting the training set with additional relevant data can help balance the model's ability to generalize.

In KNN, how does an increase in the value of K generally affect the bias and variance of the model?

  • Decreases bias, increases variance
  • Decreases both bias and variance
  • Increases bias, decreases variance
  • Increases both bias and variance
Increasing the value of K generally increases bias and decreases variance in the KNN model.

What is regression in the context of Machine Learning?

  • A type of classification
  • Analyzing computer performance
  • Predicting a categorical value
  • Predicting a numerical value
Regression in Machine Learning refers to the process of predicting a continuous numerical outcome or dependent variable based on one or more independent variables. It plays a crucial role in various areas, like economics, where it might be used to predict prices, or in biology, to forecast growth rates.

Name a popular algorithm used in classification problems.

  • Clustering
  • Decision Trees
  • Linear Regression
  • Principal Component Analysis
Decision Trees are a popular algorithm used in classification problems. They work by recursively partitioning the data into subsets based on feature values, leading to a decision on the class label.

What are some common methods to detect multicollinearity in a dataset?

  • Adding more data
  • Feature scaling
  • Regularization techniques
  • VIF, Correlation Matrix
Common methods to detect multicollinearity include calculating the Variance Inflation Factor (VIF) and examining the correlation matrix among variables.

If multicollinearity is a concern, ________ regularization can provide a solution by shrinking the coefficients.

  • ElasticNet
  • Lasso
  • Ridge
  • nan
Ridge regularization provides a solution to multicollinearity by shrinking the coefficients through the L2 penalty, which helps to stabilize the estimates.

You built a model using Lasso regularization but some important features were wrongly set to zero. How would you modify your approach to keep these features?

  • Combine with ElasticNet
  • Decrease L1 penalty
  • Increase L1 penalty
  • Switch to Ridge
Combining with ElasticNet allows for balancing between L1 and L2 penalties, thus avoiding complete elimination of important features by the L1 penalty.

In the context of building a model, the _________ are carefully selected and processed to improve the model's performance.

  • features
  • parameters
  • testing set
  • training set
"Features" are the input variables that are carefully selected and processed (e.g., through feature engineering or scaling) to enhance the model's predictive performance.

How do Precision and Recall trade-off in a classification problem, and when might you prioritize one over the other?

  • Increasing Precision decreases Recall, prioritize Precision when false positives are costly
  • Increasing Precision increases Recall, prioritize Recall when false positives are costly
  • Precision and Recall are independent, no trade-off
  • nan
Precision and Recall often trade-off; increasing one can decrease the other. You might prioritize Precision when false positives are more costly (e.g., spam detection) and Recall when false negatives are more costly (e.g., fraud detection).