Explain the role of Machine Learning in optimizing supply chain and inventory management.

  • Customer Segmentation
  • Image Recognition
  • Sentiment Analysis
  • Supply Chain Optimization
Machine Learning plays a vital role in supply chain optimization by analyzing and predicting demand, improving inventory management, optimizing logistics, and enhancing decision-making through predictive analytics.

A linear regression model's R-Squared value significantly improves after polynomial features are added. What could be the reason, and what should you be cautious about?

  • Reason: Improved fit to nonlinear patterns; Caution: Risk of overfitting
  • Reason: Increased bias; Caution: Risk of complexity
  • Reason: Increased complexity; Caution: Risk of bias
  • Reason: Reduced error; Caution: Risk of underfitting
The significant improvement in R-Squared value after adding polynomial features indicates an improved fit to potentially nonlinear patterns in the data. However, caution should be exercised as adding too many polynomial features may lead to overfitting, where the model fits the noise in the training data rather than the underlying trend. Regularization techniques and cross-validation can be used to mitigate this risk.

You are working on a project where Simple Linear Regression seems appropriate, but the independent variable is categorical. How would you handle this situation?

  • Change the Dependent Variable
  • Ignore the Variable
  • Treat as Continuous Variable
  • Use Dummy Variables
For a categorical independent variable in Simple Linear Regression, you can create dummy variables to represent the categories.

An educational institution wants to personalize its online learning platform for individual student needs. How would you leverage Machine Learning to achieve this goal?

  • Image Recognition, Fraud Detection
  • Personalized Learning Paths, Data Analysis
  • Recommender Systems, Drug Development
  • Supply Chain Management, Weather Prediction
Creating Personalized Learning Paths and analyzing student data using techniques like clustering or decision trees allows for the customization of content and resources according to individual student performance and preferences.

In regression analysis, the metric that tells you the proportion of the variance in the dependent variable that is predictable from the independent variables is called _________.

  • Adjusted R-Squared
  • Mean Squared Error
  • R-Squared
  • Root Mean Squared Error
In regression analysis, R-Squared tells you the proportion of the variance in the dependent variable that is predictable from the independent variables. It provides a measure of how well the regression line fits the data.

_________ regression adds a penalty equivalent to the square of the magnitude of coefficients, whereas _________ regression adds a penalty equivalent to the absolute value of the magnitude.

  • Lasso
  • Linear
  • Polynomial
  • Ridge
Ridge regression adds a penalty equivalent to the square of the magnitude of coefficients, whereas Lasso regression adds a penalty equivalent to the absolute value of the magnitude.

How is Recall defined in classification, and when is it an important metric to consider?

  • False Positives / Total predictions
  • True Negatives / (True Negatives + False Positives)
  • True Positives / (True Positives + False Negatives)
  • True Positives / (True Positives + False Positives)
Recall is the ratio of true positive predictions to the sum of true positives and false negatives. It measures the ability to correctly identify all relevant instances and is crucial when the cost of false negatives is high, such as in medical diagnoses.

What could be the possible consequence of choosing a very small value of K in the KNN algorithm?

  • Increased efficiency
  • Overfitting
  • Reduced complexity
  • Underfitting
Choosing a very small value of K in the KNN algorithm can lead to overfitting, where the model becomes too sensitive to noise in the training data.

What is the mathematical relationship between Eigenvalues and Eigenvectors in PCA?

  • Eigenvalues are scalar multiples of eigenvectors
  • They are inversely related
  • They are the same
  • They are unrelated
In PCA, eigenvalues and eigenvectors have a mathematical relationship where the eigenvalues are scalar multiples of the eigenvectors. They form the eigenvalue-eigenvector equation for the covariance matrix.

Explain the role of eigenvalues and eigenvectors in PCA.

  • Eigenvalues represent direction, eigenvectors variance
  • Eigenvalues represent variance, eigenvectors direction
  • Neither plays a role in PCA
  • They are used in LDA, not PCA
In PCA, eigenvectors represent the directions in which the data varies the most, while the corresponding eigenvalues give the amount of variance in those directions. These are obtained from the covariance matrix of the original data, and the eigenvectors with the largest eigenvalues become the principal components that capture the most significant patterns in the data.

In LDA, what is meant by the term "between-class variance"?

  • Variance among different classes
  • Variance among similar classes
  • Variance between individual data points
  • Variance within individual classes
"Between-class variance" in LDA refers to the "variance among different classes." It quantifies how separated the means of different classes are from each other. Maximizing this variance enhances class separation.

A colleague has built a Polynomial Regression model and suspects overfitting. What diagnostic tools and techniques would you recommend to confirm or deny this suspicion?

  • Cross-validation and visual inspection of residuals
  • Ignore the suspicion
  • Increase polynomial degree
  • Look at training data only
Cross-validation and visual inspection of residuals are common techniques to detect overfitting. They can help in assessing how well the model generalizes to new data, revealing any overfitting issues.