The ___________ regression technique can be used when the relationship between the independent and dependent variables is not linear.
- L1 Regularization
- Logistic
- Polynomial
- Simple Linear
Polynomial Regression can model non-linear relationships between independent and dependent variables by transforming the predictors into a polynomial form, allowing for more complex fits.
How does adding regularization help in avoiding overfitting?
- By adding noise to the training data
- By fitting the model closely to the training data
- By increasing model complexity
- By reducing model complexity
Regularization helps in avoiding overfitting by "reducing model complexity." It adds a penalty to the loss function, constraining the weights and preventing the model from fitting too closely to the training data.
How can you tune hyperparameters in SVM to prevent overfitting?
- Changing the color of hyperplane
- Increasing data size
- Reducing feature dimensions
- Using appropriate kernel and regularization
Tuning hyperparameters like the choice of kernel and regularization helps in controlling model complexity to prevent overfitting in SVM.
How can you evaluate the performance of an LDA model?
- By checking the size of the scatter matrices
- By comparing with PCA
- Using confusion matrix and ROC curves
- Using only accuracy
The performance of an LDA model can be evaluated using metrics like the "confusion matrix and ROC curves." These tools provide insights into the model's ability to classify instances correctly and its trade-off between sensitivity and specificity.
You applied DBSCAN and found that many points are being classified as noise. What adjustments could you make to the parameters?
- Decrease Epsilon; Increase MinPts
- Increase Epsilon; Decrease MinPts
- Increase both Epsilon and MinPts
- Use the same Epsilon and MinPts but change the clustering method
Increasing Epsilon and decreasing MinPts will make the clustering less strict, reducing the chance of points being classified as noise. Epsilon defines the neighborhood size, and MinPts defines the minimum points required to form a cluster. By adjusting them, more points can be included in clusters, reducing noise classification.
In K-Means clustering, a common approach to avoid local minima due to initial centroid selection is to run the algorithm multiple times with different _________.
- Centroid initializations
- Distance metrics
- Learning rates
- Number of clusters
Running the K-Means algorithm multiple times with different centroid initializations helps in avoiding local minima. It increases the chance of finding a more globally optimal clustering solution.
What is the role of cross-validation in detecting and preventing overfitting in Polynomial Regression?
- It assists in increasing model complexity
- It focuses on training data only
- It helps in choosing the right degree and assessing generalization
- It helps in selecting features
Cross-validation plays a key role in detecting and preventing overfitting in Polynomial Regression by helping in choosing the right degree for the polynomial and assessing how well the model generalizes to new data.
How can the 'k-distance graph' be used in selecting the optimal Epsilon for DBSCAN?
- By calculating the average distance to k-nearest neighbors
- By determining the distance between k centroids
- By displaying k clusters' distances
- By plotting the distance to the kth nearest neighbor of each point
The 'k-distance graph' can be used to select the optimal Epsilon by plotting the distance to the kth nearest neighbor for each point and looking for an "elbow" or a point of inflection. This inflection point can be a good estimate for Epsilon, helping to choose a value that balances density requirements without overly segmenting the data.
In the context of DBSCAN, if two core points are within the Epsilon distance of each other, they are said to be __________.
- directly dense reachable
- indirectly dense reachable
- separate
- unrelated
In DBSCAN, if two core points are within the Epsilon distance of each other, they are said to be directly dense reachable. This concept is used to establish connections between core points and helps in forming clusters by linking together points that are close to one another.
You have applied PCA to your dataset and the first three principal components explain 95% of the variance. What does this signify, and how would you proceed?
- This indicates an error in the PCA process
- This means that 5% of the data is missing, so you should reapply PCA
- This means that 95% of the variance is captured, so you may choose to proceed with these components
- This means that the data is uniformly distributed and PCA is not needed
The first three principal components explaining 95% of the variance means that most of the original information is captured, and you may proceed with these components if the loss of 5% is acceptable.