How do you decide on the number of Principal Components to retain during PCA?

  • All of the above
  • By calculating the cumulative explained variance
  • By checking the eigenvalues
  • By using the elbow method
The number of principal components to retain can be decided in several ways: checking the eigenvalues (typically, components with eigenvalues greater than 1 are retained), using the elbow method (looking for a clear "elbow" in the scree plot), or calculating the cumulative explained variance (often, enough components to explain at least 95% of the variance are retained).

How does the choice of significance level affect the probability of making a Type I error?

  • Higher significance level leads to higher probability of Type I error
  • Lower significance level leads to higher probability of Type I error
  • Significance level has no effect on the probability of Type I error
  • The choice of significance level affects the probability of Type II error, not Type I
The significance level (alpha) is the probability of making a Type I error. So, a higher significance level increases the chance of rejecting the null hypothesis when it's true, hence increasing the probability of a Type I error.

How does the Spearman rank correlation deal with categorical variables?

  • It assigns a numerical value to each category
  • It can't handle categorical variables
  • It groups categorical variables together
  • It transforms categorical variables into ranks
The Spearman rank correlation transforms categorical variables into ranks, which allows it to handle both continuous and ordinal (a type of categorical variable) data.

How does independence between events affect the calculation of their joint probability?

  • It makes the joint probability equal to the difference of the probabilities of each event
  • It makes the joint probability equal to the product of the probabilities of each event
  • It makes the joint probability equal to the ratio of the probabilities of each event
  • It makes the joint probability equal to the sum of the probabilities of each event
If events are independent, their joint probability equals the product of their individual probabilities. That is, P(A ∩ B) = P(A) * P(B) for independent events A and B.

What type of statistical test is the Kruskal-Wallis Test?

  • Chi-square test
  • Non-parametric
  • Parametric
  • T-test
The Kruskal-Wallis Test is a non-parametric statistical test.

The ________ is the average of a data set calculated by adding all values and then dividing by the number of values.

  • Mean
  • Median
  • Mode
  • nan
The mean, also referred to as average or arithmetic mean, is calculated by adding all values in the data set and then dividing by the number of values. The mean is often used as a summary statistic.

The probability of committing a Type I error is also known as the ______ level of the test.

  • Confidence
  • Power
  • Significance
  • Size
The probability of committing a Type I error (rejecting a true null hypothesis) is known as the significance level (often denoted by alpha) of the test. A common significance level is 0.05, indicating a 5% risk of committing a Type I error if the null hypothesis is true.

The process of testing the effect of varying one predictor at different levels of another predictor is known as ________ effect analysis.

  • Additive
  • Independent
  • Interaction
  • Subtractive
This is known as interaction effect analysis. Interaction effect analysis involves testing how the effect of one predictor on the response variable changes at different levels of another predictor. It helps in understanding how different variables interact with each other to affect the dependent variable.

The degrees of freedom for a Chi-square test for a contingency table with r rows and c columns is (r-1)*(c-1), otherwise known as ________ degrees of freedom.

  • dependent
  • independent
  • joint
  • multicollinearity
The degrees of freedom for a Chi-square test for a contingency table with r rows and c columns is calculated as (r-1)*(c-1). These are also known as independent degrees of freedom as they depend on the number of independent ways that the data can vary.

Can Pearson's Correlation Coefficient be used with non-linear relationships?

  • No, never
  • Yes, always
  • Yes, but it may not provide meaningful results
  • Yes, but only if the relationship is monotonic
While you can technically compute a Pearson correlation coefficient for non-linear relationships, it may not provide meaningful results. The Pearson correlation measures the degree of a linear relationship between variables, and does not fully capture the dynamics of a non-linear relationship. In such cases, Spearman's rank correlation or other non-parametric correlations may be more appropriate.