Why is residual analysis important in regression models?

  • To check the assumptions of the regression model
  • To determine the slope of the regression line
  • To estimate the parameters of the model
  • To predict the dependent variable
Residual analysis is important because it helps us to validate the assumptions of the regression model, such as linearity, independence, normality, and equal variance (homoscedasticity). This is crucial for the reliability and validity of the regression model.

What is the significance of the total probability rule?

  • It is a rule for determining the probability of dependent events
  • It is used to calculate conditional probabilities
  • It is used to calculate the probability of mutually exclusive events
  • It provides a way to break down probabilities of complex events into simpler ones
The Total Probability Rule provides a way to compute the probability of an event from the probabilities of that event occurring within disjoint subsets of the sample space. It essentially allows you to break down the probability of complex events into simpler or more basic component events.

What is multicollinearity and how does it affect simple linear regression?

  • It is the correlation between dependent variables and it has no effect on regression
  • It is the correlation between errors and it makes the regression model more accurate
  • It is the correlation between independent variables and it can cause instability in the regression coefficients
  • It is the correlation between residuals and it causes bias in the regression coefficients
Multicollinearity refers to a high correlation among independent variables in a regression model. It does not reduce the predictive power or reliability of the model as a whole, but it can cause instability in the estimation of individual regression coefficients, making them difficult to interpret.

The distribution of all possible sample means is known as a __________.

  • Normal Distribution
  • Population Distribution
  • Sampling Distribution
  • Uniform Distribution
The sampling distribution in statistics is the probability distribution of a given statistic based on a random sample. For a statistic that is calculated from a sample, each different sample could (and likely will) provide a different value of that statistic. The sampling distribution shows us how those calculated statistics would be distributed.

How is 'K-means' clustering different from 'hierarchical' clustering?

  • Hierarchical clustering creates a hierarchy of clusters, while K-means does not
  • Hierarchical clustering uses centroids, while K-means does not
  • K-means requires the number of clusters to be defined beforehand, while hierarchical clustering does not
  • K-means uses a distance metric to group instances, while hierarchical clustering does not
K-means clustering requires the number of clusters to be defined beforehand, while hierarchical clustering does not. Hierarchical clustering forms a dendrogram from which the user can choose the number of clusters based on the problem requirements.

Under what conditions does a binomial distribution approximate a normal distribution?

  • When the events are not independent
  • When the number of trials is large and the probability of success is not too close to 0 or 1
  • When the number of trials is small
  • When the probability of success changes with each trial
The binomial distribution approaches the normal distribution as the number of trials gets large, provided that the probability of success is not too close to 0 or 1. This is known as the De Moivre–Laplace theorem.

What does the F-statistic signify in an ANOVA test?

  • The ratio of between-group variability to within-group variability
  • The ratio of total variability to within-group variability
  • The ratio of within-group variability to between-group variability
  • The ratio of within-group variability to total variability
In an ANOVA test, the F-statistic is the ratio of the between-group variability to the within-group variability. In other words, it measures how much the means of each group vary between the groups, compared to how much they vary within each group. A larger F-statistic implies a greater degree of difference between the group means.

What assumption about the residuals of a linear regression model does homoscedasticity refer to?

  • The residuals are independent
  • The residuals are normally distributed
  • The residuals have a linear relationship with the dependent variable
  • The residuals have constant variance
Homoscedasticity refers to the assumption that the residuals (errors) have constant variance at each level of the independent variable(s). This is important for the reliability of the regression model.

How does stratified random sampling differ from simple random sampling?

  • Stratified random sampling always involves larger sample sizes than simple random sampling
  • Stratified random sampling involves dividing the population into subgroups and selecting individuals from each subgroup
  • Stratified random sampling is the same as simple random sampling
  • Stratified random sampling only selects individuals from a single subgroup
Stratified random sampling differs from simple random sampling in that it first divides the population into non-overlapping groups, or strata, based on specific characteristics, and then selects a simple random sample from each stratum. This can ensure that each subgroup is adequately represented in the sample, which can increase the precision of estimates.

Why are bar plots commonly used in data analysis?

  • To compare the frequency of categorical variables
  • To show the change of a variable over time
  • To show the distribution of a single variable
  • To show the relationship between two continuous variables
Bar plots are commonly used in data analysis to compare the frequency, count, or proportion of categorical variables. Each category is represented by a separate bar, and the length or height of the bar represents its corresponding value.