Which machine learning models are more susceptible to the issue of feature redundancy?

  • All of the above
  • Decision Trees
  • Linear Models
  • Neural Networks
Linear models are more susceptible to the issue of feature redundancy as they assume independence among features. Redundant features violate this assumption and can cause problems.

What method is typically used to handle missing categorical data by filling missing values?

  • Listwise Deletion
  • Mean Imputation
  • Median Imputation
  • Mode Imputation
'Mode Imputation' is a method that is typically used to handle missing categorical data. It fills missing values with the mode (most common value) of the available data. While it's a simple and fast method, it could introduce bias if the data is not missing completely at random.

The parameters of a Uniform Distribution are typically defined as _____ and _____, representing the minimum and maximum values respectively.

  • a, b
  • mean, standard deviation
  • median, mode
  • nan
The parameters of a Uniform Distribution are typically defined as 'a' and 'b', representing the minimum and maximum values respectively.

You are given a task to represent a complex data set in a visually appealing and easy to understand format for non-technical stakeholders. How would you approach this task?

  • Avoid using colors to keep the graph simple
  • Only use bar graphs, because everyone understands them
  • Use complex graph types to show your technical expertise
  • Use simple and familiar graph types, clear labels, and a thoughtful color scheme
When presenting complex data to non-technical stakeholders, it's best to use simple and familiar graph types, use clear labels and legends, and choose a thoughtful color scheme. Avoid using complex graphs that the audience might not understand, and remember that your goal is to make the data understandable, not to show off your technical skills.

What is the term used to describe the measure of the center of a distribution of data?

  • Central Distribution
  • Central Pattern
  • Central Position
  • Central Tendency
The term "Central Tendency" is used to describe the measure of the center of a distribution of data. Measures of central tendency aim to describe the central position of a distribution for a data set where the most typical values lie.

You're working on a medical diagnosis problem where interpretability is crucial. How might you approach feature selection?

  • By selecting random features
  • By selecting the features that contribute most to the model's performance
  • By using a black-box model
  • By using all features
When interpretability is crucial, such as in a medical diagnosis problem, it's important to select the features that contribute most to the model's performance. This can make the model more understandable and transparent, which is important in fields where decisions have significant impacts and need to be explained.

Which scaling technique is most affected by the presence of outliers?

  • Min-Max scaling
  • Robust scaling
  • Standardization
  • nan
The Min-Max scaling technique, which scales the data to a fixed range (usually 0 to 1), is highly sensitive to the presence of outliers. It shrinks the range of the feature values, so the outliers can drastically change the ranges of the attributes.

The process of converting an actual range of values in a numeric feature column into a standard range of values is known as _____.

  • Binning
  • Data Encoding
  • Data Integration
  • Data Scaling
The process of converting an actual range of values in a numeric feature column into a standard range of values is known as Data Scaling. This is a fundamental step in data preprocessing, particularly important when dealing with machine learning algorithms.

You're working on a high-dimensional dataset with many redundant features. Which feature selection methods might help reduce the dimensionality while maintaining the essential information?

  • Embedded methods
  • Filter methods
  • Principal Component Analysis (PCA)
  • Wrapper methods
Principal Component Analysis (PCA) is a dimensionality reduction technique that can be used when dealing with high-dimensional datasets with many redundant features. PCA transforms the original features into a new set of uncorrelated features, capturing the most variance in the data, thus helping to maintain the essential information while reducing the dimensionality.

When a dataset is normally distributed, the mean, median, and mode will all be _____.

  • Different
  • The same
  • Undefined
  • Zero
In a normal distribution, the "Mean", "Median", and "Mode" are all the "Same", falling at the center of the distribution.