Which process involves transforming and creating new variables to improve a machine learning model's predictive performance?
- Data preprocessing
- Feature engineering
- Hyperparameter tuning
- Model training
Feature engineering is the process of transforming and creating new variables based on the existing data to enhance a model's predictive performance. This can involve scaling, encoding, or creating new features from existing ones.
A researcher is working on a medical imaging problem with a limited amount of labeled data. To improve the performance of the deep learning model, the researcher decides to use a model pre-trained on a large generic image dataset. This approach is an example of what?
- Transfer Learning
- Reinforcement Learning
- Ensemble Learning
- Supervised Learning
Transfer learning is the practice of using a pre-trained model as a starting point to solve a new problem. In this case, it leverages prior knowledge from generic images to enhance medical image analysis.
What is the primary benefit of using transfer learning in deep learning models?
- Improved training time
- Better performance
- Reduced data requirement
- Enhanced model complexity
The primary benefit of transfer learning in deep learning is 'Better performance.' This technique leverages knowledge from pre-trained models, allowing the model to perform well even with limited data and reducing the need for lengthy training.
Which type of neural network is specifically designed to handle image data?
- Convolutional Neural Network
- Recurrent Neural Network
- Feedforward Network
- Decision Tree
Convolutional Neural Networks (CNNs) are tailored for image data processing, thanks to their ability to capture spatial patterns and features.
One common regularization technique involves adding a penalty to the loss function based on the magnitude of the coefficients, known as ________ regularization.
- L1 (Lasso)
- L2 (Ridge)
- Elastic Net
- Mean Squared Error
L2 (Ridge) regularization adds a penalty based on the sum of squared coefficients, helping to control the model's complexity and reduce overfitting.
Support Vector Machines (SVM) aim to find a ______ that best divides a dataset into classes.
- Cluster
- Decision Boundary
- Hyperplane
- Mean
Support Vector Machines aim to find a hyperplane that best divides a dataset into classes. This hyperplane maximizes the margin between the classes, making it a powerful tool for binary classification tasks. The concept of the "support vector" is crucial in SVM.
A bioinformatics researcher is trying to visualize the similarities and differences between different genes in a 2D space. The data is high dimensional. Which technique would provide a good visualization emphasizing local similarities?
- t-Distributed Stochastic Neighbor Embedding (t-SNE)
- Principal Component Analysis
- Linear Regression
- A* Search Algorithm
t-SNE is well-suited for visualizing high-dimensional data by preserving local similarities. It maps data points to a 2D space in a way that emphasizes neighborhood relationships, making it ideal for visualizing gene similarities in high-dimensional data.
What is the primary goal of exploration in reinforcement learning?
- To gather information about the environment
- To maximize immediate rewards
- To stick with known actions
- To build a policy
Exploration's primary goal is to gather information about the environment, helping an RL agent learn and make better decisions in the long run.
GRUs are often considered a middle ground between basic RNNs and ________ in terms of complexity and performance.
- LSTMs
- CNNs
- Autoencoders
- K-Means Clustering
GRUs (Gated Recurrent Units) are a compromise between basic RNNs and LSTMs, offering a balance between the complexity and performance of these two types of recurrent networks.
If a model has low bias and high variance, it is likely that the model is ________.
- Optimally Fitted
- Overfitting
- Underfitting
- Well-fitted
A model with low bias and high variance is likely overfitting. Low bias means the model fits the training data very well (potentially too well), and high variance indicates that it's very sensitive to fluctuations in the data, which can lead to poor generalization. Overfitting is a common outcome of this scenario.