In a task involving the classification of hand-written digits, the model is failing to capture intricate patterns in the data. Adding more layers seems to exacerbate the problem due to a certain issue in training deep networks. What is this issue likely called?

  • Overfitting
  • Vanishing Gradient
  • Underfitting
  • Exploding Gradient
The issue where adding more layers to a deep neural network exacerbates the training problem due to diminishing gradients is called "Vanishing Gradient." It occurs when gradients become too small during backpropagation, making it challenging for deep networks to learn intricate patterns in the data.

A common method to combat the vanishing gradient problem in RNNs is to use _______.

  • Long Short-Term Memory (LSTM)
  • Decision Trees
  • K-Means Clustering
  • Principal Component Analysis
To address the vanishing gradient problem in RNNs, one common technique is to use Long Short-Term Memory (LSTM) networks. LSTMs are a type of RNN that helps mitigate the vanishing gradient problem by preserving and updating information over long sequences. LSTMs are designed to capture long-term dependencies and are more effective than traditional RNNs for tasks where data from distant time steps is important.

You are responsible for ensuring that the data in your company's data warehouse is consistent, reliable, and easily accessible. Recently, there have been complaints about data discrepancies. Which stage in the ETL process should you primarily focus on to resolve these issues?

  • Extraction
  • Transformation
  • Loading
  • Data Ingestion
The Transformation stage is where data discrepancies are often addressed. During transformation, data is cleaned, normalized, and validated to ensure consistency and reliability. This stage is critical for data quality and consistency in the data warehouse. Extraction involves collecting data, Loading is about data loading into the warehouse, and Data Ingestion is the process of bringing data into the system.

In the context of AI ethics, what is the primary concern of "interpretability"?

  • Ensuring AI is always right
  • Making AI faster
  • Understanding how AI makes decisions
  • Controlling the cost of AI deployment
"Interpretability" in AI ethics is about understanding how AI systems make decisions. It's crucial for accountability, transparency, and identifying and addressing potential biases in AI algorithms. AI being right or fast is important but not the primary concern in this context.

For models with a large number of layers, which technique helps in improving the internal covariate shift and accelerates the training?

  • Stochastic Gradient Descent (SGD) with a small learning rate
  • Batch Normalization
  • L1 Regularization
  • DropConnect
Batch Normalization is a technique used to improve the training of deep neural networks. It addresses the internal covariate shift problem by normalizing the activations of each layer. This helps in accelerating training and allows for the use of higher learning rates without the risk of divergence. It also aids in better gradient flow.

What is the primary goal of tokenization in NLP?

  • Removing stop words
  • Splitting text into words
  • Extracting named entities
  • Translating text to other languages
The primary goal of tokenization in NLP is to split text into words or tokens. This process is essential for various NLP tasks such as text analysis, language modeling, and information retrieval. Tokenization helps in breaking down text into meaningful units for analysis.

For datasets with multiple features, EDA often involves dimensionality reduction techniques like PCA to visualize data in two or three _______.

  • Planes
  • Points
  • Dimensions
  • Directions
Exploratory Data Analysis (EDA) often employs dimensionality reduction techniques like Principal Component Analysis (PCA) to visualize data in lower-dimensional spaces (2 or 3 dimensions) for better understanding, hence the term "dimensions."

When a dataset has values ranging from 0 to 1000 in one column and 0 to 1 in another column, which transformation can be used to scale them to a similar range?

  • Normalization
  • Log Transformation
  • Standardization
  • Min-Max Scaling
Min-Max Scaling, also known as feature scaling, is used to transform values within a specific range (typically 0 to 1) for different features. It ensures that variables with different scales have a similar impact on the analysis.

In a relational database, what is used to ensure data integrity across multiple tables?

  • Primary Key
  • Foreign Key
  • Index
  • Trigger
A Foreign Key is used in a relational database to ensure data integrity by creating a link between tables. It enforces referential integrity, ensuring that values in one table match values in another. Primary Keys are used to uniquely identify records in a table, not to maintain integrity across tables. Indexes and Triggers serve different purposes.

A company uses an AI model for recruitment, and it's observed that the model is selecting more male candidates than female candidates for a tech role, even when both genders have similar qualifications. What ethical concern does this scenario highlight?

  • Data bias in AI
  • Lack of transparency in AI
  • Data security and privacy issues in AI
  • Ethical AI governance and accountability
This scenario highlights the ethical concern of "Data bias in AI." The AI model's biased selection towards male candidates indicates that the training data may be biased, leading to unfair and discriminatory outcomes. Addressing data bias is essential to ensure fairness and diversity in AI-driven recruitment.

Which type of database is ideal for handling hierarchical data and provides better scalability, MongoDB or MySQL?

  • MongoDB
  • MySQL
  • Both MongoDB and MySQL
  • Neither MongoDB nor MySQL
MongoDB is a NoSQL database that is ideal for handling hierarchical data and provides better scalability for unstructured data. MongoDB uses BSON (Binary JSON) format, which makes it a good choice for applications that require flexibility and scalability in dealing with complex data structures.

When a model performs well on training data but poorly on unseen data, what issue might it be facing?

  • Overfitting
  • Underfitting
  • Data leakage
  • Bias-variance tradeoff
The model is likely facing the issue of overfitting. Overfitting occurs when the model learns the training data too well, including noise, resulting in excellent performance on the training set but poor generalization to unseen data. It's an example of a high-variance problem in the bias-variance tradeoff. To address overfitting, techniques like regularization and more data are often used.