Which activation function can alleviate the vanishing gradient problem to some extent?
- Sigmoid
- ReLU (Rectified Linear Unit)
- Tanh (Hyperbolic Tangent)
- Leaky ReLU
The ReLU activation function is known for mitigating the vanishing gradient problem, which is a common issue in deep learning. ReLU allows gradients to flow more freely during backpropagation, making it easier to train deep neural networks.
Loading...
Related Quiz
- What is the primary purpose of transfer learning in the context of deep learning for computer vision?
- A common architecture for real-time data processing involves using ________ to ingest and process streaming data.
- Which ensemble method adjusts weights for misclassified instances in iterative training?
- When you want to visualize geographical data with customizable layers and styles, which tool is commonly used?
- Which layer type in a neural network is primarily responsible for feature extraction and spatial hierarchy?