Neural Network Activation Basics Quiz

  • 12th Grade
Reviewed by Editorial Team
The ProProfs editorial team is comprised of experienced subject matter experts. They've collectively created over 10,000 quizzes and lessons, serving over 100 million users. Our team includes in-house content moderators and subject matter experts, as well as a global network of rigorously trained contributors. All adhere to our comprehensive editorial guidelines, ensuring the delivery of high-quality content.
Learn about Our Editorial Process
| By ProProfs AI
P
ProProfs AI
Community Contributor
Quizzes Created: 81 | Total Attempts: 817
| Questions: 15 | Updated: May 1, 2026
Please wait...
Question 1 / 16
🏆 Rank #--
0 %
0/100
Score 0/100

1. What is the primary purpose of an activation function in a neural network?

Explanation

An activation function introduces non-linearity into a neural network, allowing it to learn complex patterns and relationships in the data. Without non-linearity, the network would behave like a linear model, limiting its capacity to capture intricate features necessary for tasks such as classification and regression.

Submit
Please wait...
About This Quiz
Neural Network Activation Basics Quiz - Quiz

This Neural Network Activation Basics Quiz tests your understanding of activation functions in neural networks. You'll explore how different activation functions like ReLU, sigmoid, and tanh shape network behavior, their advantages and limitations, and their applications in modern deep learning. Perfect for Grade 12 students learning foundational machine learning concepts.

2.

What first name or nickname would you like us to use?

You may optionally provide this to label your report, leaderboard, or certificate.

2. Which activation function outputs values between 0 and 1?

Explanation

The Sigmoid activation function transforms its input into a value between 0 and 1 using the formula \( \sigma(x) = \frac{1}{1 + e^{-x}} \). This characteristic makes it particularly useful for binary classification tasks, as it can represent probabilities. In contrast, other functions like ReLU and Tanh have different output ranges.

Submit

3. What does ReLU stand for?

Explanation

ReLU, or Rectified Linear Unit, is an activation function used in neural networks. It outputs the input directly if it is positive; otherwise, it outputs zero. This characteristic helps in addressing the vanishing gradient problem and allows models to learn complex patterns effectively, making it a popular choice in deep learning architectures.

Submit

4. The ReLU function is defined as f(x) = max(0, x). What is the output when x = -5?

Explanation

The ReLU (Rectified Linear Unit) function outputs the maximum value between 0 and the input x. When x is -5, the function evaluates to max(0, -5), which is 0, as it disregards negative values and returns zero instead.

Submit

5. Which activation function is commonly used in the output layer for binary classification?

Explanation

The sigmoid activation function is commonly used in the output layer for binary classification because it maps input values to a range between 0 and 1. This allows the model to predict probabilities for the two classes, making it suitable for distinguishing between binary outcomes.

Submit

6. The tanh function outputs values in the range ____.

Explanation

The tanh function, or hyperbolic tangent function, is defined as the ratio of the hyperbolic sine and cosine functions. Its output values range from -1 to 1, making it useful in various applications, particularly in neural networks, where it helps in normalizing outputs and controlling gradients during training.

Submit

7. True or False: The sigmoid function has a constant derivative across all input values.

Explanation

The sigmoid function does not have a constant derivative; instead, its derivative varies depending on the input value. The derivative is highest at the center of the sigmoid curve and approaches zero as the input moves towards the extremes, indicating that the rate of change is not uniform across all input values.

Submit

8. Which of the following is a disadvantage of the sigmoid activation function?

Explanation

The sigmoid activation function can lead to vanishing gradient problems because its output saturates at extreme values (close to 0 or 1). This saturation results in very small gradients during backpropagation, hindering the learning process and making it difficult for deep networks to update weights effectively.

Submit

9. The softmax activation function is typically used for ____.

Explanation

The softmax activation function is designed to convert raw output scores from a model into probabilities that sum to one. This makes it ideal for multiclass classification tasks, where each class's probability is calculated, allowing the model to predict the most likely class among multiple options based on the highest probability.

Submit

10. Which activation function is most commonly used in hidden layers of modern deep networks?

Explanation

ReLU, or Rectified Linear Unit, is preferred in hidden layers of deep networks due to its ability to mitigate the vanishing gradient problem, allowing for faster convergence during training. It introduces non-linearity while maintaining computational efficiency, enabling models to learn complex patterns effectively. Its simplicity and performance make it a popular choice among practitioners.

Submit

11. True or False: The linear activation function introduces non-linearity to the network.

Explanation

A linear activation function does not introduce non-linearity to a neural network because it outputs a linear combination of inputs. As a result, stacking multiple layers with linear activation functions will still produce a linear transformation, meaning the overall function remains linear. Non-linearity is introduced through non-linear activation functions like ReLU or sigmoid.

Submit

12. What is the main advantage of ReLU over sigmoid for training deep networks?

Explanation

ReLU (Rectified Linear Unit) has a simpler derivative, which allows for faster computation during backpropagation. Additionally, it mitigates the vanishing gradient problem common in sigmoid activation functions, enabling deeper networks to learn more effectively by maintaining gradient flow, thus improving overall training efficiency and performance.

Submit

13. The derivative of the sigmoid function σ(x) is σ(x)(1 - σ(x)). This means the gradient is largest when σ(x) = ____.

Submit

14. Which activation function can output negative values?

Submit

15. True or False: All activation functions must be differentiable for backpropagation.

Submit
×
Saved
Thank you for your feedback!
View My Results
Cancel
  • All
    All (15)
  • Unanswered
    Unanswered ()
  • Answered
    Answered ()
What is the primary purpose of an activation function in a neural...
Which activation function outputs values between 0 and 1?
What does ReLU stand for?
The ReLU function is defined as f(x) = max(0, x). What is the output...
Which activation function is commonly used in the output layer for...
The tanh function outputs values in the range ____.
True or False: The sigmoid function has a constant derivative across...
Which of the following is a disadvantage of the sigmoid activation...
The softmax activation function is typically used for ____.
Which activation function is most commonly used in hidden layers of...
True or False: The linear activation function introduces non-linearity...
What is the main advantage of ReLU over sigmoid for training deep...
The derivative of the sigmoid function σ(x) is σ(x)(1 - σ(x)). This...
Which activation function can output negative values?
True or False: All activation functions must be differentiable for...
play-Mute sad happy unanswered_answer up-hover down-hover success oval cancel Check box square blue
Alert!