A. ReLU is a linear function while sigmoid is non-linear.
B. ReLU is less computationally efficient than sigmoid, but it is more accurate than sigmoid.
C. ReLU and sigmoid both have a range of 0 to 1.
D. ReLU is more computationally efficient, but sigmoid is better for predicting probabilities.
Explanation:
ReLU (Rectified Linear Unit) and sigmoid are activation functions used in neural networks. According to NVIDIA’s deep learning documentation (e.g., cuDNN and TensorRT), ReLU, defined as f(x) = max(0, x), is computationally efficient because it involves simple thresholding, avoiding expensive exponential calculations required by sigmoid, f(x) = 1/(1 + e^(-x)). Sigmoid outputs values in the range [0, 1], making it suitable for predicting probabilities in binary classification tasks. ReLU, with an unbounded positive range, is less suited for direct probability prediction but accelerates training by mitigating vanishing gradient issues.
Option A is incorrect, as ReLU is non-linear (piecewise linear).
Option B is false, as ReLU is more efficient and not inherently more accurate.
Option C is wrong, as ReLU’s range is
[0, ∞), not
[0, 1].
Reference: NVIDIA cuDNN Documentation: https://docs.nvidia.com/deeplearning/cudnn/developer-guide/index.html
Goodfellow, I., et al. (2016). "Deep Learning." MIT Press.