[Fundamentals of Machine Learning and Neural Networks]
When comparing and contrasting the ReLU and sigmoid activation functions, which statement is true?
ReLU (Rectified Linear Unit) and sigmoid are activation functions used in neural networks. According to NVIDIA's deep learning documentation (e.g., cuDNN and TensorRT), ReLU, defined as f(x) = max(0, x), is computationally efficient because it involves simple thresholding, avoiding expensive exponential calculations required by sigmoid, f(x) = 1/(1 + e^(-x)). Sigmoid outputs values in the range
[0, 1], making it suitable for predicting probabilities in binary classification tasks. ReLU, with an unbounded positive range, is less suited for direct probability prediction but accelerates training by mitigating vanishing gradient issues. Option A is incorrect, as ReLU is non-linear (piecewise linear). Option B is false, as ReLU is more efficient and not inherently more accurate. Option C is wrong, as ReLU's range is
[0, ), not
[0, 1].
NVIDIA cuDNN Documentation: https://docs.nvidia.com/deeplearning/cudnn/developer-guide/index.html
Goodfellow, I., et al. (2016). 'Deep Learning.' MIT Press.
Dana
5 days agoRyan
10 days ago