Zettelkasten/Terminology Information

ReLU (Rectified Linear Unit)

Computer-Nerd 2023. 4. 19.

Information

  • ReLU (Rectified Linear Unit) is a type of activation function used in neural networks, which is defined as f(x) = max(0,x).
  • It returns 0 for all negative inputs and returns the input as it is for all positive inputs.
  • ReLU is a popular choice for activation function due to its simplicity and effectiveness in reducing vanishing gradients in deep neural networks.
  • Vanishing gradients occur when the gradient becomes very small, making it difficult to update the weights and causing the network to stop learning.
  • ReLU can help prevent vanishing gradients by ensuring that gradients can flow through the network, as it is non-linear and has a constant gradient for positive inputs.
  • ReLU is computationally efficient and easy to implement, making it a popular choice for many machine learning applications.

댓글