- The first mention I found to ReLU is in Fukushima's paper from 1975: Cognitron: A self-organizing multilayered neural network | SpringerLink
- Nair and Hinton paper from 2010 make the case that ReLU preserves information during backpropagation Rectified Linear Units Improve Restricted Boltzmann Machines
- Why is the ReLU function not differentiable at x=0?
- Deep Learning with Python - François Chollet
Created
November 24, 2022 19:11
-
-
Save svpino/8c34ecb612f9f66c13f7542a9e5043cc to your computer and use it in GitHub Desktop.
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment