sin(x)

1 Post

Graphs, images and data related to the activation function known as ReLU
sin(x)

Upgrade for ReLU

The activation function known as ReLU builds complex nonlinear functions across layers of a neural network, making functions that outline flat faces and sharp edges. But how much of the world breaks down into perfect polyhedra?
2 min read

Subscribe to The Batch

Stay updated with weekly AI News and Insights delivered to your inbox