Quantcast
Viewing all articles
Browse latest Browse all 4448

Why is ReLU a good activation function?

@ankit2106 wrote:

I understand that ReLU helps avoid dead neurons during backpropagation.

However, I am not able to understand why is ReLU used as an activation function if its output is linear?

Doesn’t the whole point of being the activation function get defeated if it won’t introduce non-linearity? Can someone explain mathematically why this is a good choice?

Posts: 1

Participants: 1

Read full topic


Viewing all articles
Browse latest Browse all 4448

Trending Articles