ReLU (Rectified Linear Unit) Example
The ReLU function is defined as:
This means:
- If x is positive, it stays the same.
- If x is negative, it becomes 0.
Real Number Examples
Input (x) | ReLU Output f(x) |
---|---|
-3 | 0 |
-1 | 0 |
0 | 0 |
2 | 2 |
5 | 5 |
In this table:
- Negative numbers become 0 🚫
- Positive numbers pass through ✅
This makes ReLU very fast and useful for deep learning models! 🤖✨