Answer Posted / Vipin Kumar Gangwar
The Swish activation function, also known as SiLU (scaled linear unit), is a smooth and differentiable alternative to ReLU. It's defined as x * sigmoid(beta * x) where beta is a learned parameter.
| Is This Answer Correct ? | 0 Yes | 0 No |
Post New Answer View All Answers
No New Questions to Answer in this Category !! You can
Post New Questions
Answer Questions in Different Category