-
-
Notifications
You must be signed in to change notification settings - Fork 46.6k
Added A General Swish Activation Function inNeural Networks #10415
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Added A General Swish Activation Function inNeural Networks #10415
Conversation
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Thanks for your contribution, but could you modify neural_network/activation_functions/sigmoid_linear_unit.py rather than making a separate file for the general swish function? Since the SiLU function is just the swish function with a parameter of 1, I think it'd make much more sense if you could rename that file to swish.py
and generalize the existing function.
@tianyizheng02 Sir, I added the general swish model in the Sigmoid Linear Unit Model and renamed it as 'swish.py'. Now in order to maintain learning through these algorithms, I didn't find it appealing enough to delete the existing SiLU function to let others know that it is a popular name for the swish function of parameter 1. Hence I just added both functions. Could you please look into this and merge my PR? |
Thanks for your contribution! |
…ithms#10415) * Added A General Swish Activation Function inNeural Networks * Added the general swish function in the SiLU function and renamed it as swish.py * [pre-commit.ci] auto fixes from pre-commit.com hooks for more information, see https://pre-commit.ci --------- Co-authored-by: Shivansh Bhatnagar <[email protected]> Co-authored-by: pre-commit-ci[bot] <66853113+pre-commit-ci[bot]@users.noreply.github.com>
Describe your change:
Checklist: