-
Notifications
You must be signed in to change notification settings - Fork 19.6k
Feature Request : Add GELU activation function #11834
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Comments
I don't think this should be merged into Keras.
Please submit your PR at keras-contrib. |
This guy uses it, and he clearly knows whats going on..
somethings wrong with that custom loss.. getting really strange predictions not going under ~-.25 |
GELU activation has started to pick up and it has been published a while ago (2016): Also been used in OpenAI's GPT-1 and 2 and Google's BERT papers. Would love to see this implemented in Keras activations. |
Code from Google's BERT:
Code from OpenAI's GPT-2:
|
it's not wrong that you are not getting below -0.25, look at the graph for the function: |
I know that It start to be very confusing but I need to make a cross Org reference tensorflow/tensorflow#33945 |
Gelu is in tensorflow tensorflow/tensorflow#41178. You can close this. |
Thank you, @bhack! I will close this issue :) |
I just realized that keras does not have a GELU activation function in activations.py. I request that it be added, because it has many applications in neural networks.
Note : I'll probably submit a pull request for it.
[y] Check that you are up-to-date with the master branch of Keras. You can update with:
pip install git+git://github.com/keras-team/keras.git --upgrade --no-deps
[y] Check that your version of TensorFlow is up-to-date. The installation instructions can be found here.
The text was updated successfully, but these errors were encountered: