Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Refactor RotaryEmbedding and GPTNeoXAttention #1101

Merged
merged 8 commits into from
Jul 6, 2023

Conversation

shivance
Copy link
Collaborator

@shivance shivance commented Jun 28, 2023

@shivance shivance changed the title Fix rotary embedding layer Fix + Refactor RotaryEmbedding and GPTNeoXAttention Jun 28, 2023
@shivance shivance requested a review from mattdangerw June 30, 2023 17:30
@shivance shivance changed the title Fix + Refactor RotaryEmbedding and GPTNeoXAttention Refactor RotaryEmbedding and GPTNeoXAttention Jun 30, 2023
Copy link
Member

@mattdangerw mattdangerw left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Nice!

@mattdangerw
Copy link
Member

/gcbrun

@shivance
Copy link
Collaborator Author

shivance commented Jul 1, 2023

Good ! All checks pass.

Copy link
Member

@mattdangerw mattdangerw left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Thanks!

@mattdangerw mattdangerw merged commit f68c256 into keras-team:master Jul 6, 2023
shivance added a commit to shivance/keras-nlp that referenced this pull request Jul 6, 2023
* fix rotary emb

* refactor + remove unnecessary typecast

* fix formatting

* refactor

* formatting fix

* refactoring rotary emb

* added a kwarg in super().__init__()
@shivance shivance deleted the fix-attention branch July 13, 2023 02:33
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants