Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

GH-512: Minor embedding fixes #520

Merged
merged 4 commits into from
Feb 19, 2019
Merged

GH-512: Minor embedding fixes #520

merged 4 commits into from
Feb 19, 2019

Conversation

alanakbik
Copy link
Collaborator

@alanakbik alanakbik commented Feb 19, 2019

This PR fixes a few minor issues in preparation for 0.4.1 release.

  • DocumentLSTMEmbeddings are marked as deprecated but not removed so that serialized models still work. The deprecation warning points to the new DocumentRNNEmbeddings class as class to be used instead.

  • BytePairEmbeddings threw an error for empty words. Now, a null vector is used for empty words instead.

  • the load_text_classification_corpus() methods of the NLPTaskDataFetcher previously tokenized all corpora by default. But some text classification corpora, such as TREC_6 are already tokenized. So we added an option to turn tokenization off.

  • added Turian embeddings which are very small and will hopefully speed up integration tests. Load with WordEmbeddings('turian')

@alanakbik alanakbik merged commit d2682f8 into release-0.4.1 Feb 19, 2019
@alanakbik alanakbik deleted the GH-512-embeddings branch February 19, 2019 15:06
@alanakbik alanakbik restored the GH-512-embeddings branch February 19, 2019 15:08
@alanakbik alanakbik deleted the GH-512-embeddings branch February 19, 2019 18:47
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

1 participant