This codebase contains a set of simple postprocessing transformations that improve the performance of word embeddings. Prior work has shown that mean subtraction and removal of early principal components can enhance performance on lexical similarity tasks. We further demonstrate that, simply by performing these transformations only on a strategic subset of the vocabulary, we can consistently achieve even further gains (up to 20% overall), while consuming less compute and memory resources. Not only does this behavior offer insights into the linguistic properties of these word representations, but the gains are considerable and hold on both static word embeddings (word2vec and GloVe) and contextual word embeddings (BERT and GPT-2) across a broad range of lexical similarity tasks.
This repository has been archived by the owner on Sep 28, 2023. It is now read-only.
-
Notifications
You must be signed in to change notification settings - Fork 0
License
google/one-weird-trick
About
No description, website, or topics provided.
Resources
License
Code of conduct
Security policy
Stars
Watchers
Forks
Releases
No releases published
Packages 0
No packages published