Read about sense2vec here:
https://spacy.io/blog/sense2vec-with-spacy
You can use an online demo of the technology here:
We're currently refining the API, to make this technology easy to use. Once we've completed that, you'll be able to download the package on PyPi. For now, the code is available to clarify the blog post.
There are three relevant files in this repository:
This script pre-processes text using spaCy, so that the sense2vec model can be trained using Gensim.
This script reads a directory of text files, and then trains a word2vec model using Gensim. The script includes its own vocabulary counting code, because Gensim's vocabulary count is a bit slow for our large, sparse vocabulary.
To serve the similarity queries, we wrote a small vector-store class in Cython. This made it easier to add an efficient cache in front of the service. It also less memory than Gensim's Word2Vec class, as it doesn't hold the keys as Python unicode strings.
Similarity queries could be faster, if we had made all vectors contiguous in memory, instead of holding them
as an array of pointers. However, we wanted to allow a .borrow()
method, so that vectors can be added to the store
by reference, without copying the data.
WARNING: sense2vec is still work-in-progress. There is no official release yet. The content of the repository might change at any time and could be in a non-working state.
Until there is a PyPI release you can install sense2vec by 1) cloning the repository 2) run pip install -r requirements.txt
3) pip install -e .
and 4) install the latest model via python -m sense2vec.download
You might also be tempted to simply run pip install -e git+git://github.com/spacy-io/sense2vec.git#egg=sense2vec
instead of steps 1-3, but it expects cython to be present.
import sense2vec
model = sense2vec.load()
freq, query_vector = model["natural_language_processing|NOUN"]
model.most_similar(query_vector, n=3)
(['natural_language_processing|NOUN', 'machine_learning|NOUN', 'computer_vision|NOUN'], <MemoryView of 'ndarray'>)
IMPORTANT The API is work-in-progress and is subject to change.
For additional performance experimental support for BLAS can be enabled by setting the USE_BLAS environment variable before installing (e.g. USE_BLAS=1 pip install ...
). This requires an up-to-date BLAS/OpenBlas/Atlas installation.
- CPython 2.6, 2.7, 3.3, 3.4, 3.5 (only 64 bit)
- OSX
- Linux
- Windows