Load pretrained word embeddings (word2vec, glove format) into torch.FloatTensor for PyTorch
PyTorch required.
pip install torchwordemb
import torch
import torchwordembread word2vec binary-format model from path.
returns (vocab, vec)
vocabis adictmapping a word to its index.vecis atorch.FloatTensorof sizeV x D, whereVis the vocabulary size andDis the dimension of word2vec.
vocab, vec = torchwordemb.load_word2vec_bin("/path/to/word2vec/model.bin")
print(vec.size())
print(vec[ w2v.vocab["apple"] ] )read word2vec text-format model from path.
read GloVe text-format model from path.