question-mark
Stuck on an issue?

Lightrun Answers was designed to reduce the constant googling that comes with debugging 3rd party libraries. It collects links to all the places you might be looking at while hunting down a tough bug.

And, if you’re still stuck at the end, we’re happy to hop on a call to see how we can help out.

ETM - Possibility of using KeyedVectors input for pre-trained W2V embeddings

See original GitHub issue
  • OCTIS version: 1.9.0
  • Python version: 3.7.6
  • Operating System: Ubuntu 20.04 LTS

Description

Hi, this is more of a question than anything else. I’ve seen that for ETM model training, we must pass an embeddings path corresponding to a “pickled” file. However, I need to execute ETM with rather large embeddings. There’s any intent on implementing a gensim.models.KeyedVectors based (or something like that) embeddings input for this model? I’ve implemented something like that for an etm package of mine, but yours’ has all I need to execute model optimization. Would a PR on this matter be accepted?

Anyway, cheers for the nice work, this package is really great!

What I Did

Gave a look at here.

Issue Analytics

  • State:closed
  • Created 2 years ago
  • Comments:8 (2 by maintainers)

github_iconTop GitHub Comments

3reactions
silviatticommented, Nov 21, 2021

Hi, I merged the PR on the branch dev_etm. Now I’m going to fix another issue and make the new release. It should be available in the next few hours. Thank you for your patience 😃

1reaction
cayalukecommented, Nov 9, 2021

@lffloyd Thank you for your response. It is much appreciated.

Read more comments on GitHub >

github_iconTop Results From Across the Web

models.keyedvectors – Store and query word vectors — gensim
models.keyedvectors – Store and query word vectors¶. This module implements word vectors, and more generally sets of vectors keyed by lookup tokens/ints,.
Read more >
Gensim Word2Vec select minor set of word vectors from ...
The use case I wanted to cover is the creation of a pre-trained embedding matrix to initialize an Embedding layer. I wanted to...
Read more >
FROM Pre-trained Word Embeddings ... - Towards Data Science
This iterative process through the blocks will help the neural network capture more complex relationships between words in the input sequence.
Read more >
Easily Access Pre-trained Word Embeddings with Gensim
Pre-trained word embeddings are vector representation of words trained on a large dataset. With pre-trained embeddings, you will essentially ...
Read more >
Do Pretrained Embeddings Give You The Extra Edge? - Kaggle
In this kernel, we shall see if pretrained embeddings like Word2Vec, GLOVE and Fasttext, which are pretrained using billions of words could improve...
Read more >

github_iconTop Related Medium Post

No results found

github_iconTop Related StackOverflow Question

No results found

github_iconTroubleshoot Live Code

Lightrun enables developers to add logs, metrics and snapshots to live code - no restarts or redeploys required.
Start Free

github_iconTop Related Reddit Thread

No results found

github_iconTop Related Hackernoon Post

No results found

github_iconTop Related Tweet

No results found

github_iconTop Related Dev.to Post

No results found

github_iconTop Related Hashnode Post

No results found