question-mark
Stuck on an issue?

Lightrun Answers was designed to reduce the constant googling that comes with debugging 3rd party libraries. It collects links to all the places you might be looking at while hunting down a tough bug.

And, if you’re still stuck at the end, we’re happy to hop on a call to see how we can help out.

text embedding dimension

See original GitHub issue

Hi. Since I tried to train Glow-TTS using Mandarin datasets, there are about 300 symbols in symbols.py. Therefore, it seems that I need to increase the text embedding depth. I notice that in your paper, you mentioned that: Screenshot 2020-06-22 at 8 33 21 PM Does the Embedding Dimension here stands for “text embedding dimension”? If it is, which parameter here should I modify, hidden_channels , or hidden_channels_enc? Screenshot 2020-06-22 at 8 36 02 PM

Thank you very much!

Issue Analytics

  • State:closed
  • Created 3 years ago
  • Comments:6

github_iconTop GitHub Comments

1reaction
shahuzicommented, Nov 8, 2020

I changed hidden_channels, hidden_channels_enc and hidden_channels_decto 512, but I still encountered the following problem: in the inference time, some words are missing. For example, if I input: Screenshot 2020-06-23 at 10 16 07 AM I cannot hear r e4 in the synthesized voice. (The number here stands for tones in mandarin). Could you please give me some advice? Thanks!

Hi, you can try the trick add blank token between any two input tokens. My experiment in Chinese shows that this trick can improve pronunciation significantly.

0reactions
shahuzicommented, Dec 9, 2020

@shahuzi 您好,方便提供几个你用glow-tts合成的音频样例吗?十分感谢

由于涉及到数据安全问题,我没法给你提供demo,见谅。目前我的结论是:对于播报式的音库,可以正常地合成,对于表现力很丰富的音库,合成会出问题。

Read more comments on GitHub >

github_iconTop Results From Across the Web

what is dimensionality in word embeddings? - Stack Overflow
Answer. A Word Embedding is just a mapping from words to vectors. Dimensionality in word embeddings refers to the length of these vectors....
Read more >
Word embeddings, what are they really? | by Kristin H. Huseby
Mathematically, each word contributes with one dimension to the resulting vector. This is a sparse encoding, and uses a lot of memory to...
Read more >
Word embeddings | Text - TensorFlow
It is common to see word embeddings that are 8-dimensional (for small datasets), up to 1024-dimensions when working with large datasets.
Read more >
NLP-101: Understanding Word Embedding - Kaggle
It is common to see word embeddings that are 8-dimensional (for small datasets), up to 1024-dimensions when working with large datasets. A higher...
Read more >
Embeddings | Machine Learning - Google Developers
An embedding is a relatively low-dimensional space into which you can translate high-dimensional vectors. Embeddings make it easier to do ...
Read more >

github_iconTop Related Medium Post

No results found

github_iconTop Related StackOverflow Question

No results found

github_iconTroubleshoot Live Code

Lightrun enables developers to add logs, metrics and snapshots to live code - no restarts or redeploys required.
Start Free

github_iconTop Related Reddit Thread

No results found

github_iconTop Related Hackernoon Post

No results found

github_iconTop Related Tweet

No results found

github_iconTop Related Dev.to Post

No results found

github_iconTop Related Hashnode Post

No results found