Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Regd Forcing Encoder Attention Alignments #96

Open
bkumardevan07 opened this issue Apr 6, 2021 · 1 comment
Open

Regd Forcing Encoder Attention Alignments #96

bkumardevan07 opened this issue Apr 6, 2021 · 1 comment

Comments

@bkumardevan07
Copy link

bkumardevan07 commented Apr 6, 2021

Hi I observed that you are forcing encoder attention too to be diagonal for some steps, and I found that after completion of training, the alignment remains diagonal. My question is then why we need more encoder layers if all it has to remain diagonal?? Did you see any issues when not forcing encoder attention diagonal? Any other observations?

  1. Also often I have seen in some papers that their mel outputs are well predicted towards higher frequency region of mel spectrogram, but in all my trainings the results come a little blurry around the top regions of mel-spectrogram. Does this has to do anything with convergence? Any ideas what might be happening wrong?
@cfrancesco
Copy link
Contributor

Hi,
in my experiments the encoder alignments are rather optional, that's why I set it to a lower number of steps than the decoder. You probably can safely set it to 0. I didn't experiment extensively, but I didn't notice a drawback. Also, without forcing this diagonality almost all the encoder heads in the aligner tend to become diagonal eventually (typically 0-1 per layer are scattered).
With fewer layers I did experience reduction in quality on the predicted mels.

In case you perform a more complete analysis it would be great to hear the results!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants