Replies: 1 comment 2 replies
-
|
Thanks for the detailed post. I love the notes :) It is not actually making W positive, more like making its determinant positive. About h and w, they are the height and width of the input. However, in our case, tensors are sequences so we use only length I hope it helps with the confusion if there is any. |
Beta Was this translation helpful? Give feedback.
2 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
While reading GlowTTS and WaveGlow papers, they refer to Glow (Generative Flow with Invertible 1x1 Convolutions) paper.
In the following Table 1 [Glow paper], we can see the log-determinant (logdet) of the invertible 1x1 convolution (highlighted in yellow).
In the 🐸 TTS repo, we use this
logdetas follows:TTS/TTS/tts/layers/glow_tts/glow.py
Lines 99 to 100 in 5094499
TTS/TTS/tts/layers/glow_tts/glow.py
Line 131 in 5094499
I was not sure about the derivation for that formula and no Google search gave me satisfactory answer (especially about the h x w term), so I came up with my own proof as show below. I hope someone who wants to learn the math behind these things can benefit from these. I would advice at least reading the Glow paper as a prerequisite.
Beta Was this translation helpful? Give feedback.
All reactions