Replies: 1 comment 1 reply
-
Hello @doyney, Thank you very much for your interest in this library. As to your questions:
Yes it is since you can define your own autoencoding architecture (as explained here and in this example). The only thing you need to ensure it that your encoder outputs the emdeddings and associated covariances with correct shape (i.e. compatible with the latent space dimension) and your decoder the reconstructed samples. You can hence choose the type of neural nets you want to work with and everything should be working.
From what I understand in the paper, the authors indeed proposed to train a second VAE in the latent space of the first one to learn a compressed representation of the features. This is indeed pretty similar to the approach used in the 2-stage VAE sampler. Please nonetheless note that the
For models trained with I hope this helps, Best, Clément |
Beta Was this translation helpful? Give feedback.
-
Hi,
I am working on a project to generate music in a specific style based on a MIDI file, following the approach outlined in Google's MidiME paper
(https://storage.googleapis.com/pub-tools-public-publication-data/pdf/c667ad30514350d65e9fa591f8b2263a8abcc9fd.pdf).
MidiME suggests training a VAE on the latent space of a larger pre-trained VAE to summarise the latent vectors and sample from vectors that are representative of a particular musical style.
I read about the TwoStageVAESampler in the pythae docs and it seems to offer a similar two-stage VAE training approach! Specifically, it allows training a secondary VAE on the latent representations produced by a primary VAE?
On the face of it this library seems perfect, but I'd like to ask to clarification questions:
I'm still getting my head around my project, so hopefully I've been clear enough. I'd really appreciate any thoughts :)
Beta Was this translation helpful? Give feedback.
All reactions