-
Notifications
You must be signed in to change notification settings - Fork 9
Open
Description
Thank you to the author for the great work.
However, it appears that the code for LA-VAE actually uses a classic AutoEncoder, and does not apply the KL loss to constrain the latent space as in a VAE.
if mode == 'train':
optimizer.zero_grad()
z, before = self.encoder(batch)
data_recon, after = self.decoder(z, length=batch.shape[-1])
recon_error = F.mse_loss(data_recon, batch)
cross_loss = F.mse_loss(before, after)
loss = recon_error + cross_loss
loss.backward()
optimizer.step()
May I ask why it is named LA-VAE instead of LA-AE? Also, can this design ensure good generation quality and meaningful reverse diffusion? From what I have seen in the original VAE-DDPM paper, if it is changed to AE-DDPM, it seems there is no theoretical guarantee.
Reactions are currently unavailable
Metadata
Metadata
Assignees
Labels
No labels