Skip to content

LA-VAE & LA-AE #2

@hulne

Description

@hulne

Thank you to the author for the great work.

However, it appears that the code for LA-VAE actually uses a classic AutoEncoder, and does not apply the KL loss to constrain the latent space as in a VAE.

if mode == 'train':
    optimizer.zero_grad()
    z, before = self.encoder(batch)
    data_recon, after = self.decoder(z, length=batch.shape[-1])
    recon_error = F.mse_loss(data_recon, batch)
    cross_loss = F.mse_loss(before, after)
    loss = recon_error + cross_loss
    loss.backward()
    optimizer.step()

May I ask why it is named LA-VAE instead of LA-AE? Also, can this design ensure good generation quality and meaningful reverse diffusion? From what I have seen in the original VAE-DDPM paper, if it is changed to AE-DDPM, it seems there is no theoretical guarantee.

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions