Skip to content

Questions about Training Details and Command for MAR-L on ImageNet 512x512 #85

@Tom-zgt

Description

@Tom-zgt

Hello, @LTH14. I am following your great work.

I noticed that you provided results for MAR-L on ImageNet 512×512, and I would like to ask for more details regarding the training setup.

Could you kindly share the specific training hyperparameters, including:

Batch learning rate (blr)
Batch size (batch_size)
Warm-up epochs (warmup_epochs)
Diffusion batch multiplier (diffusion_batch_mul)
Configuration used during testing (cfg)
Additionally, it would be extremely helpful if you could provide the exact command used for training, similar to the following example:
torchrun --nproc_per_node=8 --nnodes=4 --node_rank=${NODE_RANK} --master_addr=${MASTER_ADDR} --master_port=${MASTER_PORT}
main_mar.py
--img_size 256 --vae_path pretrained_models/vae/kl16.ckpt --vae_embed_dim 16 --vae_stride 16 --patch_size 1
--model mar_large --diffloss_d 3 --diffloss_w 1024
--epochs 400 --warmup_epochs 100 --batch_size 64 --blr 1.0e-4 --diffusion_batch_mul 4
--output_dir ${OUTPUT_DIR} --resume ${OUTPUT_DIR}
--data_path ${IMAGENET_PATH}

I have also sent an email inquiring about more detailed information. If you could kindly respond, I would be truly grateful.

Looking forward to your response!

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions