From 2f6133623ebc5f5a83757c33a9143a057a92315d Mon Sep 17 00:00:00 2001 From: ftgreat Date: Tue, 5 Aug 2025 16:00:04 +0800 Subject: [PATCH] Update AltCLIP and vit_cifar100 Signed-off-by: ftgreat --- examples/AltCLIP/altclip_finetuning.py | 4 ++-- examples/AltDiffusion/requirements.txt | 5 +++++ examples/vit_cifar100/requirements.txt | 5 +++++ examples/vit_cifar100/train_single_gpu.py | 2 +- flagai/model/vision/vit.py | 4 +++- 5 files changed, 16 insertions(+), 4 deletions(-) create mode 100644 examples/AltDiffusion/requirements.txt create mode 100644 examples/vit_cifar100/requirements.txt diff --git a/examples/AltCLIP/altclip_finetuning.py b/examples/AltCLIP/altclip_finetuning.py index 2b95fd4c..1f08690d 100644 --- a/examples/AltCLIP/altclip_finetuning.py +++ b/examples/AltCLIP/altclip_finetuning.py @@ -11,7 +11,7 @@ device = torch.device("cuda" if torch.cuda.is_available() else "cpu") -dataset_root = "./clip_benchmark_datasets" +dataset_root = "./data" dataset_name = "cifar10" batch_size = 4 @@ -62,4 +62,4 @@ def cifar10_collate_fn(batch): } if __name__ == "__main__": - trainer.train(model=model, train_dataset=dataset, collate_fn=cifar10_collate_fn) \ No newline at end of file + trainer.train(model=model, train_dataset=dataset, collate_fn=cifar10_collate_fn) diff --git a/examples/AltDiffusion/requirements.txt b/examples/AltDiffusion/requirements.txt new file mode 100644 index 00000000..3386176c --- /dev/null +++ b/examples/AltDiffusion/requirements.txt @@ -0,0 +1,5 @@ +pillow +antlr4 +pytorch-lightning==1.9.0 +taming-transformers==0.0.6 +transformers==4.30.0 diff --git a/examples/vit_cifar100/requirements.txt b/examples/vit_cifar100/requirements.txt new file mode 100644 index 00000000..3386176c --- /dev/null +++ b/examples/vit_cifar100/requirements.txt @@ -0,0 +1,5 @@ +pillow +antlr4 +pytorch-lightning==1.9.0 +taming-transformers==0.0.6 +transformers==4.30.0 diff --git a/examples/vit_cifar100/train_single_gpu.py b/examples/vit_cifar100/train_single_gpu.py index ef7e1356..05ea8169 100644 --- a/examples/vit_cifar100/train_single_gpu.py +++ b/examples/vit_cifar100/train_single_gpu.py @@ -21,7 +21,7 @@ weight_decay=1e-5, epochs=n_epochs, log_interval=100, - eval_interval=1000, + eval_interval=10000, load_dir=None, pytorch_device=device, save_dir="checkpoints_vit_cifar100_single_gpu", diff --git a/flagai/model/vision/vit.py b/flagai/model/vision/vit.py index ebe6c400..55624752 100755 --- a/flagai/model/vision/vit.py +++ b/flagai/model/vision/vit.py @@ -225,10 +225,12 @@ def __init__( norm_layer: (nn.Module): normalization layer act_layer: (nn.Module): MLP activation layer """ + config = config.json_config super().__init__(config) embed_layer=PatchEmbed block_fn=Block - config = config.json_config + if 'use_cache' in config: + del config['use_cache'] vit_config = VitConfig(**config) vit_config.num_classes = num_classes # config = vit_config