File tree 1 file changed +7
-11
lines changed
1 file changed +7
-11
lines changed Original file line number Diff line number Diff line change 1
- MODEL=EVA02-CLIP-L-14-336
1
+ MODEL=EVA02-CLIP-L-14
2
2
PRETRAINED=eva_clip
3
- python -m torch.distributed.launch --nproc_per_node=8 \
3
+ python -m torch.distributed.launch --nproc_per_node=2 \
4
4
--use_env training/main.py \
5
5
--enable-deepspeed \
6
6
--grad-checkpointing \
7
7
--name=" T_vitl336_Rcc12mR_Rcc3m_4ep" \
8
8
--save-frequency 1 \
9
9
--zeroshot-frequency 1 \
10
- --report-to=" tensorboard, wandb " \
10
+ --report-to=" " \
11
11
--wandb-project-name=" LLM2CLIP" \
12
12
--wandb-notes=" EVA02-CLIP-L-14-336" \
13
- --train-data-list " data/cc3m/cc3m-train-{00..0287}.tar;data/cc12m/cc12m-train-{00..1001}.tar" \
14
- --train-num-samples-list 2873538 10000225 \
15
- --eval-data-file=training/eval_datasets.yaml \
16
- --imagenet-val=data/eval_data/imagenet/val.zip \
17
- --imagenet-val-text=data/eval_data/imagenet/val_map.txt \
18
- --imagenet-classname-feautres data/eval_data/imagenet/im_classname_llm_features.dpt \
13
+ --train-data-list " /home/aiscuser/LLM2CLIP/llm2clip/data/cc3m/{00..00004}.tar" \
14
+ --train-num-samples-list 2873538 \
19
15
--pretrained=${PRETRAINED} \
20
16
--dataset-resampled \
21
17
--precision " fp16" \
22
18
--warmup 0 \
23
- --batch-size=512 \
19
+ --batch-size=16 \
24
20
--eval-batch-size=1024 \
25
21
--log-every-n-steps 50 \
26
22
--epochs=20 \
@@ -35,7 +31,7 @@ python -m torch.distributed.launch --nproc_per_node=8 \
35
31
--visual-ld=0.85 \
36
32
--grad-clip-norm=5.0 \
37
33
--smoothing=0. \
38
- --workers=8 \
34
+ --workers=1 \
39
35
--model=${MODEL} \
40
36
--seed 4096 \
41
37
--gather-with-grad \
You can’t perform that action at this time.
0 commit comments