File tree 1 file changed +8
-7
lines changed
1 file changed +8
-7
lines changed Original file line number Diff line number Diff line change 1
- MODEL=EVA02-CLIP-L-14
1
+ MODEL=EVA02-CLIP-L-14-336
2
2
PRETRAINED=eva_clip
3
- python -m torch.distributed.launch --nproc_per_node=2 \
3
+ python -m torch.distributed.launch --nproc_per_node=8 \
4
4
--use_env training/main.py \
5
5
--enable-deepspeed \
6
6
--grad-checkpointing \
7
7
--name=" T_vitl336_Rcc12mR_Rcc3m_4ep" \
8
8
--save-frequency 1 \
9
9
--zeroshot-frequency 1 \
10
- --report-to=" " \
10
+ --report-to=" tensorboard, wandb " \
11
11
--wandb-project-name=" LLM2CLIP" \
12
12
--wandb-notes=" EVA02-CLIP-L-14-336" \
13
- --train-data-list " /home/aiscuser/LLM2CLIP/llm2clip/data/cc3m/{00..00004}.tar" \
14
- --train-num-samples-list 2873538 \
13
+ --train-data-list " data/cc3m/cc3m-train-{00..0287}.tar;data/cc12m/cc12m-train-{00..1001}.tar" \
14
+ --train-num-samples-list 2873538 10000225 \
15
+ --eval-data-file=training/eval_datasets.yaml \
15
16
--pretrained=${PRETRAINED} \
16
17
--dataset-resampled \
17
18
--precision " fp16" \
18
19
--warmup 0 \
19
- --batch-size=16 \
20
+ --batch-size=512 \
20
21
--eval-batch-size=1024 \
21
22
--log-every-n-steps 50 \
22
23
--epochs=20 \
@@ -31,7 +32,7 @@ python -m torch.distributed.launch --nproc_per_node=2 \
31
32
--visual-ld=0.85 \
32
33
--grad-clip-norm=5.0 \
33
34
--smoothing=0. \
34
- --workers=1 \
35
+ --workers=8 \
35
36
--model=${MODEL} \
36
37
--seed 4096 \
37
38
--gather-with-grad \
You can’t perform that action at this time.
0 commit comments