Skip to content

Commit 6e1f617

Browse files
authored
Update README.md
1 parent 8a7f45d commit 6e1f617

File tree

1 file changed

+3
-3
lines changed

1 file changed

+3
-3
lines changed

README.md

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -497,8 +497,8 @@ python3 finetune.py --batch_size [batch_size] --accum_iter [grad_accumulation_st
497497
--input_row_size [input_row_size] --input_col_size [input_col_size] \
498498
--patch_size [patch_size] --print_freq [print_freq] --save_freq [save_freq]
499499
```
500-
- `batch_size`: batch size per GPU for fine-tuning.
501-
- `accum_iter`: gradient accumulation steps. The effective batch size is batch_size*accum_iter*num_GPU. <br>
500+
- `batch_size`: batch size per GPU for fine-tuning.
501+
- `accum_iter`: gradient accumulation steps. The effective batch size is batch_size*accum_iter*num_GPU. We recommend at least 256 for stable and reliable training. <br>
502502
If you have memory constraints, you can increase --accum_iter and reduce the --batch_size to trade off memory for computation.
503503
- `epochs`: number of epochs for fine-tuning. Default: 50.
504504
The performance will increase with more epochs, but 50 should be enough to have very good performances.
@@ -542,7 +542,7 @@ Please make sure you include at least **batch_size*num_gpu** examples in the tra
542542

543543
#### Example command
544544
```
545-
python3 finetune.py --batch_size 1 --accum_iter 4 \
545+
python3 finetune.py --batch_size 128 --accum_iter 4 \
546546
--epochs 50 --warmup_epochs 5 --pin_mem \
547547
--blr 1e-3 --min_lr 1e-7 --weight_decay 0.05 \
548548
--layer_decay 0.75 --model vit_large_patch16 \

0 commit comments

Comments
 (0)