Skip to content

Commit a6b3021

Browse files
committed
workaround for resume
1 parent f12d209 commit a6b3021

File tree

1 file changed

+37
-0
lines changed

1 file changed

+37
-0
lines changed

resume_clm_streaming-3e-4-256.sh

Lines changed: 37 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,37 @@
1+
#! /bin/bash
2+
./run_clm_streaming_flax_v2.py \
3+
--output_dir $HOME/gpt-code-clippy-125M-256-resumed \
4+
--model_name_or_path flax-community/gpt-neo-125M-code-clippy \
5+
--dataset_name $HOME/gpt-code-clippy/code_clippy.py \
6+
--data_dir /home/shared/code-clippy-dataset/merged-data \
7+
--text_column_name="text" \
8+
--do_train --do_eval \
9+
--block_size="2048" \
10+
--per_device_train_batch_size="8" \
11+
--per_device_eval_batch_size="16" \
12+
--preprocessing_num_workers="8" \
13+
--learning_rate="3e-4" \
14+
--max_steps 100000 \
15+
--warmup_steps 5000 \
16+
--decay_steps 100000 \
17+
--adam_beta1="0.9" \
18+
--adam_beta2="0.95" \
19+
--weight_decay="0.1" \
20+
--overwrite_output_dir \
21+
--logging_steps 100 \
22+
--eval_steps 2000 \
23+
--push_to_hub="False" \
24+
--report_to="all" \
25+
--dtype="bfloat16" \
26+
--skip_memory_metrics="True" \
27+
--save_steps 2000 \
28+
--save_total_limit 2 \
29+
--gradient_accumulation_steps 4 \
30+
--report_to="wandb" \
31+
--run_name="gpt-code-clippy-125m-3e-4-256" \
32+
--max_eval_samples 2000 \
33+
--save_optimizer true \
34+
--resume_from_checkpoint $HOME/gpt-code-clippy-125M-256 \
35+
# --adafactor \
36+
# --max_train_samples="10000" \
37+

0 commit comments

Comments
 (0)