File tree Expand file tree Collapse file tree 1 file changed +37
-0
lines changed Expand file tree Collapse file tree 1 file changed +37
-0
lines changed Original file line number Diff line number Diff line change
1
+ #! /bin/bash
2
+ ./run_clm_streaming_flax_v2.py \
3
+ --output_dir $HOME /gpt-code-clippy-125M-256-resumed \
4
+ --model_name_or_path flax-community/gpt-neo-125M-code-clippy \
5
+ --dataset_name $HOME /gpt-code-clippy/code_clippy.py \
6
+ --data_dir /home/shared/code-clippy-dataset/merged-data \
7
+ --text_column_name=" text" \
8
+ --do_train --do_eval \
9
+ --block_size=" 2048" \
10
+ --per_device_train_batch_size=" 8" \
11
+ --per_device_eval_batch_size=" 16" \
12
+ --preprocessing_num_workers=" 8" \
13
+ --learning_rate=" 3e-4" \
14
+ --max_steps 100000 \
15
+ --warmup_steps 5000 \
16
+ --decay_steps 100000 \
17
+ --adam_beta1=" 0.9" \
18
+ --adam_beta2=" 0.95" \
19
+ --weight_decay=" 0.1" \
20
+ --overwrite_output_dir \
21
+ --logging_steps 100 \
22
+ --eval_steps 2000 \
23
+ --push_to_hub=" False" \
24
+ --report_to=" all" \
25
+ --dtype=" bfloat16" \
26
+ --skip_memory_metrics=" True" \
27
+ --save_steps 2000 \
28
+ --save_total_limit 2 \
29
+ --gradient_accumulation_steps 4 \
30
+ --report_to=" wandb" \
31
+ --run_name=" gpt-code-clippy-125m-3e-4-256" \
32
+ --max_eval_samples 2000 \
33
+ --save_optimizer true \
34
+ --resume_from_checkpoint $HOME /gpt-code-clippy-125M-256 \
35
+ # --adafactor \
36
+ # --max_train_samples="10000" \
37
+
You can’t perform that action at this time.
0 commit comments