@@ -14,28 +14,25 @@ Install the `pytorch` version compatible with your version of cuda [here](https:
1414``` bash
1515conda install pytorch==1.13.1 torchvision==0.14.1 torchaudio==0.13.1 pytorch-cuda=11.6 -c pytorch -c nvidia
1616```
17- Install ` datasets ` , ` accelerate ` and ` huggingface_hub `
18-
17+ Install ` transformers ` and ` peft `
1918``` bash
20- conda install -c huggingface conda-forge datasets
21- conda install -c conda-forge accelerate
22- conda install -c conda-forge huggingface_hub
19+ conda install -c huggingface transformers
20+ pip install git+https://github.com/huggingface/peft.git
2321```
24-
25- Install the latest version of ` transformers ` by cloning the repository
22+ Note that you can install the latest stable version of transformers by using
2623
2724``` bash
28- git clone https://github.com/huggingface/transformers
29- cd transformers
30- pip install -e .
25+ pip install git+https://github.com/huggingface/transformers
3126```
32- Install the latest version of ` peft ` by cloning the repository
27+
28+ Install ` datasets ` , ` accelerate ` and ` huggingface_hub `
3329
3430``` bash
35- git clone https://github.com/ huggingface/peft
36- cd peft
37- pip install -e .
31+ conda install -c huggingface -c conda-forge datasets
32+ conda install -c conda-forge accelerate
33+ conda install -c conda-forge huggingface_hub
3834```
35+
3936Finally, install ` bitsandbytes ` and ` wandb `
4037``` bash
4138pip install bitsandbytes
@@ -66,7 +63,6 @@ To execute the fine-tuning script run the following command:
6663python finetune/finetune.py \
6764 --model_path=" bigcode/large-model" \
6865 --dataset_name=" HuggingFaceH4/CodeAlpaca_20K" \
69- --streaming=False\
7066 --seq_length 2048\
7167 --max_steps 2000\
7268 --batch_size 1\
@@ -84,7 +80,6 @@ python -m torch.distributed.launch \
8480 --nproc_per_node number_of_gpus finetune/finetune.py \
8581 --model_path=" bigcode/large-model" \
8682 --dataset_name=" HuggingFaceH4/CodeAlpaca_20K" \
87- --streaming=False\
8883 --seq_length 2048\
8984 --max_steps 2000\
9085 --batch_size 1\
@@ -106,7 +101,7 @@ python finetune/finetune.py \
106101 --subset=" data/finetune" \
107102 --split=" train" \
108103 --size_valid_set 10000\
109- --streaming True \
104+ --streaming\
110105 --seq_length 2048\
111106 --max_steps 1000\
112107 --batch_size 1\
@@ -127,7 +122,7 @@ python -m torch.distributed.launch \
127122 --subset=" data/finetune" \
128123 --split=" train" \
129124 --size_valid_set 10000\
130- --streaming True \
125+ --streaming \
131126 --seq_length 2048\
132127 --max_steps 1000\
133128 --batch_size 1\
0 commit comments