-
Notifications
You must be signed in to change notification settings - Fork 498
/
falcon.yaml
42 lines (36 loc) · 1.1 KB
/
falcon.yaml
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
resources:
accelerators: A100-80GB:1
disk_size: 1000
disk_tier: best
workdir: .
envs:
MODEL_NAME: tiiuae/falcon-7b # [ybelkada/falcon-7b-sharded-bf16, tiiuae/falcon-7b, tiiuae/falcon-40b]
WANDB_API_KEY: # TODO: Fill with your own WANDB_API_KEY, or use --env to pass.
OUTPUT_BUCKET_NAME: # Set a unique name for the bucket which will store model weights
file_mounts:
/results: # Change if the output_dir parameter is changed below
name: $OUTPUT_BUCKET_NAME
mode: MOUNT
setup: |
# Setup the environment
conda activate falcon
if [ $? -ne 0 ]; then
conda create -n falcon python=3.10 -y
conda activate falcon
fi
# Install dependencies
pip install -q -U transformers accelerate peft
pip install -q trl==0.4.6 datasets bitsandbytes einops wandb scipy torch
run: |
conda activate falcon
wandb login $WANDB_API_KEY
echo "Starting training..."
python train.py \
--model_name $MODEL_NAME \
--max_seq_len 2048 \
--bf16 \
--group_by_length \
--bnb_4bit_compute_dtype bfloat16 \
--max_steps 500 \
--dataset_name timdettmers/openassistant-guanaco \
--output_dir /results