Skip to content

Commit

Permalink
update outputs path so that we can mount workspace to /workspace/data (
Browse files Browse the repository at this point in the history
…axolotl-ai-cloud#1623)

* update outputs path so that we can mount workspace to /workspace/data

* fix ln order
  • Loading branch information
winglian authored May 15, 2024
1 parent 3319780 commit 4fde300
Show file tree
Hide file tree
Showing 70 changed files with 72 additions and 69 deletions.
4 changes: 3 additions & 1 deletion docker/Dockerfile-cloud
Original file line number Diff line number Diff line change
Expand Up @@ -21,7 +21,9 @@ RUN apt install --yes --no-install-recommends openssh-server tmux && \
printf "\n[[ -z \"\$TMUX\" ]] && { tmux attach-session -t ssh_tmux || tmux new-session -s ssh_tmux; exit; }\n" >> ~/.bashrc && \
printf "[ ! -z \"\$TERM\" -a -r /etc/motd ] && cat /etc/motd\n" >> ~/.bashrc && \
chmod +x /workspace/axolotl/scripts/cloud-entrypoint.sh && \
chmod +x /root/cloud-entrypoint.sh
chmod +x /root/cloud-entrypoint.sh && \
mkdir -p /workspace/data/axolotl-artifacts && \
ln -sf /workspace/data/axolotl-artifacts /workspace/axolotl/outputs

ENTRYPOINT ["/root/cloud-entrypoint.sh"]
CMD ["sleep", "infinity"]
2 changes: 1 addition & 1 deletion examples/cerebras/btlm-ft.yml
Original file line number Diff line number Diff line change
Expand Up @@ -38,7 +38,7 @@ wandb_watch:
wandb_name:
wandb_log_model:

output_dir: btlm-out
output_dir: ./outputs/btlm-out
gradient_accumulation_steps: 1
micro_batch_size: 1
num_epochs: 1
Expand Down
2 changes: 1 addition & 1 deletion examples/cerebras/qlora.yml
Original file line number Diff line number Diff line change
Expand Up @@ -25,7 +25,7 @@ wandb_entity:
wandb_watch:
wandb_name:
wandb_log_model:
output_dir: ./qlora-out
output_dir: ./outputs/qlora-out
batch_size: 4
micro_batch_size: 4
num_epochs: 2
Expand Down
2 changes: 1 addition & 1 deletion examples/code-llama/13b/lora.yml
Original file line number Diff line number Diff line change
Expand Up @@ -11,7 +11,7 @@ datasets:
type: alpaca
dataset_prepared_path:
val_set_size: 0.05
output_dir: ./lora-out
output_dir: ./outputs/lora-out

sequence_len: 4096
sample_packing: true
Expand Down
2 changes: 1 addition & 1 deletion examples/code-llama/13b/qlora.yml
Original file line number Diff line number Diff line change
Expand Up @@ -11,7 +11,7 @@ datasets:
type: alpaca
dataset_prepared_path:
val_set_size: 0.05
output_dir: ./qlora-out
output_dir: ./outputs/qlora-out

adapter: qlora
lora_model_dir:
Expand Down
2 changes: 1 addition & 1 deletion examples/code-llama/34b/lora.yml
Original file line number Diff line number Diff line change
Expand Up @@ -11,7 +11,7 @@ datasets:
type: alpaca
dataset_prepared_path:
val_set_size: 0.05
output_dir: ./lora-out
output_dir: ./outputs/lora-out

sequence_len: 4096
sample_packing: true
Expand Down
2 changes: 1 addition & 1 deletion examples/code-llama/34b/qlora.yml
Original file line number Diff line number Diff line change
Expand Up @@ -11,7 +11,7 @@ datasets:
type: alpaca
dataset_prepared_path:
val_set_size: 0.05
output_dir: ./qlora-out
output_dir: ./outputs/qlora-out

adapter: qlora
lora_model_dir:
Expand Down
2 changes: 1 addition & 1 deletion examples/code-llama/7b/lora.yml
Original file line number Diff line number Diff line change
Expand Up @@ -11,7 +11,7 @@ datasets:
type: alpaca
dataset_prepared_path:
val_set_size: 0.05
output_dir: ./lora-out
output_dir: ./outputs/lora-out

sequence_len: 4096
sample_packing: true
Expand Down
2 changes: 1 addition & 1 deletion examples/code-llama/7b/qlora.yml
Original file line number Diff line number Diff line change
Expand Up @@ -11,7 +11,7 @@ datasets:
type: alpaca
dataset_prepared_path:
val_set_size: 0.05
output_dir: ./qlora-out
output_dir: ./outputs/qlora-out

adapter: qlora
lora_model_dir:
Expand Down
2 changes: 1 addition & 1 deletion examples/colab-notebooks/colab-axolotl-example.ipynb
Original file line number Diff line number Diff line change
Expand Up @@ -84,7 +84,7 @@
" type: alpaca\n",
"dataset_prepared_path:\n",
"val_set_size: 0.05\n",
"output_dir: ./qlora-out\n",
"output_dir: ./outputs/qlora-out\n",
"\n",
"adapter: qlora\n",
"lora_model_dir:\n",
Expand Down
2 changes: 1 addition & 1 deletion examples/dbrx/16bit-lora.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -10,7 +10,7 @@ datasets:
type: alpaca
dataset_prepared_path: last_run_prepared
val_set_size: 0.0
output_dir: ./out
output_dir: ./outputs/out

sequence_len: 512
sample_packing: false
Expand Down
2 changes: 1 addition & 1 deletion examples/dbrx/8bit-lora.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -10,7 +10,7 @@ datasets:
type: alpaca
dataset_prepared_path: last_run_prepared
val_set_size: 0.0
output_dir: ./out
output_dir: ./outputs/out

sequence_len: 512
sample_packing: false
Expand Down
2 changes: 1 addition & 1 deletion examples/dbrx/fft-ds-zero3.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -10,7 +10,7 @@ datasets:
type: alpaca
dataset_prepared_path: last_run_prepared
val_set_size: 0.0
output_dir: ./out
output_dir: ./outputs/out

sequence_len: 512
sample_packing: false
Expand Down
2 changes: 1 addition & 1 deletion examples/falcon/config-7b-lora.yml
Original file line number Diff line number Diff line change
Expand Up @@ -28,7 +28,7 @@ wandb_entity:
wandb_watch:
wandb_name:
wandb_log_model:
output_dir: ./falcon-7b
output_dir: ./outputs/falcon-7b
batch_size: 2
micro_batch_size: 1
num_epochs: 4
Expand Down
2 changes: 1 addition & 1 deletion examples/falcon/config-7b-qlora.yml
Original file line number Diff line number Diff line change
Expand Up @@ -42,7 +42,7 @@ wandb_entity:
wandb_watch:
wandb_name:
wandb_log_model:
output_dir: ./qlora-out
output_dir: ./outputs/qlora-out

# QLoRA paper Table 9
# - 16 for 7b & 13b
Expand Down
2 changes: 1 addition & 1 deletion examples/falcon/config-7b.yml
Original file line number Diff line number Diff line change
Expand Up @@ -28,7 +28,7 @@ wandb_entity:
wandb_watch:
wandb_name:
wandb_log_model:
output_dir: ./falcon-7b
output_dir: ./outputs/falcon-7b
batch_size: 2
micro_batch_size: 1
num_epochs: 4
Expand Down
2 changes: 1 addition & 1 deletion examples/gemma/qlora.yml
Original file line number Diff line number Diff line change
Expand Up @@ -12,7 +12,7 @@ datasets:
- path: mhenrichsen/alpaca_2k_test
type: alpaca
val_set_size: 0.1
output_dir: ./out
output_dir: ./outputs/out

adapter: qlora
lora_r: 32
Expand Down
2 changes: 1 addition & 1 deletion examples/gptj/qlora.yml
Original file line number Diff line number Diff line change
Expand Up @@ -23,7 +23,7 @@ wandb_entity:
wandb_watch:
wandb_name:
wandb_log_model:
output_dir: ./qlora-out
output_dir: ./outputs/qlora-out
gradient_accumulation_steps: 2
micro_batch_size: 2
num_epochs: 2
Expand Down
2 changes: 1 addition & 1 deletion examples/jamba/qlora.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -10,7 +10,7 @@ datasets:
type: alpaca
dataset_prepared_path:
val_set_size: 0.0
output_dir: ./out
output_dir: ./outputs/out

sequence_len: 4096
sample_packing: false
Expand Down
2 changes: 1 addition & 1 deletion examples/jamba/qlora_deepspeed.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -10,7 +10,7 @@ datasets:
type: alpaca
dataset_prepared_path:
val_set_size: 0.0
output_dir: ./out
output_dir: ./outputs/out

sequence_len: 4096
sample_packing: false
Expand Down
2 changes: 1 addition & 1 deletion examples/jeopardy-bot/config.yml
Original file line number Diff line number Diff line change
Expand Up @@ -21,7 +21,7 @@ wandb_entity:
wandb_watch:
wandb_name:
wandb_log_model:
output_dir: ./jeopardy-bot-7b
output_dir: ./outputs/jeopardy-bot-7b
gradient_accumulation_steps: 1
micro_batch_size: 1
num_epochs: 4
Expand Down
2 changes: 1 addition & 1 deletion examples/llama-2/fft_optimized.yml
Original file line number Diff line number Diff line change
Expand Up @@ -11,7 +11,7 @@ datasets:
type: alpaca
dataset_prepared_path: last_run_prepared
val_set_size: 0.05
output_dir: ./out
output_dir: ./outputs/out

sequence_len: 4096
sample_packing: true
Expand Down
2 changes: 1 addition & 1 deletion examples/llama-2/gptq-lora.yml
Original file line number Diff line number Diff line change
Expand Up @@ -33,7 +33,7 @@ wandb_project:
wandb_watch:
wandb_name:
wandb_log_model:
output_dir: ./model-out
output_dir: ./outputs/model-out
gradient_accumulation_steps: 1
micro_batch_size: 1
num_epochs: 4
Expand Down
2 changes: 1 addition & 1 deletion examples/llama-2/lisa.yml
Original file line number Diff line number Diff line change
Expand Up @@ -11,7 +11,7 @@ datasets:
type: alpaca
dataset_prepared_path: last_run_prepared
val_set_size: 0.05
output_dir: ./lisa-out
output_dir: ./outputs/lisa-out

sequence_len: 4096
sample_packing: true
Expand Down
2 changes: 1 addition & 1 deletion examples/llama-2/loftq.yml
Original file line number Diff line number Diff line change
Expand Up @@ -11,7 +11,7 @@ datasets:
type: alpaca
dataset_prepared_path:
val_set_size: 0.05
output_dir: ./lora-out
output_dir: ./outputs/lora-out

sequence_len: 4096
sample_packing: true
Expand Down
2 changes: 1 addition & 1 deletion examples/llama-2/lora.yml
Original file line number Diff line number Diff line change
Expand Up @@ -11,7 +11,7 @@ datasets:
type: alpaca
dataset_prepared_path:
val_set_size: 0.05
output_dir: ./lora-out
output_dir: ./outputs/lora-out

sequence_len: 4096
sample_packing: true
Expand Down
2 changes: 1 addition & 1 deletion examples/llama-2/qlora-fsdp.yml
Original file line number Diff line number Diff line change
Expand Up @@ -11,7 +11,7 @@ datasets:
type: alpaca
dataset_prepared_path: last_run_prepared
val_set_size: 0.05
output_dir: ./qlora-out
output_dir: ./outputs/qlora-out

adapter: qlora
lora_model_dir:
Expand Down
2 changes: 1 addition & 1 deletion examples/llama-2/qlora.yml
Original file line number Diff line number Diff line change
Expand Up @@ -11,7 +11,7 @@ datasets:
type: alpaca
dataset_prepared_path:
val_set_size: 0.05
output_dir: ./qlora-out
output_dir: ./outputs/qlora-out

adapter: qlora
lora_model_dir:
Expand Down
2 changes: 1 addition & 1 deletion examples/llama-2/relora.yml
Original file line number Diff line number Diff line change
Expand Up @@ -12,7 +12,7 @@ datasets:
type: alpaca
dataset_prepared_path:
val_set_size: 0.05
output_dir: ./relora-out
output_dir: ./outputs/relora-out

adapter: qlora
lora_model_dir:
Expand Down
2 changes: 1 addition & 1 deletion examples/llama-3/fft-8b.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -11,7 +11,7 @@ datasets:
type: alpaca
dataset_prepared_path: last_run_prepared
val_set_size: 0.05
output_dir: ./out
output_dir: ./outputs/out

sequence_len: 8192
sample_packing: true
Expand Down
2 changes: 1 addition & 1 deletion examples/llama-3/lora-8b.yml
Original file line number Diff line number Diff line change
Expand Up @@ -11,7 +11,7 @@ datasets:
type: alpaca
dataset_prepared_path:
val_set_size: 0.05
output_dir: ./lora-out
output_dir: ./outputs/lora-out

sequence_len: 4096
sample_packing: true
Expand Down
2 changes: 1 addition & 1 deletion examples/llama-3/qlora-fsdp-70b.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -11,7 +11,7 @@ datasets:
type: alpaca
dataset_prepared_path: last_run_prepared
val_set_size: 0.05
output_dir: ./out/qlora-llama3-70b
output_dir: ./outputs/out/qlora-llama3-70b

adapter: qlora
lora_model_dir:
Expand Down
2 changes: 1 addition & 1 deletion examples/llama-3/qlora.yml
Original file line number Diff line number Diff line change
Expand Up @@ -11,7 +11,7 @@ datasets:
type: alpaca
dataset_prepared_path:
val_set_size: 0
output_dir: ./qlora-out
output_dir: ./outputs/qlora-out

adapter: qlora
lora_model_dir:
Expand Down
2 changes: 1 addition & 1 deletion examples/mamba/config.yml
Original file line number Diff line number Diff line change
Expand Up @@ -12,7 +12,7 @@ datasets:
type: alpaca
dataset_prepared_path:
val_set_size: 0.0
output_dir: ./out
output_dir: ./outputs/out

sequence_len: 2048
sample_packing: false
Expand Down
2 changes: 1 addition & 1 deletion examples/mistral/bigstral-ds-zero3.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -23,7 +23,7 @@ datasets:
type: alpaca
dataset_prepared_path: last_run_prepared
val_set_size: 0.05
output_dir: ./out
output_dir: ./outputs/out

sequence_len: 2048
sample_packing: true
Expand Down
2 changes: 1 addition & 1 deletion examples/mistral/config.yml
Original file line number Diff line number Diff line change
Expand Up @@ -11,7 +11,7 @@ datasets:
type: alpaca
dataset_prepared_path:
val_set_size: 0.05
output_dir: ./out
output_dir: ./outputs/out

sequence_len: 8192
sample_packing: true
Expand Down
2 changes: 1 addition & 1 deletion examples/mistral/lora-mps.yml
Original file line number Diff line number Diff line change
Expand Up @@ -11,7 +11,7 @@ datasets:
type: alpaca
dataset_prepared_path: last_run_prepared
val_set_size: 0
output_dir: ./lora-out
output_dir: ./outputs/lora-out
eval_sample_packing: false

adapter: lora
Expand Down
2 changes: 1 addition & 1 deletion examples/mistral/lora.yml
Original file line number Diff line number Diff line change
Expand Up @@ -11,7 +11,7 @@ datasets:
type: alpaca
dataset_prepared_path: last_run_prepared
val_set_size: 0.1
output_dir: ./lora-out
output_dir: ./outputs/lora-out

adapter: lora
lora_model_dir:
Expand Down
2 changes: 1 addition & 1 deletion examples/mistral/mistral-qlora-fsdp.yml
Original file line number Diff line number Diff line change
Expand Up @@ -12,7 +12,7 @@ datasets:
type: alpaca
dataset_prepared_path: last_run_prepared
val_set_size: 0.02
output_dir: ./qlora-out
output_dir: ./outputs/qlora-out

model_config:
output_router_logits: true
Expand Down
2 changes: 1 addition & 1 deletion examples/mistral/mistral-qlora-orpo.yml
Original file line number Diff line number Diff line change
Expand Up @@ -16,7 +16,7 @@ datasets:
type: chat_template.argilla
dataset_prepared_path: last_run_prepared
val_set_size: 0.1
output_dir: ./mistral-qlora-orpo-out
output_dir: ./outputs/mistral-qlora-orpo-out

adapter: qlora
lora_model_dir:
Expand Down
2 changes: 1 addition & 1 deletion examples/mistral/mixtral-8x22b-qlora-fsdp.yml
Original file line number Diff line number Diff line change
Expand Up @@ -11,7 +11,7 @@ datasets:
type: alpaca
dataset_prepared_path: last_run_prepared
val_set_size: 0.02
output_dir: ./qlora-out
output_dir: ./outputs/qlora-out

model_config:
output_router_logits: true
Expand Down
2 changes: 1 addition & 1 deletion examples/mistral/mixtral-qlora-fsdp.yml
Original file line number Diff line number Diff line change
Expand Up @@ -12,7 +12,7 @@ datasets:
type: alpaca
dataset_prepared_path: last_run_prepared
val_set_size: 0.02
output_dir: ./qlora-out
output_dir: ./outputs/qlora-out

model_config:
output_router_logits: true
Expand Down
2 changes: 1 addition & 1 deletion examples/mistral/mixtral.yml
Original file line number Diff line number Diff line change
Expand Up @@ -12,7 +12,7 @@ datasets:
type: alpaca
dataset_prepared_path: last_run_prepared
val_set_size: 0.0
output_dir: ./qlora-out
output_dir: ./outputs/qlora-out

## You can optionally freeze the entire model and unfreeze a subset of parameters
unfrozen_parameters:
Expand Down
2 changes: 1 addition & 1 deletion examples/mistral/mixtral_22.yml
Original file line number Diff line number Diff line change
Expand Up @@ -21,7 +21,7 @@ model_config:
datasets:
- path: yahma/alpaca-cleaned
type: alpaca
output_dir: ./out
output_dir: ./outputs/out

sequence_len: 8000
sample_packing: true
Expand Down
Loading

0 comments on commit 4fde300

Please sign in to comment.