2023-06-29 22:05:34 +08:00
|
|
|
export PYTHONPATH=$PYTHONPATH:./
|
2023-07-25 17:32:46 +08:00
|
|
|
torchrun examples/pytorch/llama/finetune_llama.py \
|
2023-06-29 22:05:34 +08:00
|
|
|
--work_dir './tmp' \
|
|
|
|
|
--model 'skyline2006/llama-7b' \
|
2023-08-07 19:48:36 +08:00
|
|
|
--train_dataset_name 'alpaca-gpt4-data-zh' \
|
|
|
|
|
--train_subset_name 'default' \
|
|
|
|
|
--train_split 'train' \
|
|
|
|
|
--train_dataset_namespace 'AI-ModelScope' \
|
|
|
|
|
--per_device_train_batch_size 4 \
|
|
|
|
|
--per_device_eval_batch_size 4 \
|
|
|
|
|
--eval_strategy 'by_epoch' \
|
|
|
|
|
--eval_interval 1 \
|
|
|
|
|
--eval_metrics 'ppl' \
|
|
|
|
|
--lr 2e-5 \
|
|
|
|
|
--save_strategy no \
|
|
|
|
|
--save_best true \
|
|
|
|
|
--metric_for_best_model ppl \
|
|
|
|
|
--metric_rule_for_best_model min \
|
2023-06-29 22:05:34 +08:00
|
|
|
--use_lora 1 \
|
2023-07-25 17:32:46 +08:00
|
|
|
--device_map 'auto' \
|
2023-08-07 19:48:36 +08:00
|
|
|
--task 'text-generation' \
|
|
|
|
|
--model.type 'llama' \
|
|
|
|
|
--max_epochs 3 \
|