Using FastChat
Fine-tuning FastChat-T5
You can use the following command to train FastChat-T5 with 4 x A100 (40GB).
After training, please use our post-processing function to update the saved model weight. Additional discussions can be found here.
Fine-tuning using (Q)LoRA
You can use the following command to train Vicuna-7B using QLoRA using ZeRO2. Note that ZeRO3 is not currently supported with QLoRA but ZeRO3 does support LoRA, which has a reference configuraiton under playground/deepspeed_config_s3.json. To use QLoRA, you must have bitsandbytes>=0.39.0 and transformers>=4.30.0 installed.
For T5-XL or XXL
Fine-tuning Vicuna-7B with Local NPUs
You can use the following command to train Vicuna-7B with 8 x 910B (60GB). Use --nproc_per_node
to specify the number of NPUs.