We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
llamafactory version: 0.9.2.dev0
llamafactory
‘’‘ llamafactory-cli train --stage sft --do_train True --model_name_or_path /LLaMA-Factory-main/model/Qwen2.5-7B-Instruct --preprocessing_num_workers 2 --finetuning_type lora --template qwen --flash_attn fa2 --dataset_dir /LLaMA-Factory-main/data --dataset alpaca_zh_demo --cutoff_len 1024 --learning_rate 0.000001 --num_train_epochs 5 --max_samples 1000 --per_device_train_batch_size 4 --gradient_accumulation_steps 2 --lr_scheduler_type cosine --logging_steps 5 --save_steps 100 --output_dir saves/Qwen2.5-7B-Instruct/qlora/sft/train_2024-12-26-01 --quantization_bit 4 --quantization_method bitsandbytes --deepspeed cache/ds_z3_config.json ’‘’
异常堆栈:
根据bitsandbytes官方安装指南(https://huggingface.co/docs/bitsandbytes/v0.45.0/en/installation?backend=Ascend+NPU#multi-backend-compile) 安装了NPU版本的,但QLora训练依旧报错。 请问目前支持在NPU上进行QLora训练吗?如果支持的话,需要安装哪些依赖?
No response
The text was updated successfully, but these errors were encountered:
目前昇腾不支持QLora
Sorry, something went wrong.
@statelesshz Could you help with this problem?
@sunxiaoyu12 源码安装一下main分支的transformers
No branches or pull requests
Reminder
System Info
llamafactory
version: 0.9.2.dev0Reproduction
‘’‘
llamafactory-cli train
--stage sft
--do_train True
--model_name_or_path /LLaMA-Factory-main/model/Qwen2.5-7B-Instruct
--preprocessing_num_workers 2
--finetuning_type lora
--template qwen
--flash_attn fa2
--dataset_dir /LLaMA-Factory-main/data
--dataset alpaca_zh_demo
--cutoff_len 1024
--learning_rate 0.000001
--num_train_epochs 5
--max_samples 1000
--per_device_train_batch_size 4
--gradient_accumulation_steps 2
--lr_scheduler_type cosine
--logging_steps 5
--save_steps 100
--output_dir saves/Qwen2.5-7B-Instruct/qlora/sft/train_2024-12-26-01
--quantization_bit 4
--quantization_method bitsandbytes
--deepspeed cache/ds_z3_config.json
’‘’
异常堆栈:
Expected behavior
根据bitsandbytes官方安装指南(https://huggingface.co/docs/bitsandbytes/v0.45.0/en/installation?backend=Ascend+NPU#multi-backend-compile) 安装了NPU版本的,但QLora训练依旧报错。
请问目前支持在NPU上进行QLora训练吗?如果支持的话,需要安装哪些依赖?
Others
No response
The text was updated successfully, but these errors were encountered: