* Add Axolotl 0.4.0, remove legacy 0.3.0 support. * replace is_torch_bf16_gpu_available * Add HF_HUB_OFFLINE=1 * Move transformers out of requirement * Refine readme and qlora.yml |
||
|---|---|---|
| .. | ||
| axolotl | ||
| common | ||
| DPO | ||
| HF-PEFT | ||
| LoRA | ||
| QA-LoRA | ||
| QLoRA | ||
| ReLora | ||
| README.md | ||
Running LLM Finetuning using IPEX-LLM on Intel GPU
This folder contains examples of running different training mode with IPEX-LLM on Intel GPU:
- LoRA: examples of running LoRA finetuning
- QLoRA: examples of running QLoRA finetuning
- QA-LoRA: examples of running QA-LoRA finetuning
- ReLora: examples of running ReLora finetuning
- DPO: examples of running DPO finetuning
- common: common templates and utility classes in finetuning examples
- HF-PEFT: run finetuning on Intel GPU using Hugging Face PEFT code without modification
- axolotl: LLM finetuning on Intel GPU using axolotl without writing code
Troubleshooting
-
If you fail to finetune on multi cards because of following error message:
RuntimeError: oneCCL: comm_selector.cpp:57 create_comm_impl: EXCEPTION: ze_data was not initializedPlease try
sudo apt install level-zero-devto fix it. -
Please raise the system open file limit using
ulimit -n 1048576. Otherwise, there may exist errorToo many open files.