From a0a80d232ee218ae5b6401d92883ac5afeef6169 Mon Sep 17 00:00:00 2001 From: "Wang, Jian4" <61138589+hzjane@users.noreply.github.com> Date: Thu, 30 Nov 2023 13:42:30 +0800 Subject: [PATCH] LLM: Add qlora cpu distributed readme (#9561) * init readme * add distributed guide * update --- .../llm/example/CPU/QLoRA-FineTuning/README.md | 6 ++++++ .../QLoRA-FineTuning/alpaca-qlora/README.md | 14 ++++++++++++++ .../finetune_one_node_two_sockets.sh | 18 ++++++++++++++++++ 3 files changed, 38 insertions(+) create mode 100644 python/llm/example/CPU/QLoRA-FineTuning/alpaca-qlora/finetune_one_node_two_sockets.sh diff --git a/python/llm/example/CPU/QLoRA-FineTuning/README.md b/python/llm/example/CPU/QLoRA-FineTuning/README.md index 7f6a5e79..02b482c9 100644 --- a/python/llm/example/CPU/QLoRA-FineTuning/README.md +++ b/python/llm/example/CPU/QLoRA-FineTuning/README.md @@ -3,6 +3,12 @@ This example demonstrates how to finetune a llama2-7b model using Big-LLM 4bit optimizations on [Intel CPUs](../README.md). +## Distributed Training Guide +1. Single node with single socket: [simple example](https://github.com/intel-analytics/BigDL/tree/main/python/llm/example/CPU/QLoRA-FineTuning#example-finetune-llama2-7b-using-qlora) +or [alpaca example](https://github.com/intel-analytics/BigDL/tree/main/python/llm/example/CPU/QLoRA-FineTuning/alpaca-qlora) +2. [Single node with multiple sockets](https://github.com/intel-analytics/BigDL/tree/main/python/llm/example/CPU/QLoRA-FineTuning/alpaca-qlora#guide-to-finetuning-qlora-on-one-node-with-multiple-sockets) +3. multiple nodes with multiple sockets + ## Example: Finetune llama2-7b using QLoRA This example is ported from [bnb-4bit-training](https://colab.research.google.com/drive/1VoYNfYDKcKRQRor98Zbf2-9VQTtGJ24k). diff --git a/python/llm/example/CPU/QLoRA-FineTuning/alpaca-qlora/README.md b/python/llm/example/CPU/QLoRA-FineTuning/alpaca-qlora/README.md index a256fcc2..b920d66d 100644 --- a/python/llm/example/CPU/QLoRA-FineTuning/alpaca-qlora/README.md +++ b/python/llm/example/CPU/QLoRA-FineTuning/alpaca-qlora/README.md @@ -44,6 +44,20 @@ python ./alpaca_qlora_finetuning_cpu.py \ 1%|█ | 8/1164 [xx:xx