diff --git a/python/llm/example/CPU/Applications/hf-agent/README.md b/python/llm/example/CPU/Applications/hf-agent/README.md index 46bb8aa5..2bfe5d1e 100644 --- a/python/llm/example/CPU/Applications/hf-agent/README.md +++ b/python/llm/example/CPU/Applications/hf-agent/README.md @@ -41,8 +41,8 @@ For optimal performance on server, it is recommended to set several environment E.g. on Linux, ```bash -# set BigDL-Nano env variables -source bigdl-nano-init +# set BigDL-LLM env variables +source bigdl-llm-init # e.g. for a server with 48 cores per socket export OMP_NUM_THREADS=48 diff --git a/python/llm/example/CPU/HF-Transformers-AutoModels/Advanced-Quantizations/AWQ/README.md b/python/llm/example/CPU/HF-Transformers-AutoModels/Advanced-Quantizations/AWQ/README.md index 2785323f..f3f8755b 100644 --- a/python/llm/example/CPU/HF-Transformers-AutoModels/Advanced-Quantizations/AWQ/README.md +++ b/python/llm/example/CPU/HF-Transformers-AutoModels/Advanced-Quantizations/AWQ/README.md @@ -48,7 +48,7 @@ For optimal performance on server, it is recommended to set several environment E.g. on Linux, ```bash -# set BigDL-Nano env variables +# set BigDL-LLM env variables source bigdl-llm-init # e.g. for a server with 48 cores per socket diff --git a/python/llm/example/CPU/HF-Transformers-AutoModels/Advanced-Quantizations/GPTQ/README.md b/python/llm/example/CPU/HF-Transformers-AutoModels/Advanced-Quantizations/GPTQ/README.md index 7b38158f..52b05c3c 100644 --- a/python/llm/example/CPU/HF-Transformers-AutoModels/Advanced-Quantizations/GPTQ/README.md +++ b/python/llm/example/CPU/HF-Transformers-AutoModels/Advanced-Quantizations/GPTQ/README.md @@ -43,8 +43,8 @@ For optimal performance on server, it is recommended to set several environment E.g. on Linux, ```bash -# set BigDL-Nano env variables -source bigdl-nano-init +# set BigDL-LLM env variables +source bigdl-llm-init # e.g. for a server with 48 cores per socket export OMP_NUM_THREADS=48 diff --git a/python/llm/example/CPU/HF-Transformers-AutoModels/Model/README.md b/python/llm/example/CPU/HF-Transformers-AutoModels/Model/README.md index b319f89b..8bb8d592 100644 --- a/python/llm/example/CPU/HF-Transformers-AutoModels/Model/README.md +++ b/python/llm/example/CPU/HF-Transformers-AutoModels/Model/README.md @@ -7,8 +7,8 @@ To run the examples, we recommend using Intel® Xeon® processors (server), or > For OS, BigDL-LLM supports Ubuntu 20.04 or later, CentOS 7 or later, and Windows 10/11. ## Best Known Configuration on Linux -For better performance, it is recommended to set environment variables on Linux with the help of BigDL-Nano: +For better performance, it is recommended to set environment variables on Linux with the help of BigDL-LLM: ```bash -pip install bigdl-nano -source bigdl-nano-init +pip install bigdl-llm +source bigdl-llm-init ``` diff --git a/python/llm/example/CPU/HF-Transformers-AutoModels/Model/aquila/README.md b/python/llm/example/CPU/HF-Transformers-AutoModels/Model/aquila/README.md index 2e4a81d0..0a5914ed 100644 --- a/python/llm/example/CPU/HF-Transformers-AutoModels/Model/aquila/README.md +++ b/python/llm/example/CPU/HF-Transformers-AutoModels/Model/aquila/README.md @@ -41,8 +41,8 @@ For optimal performance on server, it is recommended to set several environment E.g. on Linux, ```bash -# set BigDL-Nano env variables -source bigdl-nano-init +# set BigDL-LLM env variables +source bigdl-llm-init # e.g. for a server with 48 cores per socket export OMP_NUM_THREADS=48 diff --git a/python/llm/example/CPU/HF-Transformers-AutoModels/Model/aquila2/README.md b/python/llm/example/CPU/HF-Transformers-AutoModels/Model/aquila2/README.md index cbdd560a..07bb91b5 100644 --- a/python/llm/example/CPU/HF-Transformers-AutoModels/Model/aquila2/README.md +++ b/python/llm/example/CPU/HF-Transformers-AutoModels/Model/aquila2/README.md @@ -41,8 +41,8 @@ For optimal performance on server, it is recommended to set several environment E.g. on Linux, ```bash -# set BigDL-Nano env variables -source bigdl-nano-init +# set BigDL-LLM env variables +source bigdl-llm-init # e.g. for a server with 48 cores per socket export OMP_NUM_THREADS=48 diff --git a/python/llm/example/CPU/HF-Transformers-AutoModels/Model/baichuan/README.md b/python/llm/example/CPU/HF-Transformers-AutoModels/Model/baichuan/README.md index 34777642..4f15618b 100644 --- a/python/llm/example/CPU/HF-Transformers-AutoModels/Model/baichuan/README.md +++ b/python/llm/example/CPU/HF-Transformers-AutoModels/Model/baichuan/README.md @@ -41,8 +41,8 @@ For optimal performance on server, it is recommended to set several environment E.g. on Linux, ```bash -# set BigDL-Nano env variables -source bigdl-nano-init +# set BigDL-LLM env variables +source bigdl-llm-init # e.g. for a server with 48 cores per socket export OMP_NUM_THREADS=48 diff --git a/python/llm/example/CPU/HF-Transformers-AutoModels/Model/baichuan2/README.md b/python/llm/example/CPU/HF-Transformers-AutoModels/Model/baichuan2/README.md index 79327abf..ff0347b7 100644 --- a/python/llm/example/CPU/HF-Transformers-AutoModels/Model/baichuan2/README.md +++ b/python/llm/example/CPU/HF-Transformers-AutoModels/Model/baichuan2/README.md @@ -41,8 +41,8 @@ For optimal performance on server, it is recommended to set several environment E.g. on Linux, ```bash -# set BigDL-Nano env variables -source bigdl-nano-init +# set BigDL-LLM env variables +source bigdl-llm-init # e.g. for a server with 48 cores per socket export OMP_NUM_THREADS=48 diff --git a/python/llm/example/CPU/HF-Transformers-AutoModels/Model/chatglm/README.md b/python/llm/example/CPU/HF-Transformers-AutoModels/Model/chatglm/README.md index 861ee12a..f549ae78 100644 --- a/python/llm/example/CPU/HF-Transformers-AutoModels/Model/chatglm/README.md +++ b/python/llm/example/CPU/HF-Transformers-AutoModels/Model/chatglm/README.md @@ -41,8 +41,8 @@ For optimal performance on server, it is recommended to set several environment E.g. on Linux, ```bash -# set BigDL-Nano env variables -source bigdl-nano-init +# set BigDL-LLM env variables +source bigdl-llm-init # e.g. for a server with 48 cores per socket export OMP_NUM_THREADS=48 diff --git a/python/llm/example/CPU/HF-Transformers-AutoModels/Model/chatglm2/README.md b/python/llm/example/CPU/HF-Transformers-AutoModels/Model/chatglm2/README.md index ff9b051b..969db1f8 100644 --- a/python/llm/example/CPU/HF-Transformers-AutoModels/Model/chatglm2/README.md +++ b/python/llm/example/CPU/HF-Transformers-AutoModels/Model/chatglm2/README.md @@ -41,8 +41,8 @@ For optimal performance on server, it is recommended to set several environment E.g. on Linux, ```bash -# set BigDL-Nano env variables -source bigdl-nano-init +# set BigDL-LLM env variables +source bigdl-llm-init # e.g. for a server with 48 cores per socket export OMP_NUM_THREADS=48 @@ -118,8 +118,8 @@ For optimal performance on server, it is recommended to set several environment E.g. on Linux, ```bash -# set BigDL-Nano env variables -source bigdl-nano-init +# set BigDL-LLM env variables +source bigdl-llm-init # e.g. for a server with 48 cores per socket export OMP_NUM_THREADS=48 diff --git a/python/llm/example/CPU/HF-Transformers-AutoModels/Model/chatglm3/README.md b/python/llm/example/CPU/HF-Transformers-AutoModels/Model/chatglm3/README.md index a1df8bff..93e8c9c1 100644 --- a/python/llm/example/CPU/HF-Transformers-AutoModels/Model/chatglm3/README.md +++ b/python/llm/example/CPU/HF-Transformers-AutoModels/Model/chatglm3/README.md @@ -41,8 +41,8 @@ For optimal performance on server, it is recommended to set several environment E.g. on Linux, ```bash -# set BigDL-Nano env variables -source bigdl-nano-init +# set BigDL-LLM env variables +source bigdl-llm-init # e.g. for a server with 48 cores per socket export OMP_NUM_THREADS=48 @@ -119,8 +119,8 @@ For optimal performance on server, it is recommended to set several environment E.g. on Linux, ```bash -# set BigDL-Nano env variables -source bigdl-nano-init +# set BigDL-LLM env variables +source bigdl-llm-init # e.g. for a server with 48 cores per socket export OMP_NUM_THREADS=48 diff --git a/python/llm/example/CPU/HF-Transformers-AutoModels/Model/codellama/README.md b/python/llm/example/CPU/HF-Transformers-AutoModels/Model/codellama/README.md index 8fc01957..69eb1f3f 100644 --- a/python/llm/example/CPU/HF-Transformers-AutoModels/Model/codellama/README.md +++ b/python/llm/example/CPU/HF-Transformers-AutoModels/Model/codellama/README.md @@ -40,8 +40,8 @@ For optimal performance on server, it is recommended to set several environment E.g. on Linux, ```bash -# set BigDL-Nano env variables -source bigdl-nano-init +# set BigDL-LLM env variables +source bigdl-llm-init # e.g. for a server with 48 cores per socket export OMP_NUM_THREADS=48 diff --git a/python/llm/example/CPU/HF-Transformers-AutoModels/Model/codeshell/README.md b/python/llm/example/CPU/HF-Transformers-AutoModels/Model/codeshell/README.md index 24f2be64..9b2d976d 100644 --- a/python/llm/example/CPU/HF-Transformers-AutoModels/Model/codeshell/README.md +++ b/python/llm/example/CPU/HF-Transformers-AutoModels/Model/codeshell/README.md @@ -41,8 +41,8 @@ For optimal performance on server, it is recommended to set several environment E.g. on Linux, ```bash -# set BigDL-Nano env variables -source bigdl-nano-init +# set BigDL-LLM env variables +source bigdl-llm-init # e.g. for a server with 48 cores per socket export OMP_NUM_THREADS=48 diff --git a/python/llm/example/CPU/HF-Transformers-AutoModels/Model/distil-whisper/README.md b/python/llm/example/CPU/HF-Transformers-AutoModels/Model/distil-whisper/README.md index d04de208..35086459 100644 --- a/python/llm/example/CPU/HF-Transformers-AutoModels/Model/distil-whisper/README.md +++ b/python/llm/example/CPU/HF-Transformers-AutoModels/Model/distil-whisper/README.md @@ -47,8 +47,8 @@ For optimal performance on server, it is recommended to set several environment E.g. on Linux, ```bash -# set BigDL-Nano env variables -source bigdl-nano-init +# set BigDL-LLM env variables +source bigdl-llm-init # e.g. for a server with 48 cores per socket export OMP_NUM_THREADS=48 diff --git a/python/llm/example/CPU/HF-Transformers-AutoModels/Model/dolly_v1/README.md b/python/llm/example/CPU/HF-Transformers-AutoModels/Model/dolly_v1/README.md index ddc16e07..684c3f81 100644 --- a/python/llm/example/CPU/HF-Transformers-AutoModels/Model/dolly_v1/README.md +++ b/python/llm/example/CPU/HF-Transformers-AutoModels/Model/dolly_v1/README.md @@ -40,8 +40,8 @@ For optimal performance on server, it is recommended to set several environment E.g. on Linux, ```bash -# set BigDL-Nano env variables -source bigdl-nano-init +# set BigDL-LLM env variables +source bigdl-llm-init # e.g. for a server with 48 cores per socket export OMP_NUM_THREADS=48 diff --git a/python/llm/example/CPU/HF-Transformers-AutoModels/Model/dolly_v2/README.md b/python/llm/example/CPU/HF-Transformers-AutoModels/Model/dolly_v2/README.md index 7ef54e16..26a66020 100644 --- a/python/llm/example/CPU/HF-Transformers-AutoModels/Model/dolly_v2/README.md +++ b/python/llm/example/CPU/HF-Transformers-AutoModels/Model/dolly_v2/README.md @@ -40,8 +40,8 @@ For optimal performance on server, it is recommended to set several environment E.g. on Linux, ```bash -# set BigDL-Nano env variables -source bigdl-nano-init +# set BigDL-LLM env variables +source bigdl-llm-init # e.g. for a server with 48 cores per socket export OMP_NUM_THREADS=48 diff --git a/python/llm/example/CPU/HF-Transformers-AutoModels/Model/falcon/README.md b/python/llm/example/CPU/HF-Transformers-AutoModels/Model/falcon/README.md index eb4aacb5..3c008f8f 100644 --- a/python/llm/example/CPU/HF-Transformers-AutoModels/Model/falcon/README.md +++ b/python/llm/example/CPU/HF-Transformers-AutoModels/Model/falcon/README.md @@ -70,8 +70,8 @@ For optimal performance on server, it is recommended to set several environment E.g. on Linux, ```bash -# set BigDL-Nano env variables -source bigdl-nano-init +# set BigDL-LLM env variables +source bigdl-llm-init # e.g. for a server with 48 cores per socket export OMP_NUM_THREADS=48 diff --git a/python/llm/example/CPU/HF-Transformers-AutoModels/Model/flan-t5/README.md b/python/llm/example/CPU/HF-Transformers-AutoModels/Model/flan-t5/README.md index 0b6ab57a..9af2109e 100644 --- a/python/llm/example/CPU/HF-Transformers-AutoModels/Model/flan-t5/README.md +++ b/python/llm/example/CPU/HF-Transformers-AutoModels/Model/flan-t5/README.md @@ -37,8 +37,8 @@ For optimal performance on server, it is recommended to set several environment E.g. on Linux, ```bash -# set BigDL-Nano env variables -source bigdl-nano-init +# set BigDL-LLM env variables +source bigdl-llm-init # e.g. for a server with 48 cores per socket export OMP_NUM_THREADS=48 diff --git a/python/llm/example/CPU/HF-Transformers-AutoModels/Model/fuyu/README.md b/python/llm/example/CPU/HF-Transformers-AutoModels/Model/fuyu/README.md index 410c703c..5121474e 100644 --- a/python/llm/example/CPU/HF-Transformers-AutoModels/Model/fuyu/README.md +++ b/python/llm/example/CPU/HF-Transformers-AutoModels/Model/fuyu/README.md @@ -38,8 +38,8 @@ For optimal performance on server, it is recommended to set several environment E.g. on Linux, ```bash -# set BigDL-Nano env variables -source bigdl-nano-init +# set BigDL-LLM env variables +source bigdl-llm-init # e.g. for a server with 48 cores per socket export OMP_NUM_THREADS=48 diff --git a/python/llm/example/CPU/HF-Transformers-AutoModels/Model/internlm-xcomposer/README.md b/python/llm/example/CPU/HF-Transformers-AutoModels/Model/internlm-xcomposer/README.md index f0783ed2..0b608b34 100644 --- a/python/llm/example/CPU/HF-Transformers-AutoModels/Model/internlm-xcomposer/README.md +++ b/python/llm/example/CPU/HF-Transformers-AutoModels/Model/internlm-xcomposer/README.md @@ -59,8 +59,8 @@ For optimal performance on server, it is recommended to set several environment E.g. on Linux, ```bash -# set BigDL-Nano env variables -source bigdl-nano-init +# set BigDL-LLM env variables +source bigdl-llm-init # e.g. for a server with 48 cores per socket export OMP_NUM_THREADS=48 diff --git a/python/llm/example/CPU/HF-Transformers-AutoModels/Model/internlm/README.md b/python/llm/example/CPU/HF-Transformers-AutoModels/Model/internlm/README.md index d9b77293..3f73833b 100644 --- a/python/llm/example/CPU/HF-Transformers-AutoModels/Model/internlm/README.md +++ b/python/llm/example/CPU/HF-Transformers-AutoModels/Model/internlm/README.md @@ -41,8 +41,8 @@ For optimal performance on server, it is recommended to set several environment E.g. on Linux, ```bash -# set BigDL-Nano env variables -source bigdl-nano-init +# set BigDL-LLM env variables +source bigdl-llm-init # e.g. for a server with 48 cores per socket export OMP_NUM_THREADS=48 diff --git a/python/llm/example/CPU/HF-Transformers-AutoModels/Model/llama2/README.md b/python/llm/example/CPU/HF-Transformers-AutoModels/Model/llama2/README.md index dc61777c..e4c7b899 100644 --- a/python/llm/example/CPU/HF-Transformers-AutoModels/Model/llama2/README.md +++ b/python/llm/example/CPU/HF-Transformers-AutoModels/Model/llama2/README.md @@ -40,8 +40,8 @@ For optimal performance on server, it is recommended to set several environment E.g. on Linux, ```bash -# set BigDL-Nano env variables -source bigdl-nano-init +# set BigDL-LLM env variables +source bigdl-llm-init # e.g. for a server with 48 cores per socket export OMP_NUM_THREADS=48 diff --git a/python/llm/example/CPU/HF-Transformers-AutoModels/Model/mistral/README.md b/python/llm/example/CPU/HF-Transformers-AutoModels/Model/mistral/README.md index dc37c835..13e80deb 100644 --- a/python/llm/example/CPU/HF-Transformers-AutoModels/Model/mistral/README.md +++ b/python/llm/example/CPU/HF-Transformers-AutoModels/Model/mistral/README.md @@ -41,8 +41,8 @@ For optimal performance on server, it is recommended to set several environment E.g. on Linux, ```bash -# set BigDL-Nano env variables -source bigdl-nano-init +# set BigDL-LLM env variables +source bigdl-llm-init # e.g. for a server with 48 cores per socket export OMP_NUM_THREADS=48 diff --git a/python/llm/example/CPU/HF-Transformers-AutoModels/Model/moss/README.md b/python/llm/example/CPU/HF-Transformers-AutoModels/Model/moss/README.md index 2e649f53..298a5bfa 100644 --- a/python/llm/example/CPU/HF-Transformers-AutoModels/Model/moss/README.md +++ b/python/llm/example/CPU/HF-Transformers-AutoModels/Model/moss/README.md @@ -41,8 +41,8 @@ For optimal performance on server, it is recommended to set several environment E.g. on Linux, ```bash -# set BigDL-Nano env variables -source bigdl-nano-init +# set BigDL-LLM env variables +source bigdl-llm-init # e.g. for a server with 48 cores per socket export OMP_NUM_THREADS=48 diff --git a/python/llm/example/CPU/HF-Transformers-AutoModels/Model/mpt/README.md b/python/llm/example/CPU/HF-Transformers-AutoModels/Model/mpt/README.md index b082bde7..aee9f9a6 100644 --- a/python/llm/example/CPU/HF-Transformers-AutoModels/Model/mpt/README.md +++ b/python/llm/example/CPU/HF-Transformers-AutoModels/Model/mpt/README.md @@ -41,8 +41,8 @@ For optimal performance on server, it is recommended to set several environment E.g. on Linux, ```bash -# set BigDL-Nano env variables -source bigdl-nano-init +# set BigDL-LLM env variables +source bigdl-llm-init # e.g. for a server with 48 cores per socket export OMP_NUM_THREADS=48 diff --git a/python/llm/example/CPU/HF-Transformers-AutoModels/Model/phi-1_5/README.md b/python/llm/example/CPU/HF-Transformers-AutoModels/Model/phi-1_5/README.md index 013704c8..a6f6017b 100644 --- a/python/llm/example/CPU/HF-Transformers-AutoModels/Model/phi-1_5/README.md +++ b/python/llm/example/CPU/HF-Transformers-AutoModels/Model/phi-1_5/README.md @@ -42,8 +42,8 @@ For optimal performance on server, it is recommended to set several environment E.g. on Linux, ```bash -# set BigDL-Nano env variables -source bigdl-nano-init +# set BigDL-LLM env variables +source bigdl-llm-init # e.g. for a server with 48 cores per socket export OMP_NUM_THREADS=48 diff --git a/python/llm/example/CPU/HF-Transformers-AutoModels/Model/phoenix/README.md b/python/llm/example/CPU/HF-Transformers-AutoModels/Model/phoenix/README.md index 5efdd658..accd6016 100644 --- a/python/llm/example/CPU/HF-Transformers-AutoModels/Model/phoenix/README.md +++ b/python/llm/example/CPU/HF-Transformers-AutoModels/Model/phoenix/README.md @@ -41,8 +41,8 @@ For optimal performance on server, it is recommended to set several environment E.g. on Linux, ```bash -# set BigDL-Nano env variables -source bigdl-nano-init +# set BigDL-LLM env variables +source bigdl-llm-init # e.g. for a server with 48 cores per socket export OMP_NUM_THREADS=48 diff --git a/python/llm/example/CPU/HF-Transformers-AutoModels/Model/qwen-vl/README.md b/python/llm/example/CPU/HF-Transformers-AutoModels/Model/qwen-vl/README.md index b65be069..fb46bd56 100644 --- a/python/llm/example/CPU/HF-Transformers-AutoModels/Model/qwen-vl/README.md +++ b/python/llm/example/CPU/HF-Transformers-AutoModels/Model/qwen-vl/README.md @@ -35,8 +35,8 @@ For optimal performance on server, it is recommended to set several environment E.g. on Linux, ```bash -# set BigDL-Nano env variables -source bigdl-nano-init +# set BigDL-LLM env variables +source bigdl-llm-init # e.g. for a server with 48 cores per socket export OMP_NUM_THREADS=48 diff --git a/python/llm/example/CPU/HF-Transformers-AutoModels/Model/qwen/README.md b/python/llm/example/CPU/HF-Transformers-AutoModels/Model/qwen/README.md index d2f1c97f..14ccec88 100644 --- a/python/llm/example/CPU/HF-Transformers-AutoModels/Model/qwen/README.md +++ b/python/llm/example/CPU/HF-Transformers-AutoModels/Model/qwen/README.md @@ -41,8 +41,8 @@ For optimal performance on server, it is recommended to set several environment E.g. on Linux, ```bash -# set BigDL-Nano env variables -source bigdl-nano-init +# set BigDL-LLM env variables +source bigdl-llm-init # e.g. for a server with 48 cores per socket export OMP_NUM_THREADS=48 diff --git a/python/llm/example/CPU/HF-Transformers-AutoModels/Model/redpajama/README.md b/python/llm/example/CPU/HF-Transformers-AutoModels/Model/redpajama/README.md index f4ef73a3..003e76d5 100644 --- a/python/llm/example/CPU/HF-Transformers-AutoModels/Model/redpajama/README.md +++ b/python/llm/example/CPU/HF-Transformers-AutoModels/Model/redpajama/README.md @@ -41,8 +41,8 @@ For optimal performance on server, it is recommended to set several environment E.g. on Linux, ```bash -# set BigDL-Nano env variables -source bigdl-nano-init +# set BigDL-LLM env variables +source bigdl-llm-init # e.g. for a server with 48 cores per socket export OMP_NUM_THREADS=48 diff --git a/python/llm/example/CPU/HF-Transformers-AutoModels/Model/replit/README.md b/python/llm/example/CPU/HF-Transformers-AutoModels/Model/replit/README.md index b191e32d..b25e602d 100644 --- a/python/llm/example/CPU/HF-Transformers-AutoModels/Model/replit/README.md +++ b/python/llm/example/CPU/HF-Transformers-AutoModels/Model/replit/README.md @@ -32,8 +32,8 @@ For optimal performance on server, it is recommended to set several environment E.g. on Linux, ```bash -# set BigDL-Nano env variables -source bigdl-nano-init +# set BigDL-LLM env variables +source bigdl-llm-init # e.g. for a server with 48 cores per socket export OMP_NUM_THREADS=48 diff --git a/python/llm/example/CPU/HF-Transformers-AutoModels/Model/skywork/README.md b/python/llm/example/CPU/HF-Transformers-AutoModels/Model/skywork/README.md index 7e0ab293..a717b931 100644 --- a/python/llm/example/CPU/HF-Transformers-AutoModels/Model/skywork/README.md +++ b/python/llm/example/CPU/HF-Transformers-AutoModels/Model/skywork/README.md @@ -40,8 +40,8 @@ For optimal performance on server, it is recommended to set several environment E.g. on Linux, ```bash -# set BigDL-Nano env variables -source bigdl-nano-init +# set BigDL-LLM env variables +source bigdl-llm-init # e.g. for a server with 48 cores per socket export OMP_NUM_THREADS=48 diff --git a/python/llm/example/CPU/HF-Transformers-AutoModels/Model/starcoder/README.md b/python/llm/example/CPU/HF-Transformers-AutoModels/Model/starcoder/README.md index a0d24928..1509f5cb 100644 --- a/python/llm/example/CPU/HF-Transformers-AutoModels/Model/starcoder/README.md +++ b/python/llm/example/CPU/HF-Transformers-AutoModels/Model/starcoder/README.md @@ -40,8 +40,8 @@ For optimal performance on server, it is recommended to set several environment E.g. on Linux, ```bash -# set BigDL-Nano env variables -source bigdl-nano-init +# set BigDL-LLM env variables +source bigdl-llm-init # e.g. for a server with 48 cores per socket export OMP_NUM_THREADS=48 diff --git a/python/llm/example/CPU/HF-Transformers-AutoModels/Model/vicuna/README.md b/python/llm/example/CPU/HF-Transformers-AutoModels/Model/vicuna/README.md index 3fd95265..3abd2c43 100644 --- a/python/llm/example/CPU/HF-Transformers-AutoModels/Model/vicuna/README.md +++ b/python/llm/example/CPU/HF-Transformers-AutoModels/Model/vicuna/README.md @@ -40,8 +40,8 @@ For optimal performance on server, it is recommended to set several environment E.g. on Linux, ```bash -# set BigDL-Nano env variables -source bigdl-nano-init +# set BigDL-LLM env variables +source bigdl-llm-init # e.g. for a server with 48 cores per socket export OMP_NUM_THREADS=48 diff --git a/python/llm/example/CPU/HF-Transformers-AutoModels/Model/whisper/readme.md b/python/llm/example/CPU/HF-Transformers-AutoModels/Model/whisper/readme.md index ac08a094..5dca740f 100644 --- a/python/llm/example/CPU/HF-Transformers-AutoModels/Model/whisper/readme.md +++ b/python/llm/example/CPU/HF-Transformers-AutoModels/Model/whisper/readme.md @@ -43,8 +43,8 @@ For optimal performance on server, it is recommended to set several environment E.g. on Linux, ```bash -# set BigDL-Nano env variables -source bigdl-nano-init +# set BigDL-LLM env variables +source bigdl-llm-init # e.g. for a server with 48 cores per socket export OMP_NUM_THREADS=48 @@ -102,8 +102,8 @@ For optimal performance on server, it is recommended to set several environment E.g. on Linux, ```bash -# set BigDL-Nano env variables -source bigdl-nano-init +# set BigDL-LLM env variables +source bigdl-llm-init # e.g. long segment recognize for a server with 48 cores per socket export OMP_NUM_THREADS=48 diff --git a/python/llm/example/CPU/HF-Transformers-AutoModels/Model/wizardcoder-python/README.md b/python/llm/example/CPU/HF-Transformers-AutoModels/Model/wizardcoder-python/README.md index 21d08a13..65d052a0 100644 --- a/python/llm/example/CPU/HF-Transformers-AutoModels/Model/wizardcoder-python/README.md +++ b/python/llm/example/CPU/HF-Transformers-AutoModels/Model/wizardcoder-python/README.md @@ -40,8 +40,8 @@ For optimal performance on server, it is recommended to set several environment E.g. on Linux, ```bash -# set BigDL-Nano env variables -source bigdl-nano-init +# set BigDL-LLM env variables +source bigdl-llm-init # e.g. for a server with 48 cores per socket export OMP_NUM_THREADS=48 diff --git a/python/llm/example/CPU/HF-Transformers-AutoModels/Model/yi/README.md b/python/llm/example/CPU/HF-Transformers-AutoModels/Model/yi/README.md index 79a40962..2ac2b82e 100644 --- a/python/llm/example/CPU/HF-Transformers-AutoModels/Model/yi/README.md +++ b/python/llm/example/CPU/HF-Transformers-AutoModels/Model/yi/README.md @@ -43,8 +43,8 @@ For optimal performance on server, it is recommended to set several environment E.g. on Linux, ```bash -# set BigDL-Nano env variables -source bigdl-nano-init +# set BigDL-LLM env variables +source bigdl-llm-init # e.g. for a server with 48 cores per socket export OMP_NUM_THREADS=48 diff --git a/python/llm/example/CPU/PyTorch-Models/Model/README.md b/python/llm/example/CPU/PyTorch-Models/Model/README.md index b89ad944..0a09bf2e 100644 --- a/python/llm/example/CPU/PyTorch-Models/Model/README.md +++ b/python/llm/example/CPU/PyTorch-Models/Model/README.md @@ -7,8 +7,8 @@ To run the examples, we recommend using Intel® Xeon® processors (server), or > For OS, BigDL-LLM supports Ubuntu 20.04 or later, CentOS 7 or later, and Windows 10/11. ## Best Known Configuration on Linux -For better performance, it is recommended to set environment variables on Linux with the help of BigDL-Nano: +For better performance, it is recommended to set environment variables on Linux with the help of BigDL-LLM: ```bash -pip install bigdl-nano -source bigdl-nano-init +pip install bigdl-llm +source bigdl-llm-init ``` diff --git a/python/llm/example/CPU/PyTorch-Models/Model/aquila2/README.md b/python/llm/example/CPU/PyTorch-Models/Model/aquila2/README.md index 55f9ed11..82c8c15d 100644 --- a/python/llm/example/CPU/PyTorch-Models/Model/aquila2/README.md +++ b/python/llm/example/CPU/PyTorch-Models/Model/aquila2/README.md @@ -32,8 +32,8 @@ For optimal performance on server, it is recommended to set several environment E.g. on Linux, ```bash -# set BigDL-Nano env variables -source bigdl-nano-init +# set BigDL-LLM env variables +source bigdl-llm-init # e.g. for a server with 48 cores per socket export OMP_NUM_THREADS=48 diff --git a/python/llm/example/CPU/PyTorch-Models/Model/bark/README.md b/python/llm/example/CPU/PyTorch-Models/Model/bark/README.md index 4be37f75..da660e7c 100644 --- a/python/llm/example/CPU/PyTorch-Models/Model/bark/README.md +++ b/python/llm/example/CPU/PyTorch-Models/Model/bark/README.md @@ -45,8 +45,8 @@ For optimal performance on server, it is recommended to set several environment E.g. on Linux, ```bash -# set BigDL-Nano env variables -source bigdl-nano-init +# set BigDL-LLM env variables +source bigdl-llm-init # e.g. for a server with 48 cores per socket export OMP_NUM_THREADS=48 diff --git a/python/llm/example/CPU/PyTorch-Models/Model/bert/README.md b/python/llm/example/CPU/PyTorch-Models/Model/bert/README.md index 522d1cb8..d5a093f0 100644 --- a/python/llm/example/CPU/PyTorch-Models/Model/bert/README.md +++ b/python/llm/example/CPU/PyTorch-Models/Model/bert/README.md @@ -32,8 +32,8 @@ For optimal performance on server, it is recommended to set several environment E.g. on Linux, ```bash -# set BigDL-Nano env variables -source bigdl-nano-init +# set BigDL-LLM env variables +source bigdl-llm-init # e.g. for a server with 48 cores per socket export OMP_NUM_THREADS=48 diff --git a/python/llm/example/CPU/PyTorch-Models/Model/chatglm/README.md b/python/llm/example/CPU/PyTorch-Models/Model/chatglm/README.md index c71bd436..eeec312a 100644 --- a/python/llm/example/CPU/PyTorch-Models/Model/chatglm/README.md +++ b/python/llm/example/CPU/PyTorch-Models/Model/chatglm/README.md @@ -32,8 +32,8 @@ For optimal performance on server, it is recommended to set several environment E.g. on Linux, ```bash -# set BigDL-Nano env variables -source bigdl-nano-init +# set BigDL-LLM env variables +source bigdl-llm-init # e.g. for a server with 48 cores per socket export OMP_NUM_THREADS=48 diff --git a/python/llm/example/CPU/PyTorch-Models/Model/chatglm3/README.md b/python/llm/example/CPU/PyTorch-Models/Model/chatglm3/README.md index b41bbbe1..56c589fa 100644 --- a/python/llm/example/CPU/PyTorch-Models/Model/chatglm3/README.md +++ b/python/llm/example/CPU/PyTorch-Models/Model/chatglm3/README.md @@ -32,8 +32,8 @@ For optimal performance on server, it is recommended to set several environment E.g. on Linux, ```bash -# set BigDL-Nano env variables -source bigdl-nano-init +# set BigDL-LLM env variables +source bigdl-llm-init # e.g. for a server with 48 cores per socket export OMP_NUM_THREADS=48 diff --git a/python/llm/example/CPU/PyTorch-Models/Model/codellama/README.md b/python/llm/example/CPU/PyTorch-Models/Model/codellama/README.md index b10746f1..8a788ebc 100644 --- a/python/llm/example/CPU/PyTorch-Models/Model/codellama/README.md +++ b/python/llm/example/CPU/PyTorch-Models/Model/codellama/README.md @@ -32,8 +32,8 @@ For optimal performance on server, it is recommended to set several environment E.g. on Linux, ```bash -# set BigDL-Nano env variables -source bigdl-nano-init +# set BigDL-LLM env variables +source bigdl-llm-init # e.g. for a server with 48 cores per socket export OMP_NUM_THREADS=48 diff --git a/python/llm/example/CPU/PyTorch-Models/Model/codeshell/README.md b/python/llm/example/CPU/PyTorch-Models/Model/codeshell/README.md index de580297..d00b2d17 100644 --- a/python/llm/example/CPU/PyTorch-Models/Model/codeshell/README.md +++ b/python/llm/example/CPU/PyTorch-Models/Model/codeshell/README.md @@ -32,8 +32,8 @@ For optimal performance on server, it is recommended to set several environment E.g. on Linux, ```bash -# set BigDL-Nano env variables -source bigdl-nano-init +# set BigDL-LLM env variables +source bigdl-llm-init # e.g. for a server with 48 cores per socket export OMP_NUM_THREADS=48 diff --git a/python/llm/example/CPU/PyTorch-Models/Model/distil-whisper/README.md b/python/llm/example/CPU/PyTorch-Models/Model/distil-whisper/README.md index 0c3df9ca..9aeed650 100644 --- a/python/llm/example/CPU/PyTorch-Models/Model/distil-whisper/README.md +++ b/python/llm/example/CPU/PyTorch-Models/Model/distil-whisper/README.md @@ -38,8 +38,8 @@ For optimal performance on server, it is recommended to set several environment E.g. on Linux, ```bash -# set BigDL-Nano env variables -source bigdl-nano-init +# set BigDL-LLM env variables +source bigdl-llm-init # e.g. for a server with 48 cores per socket export OMP_NUM_THREADS=48 diff --git a/python/llm/example/CPU/PyTorch-Models/Model/flan-t5/README.md b/python/llm/example/CPU/PyTorch-Models/Model/flan-t5/README.md index 0b6ab57a..9af2109e 100644 --- a/python/llm/example/CPU/PyTorch-Models/Model/flan-t5/README.md +++ b/python/llm/example/CPU/PyTorch-Models/Model/flan-t5/README.md @@ -37,8 +37,8 @@ For optimal performance on server, it is recommended to set several environment E.g. on Linux, ```bash -# set BigDL-Nano env variables -source bigdl-nano-init +# set BigDL-LLM env variables +source bigdl-llm-init # e.g. for a server with 48 cores per socket export OMP_NUM_THREADS=48 diff --git a/python/llm/example/CPU/PyTorch-Models/Model/fuyu/README.md b/python/llm/example/CPU/PyTorch-Models/Model/fuyu/README.md index 410c703c..5121474e 100644 --- a/python/llm/example/CPU/PyTorch-Models/Model/fuyu/README.md +++ b/python/llm/example/CPU/PyTorch-Models/Model/fuyu/README.md @@ -38,8 +38,8 @@ For optimal performance on server, it is recommended to set several environment E.g. on Linux, ```bash -# set BigDL-Nano env variables -source bigdl-nano-init +# set BigDL-LLM env variables +source bigdl-llm-init # e.g. for a server with 48 cores per socket export OMP_NUM_THREADS=48 diff --git a/python/llm/example/CPU/PyTorch-Models/Model/internlm-xcomposer/README.md b/python/llm/example/CPU/PyTorch-Models/Model/internlm-xcomposer/README.md index a96db75d..05ef602e 100644 --- a/python/llm/example/CPU/PyTorch-Models/Model/internlm-xcomposer/README.md +++ b/python/llm/example/CPU/PyTorch-Models/Model/internlm-xcomposer/README.md @@ -59,8 +59,8 @@ For optimal performance on server, it is recommended to set several environment E.g. on Linux, ```bash -# set BigDL-Nano env variables -source bigdl-nano-init +# set BigDL-LLM env variables +source bigdl-llm-init # e.g. for a server with 48 cores per socket export OMP_NUM_THREADS=48 diff --git a/python/llm/example/CPU/PyTorch-Models/Model/llama2/README.md b/python/llm/example/CPU/PyTorch-Models/Model/llama2/README.md index fde21def..65c7947a 100644 --- a/python/llm/example/CPU/PyTorch-Models/Model/llama2/README.md +++ b/python/llm/example/CPU/PyTorch-Models/Model/llama2/README.md @@ -32,8 +32,8 @@ For optimal performance on server, it is recommended to set several environment E.g. on Linux, ```bash -# set BigDL-Nano env variables -source bigdl-nano-init +# set BigDL-LLM env variables +source bigdl-llm-init # e.g. for a server with 48 cores per socket export OMP_NUM_THREADS=48 diff --git a/python/llm/example/CPU/PyTorch-Models/Model/llava/README.md b/python/llm/example/CPU/PyTorch-Models/Model/llava/README.md index dc59435c..32724ffb 100644 --- a/python/llm/example/CPU/PyTorch-Models/Model/llava/README.md +++ b/python/llm/example/CPU/PyTorch-Models/Model/llava/README.md @@ -44,8 +44,8 @@ For optimal performance on server, it is recommended to set several environment E.g. on Linux, ```bash -# set BigDL-Nano env variables -source bigdl-nano-init +# set BigDL-LLM env variables +source bigdl-llm-init # e.g. for a server with 48 cores per socket export OMP_NUM_THREADS=48 diff --git a/python/llm/example/CPU/PyTorch-Models/Model/mistral/README.md b/python/llm/example/CPU/PyTorch-Models/Model/mistral/README.md index 900815d0..39ad20bd 100644 --- a/python/llm/example/CPU/PyTorch-Models/Model/mistral/README.md +++ b/python/llm/example/CPU/PyTorch-Models/Model/mistral/README.md @@ -37,8 +37,8 @@ For optimal performance on server, it is recommended to set several environment E.g. on Linux, ```bash -# set BigDL-Nano env variables -source bigdl-nano-init +# set BigDL-LLM env variables +source bigdl-llm-init # e.g. for a server with 48 cores per socket export OMP_NUM_THREADS=48 diff --git a/python/llm/example/CPU/PyTorch-Models/Model/openai-whisper/readme.md b/python/llm/example/CPU/PyTorch-Models/Model/openai-whisper/readme.md index 148f603a..8fa618fc 100644 --- a/python/llm/example/CPU/PyTorch-Models/Model/openai-whisper/readme.md +++ b/python/llm/example/CPU/PyTorch-Models/Model/openai-whisper/readme.md @@ -44,8 +44,8 @@ For optimal performance on server, it is recommended to set several environment E.g. on Linux, ```bash -# set BigDL-Nano env variables -source bigdl-nano-init +# set BigDL-LLM env variables +source bigdl-llm-init # e.g. for a server with 48 cores per socket export OMP_NUM_THREADS=48 diff --git a/python/llm/example/CPU/PyTorch-Models/Model/phi-1_5/README.md b/python/llm/example/CPU/PyTorch-Models/Model/phi-1_5/README.md index f925d0d6..f870838c 100644 --- a/python/llm/example/CPU/PyTorch-Models/Model/phi-1_5/README.md +++ b/python/llm/example/CPU/PyTorch-Models/Model/phi-1_5/README.md @@ -33,8 +33,8 @@ For optimal performance on server, it is recommended to set several environment E.g. on Linux, ```bash -# set BigDL-Nano env variables -source bigdl-nano-init +# set BigDL-LLM env variables +source bigdl-llm-init # e.g. for a server with 48 cores per socket export OMP_NUM_THREADS=48 diff --git a/python/llm/example/CPU/PyTorch-Models/Model/qwen-vl/README.md b/python/llm/example/CPU/PyTorch-Models/Model/qwen-vl/README.md index 17af8bf8..83acd7d3 100644 --- a/python/llm/example/CPU/PyTorch-Models/Model/qwen-vl/README.md +++ b/python/llm/example/CPU/PyTorch-Models/Model/qwen-vl/README.md @@ -35,8 +35,8 @@ For optimal performance on server, it is recommended to set several environment E.g. on Linux, ```bash -# set BigDL-Nano env variables -source bigdl-nano-init +# set BigDL-LLM env variables +source bigdl-llm-init # e.g. for a server with 48 cores per socket export OMP_NUM_THREADS=48 diff --git a/python/llm/example/CPU/PyTorch-Models/Model/skywork/README.md b/python/llm/example/CPU/PyTorch-Models/Model/skywork/README.md index 70834a77..9b1f7d8f 100644 --- a/python/llm/example/CPU/PyTorch-Models/Model/skywork/README.md +++ b/python/llm/example/CPU/PyTorch-Models/Model/skywork/README.md @@ -32,8 +32,8 @@ For optimal performance on server, it is recommended to set several environment E.g. on Linux, ```bash -# set BigDL-Nano env variables -source bigdl-nano-init +# set BigDL-LLM env variables +source bigdl-llm-init # e.g. for a server with 48 cores per socket export OMP_NUM_THREADS=48 diff --git a/python/llm/example/CPU/PyTorch-Models/Model/wizardcoder-python/README.md b/python/llm/example/CPU/PyTorch-Models/Model/wizardcoder-python/README.md index 917ca094..9ebabe22 100644 --- a/python/llm/example/CPU/PyTorch-Models/Model/wizardcoder-python/README.md +++ b/python/llm/example/CPU/PyTorch-Models/Model/wizardcoder-python/README.md @@ -32,8 +32,8 @@ For optimal performance on server, it is recommended to set several environment E.g. on Linux, ```bash -# set BigDL-Nano env variables -source bigdl-nano-init +# set BigDL-LLM env variables +source bigdl-llm-init # e.g. for a server with 48 cores per socket export OMP_NUM_THREADS=48 diff --git a/python/llm/example/CPU/PyTorch-Models/Model/yi/README.md b/python/llm/example/CPU/PyTorch-Models/Model/yi/README.md index fac47f53..04a76046 100644 --- a/python/llm/example/CPU/PyTorch-Models/Model/yi/README.md +++ b/python/llm/example/CPU/PyTorch-Models/Model/yi/README.md @@ -38,8 +38,8 @@ For optimal performance on server, it is recommended to set several environment E.g. on Linux, ```bash -# set BigDL-Nano env variables -source bigdl-nano-init +# set BigDL-LLM env variables +source bigdl-llm-init # e.g. for a server with 48 cores per socket export OMP_NUM_THREADS=48