Commit graph

811 commits

Author SHA1 Message Date
Yuwen Hu
d127a8654c
Small typo fixes (#12558) 2024-12-17 13:54:13 +08:00
binbin Deng
680ea7e4a8
[NPU doc] Update configuration for different platforms (#12554) 2024-12-17 10:15:09 +08:00
binbin Deng
caf15cc5ef
[NPU] Add IPEX_LLM_NPU_MTL to enable support on mtl (#12543) 2024-12-13 17:01:13 +08:00
SONG Ge
5402fc65c8
[Ollama] Update ipex-llm ollama readme to v0.4.6 (#12542)
* Update ipex-llm ollama readme to v0.4.6
2024-12-13 16:26:12 +08:00
Yuwen Hu
b747f3f6b8
Small fix to GPU installation guide (#12536) 2024-12-13 10:02:47 +08:00
binbin Deng
6fc27da9c1
[NPU] Update glm-edge support in docs (#12529) 2024-12-12 11:14:09 +08:00
Jinhe
5e1416c9aa
fix readme for npu cpp examples and llama.cpp (#12505)
* fix cpp readme

* fix cpp readme

* fix cpp readme
2024-12-05 12:32:42 +08:00
joan726
ae9c2154f4
Added cross-links (#12494)
* Update install_linux_gpu.zh-CN.md

Add the link for guide of windows installation.

* Update install_windows_gpu.zh-CN.md

Add the link for guide of linux installation.

* Update install_windows_gpu.md

Add the link for guide of Linux installation.

* Update install_linux_gpu.md

Add the link for guide of Windows installation.

* Update install_linux_gpu.md

Modify based on comments.

* Update install_windows_gpu.md

Modify based on comments
2024-12-04 16:53:13 +08:00
Yuwen Hu
aee9acb303
Add NPU QuickStart & update example links (#12470)
* Add initial NPU quickstart (c++ part unfinished)

* Small update

* Update based on comments

* Update main readme

* Remove LLaMA description

* Small fix

* Small fix

* Remove subsection link in main README

* Small fix

* Update based on comments

* Small fix

* TOC update and other small fixes

* Update for Chinese main readme

* Update based on comments and other small fixes

* Change order
2024-12-02 17:03:10 +08:00
Yuwen Hu
a2272b70d3
Small fix in llama.cpp troubleshooting guide (#12457) 2024-11-27 19:22:11 +08:00
Chu,Youcheng
acd77d9e87
Remove env variable BIGDL_LLM_XMX_DISABLED in documentation (#12445)
* fix: remove BIGDL_LLM_XMX_DISABLED in mddocs

* fix: remove set SYCL_CACHE_PERSISTENT=1 in example

* fix: remove BIGDL_LLM_XMX_DISABLED in workflows

* fix: merge igpu and A-series Graphics

* fix: remove set BIGDL_LLM_XMX_DISABLED=1 in example

* fix: remove BIGDL_LLM_XMX_DISABLED in workflows

* fix: merge igpu and A-series Graphics

* fix: textual adjustment

* fix: textual adjustment

* fix: textual adjustment
2024-11-27 11:16:36 +08:00
Jun Wang
cb7b08948b
update vllm-docker-quick-start for vllm0.6.2 (#12392)
* update vllm-docker-quick-start for vllm0.6.2

* [UPDATE] rm max-num-seqs parameter in vllm-serving script
2024-11-27 08:47:03 +08:00
joan726
a9cb70a71c
Add install_windows_gpu.zh-CN.md and install_linux_gpu.zh-CN.md (#12409)
* Add install_linux_gpu.zh-CN.md

* Add install_windows_gpu.zh-CN.md

* Update llama_cpp_quickstart.zh-CN.md

Related links updated to zh-CN version.

* Update install_linux_gpu.zh-CN.md

Added link to English version.

* Update install_windows_gpu.zh-CN.md

Add the link to English version.

* Update install_windows_gpu.md

Add the link to CN version.

* Update install_linux_gpu.md

Add the link to CN version.

* Update README.zh-CN.md

Modified the related link to zh-CN version.
2024-11-19 14:39:53 +08:00
Yuwen Hu
d1cde7fac4
Tiny doc fix (#12405) 2024-11-15 10:28:38 +08:00
Xu, Shuo
6726b198fd
Update readme & doc for the vllm upgrade to v0.6.2 (#12399)
Co-authored-by: ATMxsp01 <shou.xu@intel.com>
2024-11-14 10:28:15 +08:00
Jun Wang
4376fdee62
Decouple the openwebui and the ollama. in inference-cpp-xpu dockerfile (#12382)
* remove the openwebui in inference-cpp-xpu dockerfile

* update docker_cpp_xpu_quickstart.md

* add sample output in inference-cpp/readme

* remove the openwebui in main readme

* remove the openwebui in main readme
2024-11-12 20:15:23 +08:00
Shaojun Liu
fad15c8ca0
Update fastchat demo script (#12367)
* Update README.md

* Update vllm_docker_quickstart.md
2024-11-08 15:42:17 +08:00
Xin Qiu
7ef7696956
update linux installation doc (#12365)
* update linux doc

* update
2024-11-08 09:44:58 +08:00
Xin Qiu
520af4e9b5
Update install_linux_gpu.md (#12353) 2024-11-07 16:08:01 +08:00
Jinhe
71ea539351
Add troubleshootings for ollama and llama.cpp (#12358)
* add ollama troubleshoot en

* zh ollama troubleshoot

* llamacpp trouble shoot

* llamacpp trouble shoot

* fix

* save gpu memory
2024-11-07 15:49:20 +08:00
Xu, Shuo
ce0c6ae423
Update Readme for FastChat docker demo (#12354)
* update Readme for FastChat docker demo

* update readme

* add 'Serving with FastChat' part in docs

* polish docs

---------

Co-authored-by: ATMxsp01 <shou.xu@intel.com>
2024-11-07 15:22:42 +08:00
Jin, Qiao
3df6195cb0
Fix application quickstart (#12305)
* fix graphrag quickstart

* fix axolotl quickstart

* fix ragflow quickstart

* fix ragflow quickstart

* fix graphrag toc

* fix comments

* fix comment

* fix comments
2024-10-31 16:57:35 +08:00
joan726
0bbc04b5ec
Add ollama_quickstart.zh-CN.md (#12284)
* Add ollama_quickstart.zh-CN.md

Add ollama_quickstart.zh-CN.md

* Update ollama_quickstart.zh-CN.md

Add Chinese and English switching

* Update ollama_quickstart.md

Add Chinese and English switching

* Update README.zh-CN.md

Modify the related link to ollama_quickstart.zh-CN.md

* Update ollama_quickstart.zh-CN.md

Modified based on comments.

* Update ollama_quickstart.zh-CN.md

Modified based on comments
2024-10-29 15:12:44 +08:00
Yuwen Hu
42a528ded9
Small update to MTL iGPU Linux Prerequisites installation guide (#12281)
* Small update MTL iGPU Linux Prerequisites installation guide

* Small fix
2024-10-28 14:12:07 +08:00
Yuwen Hu
16074ae2a4
Update Linux prerequisites installation guide for MTL iGPU (#12263)
* Update Linux prerequisites installation guide for MTL iGPU

* Further link update

* Small fixes

* Small fix

* Update based on comments

* Small fix

* Make oneAPI installation a shared section for both MTL iGPU and other GPU

* Small fix

* Small fix

* Clarify description
2024-10-28 09:27:14 +08:00
Yuwen Hu
94c4568988
Update windows installation guide regarding troubleshooting (#12270) 2024-10-25 14:32:38 +08:00
joan726
e0a95eb2d6
Add llama_cpp_quickstart.zh-CN.md (#12221) 2024-10-24 16:08:31 +08:00
Jun Wang
aedc4edfba
[ADD] add open webui + vllm serving (#12246) 2024-10-23 10:13:14 +08:00
Jun Wang
fe3b5cd89b
[Update] mmdocs/dockerguide vllm-quick-start awq,gptq online serving document (#12227)
* [FIX] fix the docker start script error

* [ADD] add awq online serving doc

* [ADD] add gptq online serving doc

* [Fix] small fix
2024-10-18 09:46:59 +08:00
Yuwen Hu
a768d71581
Small fix to LNL installation guide (#12192) 2024-10-14 12:03:03 +08:00
Shaojun Liu
49eb20613a
add --blocksize to doc and script (#12187) 2024-10-12 09:17:42 +08:00
Jun Wang
6ffaec66a2
[UPDATE] add prefix caching document into vllm_docker_quickstart.md (#12173)
* [ADD] rewrite new vllm docker quick start

* [ADD] lora adapter doc finished

* [ADD] mulit lora adapter test successfully

* [ADD] add ipex-llm quantization doc

* [Merge] rebase main

* [REMOVE] rm tmp file

* [Merge] rebase main

* [ADD] add prefix caching experiment and result

* [REMOVE] rm cpu offloading chapter
2024-10-11 19:12:22 +08:00
Yuwen Hu
ddcdf47539
Support Windows ARL release (#12183)
* Support release for ARL

* Small fix

* Small fix to doc

* Temp for test

* Remove temp commit for test
2024-10-11 18:30:52 +08:00
Yuwen Hu
ac44e98b7d
Update Windows guide regarding LNL support (#12178)
* Update windows guide regarding LNL support

* Update based on comments
2024-10-11 09:20:08 +08:00
Guancheng Fu
0ef7e1d101
fix vllm docs (#12176) 2024-10-10 15:44:36 +08:00
Jun Wang
412cf8e20c
[UPDATE] update mddocs/DockerGuides/vllm_docker_quickstart.md (#12166)
* [ADD] rewrite new vllm docker quick start

* [ADD] lora adapter doc finished

* [ADD] mulit lora adapter test successfully

* [ADD] add ipex-llm quantization doc

* [UPDATE] update mmdocs vllm_docker_quickstart content

* [REMOVE] rm tmp file

* [UPDATE] tp and pp explaination and readthedoc link change

* [FIX] fix the error description of tp+pp and quantization part

* [FIX] fix the table of verifed model

* [UPDATE] add full low bit para list

* [UPDATE] update the load_in_low_bit params to verifed dtype
2024-10-09 11:19:32 +08:00
Shaojun Liu
e2ef9e938e
Delete deprecated docs/readthedocs directory (#12164) 2024-10-08 14:48:02 +08:00
Ch1y0q
9b75806d14
Update Windows GPU quickstart regarding demo (#12124)
* use Qwen2-1.5B-Instruct in demo

* update

* add reference link

* update

* update
2024-09-29 18:08:49 +08:00
Ruonan Wang
a767438546
fix typo (#12076)
* fix typo

* fix
2024-09-13 11:44:42 +08:00
Ruonan Wang
3f0b24ae2b
update cpp quickstart (#12075)
* update cpp quickstart

* fix style
2024-09-13 11:35:32 +08:00
Ruonan Wang
48d9092b5a
upgrade OneAPI version for cpp Windows (#12063)
* update version

* update quickstart
2024-09-12 11:12:12 +08:00
Shaojun Liu
e5581e6ded
Select the Appropriate APT Repository Based on CPU Type (#12023) 2024-09-05 17:06:07 +08:00
Yuwen Hu
643458d8f0
Update GraphRAG QuickStart (#11995)
* Update GraphRAG QuickStart

* Further updates

* Small fixes

* Small fix
2024-09-03 15:52:08 +08:00
Jinhe
e895e1b4c5
modification on llamacpp readme after Ipex-llm latest update (#11971)
* update on readme after ipex-llm update

* update on readme after ipex-llm update

* rebase & delete redundancy

* revise

* add numbers for troubleshooting
2024-08-30 11:36:45 +08:00
Ch1y0q
77b04efcc5
add notes for SYCL_PI_LEVEL_ZERO_USE_IMMEDIATE_COMMANDLISTS (#11936)
* add notes for `SYCL_PI_LEVEL_ZERO_USE_IMMEDIATE_COMMANDLISTS`

* also update other quickstart
2024-08-30 09:26:47 +08:00
Jinhe
6fc9340d53
restore ollama webui quickstart (#11955) 2024-08-29 17:53:19 +08:00
Jinhe
ec67ee7177
added accelerate version specification in open webui quickstart(#11948) 2024-08-28 15:02:39 +08:00
Ruonan Wang
460bc96d32
update version of llama.cpp / ollama (#11930)
* update version

* fix version
2024-08-27 21:21:44 +08:00
Ch1y0q
5a8fc1baa2
update troubleshooting for llama.cpp and ollama (#11890)
* update troubleshooting for llama.cpp and ollama

* update

* update
2024-08-26 20:55:23 +08:00
Jinhe
dbd14251dd
Troubleshoot for sycl not found (#11774)
* added troubleshoot for sycl not found problem

* added troubleshoot for sycl not found problem

* revision on troubleshoot

* revision on troubleshoot
2024-08-14 10:26:01 +08:00