Commit graph

730 commits

Author SHA1 Message Date
Yuwen Hu
d9dd1b70bd
Remove example page in mddocs (#11373) 2024-06-20 14:23:43 +08:00
Yuwen Hu
769728c1eb
Add initial md docs (#11371) 2024-06-20 13:47:49 +08:00
Shengsheng Huang
9601fae5d5
fix system note (#11368) 2024-06-20 11:09:53 +08:00
Shengsheng Huang
ed4c439497
small fix (#11366) 2024-06-20 10:38:20 +08:00
Shengsheng Huang
a721c1ae43
minor fix of ragflow_quickstart.md (#11364) 2024-06-19 22:30:33 +08:00
Shengsheng Huang
13727635e8
revise ragflow quickstart (#11363)
* revise ragflow quickstart

* update titles and split the quickstart into sections

* update
2024-06-19 22:24:31 +08:00
Zijie Li
5283df0078
LLM: Add RAGFlow with Ollama Example QuickStart (#11338)
* Create ragflow.md

* Update ragflow.md

* Update ragflow_quickstart

* Update ragflow_quickstart.md

* Upload RAGFlow quickstart without images

* Update ragflow_quickstart.md

* Update ragflow_quickstart.md

* Update ragflow_quickstart.md

* Update ragflow_quickstart.md

* fix typos in readme

* Fix typos in quickstart readme
2024-06-19 20:00:50 +08:00
Jason Dai
271d82a4fc
Update readme (#11357) 2024-06-19 10:05:42 +08:00
Xiangyu Tian
f6cd628cd8
Fix script usage in vLLM CPU Quickstart (#11353) 2024-06-18 16:50:48 +08:00
Guancheng Fu
c9b4cadd81
fix vLLM/docker issues (#11348)
* fix

* fix

* ffix
2024-06-18 16:23:53 +08:00
hxsz1997
44f22cba70
add config and default value (#11344)
* add config and default value

* add config in taml

* remove lookahead and max_matching_ngram_size in config

* remove streaming and use_fp16_torch_dtype in test yaml

* update task in readme

* update commit of task
2024-06-18 15:28:57 +08:00
Shengsheng Huang
1f39bb84c7
update readthedocs perf data (#11345) 2024-06-18 13:23:47 +08:00
Qiyuan Gong
de4bb97b4f
Remove accelerate 0.23.0 install command in readme and docker (#11333)
*ipex-llm's accelerate has been upgraded to 0.23.0. Remove accelerate 0.23.0 install command in README and docker。
2024-06-17 17:52:12 +08:00
Yuwen Hu
9e4d87a696
Langchain-chatchat QuickStart small link fix (#11317) 2024-06-14 14:02:17 +08:00
Yuwen Hu
bfab294f08
Update langchain-chatchat QuickStart to include Core Ultra iGPU Linux Guide (#11302) 2024-06-13 15:09:55 +08:00
Shengsheng Huang
ea372cc472
update demos section (#11298)
* update demos section

* update format
2024-06-13 11:58:19 +08:00
Jin Qiao
f224e98297
Add GLM-4 CPU example (#11223)
* Add GLM-4 example

* add tiktoken dependency

* fix

* fix
2024-06-12 15:30:51 +08:00
Yuwen Hu
8c36b5bdde
Add qwen2 example (#11252)
* Add GPU example for Qwen2

* Update comments in README

* Update README for Qwen2 GPU example

* Add CPU example for Qwen2

Sample Output under README pending

* Update generate.py and README for CPU Qwen2

* Update GPU example for Qwen2

* Small update

* Small fix

* Add Qwen2 table

* Update README for Qwen2 CPU and GPU

Update sample output under README

---------

Co-authored-by: Zijie Li <michael20001122@gmail.com>
2024-06-07 10:29:33 +08:00
Zijie Li
bfa1367149
Add CPU and GPU example for MiniCPM (#11202)
* Change installation address

Change former address: "https://docs.conda.io/en/latest/miniconda.html#" to new address: "https://conda-forge.org/download/" for 63 occurrences under python\llm\example

* Change Prompt

Change "Anaconda Prompt" to "Miniforge Prompt" for 1 occurrence

* Create and update model minicpm

* Update model minicpm

Update model minicpm under GPU/PyTorch-Models

* Update readme and generate.py

change "prompt = tokenizer.apply_chat_template(chat, tokenize=False, add_generation_prompt=False)" and delete "pip install transformers==4.37.0
"

* Update comments for minicpm GPU

Update comments for generate.py at minicpm GPU

* Add CPU example for MiniCPM

* Update minicpm README for CPU

* Update README for MiniCPM and Llama3

* Update Readme for Llama3 CPU Pytorch

* Update and fix comments for MiniCPM
2024-06-05 18:09:53 +08:00
Xu, Shuo
a27a559650
Add some information in FAQ to help users solve "RuntimeError: could not create a primitive" error on Windows (#11221)
* Add some information to help users to solve "could not create a primitive" error in Windows.

* Small update

---------

Co-authored-by: ATMxsp01 <shou.xu@intel.com>
Co-authored-by: Yuwen Hu <yuwen.hu@intel.com>
2024-06-05 17:57:42 +08:00
Guancheng Fu
3ef4aa98d1
Refine vllm_quickstart doc (#11199)
* refine doc

* refine
2024-06-04 18:46:27 +08:00
Xiangyu Tian
ff83fad400
Fix typo in vLLM CPU docker guide (#11188) 2024-06-03 15:55:27 +08:00
Shaojun Liu
401013a630
Remove chatglm_C Module to Eliminate LGPL Dependency (#11178)
* remove chatglm_C.**.pyd to solve ngsolve weak copyright vunl

* fix style check error

* remove chatglm native int4 from langchain
2024-05-31 17:03:11 +08:00
Yuwen Hu
f0aaa130a9
Update miniconda/anaconda -> miniforge in documentation (#11176)
* Update miniconda/anaconda -> miniforge in installation guide

* Update for all Quickstart

* further fix for docs
2024-05-30 17:40:18 +08:00
Jin Qiao
dcbf4d3d0a
Add phi-3-vision example (#11156)
* Add phi-3-vision example (HF-Automodels)

* fix

* fix

* fix

* Add phi-3-vision CPU example (HF-Automodels)

* add in readme

* fix

* fix

* fix

* fix

* use fp8 for gpu example

* remove eval
2024-05-30 10:02:47 +08:00
Wang, Jian4
8e25de1126
LLM: Add codegeex2 example (#11143)
* add codegeex example

* update

* update cpu

* add GPU

* add gpu

* update readme
2024-05-29 10:00:26 +08:00
Ruonan Wang
83bd9cb681
add new version for cpp quickstart and keep an old version (#11151)
* add new version

* meet review
2024-05-28 15:29:34 +08:00
Guancheng Fu
daf7b1cd56
[Docker] Fix image using two cards error (#11144)
* fix all

* done
2024-05-27 16:20:13 +08:00
Jason Dai
34dab3b4ef
Update readme (#11141) 2024-05-27 15:41:02 +08:00
Guancheng Fu
fabc395d0d
add langchain vllm interface (#11121)
* done

* fix

* fix

* add vllm

* add langchain vllm exampels

* add docs

* temp
2024-05-24 17:19:27 +08:00
Shaojun Liu
85491907f3
Update GIF link (#11119) 2024-05-24 14:26:18 +08:00
Xiangyu Tian
1291165720
LLM: Add quickstart for vLLM cpu (#11122)
Add quickstart for vLLM cpu.
2024-05-24 10:21:21 +08:00
Xiangyu Tian
b3f6faa038
LLM: Add CPU vLLM entrypoint (#11083)
Add CPU vLLM entrypoint and update CPU vLLM serving example.
2024-05-24 09:16:59 +08:00
Shengsheng Huang
7ed270a4d8
update readme docker section, fix quickstart title, remove chs figure (#11044)
* update readme and fix quickstart title, remove chs figure

* update readme according to comment

* reorganize the docker guide structure
2024-05-24 00:18:20 +08:00
Zhao Changmin
15d906a97b
Update linux igpu run script (#11098)
* update run script
2024-05-22 17:18:07 +08:00
Guancheng Fu
4fd1df9cf6
Add toc for docker quickstarts (#11095)
* fix

* fix
2024-05-22 11:23:22 +08:00
Zhao Changmin
bf0f904e66
Update level_zero on MTL linux (#11085)
* Update level_zero on MTL
---------

Co-authored-by: Yuwen Hu <yuwen.hu@intel.com>
2024-05-22 11:01:56 +08:00
Shaojun Liu
8fdc8fb197
Quickstart: Run/Develop PyTorch in VSCode with Docker on Intel GPU (#11070)
* add quickstart: Run/Develop PyTorch in VSCode with Docker on Intel GPU

* add gif

* update index.rst

* update link

* update GIFs
2024-05-22 09:29:42 +08:00
Guancheng Fu
f654f7e08c
Add serving docker quickstart (#11072)
* add temp file

* add initial docker readme

* temp

* done

* add fastchat service

* fix

* fix

* fix

* fix

* remove stale file
2024-05-21 17:00:58 +08:00
binbin Deng
7170dd9192
Update guide for running qwen with AutoTP (#11065) 2024-05-20 10:53:17 +08:00
Wang, Jian4
a2e1578fd9
Merge tgi_api_server to main (#11036)
* init

* fix style

* speculative can not use benchmark

* add tgi server readme
2024-05-20 09:15:03 +08:00
Yuwen Hu
f60565adc7
Fix toc for vllm serving quickstart (#11068) 2024-05-17 17:12:48 +08:00
Guancheng Fu
dfac168d5f
fix format/typo (#11067) 2024-05-17 16:52:17 +08:00
Guancheng Fu
67db925112
Add vllm quickstart (#10978)
* temp

* add doc

* finish

* done

* fix

* add initial docker readme

* temp

* done fixing vllm_quickstart

* done

* remove not used file

* add

* fix
2024-05-17 16:16:42 +08:00
ZehuaCao
56cb992497
LLM: Modify CPU Installation Command for most examples (#11049)
* init

* refine

* refine

* refine

* modify hf-agent example

* modify all CPU model example

* remove readthedoc modify

* replace powershell with cmd

* fix repo

* fix repo

* update

* remove comment on windows code block

* update

* update

* update

* update

---------

Co-authored-by: xiangyuT <xiangyu.tian@intel.com>
2024-05-17 15:52:20 +08:00
Shaojun Liu
84239d0bd3
Update docker image tags in Docker Quickstart (#11061)
* update docker image tag to latest

* add note

* simplify note

* add link in reStructuredText

* minor fix

* update tag
2024-05-17 11:06:11 +08:00
Xiangyu Tian
d963e95363
LLM: Modify CPU Installation Command for documentation (#11042)
* init

* refine

* refine

* refine

* refine comments
2024-05-17 10:14:00 +08:00
Wang, Jian4
00d4410746
Update cpp docker quickstart (#11040)
* add sample output

* update link

* update

* update header

* update
2024-05-16 14:55:13 +08:00
Ruonan Wang
1d73fc8106
update cpp quickstart (#11031) 2024-05-15 14:33:36 +08:00
Wang, Jian4
86cec80b51
LLM: Add llm inference_cpp_xpu_docker (#10933)
* test_cpp_docker

* update

* update

* update

* update

* add sudo

* update nodejs version

* no need npm

* remove blinker

* new cpp docker

* restore

* add line

* add manually_build

* update and add mtl

* update for workdir llm

* add benchmark part

* update readme

* update 1024-128

* update readme

* update

* fix

* update

* update

* update readme too

* update readme

* no change

* update dir_name

* update readme
2024-05-15 11:10:22 +08:00