* Update install_linux_gpu.zh-CN.md
Add the link for guide of windows installation.
* Update install_windows_gpu.zh-CN.md
Add the link for guide of linux installation.
* Update install_windows_gpu.md
Add the link for guide of Linux installation.
* Update install_linux_gpu.md
Add the link for guide of Windows installation.
* Update install_linux_gpu.md
Modify based on comments.
* Update install_windows_gpu.md
Modify based on comments
* Add initial NPU quickstart (c++ part unfinished)
* Small update
* Update based on comments
* Update main readme
* Remove LLaMA description
* Small fix
* Small fix
* Remove subsection link in main README
* Small fix
* Update based on comments
* Small fix
* TOC update and other small fixes
* Update for Chinese main readme
* Update based on comments and other small fixes
* Change order
* Add install_linux_gpu.zh-CN.md
* Add install_windows_gpu.zh-CN.md
* Update llama_cpp_quickstart.zh-CN.md
Related links updated to zh-CN version.
* Update install_linux_gpu.zh-CN.md
Added link to English version.
* Update install_windows_gpu.zh-CN.md
Add the link to English version.
* Update install_windows_gpu.md
Add the link to CN version.
* Update install_linux_gpu.md
Add the link to CN version.
* Update README.zh-CN.md
Modified the related link to zh-CN version.
* remove the openwebui in inference-cpp-xpu dockerfile
* update docker_cpp_xpu_quickstart.md
* add sample output in inference-cpp/readme
* remove the openwebui in main readme
* remove the openwebui in main readme
* Add ollama_quickstart.zh-CN.md
Add ollama_quickstart.zh-CN.md
* Update ollama_quickstart.zh-CN.md
Add Chinese and English switching
* Update ollama_quickstart.md
Add Chinese and English switching
* Update README.zh-CN.md
Modify the related link to ollama_quickstart.zh-CN.md
* Update ollama_quickstart.zh-CN.md
Modified based on comments.
* Update ollama_quickstart.zh-CN.md
Modified based on comments
* Update Linux prerequisites installation guide for MTL iGPU
* Further link update
* Small fixes
* Small fix
* Update based on comments
* Small fix
* Make oneAPI installation a shared section for both MTL iGPU and other GPU
* Small fix
* Small fix
* Clarify description
* [ADD] rewrite new vllm docker quick start
* [ADD] lora adapter doc finished
* [ADD] mulit lora adapter test successfully
* [ADD] add ipex-llm quantization doc
* [UPDATE] update mmdocs vllm_docker_quickstart content
* [REMOVE] rm tmp file
* [UPDATE] tp and pp explaination and readthedoc link change
* [FIX] fix the error description of tp+pp and quantization part
* [FIX] fix the table of verifed model
* [UPDATE] add full low bit para list
* [UPDATE] update the load_in_low_bit params to verifed dtype
* update on readme after ipex-llm update
* update on readme after ipex-llm update
* rebase & delete redundancy
* revise
* add numbers for troubleshooting
* Revert to use out-of-tree GPU driver since the performance with out-of-tree driver is better than upsteam's
* add spaces
* add troubleshooting case
* update Troubleshooting
* Small tests for demo video workaround
* Small fix
* Add workaround for langchain-chatchat demo video
* Small fix
* Small fix
* Update for other demo videos in quickstart
* Add missing for text-generation-webui quickstart
* transfer files in DockerGuides from rst to md
* add some dividing lines
* adjust the title hierarchy in docker_cpp_xpu_quickstart.md
* restore
* switch to the correct branch
* small change
---------
Co-authored-by: ATMxsp01 <shou.xu@intel.com>
* updated link
* converted to md format, need to be reviewed
* converted to md format, need to be reviewed
* converted to md format, need to be reviewed
* converted to md format, need to be reviewed
* converted to md format, need to be reviewed
* converted to md format, need to be reviewed
* converted to md format, need to be reviewed
* converted to md format, need to be reviewed
* converted to md format, need to be reviewed
* converted to md format, need to be reviewed, deleted some leftover texts
* converted to md file type, need to be reviewed
* converted to md file type, need to be reviewed
* testing Github Tags
* testing Github Tags
* added Github Tags
* added Github Tags
* added Github Tags
* Small fix
* Small fix
* Small fix
* Small fix
* Small fix
* Further fix
* Fix index
* Small fix
* Fix
---------
Co-authored-by: Yuwen Hu <yuwen.hu@intel.com>
* Quickstart index.rst -> index.md
* Update for Linux Install Quickstart
* Update md docs for Windows Install QuickStart
* Small fix
* Add blank lines
* Update mddocs for llama cpp quickstart
* Update mddocs for llama3 llama-cpp and ollama quickstart
* Update mddocs for ollama quickstart
* Update mddocs for openwebui quickstart
* Update mddocs for privateGPT quickstart
* Update mddocs for vllm quickstart
* Small fix
* Update mddocs for text-generation-webui quickstart
* Update for video links
* add config and default value
* add config in taml
* remove lookahead and max_matching_ngram_size in config
* remove streaming and use_fp16_torch_dtype in test yaml
* update task in readme
* update commit of task
* Add GPU example for Qwen2
* Update comments in README
* Update README for Qwen2 GPU example
* Add CPU example for Qwen2
Sample Output under README pending
* Update generate.py and README for CPU Qwen2
* Update GPU example for Qwen2
* Small update
* Small fix
* Add Qwen2 table
* Update README for Qwen2 CPU and GPU
Update sample output under README
---------
Co-authored-by: Zijie Li <michael20001122@gmail.com>
* Change installation address
Change former address: "https://docs.conda.io/en/latest/miniconda.html#" to new address: "https://conda-forge.org/download/" for 63 occurrences under python\llm\example
* Change Prompt
Change "Anaconda Prompt" to "Miniforge Prompt" for 1 occurrence
* Create and update model minicpm
* Update model minicpm
Update model minicpm under GPU/PyTorch-Models
* Update readme and generate.py
change "prompt = tokenizer.apply_chat_template(chat, tokenize=False, add_generation_prompt=False)" and delete "pip install transformers==4.37.0
"
* Update comments for minicpm GPU
Update comments for generate.py at minicpm GPU
* Add CPU example for MiniCPM
* Update minicpm README for CPU
* Update README for MiniCPM and Llama3
* Update Readme for Llama3 CPU Pytorch
* Update and fix comments for MiniCPM
* Add some information to help users to solve "could not create a primitive" error in Windows.
* Small update
---------
Co-authored-by: ATMxsp01 <shou.xu@intel.com>
Co-authored-by: Yuwen Hu <yuwen.hu@intel.com>