* Revert to use out-of-tree GPU driver since the performance with out-of-tree driver is better than upsteam's
* add spaces
* add troubleshooting case
* update Troubleshooting
* Small tests for demo video workaround
* Small fix
* Add workaround for langchain-chatchat demo video
* Small fix
* Small fix
* Update for other demo videos in quickstart
* Add missing for text-generation-webui quickstart
* transfer files in DockerGuides from rst to md
* add some dividing lines
* adjust the title hierarchy in docker_cpp_xpu_quickstart.md
* restore
* switch to the correct branch
* small change
---------
Co-authored-by: ATMxsp01 <shou.xu@intel.com>
* updated link
* converted to md format, need to be reviewed
* converted to md format, need to be reviewed
* converted to md format, need to be reviewed
* converted to md format, need to be reviewed
* converted to md format, need to be reviewed
* converted to md format, need to be reviewed
* converted to md format, need to be reviewed
* converted to md format, need to be reviewed
* converted to md format, need to be reviewed
* converted to md format, need to be reviewed, deleted some leftover texts
* converted to md file type, need to be reviewed
* converted to md file type, need to be reviewed
* testing Github Tags
* testing Github Tags
* added Github Tags
* added Github Tags
* added Github Tags
* Small fix
* Small fix
* Small fix
* Small fix
* Small fix
* Further fix
* Fix index
* Small fix
* Fix
---------
Co-authored-by: Yuwen Hu <yuwen.hu@intel.com>
* Quickstart index.rst -> index.md
* Update for Linux Install Quickstart
* Update md docs for Windows Install QuickStart
* Small fix
* Add blank lines
* Update mddocs for llama cpp quickstart
* Update mddocs for llama3 llama-cpp and ollama quickstart
* Update mddocs for ollama quickstart
* Update mddocs for openwebui quickstart
* Update mddocs for privateGPT quickstart
* Update mddocs for vllm quickstart
* Small fix
* Update mddocs for text-generation-webui quickstart
* Update for video links
* add config and default value
* add config in taml
* remove lookahead and max_matching_ngram_size in config
* remove streaming and use_fp16_torch_dtype in test yaml
* update task in readme
* update commit of task
* Add GPU example for Qwen2
* Update comments in README
* Update README for Qwen2 GPU example
* Add CPU example for Qwen2
Sample Output under README pending
* Update generate.py and README for CPU Qwen2
* Update GPU example for Qwen2
* Small update
* Small fix
* Add Qwen2 table
* Update README for Qwen2 CPU and GPU
Update sample output under README
---------
Co-authored-by: Zijie Li <michael20001122@gmail.com>
* Change installation address
Change former address: "https://docs.conda.io/en/latest/miniconda.html#" to new address: "https://conda-forge.org/download/" for 63 occurrences under python\llm\example
* Change Prompt
Change "Anaconda Prompt" to "Miniforge Prompt" for 1 occurrence
* Create and update model minicpm
* Update model minicpm
Update model minicpm under GPU/PyTorch-Models
* Update readme and generate.py
change "prompt = tokenizer.apply_chat_template(chat, tokenize=False, add_generation_prompt=False)" and delete "pip install transformers==4.37.0
"
* Update comments for minicpm GPU
Update comments for generate.py at minicpm GPU
* Add CPU example for MiniCPM
* Update minicpm README for CPU
* Update README for MiniCPM and Llama3
* Update Readme for Llama3 CPU Pytorch
* Update and fix comments for MiniCPM