* Add quickstart for install IPEX-LLM with PyTorch 2.6 on Intel GPUs
* Add jump links
* Rename
* Small fix
* Small fix
* Update based on comments
* Small fix
* Add llamacpp_portable_zip_gpu_quickstart.zh-CN.md
Add llamacpp_portable_zip_gpu_quickstart.zh-CN.md
* Update README.zh-CN.md
Changed and Linked to llamacpp portable zip.zh-CN.md.
* Update llamacpp_portable_zip_gpu_quickstart.md
Added CN version link
* Update README.zh-CN.md
Update all links to "llamacpp_portable_zip_gpu_quickstart.zh-CN.md
* Update llama_cpp_quickstart.zh-CN.md
* Update llamacpp_portable_zip_gpu_quickstart.zh-CN.md
Modify based on comments.
* Update llamacpp_portable_zip_gpu_quickstart.zh-CN.md
Modify based on comments.
* Update llamacpp_portable_zip_gpu_quickstart.zh-CN.md
Update the doc based on #12928
* Update llamacpp_portable_zip_gpu_quickstart.zh-CN.md
Add “More Details” on Table of Contents
* Update README.zh-CN.md
Update llamacpp_portable_zip_gpu_quickstart CN link
* Update README.zh-CN.md
Change llama.cpp link
* Update README.zh-CN.md
* Update README.md
* linux doc
* update
* Update ollama_portablze_zip_quickstart.md
* Update ollama_portablze_zip_quickstart.md
* Update ollama_portablze_zip_quickstart.zh-CN.md
* Update ollama_portablze_zip_quickstart.md
* meet code review
* update
* Add tips & troubleshooting sections for both Linux & Windows
* Rebase
* Fix based on comments
* Small fix
* Fix img
* Update table for linux
* Small fix
---------
Co-authored-by: Yuwen Hu <yuwen.hu@intel.com>
* Add initial quickstart for Ollama portable zip
* Small fix
* Fixed based on comments
* Small fix
* Add demo image for run ollama
* Update download link
* Remove model with optimize_model=False in NPU verified models tables, and remove related example
* Remove experimental in run optimized model section title
* Unify model table order & example cmd
* Move embedding example to separate folder & update quickstart example link
* Add Quickstart reference in main NPU readme
* Small fix
* Small fix
* Move save/load examples under NPU/HF-Transformers-AutoModels
* Add low-bit and polish arguments for LLM Python examples
* Small fix
* Add low-bit and polish arguments for Multi-Model examples
* Polish argument for Embedding models
* Polish argument for LLM CPP examples
* Add low-bit and polish argument for Save-Load examples
* Add accuracy tuning tips for examples
* Update NPU qucikstart accuracy tuning with low-bit optimizations
* Add save/load section to qucikstart
* Update CPP example sample output to EN
* Add installation regarding cmake for CPP examples
* Small fix
* Small fix
* Small fix
* Small fix
* Small fix
* Small fix
* Unify max prompt length to 512
* Change recommended low-bit for Qwen2.5-3B-Instruct to asym_int4
* Update based on comments
* Small fix