Update main readme for missing quickstarts (#11427)
* Update main readme to add missing quckstart * Update quickstart index page * Small fixes * Small fix
This commit is contained in:
		
							parent
							
								
									40fa23560e
								
							
						
					
					
						commit
						a45ceac4e4
					
				
					 2 changed files with 6 additions and 3 deletions
				
			
		| 
						 | 
					@ -160,17 +160,21 @@ Please see the **Perplexity** result below (tested on Wikitext dataset using the
 | 
				
			||||||
## `ipex-llm` Quickstart
 | 
					## `ipex-llm` Quickstart
 | 
				
			||||||
 | 
					
 | 
				
			||||||
### Docker
 | 
					### Docker
 | 
				
			||||||
 | 
					- [Installation on GPU](docs/mddocs/DockerGuides/docker_windows_gpu.md): installing `ipex-llm` in docker with Intel GPU
 | 
				
			||||||
- [GPU Inference in C++](docs/mddocs/DockerGuides/docker_cpp_xpu_quickstart.md): running `llama.cpp`, `ollama`, `OpenWebUI`, etc., with `ipex-llm` on Intel GPU
 | 
					- [GPU Inference in C++](docs/mddocs/DockerGuides/docker_cpp_xpu_quickstart.md): running `llama.cpp`, `ollama`, `OpenWebUI`, etc., with `ipex-llm` on Intel GPU
 | 
				
			||||||
- [GPU Inference in Python](docs/mddocs/DockerGuides/docker_pytorch_inference_gpu.md) : running HuggingFace `transformers`, `LangChain`, `LlamaIndex`, `ModelScope`, etc. with `ipex-llm` on Intel GPU
 | 
					- [GPU Inference in Python](docs/mddocs/DockerGuides/docker_pytorch_inference_gpu.md) : running HuggingFace `transformers`, `LangChain`, `LlamaIndex`, `ModelScope`, etc. with `ipex-llm` on Intel GPU
 | 
				
			||||||
- [VSCode Guide on GPU](docs/readthedocs/source/doc/LLM/DockerGuides/docker_run_pytorch_inference_in_vscode.md): running and developing Python LLM applications using VSCode on Intel GPU
 | 
					- [VSCode Guide on GPU](docs/readthedocs/source/doc/LLM/DockerGuides/docker_run_pytorch_inference_in_vscode.md): running and developing Python LLM applications using VSCode on Intel GPU
 | 
				
			||||||
- [vLLM on GPU](docs/mddocs/DockerGuides/vllm_docker_quickstart.md): running `vLLM` serving with `ipex-llm` on Intel GPU
 | 
					- [vLLM on GPU](docs/mddocs/DockerGuides/vllm_docker_quickstart.md): running `vLLM` serving with `ipex-llm` on Intel GPU
 | 
				
			||||||
 | 
					- [vLLM on CPU](docs/mddocs/DockerGuides/vllm_cpu_docker_quickstart.md): running `vLLM` serving with `ipex-llm` on Intel CPU  
 | 
				
			||||||
- [FastChat on GPU](docs/mddocs/DockerGuides/fastchat_docker_quickstart.md): running `FastChat` serving with `ipex-llm` on Intel GPU
 | 
					- [FastChat on GPU](docs/mddocs/DockerGuides/fastchat_docker_quickstart.md): running `FastChat` serving with `ipex-llm` on Intel GPU
 | 
				
			||||||
 | 
					
 | 
				
			||||||
### Use
 | 
					### Use
 | 
				
			||||||
- [llama.cpp](docs/mddocs/Quickstart/llama_cpp_quickstart.md): running **llama.cpp** (*using C++ interface of `ipex-llm` as an accelerated backend for `llama.cpp`*) on Intel GPU
 | 
					- [llama.cpp](docs/mddocs/Quickstart/llama_cpp_quickstart.md): running **llama.cpp** (*using C++ interface of `ipex-llm` as an accelerated backend for `llama.cpp`*) on Intel GPU
 | 
				
			||||||
- [Ollama](docs/mddocs/Quickstart/ollama_quickstart.md): running **ollama** (*using C++ interface of `ipex-llm` as an accelerated backend for `ollama`*) on Intel GPU
 | 
					- [Ollama](docs/mddocs/Quickstart/ollama_quickstart.md): running **ollama** (*using C++ interface of `ipex-llm` as an accelerated backend for `ollama`*) on Intel GPU
 | 
				
			||||||
 | 
					- [Llama 3 with `llama.cpp` and `ollama`](docs/mddocs/Quickstart/llama3_llamacpp_ollama_quickstart.md): running **Llama 3** on Intel GPU using `llama.cpp` and `ollama` with `ipex-llm`
 | 
				
			||||||
- [vLLM](docs/mddocs/Quickstart/vLLM_quickstart.md): running `ipex-llm` in **vLLM** on both Intel [GPU](python/llm/example/GPU/vLLM-Serving) and [CPU](python/llm/example/CPU/vLLM-Serving)
 | 
					- [vLLM](docs/mddocs/Quickstart/vLLM_quickstart.md): running `ipex-llm` in **vLLM** on both Intel [GPU](python/llm/example/GPU/vLLM-Serving) and [CPU](python/llm/example/CPU/vLLM-Serving)
 | 
				
			||||||
- [FastChat](docs/mddocs/Quickstart/fastchat_quickstart.md): running `ipex-llm` in **FastChat** serving on on both Intel GPU and CPU
 | 
					- [FastChat](docs/mddocs/Quickstart/fastchat_quickstart.md): running `ipex-llm` in **FastChat** serving on on both Intel GPU and CPU
 | 
				
			||||||
 | 
					- [Serving on multiple Intel GPUs](docs/mddocs/Quickstart/deepspeed_autotp_fastapi_quickstart.md): running `ipex-llm` **serving on multiple Intel GPUs** by leveraging DeepSpeed AutoTP and FastAPI
 | 
				
			||||||
- [Text-Generation-WebUI](docs/mddocs/Quickstart/webui_quickstart.md): running `ipex-llm` in `oobabooga` **WebUI**
 | 
					- [Text-Generation-WebUI](docs/mddocs/Quickstart/webui_quickstart.md): running `ipex-llm` in `oobabooga` **WebUI**
 | 
				
			||||||
- [Axolotl](docs/mddocs/Quickstart/axolotl_quickstart.md): running `ipex-llm` in **Axolotl** for LLM finetuning
 | 
					- [Axolotl](docs/mddocs/Quickstart/axolotl_quickstart.md): running `ipex-llm` in **Axolotl** for LLM finetuning
 | 
				
			||||||
- [Benchmarking](docs/mddocs/Quickstart/benchmark_quickstart.md): running  (latency and throughput) **benchmarks** for `ipex-llm` on Intel CPU and GPU
 | 
					- [Benchmarking](docs/mddocs/Quickstart/benchmark_quickstart.md): running  (latency and throughput) **benchmarks** for `ipex-llm` on Intel CPU and GPU
 | 
				
			||||||
| 
						 | 
					@ -181,6 +185,7 @@ Please see the **Perplexity** result below (tested on Wikitext dataset using the
 | 
				
			||||||
- [Open WebUI](docs/mddocs/Quickstart/open_webui_with_ollama_quickstart.md): running `Open WebUI` with `ipex-llm`
 | 
					- [Open WebUI](docs/mddocs/Quickstart/open_webui_with_ollama_quickstart.md): running `Open WebUI` with `ipex-llm`
 | 
				
			||||||
- [PrivateGPT](docs/mddocs/Quickstart/privateGPT_quickstart.md): running `PrivateGPT` to interact with documents with `ipex-llm`
 | 
					- [PrivateGPT](docs/mddocs/Quickstart/privateGPT_quickstart.md): running `PrivateGPT` to interact with documents with `ipex-llm`
 | 
				
			||||||
- [Dify platform](docs/mddocs/Quickstart/dify_quickstart.md): running `ipex-llm` in `Dify`(*production-ready LLM app development platform*)
 | 
					- [Dify platform](docs/mddocs/Quickstart/dify_quickstart.md): running `ipex-llm` in `Dify`(*production-ready LLM app development platform*)
 | 
				
			||||||
 | 
					- [RAGFlow](docs/mddocs/Quickstart/ragflow_quickstart.md): running `RAGFlow` (*an open-source RAG engine*) with `ipex-llm`
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					
 | 
				
			||||||
### Install 
 | 
					### Install 
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
| 
						 | 
					@ -10,12 +10,10 @@ This section includes efficient guide to show you how to:
 | 
				
			||||||
- [``bigdl-llm`` Migration Guide](./bigdl_llm_migration.md)
 | 
					- [``bigdl-llm`` Migration Guide](./bigdl_llm_migration.md)
 | 
				
			||||||
- [Install IPEX-LLM on Linux with Intel GPU](./install_linux_gpu.md)
 | 
					- [Install IPEX-LLM on Linux with Intel GPU](./install_linux_gpu.md)
 | 
				
			||||||
- [Install IPEX-LLM on Windows with Intel GPU](./install_windows_gpu.md)
 | 
					- [Install IPEX-LLM on Windows with Intel GPU](./install_windows_gpu.md)
 | 
				
			||||||
- [Install IPEX-LLM in Docker on Windows with Intel GPU](./docker_windows_gpu.md)
 | 
					 | 
				
			||||||
 | 
					
 | 
				
			||||||
## Inference
 | 
					## Inference
 | 
				
			||||||
 | 
					
 | 
				
			||||||
- [Run Performance Benchmarking with IPEX-LLM](./benchmark_quickstart.md)
 | 
					- [Run Performance Benchmarking with IPEX-LLM](./benchmark_quickstart.md)
 | 
				
			||||||
- [Run PyTorch Inference on Intel GPU using Docker (on Linux or WSL)](./docker_benchmark_quickstart.md)
 | 
					 | 
				
			||||||
- [Run Local RAG using Langchain-Chatchat on Intel GPU](./chatchat_quickstart.md)
 | 
					- [Run Local RAG using Langchain-Chatchat on Intel GPU](./chatchat_quickstart.md)
 | 
				
			||||||
- [Run Text Generation WebUI on Intel GPU](./webui_quickstart.md)
 | 
					- [Run Text Generation WebUI on Intel GPU](./webui_quickstart.md)
 | 
				
			||||||
- [Run Open WebUI on Intel GPU](./open_webui_with_ollama_quickstart.md)
 | 
					- [Run Open WebUI on Intel GPU](./open_webui_with_ollama_quickstart.md)
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
		Loading…
	
		Reference in a new issue