LLM: add initial FAQ page (#10055)
This commit is contained in:
		
							parent
							
								
									d2d3f6b091
								
							
						
					
					
						commit
						4b92235bdb
					
				
					 4 changed files with 37 additions and 0 deletions
				
			
		| 
						 | 
					@ -64,6 +64,14 @@ subtrees:
 | 
				
			||||||
          #   title: "Tips and Known Issues"
 | 
					          #   title: "Tips and Known Issues"
 | 
				
			||||||
          - file: doc/PythonAPI/LLM/index
 | 
					          - file: doc/PythonAPI/LLM/index
 | 
				
			||||||
            title: "API Reference"
 | 
					            title: "API Reference"
 | 
				
			||||||
 | 
					          - file: doc/LLM/Overview/FAQ/index
 | 
				
			||||||
 | 
					            title: "FAQ"
 | 
				
			||||||
 | 
					            subtrees:
 | 
				
			||||||
 | 
					              - entries:
 | 
				
			||||||
 | 
					                - file: doc/LLM/Overview/FAQ/general_info
 | 
				
			||||||
 | 
					                  title: "General Info & Concepts"
 | 
				
			||||||
 | 
					                - file: doc/LLM/Overview/FAQ/resolve_error
 | 
				
			||||||
 | 
					                  title: "How to Resolve Errors"
 | 
				
			||||||
 | 
					
 | 
				
			||||||
  - entries:
 | 
					  - entries:
 | 
				
			||||||
    - file: doc/Orca/index
 | 
					    - file: doc/Orca/index
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
							
								
								
									
										10
									
								
								docs/readthedocs/source/doc/LLM/Overview/FAQ/general_info.md
									
									
									
									
									
										Normal file
									
								
							
							
						
						
									
										10
									
								
								docs/readthedocs/source/doc/LLM/Overview/FAQ/general_info.md
									
									
									
									
									
										Normal file
									
								
							| 
						 | 
					@ -0,0 +1,10 @@
 | 
				
			||||||
 | 
					# FAQ: General Info & Concepts
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					Refer to this section for general information about BigDL-LLM.
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					## BigDL-LLM Support
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					### GGUF format usage with BigDL-LLM?
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					BigDL-LLM supports running GGUF/AWQ/GPTQ models on both [CPU](https://github.com/intel-analytics/BigDL/tree/main/python/llm/example/CPU/HF-Transformers-AutoModels/Advanced-Quantizations) and [GPU](https://github.com/intel-analytics/BigDL/tree/main/python/llm/example/GPU/HF-Transformers-AutoModels/Advanced-Quantizations).
 | 
				
			||||||
 | 
					Please also refer to [here](https://github.com/intel-analytics/BigDL?tab=readme-ov-file#latest-update-) for our latest support.
 | 
				
			||||||
							
								
								
									
										7
									
								
								docs/readthedocs/source/doc/LLM/Overview/FAQ/index.rst
									
									
									
									
									
										Normal file
									
								
							
							
						
						
									
										7
									
								
								docs/readthedocs/source/doc/LLM/Overview/FAQ/index.rst
									
									
									
									
									
										Normal file
									
								
							| 
						 | 
					@ -0,0 +1,7 @@
 | 
				
			||||||
 | 
					Frequently Asked Questions (FAQ)
 | 
				
			||||||
 | 
					================================
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					You could refer to corresponding page to find solutions of your requirement:
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					* `General Info & Concepts <./general_info.html>`_
 | 
				
			||||||
 | 
					* `How to Resolve Errors <./resolve_error.html>`_
 | 
				
			||||||
| 
						 | 
					@ -0,0 +1,12 @@
 | 
				
			||||||
 | 
					# FAQ: How to Resolve Errors
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					Refer to this section for common issues faced while using BigDL-LLM.
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					## Runtime Error
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					### PyTorch is not linked with support for xpu devices
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					1. Before running on Intel GPUs, please make sure you've prepared environment follwing [installation instruction](https://bigdl.readthedocs.io/en/latest/doc/LLM/Overview/install_gpu.html).
 | 
				
			||||||
 | 
					2. If you are using an older version of `bigdl-llm` (specifically, older than 2.5.0b20240104), you need to manually add `import intel_extension_for_pytorch as ipex` at the beginning of your code.
 | 
				
			||||||
 | 
					3. After optimizing the model with BigDL-LLM, you need to move model to GPU through `model = model.to('xpu')`.
 | 
				
			||||||
 | 
					4. If you have mutil GPUs, you could refer to [here](https://bigdl.readthedocs.io/en/latest/doc/LLM/Overview/KeyFeatures/multi_gpus_selection.html) for details about GPU selection.
 | 
				
			||||||
		Loading…
	
		Reference in a new issue