Update AWQ and GPTQ GPU example (#12300)
This commit is contained in:
		
							parent
							
								
									0763268e4c
								
							
						
					
					
						commit
						6f22133efc
					
				
					 2 changed files with 0 additions and 6 deletions
				
			
		| 
						 | 
					@ -37,15 +37,10 @@ conda create -n llm python=3.11
 | 
				
			||||||
conda activate llm
 | 
					conda activate llm
 | 
				
			||||||
# below command will install intel_extension_for_pytorch==2.1.10+xpu as default
 | 
					# below command will install intel_extension_for_pytorch==2.1.10+xpu as default
 | 
				
			||||||
pip install --pre --upgrade ipex-llm[xpu] --extra-index-url https://pytorch-extension.intel.com/release-whl/stable/xpu/us/
 | 
					pip install --pre --upgrade ipex-llm[xpu] --extra-index-url https://pytorch-extension.intel.com/release-whl/stable/xpu/us/
 | 
				
			||||||
pip install transformers==4.35.0
 | 
					 | 
				
			||||||
pip install autoawq==0.1.8 --no-deps
 | 
					pip install autoawq==0.1.8 --no-deps
 | 
				
			||||||
pip install accelerate==0.25.0
 | 
					pip install accelerate==0.25.0
 | 
				
			||||||
pip install einops
 | 
					pip install einops
 | 
				
			||||||
```
 | 
					```
 | 
				
			||||||
**Note: For Mixtral model, please use transformers 4.36.0:**
 | 
					 | 
				
			||||||
```bash
 | 
					 | 
				
			||||||
pip install transformers==4.36.0
 | 
					 | 
				
			||||||
```
 | 
					 | 
				
			||||||
 | 
					
 | 
				
			||||||
### 2. Configures OneAPI environment variables
 | 
					### 2. Configures OneAPI environment variables
 | 
				
			||||||
 | 
					
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
| 
						 | 
					@ -13,7 +13,6 @@ conda create -n llm python=3.11
 | 
				
			||||||
conda activate llm
 | 
					conda activate llm
 | 
				
			||||||
# below command will install intel_extension_for_pytorch==2.1.10+xpu as default
 | 
					# below command will install intel_extension_for_pytorch==2.1.10+xpu as default
 | 
				
			||||||
pip install --pre --upgrade ipex-llm[xpu] --extra-index-url https://pytorch-extension.intel.com/release-whl/stable/xpu/us/
 | 
					pip install --pre --upgrade ipex-llm[xpu] --extra-index-url https://pytorch-extension.intel.com/release-whl/stable/xpu/us/
 | 
				
			||||||
pip install transformers==4.37.0
 | 
					 | 
				
			||||||
pip install auto_gptq==0.7.1
 | 
					pip install auto_gptq==0.7.1
 | 
				
			||||||
pip install optimum==1.14.0
 | 
					pip install optimum==1.14.0
 | 
				
			||||||
```
 | 
					```
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
		Loading…
	
		Reference in a new issue