update transformers version in example of glm4 (#12453)
* fix: update transformers version in example of glm4 * fix: textual adjustments * fix: texual adjustment
This commit is contained in:
		
							parent
							
								
									effb9bb41c
								
							
						
					
					
						commit
						ce6fcaa9ba
					
				
					 2 changed files with 6 additions and 6 deletions
				
			
		| 
						 | 
					@ -17,8 +17,8 @@ conda activate llm
 | 
				
			||||||
# install the latest ipex-llm nightly build with 'all' option
 | 
					# install the latest ipex-llm nightly build with 'all' option
 | 
				
			||||||
pip install --pre --upgrade ipex-llm[all] --extra-index-url https://download.pytorch.org/whl/cpu
 | 
					pip install --pre --upgrade ipex-llm[all] --extra-index-url https://download.pytorch.org/whl/cpu
 | 
				
			||||||
 | 
					
 | 
				
			||||||
# install packages required for GLM-4
 | 
					# install packages required for GLM-4, it is recommended to use transformers>=4.44 for THUDM/glm-4-9b-chat updated after August 12, 2024
 | 
				
			||||||
pip install "tiktoken>=0.7.0" transformers==4.42.4 "trl<0.12.0"
 | 
					pip install "tiktoken>=0.7.0" transformers==4.44 "trl<0.12.0"
 | 
				
			||||||
```
 | 
					```
 | 
				
			||||||
 | 
					
 | 
				
			||||||
On Windows:
 | 
					On Windows:
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
| 
						 | 
					@ -13,8 +13,8 @@ conda activate llm
 | 
				
			||||||
# below command will install intel_extension_for_pytorch==2.1.10+xpu as default
 | 
					# below command will install intel_extension_for_pytorch==2.1.10+xpu as default
 | 
				
			||||||
pip install --pre --upgrade ipex-llm[xpu] --extra-index-url https://pytorch-extension.intel.com/release-whl/stable/xpu/us/
 | 
					pip install --pre --upgrade ipex-llm[xpu] --extra-index-url https://pytorch-extension.intel.com/release-whl/stable/xpu/us/
 | 
				
			||||||
 | 
					
 | 
				
			||||||
# install packages required for GLM-4
 | 
					# install packages required for GLM-4, it is recommended to use transformers>=4.44 for THUDM/glm-4-9b-chat updated after August 12, 2024
 | 
				
			||||||
pip install "tiktoken>=0.7.0" transformers==4.42.4 "trl<0.12.0"
 | 
					pip install "tiktoken>=0.7.0" transformers==4.44 "trl<0.12.0"
 | 
				
			||||||
```
 | 
					```
 | 
				
			||||||
 | 
					
 | 
				
			||||||
### 1.2 Installation on Windows
 | 
					### 1.2 Installation on Windows
 | 
				
			||||||
| 
						 | 
					@ -26,8 +26,8 @@ conda activate llm
 | 
				
			||||||
# below command will install intel_extension_for_pytorch==2.1.10+xpu as default
 | 
					# below command will install intel_extension_for_pytorch==2.1.10+xpu as default
 | 
				
			||||||
pip install --pre --upgrade ipex-llm[xpu] --extra-index-url https://pytorch-extension.intel.com/release-whl/stable/xpu/us/
 | 
					pip install --pre --upgrade ipex-llm[xpu] --extra-index-url https://pytorch-extension.intel.com/release-whl/stable/xpu/us/
 | 
				
			||||||
 | 
					
 | 
				
			||||||
# install packages required for GLM-4
 | 
					# install packages required for GLM-4, it is recommended to use transformers>=4.44 for THUDM/glm-4-9b-chat updated after August 12, 2024
 | 
				
			||||||
pip install "tiktoken>=0.7.0" transformers==4.42.4 "trl<0.12.0"
 | 
					pip install "tiktoken>=0.7.0" transformers==4.44 "trl<0.12.0"
 | 
				
			||||||
```
 | 
					```
 | 
				
			||||||
 | 
					
 | 
				
			||||||
## 2. Configures OneAPI environment variables for Linux
 | 
					## 2. Configures OneAPI environment variables for Linux
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
		Loading…
	
		Reference in a new issue