update deepspeed install and run guide in README (#9441)
This commit is contained in:
		
							parent
							
								
									30abd304a7
								
							
						
					
					
						commit
						921b263d6a
					
				
					 1 changed files with 3 additions and 1 deletions
				
			
		| 
						 | 
					@ -8,6 +8,8 @@ Install necessary packages (here Python 3.9 is our test environment):
 | 
				
			||||||
bash install.sh
 | 
					bash install.sh
 | 
				
			||||||
```
 | 
					```
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					The first step in the script is to install oneCCL (wrapper for Intel MPI) to enable distributed communication between deepspeed instances, which can be skipped if Inte MPI/oneCCL/oneAPI has already been prepared on your machine. Please refer to [oneCCL](https://github.com/oneapi-src/oneCCL) if any related issue when install or import.
 | 
				
			||||||
 | 
					
 | 
				
			||||||
#### 2. Initialize Deepspeed Distributed Context
 | 
					#### 2. Initialize Deepspeed Distributed Context
 | 
				
			||||||
 | 
					
 | 
				
			||||||
Like shown in example code `deepspeed_autotp.py`, you can construct parallel model with Python API:
 | 
					Like shown in example code `deepspeed_autotp.py`, you can construct parallel model with Python API:
 | 
				
			||||||
| 
						 | 
					@ -57,7 +59,7 @@ If you want to run your own application, there are **necessary configurations in
 | 
				
			||||||
 | 
					
 | 
				
			||||||
```bash
 | 
					```bash
 | 
				
			||||||
# run.sh
 | 
					# run.sh
 | 
				
			||||||
source bigdl-nano-init
 | 
					source bigdl-llm-init
 | 
				
			||||||
unset OMP_NUM_THREADS # deepspeed will set it for each instance automatically
 | 
					unset OMP_NUM_THREADS # deepspeed will set it for each instance automatically
 | 
				
			||||||
source /opt/intel/oneccl/env/setvars.sh
 | 
					source /opt/intel/oneccl/env/setvars.sh
 | 
				
			||||||
......
 | 
					......
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
		Loading…
	
		Reference in a new issue