Update fastchat demo script (#12367)
* Update README.md * Update vllm_docker_quickstart.md
This commit is contained in:
		
							parent
							
								
									51f7f87768
								
							
						
					
					
						commit
						fad15c8ca0
					
				
					 2 changed files with 6 additions and 2 deletions
				
			
		| 
						 | 
					@ -102,7 +102,9 @@ To set up model serving using `IPEX-LLM` as backend using FastChat, you can refe
 | 
				
			||||||
    
 | 
					    
 | 
				
			||||||
    # start controller
 | 
					    # start controller
 | 
				
			||||||
    python -m fastchat.serve.controller &
 | 
					    python -m fastchat.serve.controller &
 | 
				
			||||||
    
 | 
					
 | 
				
			||||||
 | 
					    export USE_XETLA=OFF
 | 
				
			||||||
 | 
					    export SYCL_PI_LEVEL_ZERO_USE_IMMEDIATE_COMMANDLISTS=2
 | 
				
			||||||
    
 | 
					    
 | 
				
			||||||
    export TORCH_LLM_ALLREDUCE=0
 | 
					    export TORCH_LLM_ALLREDUCE=0
 | 
				
			||||||
    export CCL_DG2_ALLREDUCE=1
 | 
					    export CCL_DG2_ALLREDUCE=1
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
| 
						 | 
					@ -831,7 +831,9 @@ We can set up model serving using `IPEX-LLM` as backend using FastChat, the foll
 | 
				
			||||||
    
 | 
					    
 | 
				
			||||||
    # start controller
 | 
					    # start controller
 | 
				
			||||||
    python -m fastchat.serve.controller &
 | 
					    python -m fastchat.serve.controller &
 | 
				
			||||||
    
 | 
					
 | 
				
			||||||
 | 
					    export USE_XETLA=OFF
 | 
				
			||||||
 | 
					    export SYCL_PI_LEVEL_ZERO_USE_IMMEDIATE_COMMANDLISTS=2
 | 
				
			||||||
    
 | 
					    
 | 
				
			||||||
    export TORCH_LLM_ALLREDUCE=0
 | 
					    export TORCH_LLM_ALLREDUCE=0
 | 
				
			||||||
    export CCL_DG2_ALLREDUCE=1
 | 
					    export CCL_DG2_ALLREDUCE=1
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
		Loading…
	
		Reference in a new issue