Fix vLLM CPU /chat endpoint (#11748)
This commit is contained in:
		
							parent
							
								
									27b4b104ed
								
							
						
					
					
						commit
						044e486480
					
				
					 1 changed files with 0 additions and 1 deletions
				
			
		| 
						 | 
					@ -114,7 +114,6 @@ async def create_chat_completion(request: ChatCompletionRequest,
 | 
				
			||||||
        return StreamingResponse(content=generator,
 | 
					        return StreamingResponse(content=generator,
 | 
				
			||||||
                                 media_type="text/event-stream")
 | 
					                                 media_type="text/event-stream")
 | 
				
			||||||
    else:
 | 
					    else:
 | 
				
			||||||
        invalidInputError(isinstance(generator, ChatCompletionResponse))
 | 
					 | 
				
			||||||
        return JSONResponse(content=generator.model_dump())
 | 
					        return JSONResponse(content=generator.model_dump())
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
		Loading…
	
		Reference in a new issue