Fix vLLM CPU /chat endpoint (#11748)
This commit is contained in:
parent
27b4b104ed
commit
044e486480
1 changed files with 0 additions and 1 deletions
|
|
@ -114,7 +114,6 @@ async def create_chat_completion(request: ChatCompletionRequest,
|
|||
return StreamingResponse(content=generator,
|
||||
media_type="text/event-stream")
|
||||
else:
|
||||
invalidInputError(isinstance(generator, ChatCompletionResponse))
|
||||
return JSONResponse(content=generator.model_dump())
|
||||
|
||||
|
||||
|
|
|
|||
Loading…
Reference in a new issue