Fix vllm api_server v1/models error (#12867)
This commit is contained in:
		
							parent
							
								
									8077850452
								
							
						
					
					
						commit
						3ea5389a99
					
				
					 1 changed files with 7 additions and 3 deletions
				
			
		| 
						 | 
				
			
			@ -278,6 +278,10 @@ def base(request: Request) -> OpenAIServing:
 | 
			
		|||
    return tokenization(request)
 | 
			
		||||
 | 
			
		||||
 | 
			
		||||
def models(request: Request) -> OpenAIServingModels:
 | 
			
		||||
    return request.app.state.openai_serving_models
 | 
			
		||||
 | 
			
		||||
 | 
			
		||||
def chat(request: Request) -> Optional[OpenAIServingChat]:
 | 
			
		||||
    return request.app.state.openai_serving_chat
 | 
			
		||||
 | 
			
		||||
| 
						 | 
				
			
			@ -345,10 +349,10 @@ async def detokenize(request: DetokenizeRequest, raw_request: Request):
 | 
			
		|||
 | 
			
		||||
@router.get("/v1/models")
 | 
			
		||||
async def show_available_models(raw_request: Request):
 | 
			
		||||
    handler = base(raw_request)
 | 
			
		||||
    handler = models(raw_request)
 | 
			
		||||
 | 
			
		||||
    models = await handler.show_available_models()
 | 
			
		||||
    return JSONResponse(content=models.model_dump())
 | 
			
		||||
    models_ = await handler.show_available_models()
 | 
			
		||||
    return JSONResponse(content=models_.model_dump())
 | 
			
		||||
 | 
			
		||||
 | 
			
		||||
@router.get("/version")
 | 
			
		||||
| 
						 | 
				
			
			
 | 
			
		|||
		Loading…
	
		Reference in a new issue