if request.response_format and request.response_format.type != 'text':
if VariableInterface.async_engine.backend != 'pytorch':
return create_error_response(
HTTPStatus.BAD_REQUEST,
'only pytorch backend can use response_format now')
response_format = request.response_format.model_dump()
Motivation
I'm using TurboMind engine and I got an error while requesting
response_format
withjson_schema
. Code here: https://github.com/InternLM/lmdeploy/blob/main/lmdeploy/serve/openai/api_server.py#L367Is there any plan to support this for TurboMind?
Related resources
Additional context
No response