vllm-project / vllm

A high-throughput and memory-efficient inference and serving engine for LLMs
https://docs.vllm.ai
Apache License 2.0
29.14k stars 4.35k forks source link

[Usage]: Passing a guided_json in offline inference #4899

Open ccdv-ai opened 5 months ago

ccdv-ai commented 5 months ago

Your current environment

vllm 0.4.2

How would you like to use vllm

I'm trying to force a json generation using outlines in offline inference. I don't see anything related in the documentation.

I haven't found an example of chat completion for offline inference, but I've managed to mimic it using chat templates, this is why I need to force a json generation.

amritap-ef commented 3 months ago

Also interested, any update on this @ccdv-ai ?

github-actions[bot] commented 9 hours ago

This issue has been automatically marked as stale because it has not had any activity within 90 days. It will be automatically closed if no further activity occurs within 30 days. Leave a comment if you feel this issue should remain open. Thank you!