microsoft / LLMLingua

To speed up LLMs' inference and enhance LLM's perceive of key information, compress the prompt and KV-Cache, which achieves up to 20x compression with minimal performance loss.
https://llmlingua.com/
MIT License
4.42k stars 241 forks source link

prompt中的结构化数据 #50

Open growmuye opened 7 months ago

growmuye commented 7 months ago

prompt中包含json、xml无法正常压缩

iofu728 commented 7 months ago

Hi @growmuye, thank you for your interest in LLMLingua.

In the future, we plan to support a new feature that allows users to tag specify tokens that need to be preserved, particularly for maintaining formats such as JSON and XML

CEfanmin commented 3 months ago

返回的中文也是乱码