About Anaconda Help Download Anaconda

To speed up LLMs' inference and enhance LLM's perceive of key information, compress the prompt and KV-Cache, which achieves up to 20x compression with minimal performance loss.

copied from cf-staging / llmlingua
Label Latest Version
main 0.2.2

© 2025 Anaconda, Inc. All Rights Reserved. (v4.2.2) Legal | Privacy Policy