neuralmagic 's Collections

INT8 LLMs for vLLM

Accurate INT8 quantized models by Neural Magic, ready for use with vLLM!