SwiftKV optimizations developed and integrated into vLLM can improve LLM inference throughput by up to 50%, the company said.