Hugging Face: Continuous batching capability in LLM inference — from attention to KV caching | SignalBreak | SignalBreak