A High-Performance LLM Inference Engine with vLLM-Style Continuous Batching
-
Updated
Jan 2, 2026 - C++
A High-Performance LLM Inference Engine with vLLM-Style Continuous Batching
gLLM: Global Balanced Pipeline Parallelism System for Distributed LLM Serving with Token Throttling
OpenAI-compatible server with continuous batching for MLX on Apple Silicon
Fork of OpenAI and Anthropic compatible server for Apple Silicon. Native MLX backend, 500+ tok/s. Run LLMs and vision-language models with continuous batching, MCP tool calling, and multimodal support.
PagedAttention + Continuous Batching Inference Engine Prototype (Rust): Paged KV Cache & Dynamic Scheduling | PagedAttention + Continuous Batching 推理引擎原型(Rust),KV Cache 分页管理与动态调度
Add a description, image, and links to the continuous-batching topic page so that developers can more easily learn about it.
To associate your repository with the continuous-batching topic, visit your repo's landing page and select "manage topics."