Skip to content

Pinned Loading

  1. vllm vllm Public

    A high-throughput and memory-efficient inference and serving engine for LLMs

    Python 65.5k 12k

  2. llm-compressor llm-compressor Public

    Transformers-compatible library for applying various compression algorithms to LLMs for optimized deployment with vLLM

    Python 2.4k 316

  3. recipes recipes Public

    Common recipes to run vLLM

    Jupyter Notebook 281 103

  4. speculators speculators Public

    A unified library for building, evaluating, and storing speculative decoding algorithms for LLM inference in vLLM

    Python 159 21

  5. semantic-router semantic-router Public

    Intelligent Router for Mixture-of-Models

    Go 2.4k 312

Repositories

Showing 10 of 30 repositories
  • vllm Public

    A high-throughput and memory-efficient inference and serving engine for LLMs

    vllm-project/vllm’s past year of commit activity
    Python 65,548 Apache-2.0 12,000 1,863 (37 issues need help) 1,287 Updated Dec 16, 2025
  • tpu-inference Public

    TPU inference for vLLM, with unified JAX and PyTorch support.

    vllm-project/tpu-inference’s past year of commit activity
    Python 196 Apache-2.0 60 16 (1 issue needs help) 72 Updated Dec 16, 2025
  • vllm-spyre Public

    Community maintained hardware plugin for vLLM on Spyre

    vllm-project/vllm-spyre’s past year of commit activity
    Python 38 Apache-2.0 31 4 13 Updated Dec 16, 2025
  • llm-compressor Public

    Transformers-compatible library for applying various compression algorithms to LLMs for optimized deployment with vLLM

    vllm-project/llm-compressor’s past year of commit activity
    Python 2,404 Apache-2.0 316 72 (16 issues need help) 50 Updated Dec 16, 2025
  • compressed-tensors Public

    A safetensors extension to efficiently store sparse quantized tensors on disk

    vllm-project/compressed-tensors’s past year of commit activity
    Python 219 Apache-2.0 45 3 (1 issue needs help) 12 Updated Dec 16, 2025
  • vllm-omni Public

    A framework for efficient model inference with omni-modality models

    vllm-project/vllm-omni’s past year of commit activity
    Python 938 Apache-2.0 126 62 (29 issues need help) 39 Updated Dec 17, 2025
  • ci-infra Public

    This repo hosts code for vLLM CI & Performance Benchmark infrastructure.

    vllm-project/ci-infra’s past year of commit activity
    HCL 27 Apache-2.0 51 0 29 Updated Dec 16, 2025
  • semantic-router Public

    Intelligent Router for Mixture-of-Models

    vllm-project/semantic-router’s past year of commit activity
    Go 2,426 Apache-2.0 312 93 (14 issues need help) 34 Updated Dec 16, 2025
  • vllm-gaudi Public

    Community maintained hardware plugin for vLLM on Intel Gaudi

    vllm-project/vllm-gaudi’s past year of commit activity
    Python 20 Apache-2.0 82 1 66 Updated Dec 16, 2025
  • speculators Public

    A unified library for building, evaluating, and storing speculative decoding algorithms for LLM inference in vLLM

    vllm-project/speculators’s past year of commit activity
    Python 159 Apache-2.0 21 8 (4 issues need help) 11 Updated Dec 16, 2025