vLLM’s cover photo
vLLM

vLLM

Software Development

An open source, high-throughput and memory-efficient inference and serving engine for LLMs.

About us

vLLM is a high-throughput and memory-efficient inference and serving engine for LLMs

Website
https://github.com/vllm-project/vllm
Industry
Software Development
Company size
51-200 employees
Type
Nonprofit

Employees at vLLM

Updates

  • vLLM reposted this

    View profile for Daniel van Strien

    Machine Learning Librarian at Hugging Face 🤗 | Making AI work for libraries, archives, and their communities

    I just ran batch inference on a 30B parameter LLM across 4 GPUs with a single Python command! The secret? Modern AI infrastructure where everyone handles their specialty: 📦 UV (by Astral) handles dependencies via uv scripts 🖥️ Hugging Face Jobs handles GPU orchestration   🧠 Qwen AI team handles the model (Qwen3-30B-A3B-Instruct-2507) ⚡ vLLM handles efficient batched inference I'm very excited about using uv scripts as a nice way of packaging fairly simple but useful ML tasks in a somewhat reproducible way. This combined with Jobs opens up some nice oppertunities for making pipelines that require different types of compute. Technical deep dive and code examples: https://lnkd.in/e5BEBU95

  • vLLM reposted this

    View organization page for Anyscale

    51,628 followers

    🚨 Attention vLLM users – last call! 🚨 The Call for Proposals for our vLLM Featured Track at Ray Summit closes this Wednesday, July 30. If you're building with vLLM in production, optimizing inference, or exploring advanced use cases — we want to see it. This track is all about showcasing real-world implementations and hard-won lessons from the vLLM community. Need inspiration? Check out last year's top vLLM talks: https://lnkd.in/gmRhSbHk Submit your proposal here: https://lnkd.in/gjvKdvFF

    • No alternative text description for this image
  • vLLM reposted this

    View profile for Raushan Turganbay

    ML engineer at 🤗 | Multimodality and Generation | Erasmus Mundus MSc

    🚀 Big big news for multimodal devs! The transformers ↔️ vLLM integration just leveled up: Vision-Language Models are now supported out of the box If the model is integrated into Transformers, you can now run it directly with vLLM — no need to rewrite or duplicate code. Just plug it in and go. Zero extra effort Performance might differ model to model (we’re working on that!), but functional support is guaranteed Curious how to serve Transformers models with vLLM? Full docs here 👉 https://lnkd.in/d-KjqbmU #multimodal #transformers #vLLM #VLM #opensource

    • No alternative text description for this image
  • vLLM reposted this

    View organization page for NVIDIA AI

    1,325,015 followers

    🎉Congratulations to Microsoft for the new Phi-4-mini-flash-reasoning model trained on NVIDIA H100 and A100 GPUs. This latest edition to the Phi family provides developers with a new model optimized for high-throughput and low-latency reasoning in resource-constrained environments. Bring your data and try out demos on the multimodal playground for Phi on the NVIDIA API Catalog ➡️ https://lnkd.in/geuGhZsS 📷 The first plot shows average inference latency as a function of generation length, while the second plot illustrates how inference latency varies with throughput. Both experiments were conducted using the vLLM inference framework on a single A100-80GB GPU over varying concurrency levels of user requests. 🤗 https://lnkd.in/gswYMYt9

    • No alternative text description for this image
    • No alternative text description for this image
  • vLLM reposted this

    View profile for Erik Kaunismäki

    Software Engineer @Hugging Face 🤗

    We just released native support for SGLang and vLLM in Inference Endpoints 🔥 Inference Endpoints is becoming the central place where you deploy high performance Inference Engines. And that provides the managed infra for it. Instead of spending weeks configuring infrastructure, managing servers, and debugging deployment issues, you can focus on what matters most: your AI model and your users 🙌

    • No alternative text description for this image
  • View organization page for vLLM

    2,991 followers

    Calling vLLM users! In partnership with the Anyscale, we’re opening a special call for proposals for our first ever Featured Track – dedicated entirely to the most exciting inference work happening today. If you’re building with vLLM, we want to see what you’ve got. Last year's vLLM sessions were among our most popular – now we're giving this ecosystem the spotlight it deserves. Check out our Ray Summit 2024 vLLM sessions [linked in comments] for inspiration, then show us what you're building next. Submit your proposal by July 29 to be considered for this Featured Track. Submit here: https://lnkd.in/gjvKdvFF

  • vLLM reposted this

    Calling vLLM users! In partnership with the vLLM team, we’re opening a special call for proposals for our first ever Featured Track – dedicated entirely to the most exciting inference work happening today. If you’re building with vLLM, we want to see what you’ve got.Last year's vLLM sessions were among our most popular – now we're giving this ecosystem the spotlight it deserves. Check out our Ray Summit 2024 vLLM sessions [linked in comments] for inspiration, then show us what you're building next. Submit your proposal by July 29 to be considered for this Featured Track. Submit here: https://lnkd.in/gjvKdvFF

    • No alternative text description for this image
  • vLLM reposted this

    Calling vLLM users! In partnership with the vLLM team, we’re opening a special call for proposals for our first ever Featured Track – dedicated entirely to the most exciting inference work happening today. If you’re building with vLLM, we want to see what you’ve got.Last year's vLLM sessions were among our most popular – now we're giving this ecosystem the spotlight it deserves. Check out our Ray Summit 2024 vLLM sessions [linked in comments] for inspiration, then show us what you're building next. Submit your proposal by July 29 to be considered for this Featured Track. Submit here: https://lnkd.in/gjvKdvFF

    • No alternative text description for this image
  • vLLM reposted this

    A Pro-Tip for vLLM Users: Free ~90% of Your VRAM in Seconds Struggling with GPU memory while juggling multiple models? There's a lesser-known vLLM feature that lets your server "power-nap" and wake up in seconds—no restarts needed. Imagine you need to: - Hot-swap different models or checkpoints on the same GPU without a full server restart. - Run multiple models on a single GPU, cycling through them as needed for different tasks. - Optimize batch processing jobs that require different LLMs for various stages. - Free up GPU memory during the training phase in RLHF/GRPO loops, making the entire process more efficient. Quick Setup: - Enable sleep mode: export VLLM_SERVER_DEV_MODE=1 vllm serve $MODEL --enable-sleep-mode - Toggle with a simple API call: # Put the model to sleep curl -X POST :8000/sleep -d'{"level":1}' # Wake it up curl -X POST :8000/wake_up Heads up: 🔒 These are dev endpoints; keep them internal. 🧠 Level 1 sleep uses host RAM; Level 2 is slower to wake.

    • No alternative text description for this image
  • vLLM reposted this

    View organization page for MiniMax

    9,036 followers

    🎉 Join MiniMax-M1 Technical Seminar We’re excited to announce our first official seminar on MiniMax-M1 — the world’s first open-weight, large-scale hybrid-attention reasoning model, setting new standards in long-context reasoning with a 1M-token input and 80K-token output window. This online event brings together leading voices in AI from Anthropic, Hugging Face, vLLM, SGLang, MIT CSAIL, HKUST, University of Waterloo, and more — alongside the MiniMax technical team. 🔍 What to Expect: • Behind-the-scenes of M1’s architecture & algorithm design • Inference performance & real-world applications • Expert panel discussions • Live Q&A with the global AI community 📅 Date: Thursday, July 10, 2025 🕓 Time: 4 PM PST / 7 PM EST / 7 AM CST (July 11) 💻 Format: Zoom (Online only, limited seats) Whether you’re a researcher, developer, or AI enthusiast, we welcome you to join the discussion. Innovation begins with conversation. 👉 Register here: https://lu.ma/d7ptaky2

    • No alternative text description for this image

Similar pages