#sglang — Public Fediverse posts
Live and recent posts from across the Fediverse tagged #sglang, aggregated by home.social.
-
Install SGLang with uv, pip, or Docker; configure YAML and server flags; then serve Hugging Face LLMs with an OpenAI-compatible API plus native /generate and offline Engine examples.
#Cheatsheet #Self-Hosting #LLM #AI #AI Coding #DevOps #Docker #sglang #openai #SelfHosting
-
SGLang and vLLM Workshops Coming to GOSIM Paris 2026!
The GOSIM Workshops have long been known for their diversity, hands-on learning, and interactivity, making them one of the most popular segments of the conference.
This May, the SGLang Workshop and vLLM Workshop will arrive at GOSIM Paris 2026, bringing together AI infrastructure developers from around the world to explore the latest advances in LLM inference systems.
Ticket purchase link:
https://eventbrite.com/e/gosim-paris-2026-tickets-1984013840806?aff=oddtdtcreator -
🚀 Big news!
The SGLang Workshop & vLLM Workshop are coming to GOSIM Paris 2026! 🎉
🌐 A must-attend event for AI developers and open-source contributors worldwide
💡 Dive into cutting-edge topics: large model inference, agentic AI, and more
🎓 Hands-on sessions and discussions to bring high-value learning and networkingGet your early bird tickets now and enjoy the discount: https://eventbrite.com/e/gosim-paris-2026-tickets-1984013840806?aff=oddtdtcreator 🚀
-
📦 Model weights for base and chat variants coming soon to #HuggingFace and #ModelScope with support for #vLLM and #SGLang inference frameworks
📋 Complete evaluation details and trajectory data publicly available for community research on HuggingFace datasets
-
📦 Model weights for base and chat variants coming soon to #HuggingFace and #ModelScope with support for #vLLM and #SGLang inference frameworks
📋 Complete evaluation details and trajectory data publicly available for community research on HuggingFace datasets
-
📦 Model weights for base and chat variants coming soon to #HuggingFace and #ModelScope with support for #vLLM and #SGLang inference frameworks
📋 Complete evaluation details and trajectory data publicly available for community research on HuggingFace datasets
-
📦 Model weights for base and chat variants coming soon to #HuggingFace and #ModelScope with support for #vLLM and #SGLang inference frameworks
📋 Complete evaluation details and trajectory data publicly available for community research on HuggingFace datasets
-
📦 Model weights for base and chat variants coming soon to #HuggingFace and #ModelScope with support for #vLLM and #SGLang inference frameworks
📋 Complete evaluation details and trajectory data publicly available for community research on HuggingFace datasets
-
Как запустить свою LLM для инференса. Руководство по запуску: Ollama, vLLM, Triton, LM Studio, llama.cpp, SGLang
В этой статье будет приведено практическое руководство по базовой настройке и запуску следующих инструментов для работы с LLM: Ollama, LM Studio, vLLM, Triton, llama.cpp, SGLang. 🔥 Начинаем? 🔥
https://habr.com/ru/articles/948934/
#ollama #vllm #triton #lm_studio #llamacpp #sglang #запуск_llm
-
🤖 Oh joy, another thrilling journey through the riveting world of Flash Attention in SGLang! 🌟 Because clearly, the universe was desperately yearning for a detailed breakdown of yet another backend implementation. 🤯 Guess #SGLang 0.4.6 just wouldn’t be the same without it! 🥳
https://hebiao064.github.io/fa3-attn-backend-basic #FlashAttention #BackendImplementation #TechNews #Innovation #Excitement #HackerNews #ngated -
🤖 Oh joy, another thrilling journey through the riveting world of Flash Attention in SGLang! 🌟 Because clearly, the universe was desperately yearning for a detailed breakdown of yet another backend implementation. 🤯 Guess #SGLang 0.4.6 just wouldn’t be the same without it! 🥳
https://hebiao064.github.io/fa3-attn-backend-basic #FlashAttention #BackendImplementation #TechNews #Innovation #Excitement #HackerNews #ngated -
🤖 Oh joy, another thrilling journey through the riveting world of Flash Attention in SGLang! 🌟 Because clearly, the universe was desperately yearning for a detailed breakdown of yet another backend implementation. 🤯 Guess #SGLang 0.4.6 just wouldn’t be the same without it! 🥳
https://hebiao064.github.io/fa3-attn-backend-basic #FlashAttention #BackendImplementation #TechNews #Innovation #Excitement #HackerNews #ngated -
🤖 Oh joy, another thrilling journey through the riveting world of Flash Attention in SGLang! 🌟 Because clearly, the universe was desperately yearning for a detailed breakdown of yet another backend implementation. 🤯 Guess #SGLang 0.4.6 just wouldn’t be the same without it! 🥳
https://hebiao064.github.io/fa3-attn-backend-basic #FlashAttention #BackendImplementation #TechNews #Innovation #Excitement #HackerNews #ngated -
Implement Flash Attention Back End in SGLang – Basics and KV Cache
https://hebiao064.github.io/fa3-attn-backend-basic
#HackerNews #ImplementFlashAttention #SGLang #KVCache #BackEnd #AIResearch #TechTutorial