#mixtureofexperts — Public Fediverse posts
Live and recent posts from across the Fediverse tagged #mixtureofexperts, aggregated by home.social.
-
Nemotron 3 Super pushes the frontier with 40 M supervised & alignment samples, leveraging a Mamba‑Transformer backbone and Mixture‑of‑Experts scaling. The model shows stronger agent reasoning, RL‑based fine‑tuning, and tighter AI alignment. Dive into the details to see how this LLM reshapes open‑source AI. #Nemotron3 #MixtureOfExperts #AIAlignment #SupervisedFineTuning
🔗 https://aidailypost.com/news/nemotron-3-super-incorporates-40-million-supervised-alignment-samples
-
MiniMax's new M2.5 model slashes costs to 1/20 of Claude Opus while handling 30% of HQ tasks. Built on a Mixture‑of‑Experts sparse architecture, it delivers strong code‑generation and LLM performance—all open‑source. Discover how this AI agent could boost productivity in your projects. #MiniMaxM2_5 #MixtureOfExperts #OpenSourceAI #AIProductivity
🔗 https://aidailypost.com/news/minimaxs-m25-costs-120-claude-opus-covers-30-hq-tasks
-
DeepSeek-Math-V2: Open-Source AI Earns IMO Gold, Tops Putnam Exam Chinese startup DeepSeek has released DeepSeek-Math-V2, an open-source AI model that solved five of six 2024 IMO problems, earning ...
#ChinaRevolutionUpdate #GenAIPro #AI #mathematical #reasoning #DeepSeek-Math-V2 #International #Mathematical #Olympiad #mixture-of-experts #system
Origin | Interest | Match -
Starting the new year with a blog post on #MixtureOfExperts. Most folks I talked with had a fundamental gap and they are a game-changer in AI.
"What are #MoEs and why is it a game-changer for AI? 🤔 Find out in my latest blog post! 👉 https://blog.desigeek.com/post/2025/01/intro-to-mixture-of-experts/ #AI #GenAI #DL