home.social

#mixtureofexperts — Public Fediverse posts

Live and recent posts from across the Fediverse tagged #mixtureofexperts, aggregated by home.social.

  1. Nemotron 3 Super pushes the frontier with 40 M supervised & alignment samples, leveraging a Mamba‑Transformer backbone and Mixture‑of‑Experts scaling. The model shows stronger agent reasoning, RL‑based fine‑tuning, and tighter AI alignment. Dive into the details to see how this LLM reshapes open‑source AI. #Nemotron3 #MixtureOfExperts #AIAlignment #SupervisedFineTuning

    🔗 aidailypost.com/news/nemotron-

  2. MiniMax's new M2.5 model slashes costs to 1/20 of Claude Opus while handling 30% of HQ tasks. Built on a Mixture‑of‑Experts sparse architecture, it delivers strong code‑generation and LLM performance—all open‑source. Discover how this AI agent could boost productivity in your projects. #MiniMaxM2_5 #MixtureOfExperts #OpenSourceAI #AIProductivity

    🔗 aidailypost.com/news/minimaxs-

  3. DeepSeek-Math-V2: Open-Source AI Earns IMO Gold, Tops Putnam Exam Chinese startup DeepSeek has released DeepSeek-Math-V2, an open-source AI model that solved five of six 2024 IMO problems, earning ...

    #ChinaRevolutionUpdate #GenAIPro #AI #mathematical #reasoning #DeepSeek-Math-V2 #International #Mathematical #Olympiad #mixture-of-experts #system

    Origin | Interest | Match
  4. Starting the new year with a blog post on #MixtureOfExperts. Most folks I talked with had a fundamental gap and they are a game-changer in AI.
    "What are #MoEs and why is it a game-changer for AI? 🤔 Find out in my latest blog post! 👉 blog.desigeek.com/post/2025/01 #AI #GenAI #DL