About 55,700 results
Open links in new tab
  1. INTELLECT-3: Prime Intellect's 106B MoE Model Trained End-to ...

    3 days ago · Prime Intellect just released INTELLECT-3, a 106B-parameter Mixture-of-Experts (MoE) model that utilizes only 12B active parameters at inference time. This model is trained …

  2. INTELLECT-3: A 100B+ MoE trained with large-scale RL

    6 days ago · INTELLECT-3 is a 106B parameter Mixture-of-Experts model trained with both SFT and RL on top of the GLM 4.5 Air base model. It achieves state-of-the-art performance for its …

  3. PrimeIntellect/INTELLECT-3 · Hugging Face

    6 days ago · Trained with prime-rl and verifiers Environments released on Environments Hub Read the Blog & Technical Report X | Discord | Prime Intellect Platform Introduction …

  4. Prime Intellect: INTELLECT-3 – NextAutomatica

    INTELLECT-3 is a 106B-parameter Mixture-of-Experts model (12B active) post-trained from GLM-4.5-Air-Base using supervised fine-tuning (SFT) followed by large-scale reinforcement learning …

  5. prime-intellect | OpenRouter

    INTELLECT-3 is a 106B-parameter Mixture-of-Experts model (12B active) post-trained from GLM-4.5-Air-Base using supervised fine-tuning (SFT) followed by large-scale reinforcement learning …

  6. INTELLECT-3: The new 106B MoE model that revolutionizes ...

    3 days ago · Learn more Discover INTELLECT-3, a 106B parameter Mixture-of-Experts model, trained with SFT + RL on GLM-4.5-Air and built entirely with open-source tools.

  7. Scaling RL and Self-Verifiable Reasoning: INTELLECT-3 and ...

    5 days ago · The Weekly Kaitchup #120GLM models are very popular now as they perform well on most tasks. Among open-weight models, I prefer them over recent DeepSeek and Kimi …