Cerebras Releases MiniMax-M2-REAP-162B-A10B: A Memory Efficient Version of MiniMax-M2 for Long Context Coding Agents
Cerebras has launched MiniMax-M2-REAP-162B-A10B, a compressed Sparse Mixture-of-Experts (SMoE) Causal Language Model derived from MiniMax-M2, utilizing the brand new Router weighted Expert Activation Pruning (REAP) technique. The mannequin retains the habits of the unique 230B complete, 10B energetic MiniMax M2, whereas pruning specialists and decreasing reminiscence for deployment targeted workloads resembling coding brokers and power…
