Baidu Releases ERNIE-4.5-21B-A3B-Thinking: A Compact MoE Model for Deep Reasoning
Baidu AI Research crew has simply launched ERNIE-4.5-21B-A3B-Thinking, a brand new reasoning-focused giant language mannequin designed round effectivity, long-context reasoning, and gear integration. Being a part of the ERNIE-4.5 household, this mannequin is a Mixture-of-Experts (MoE) structure with 21B whole parameters however solely 3B energetic parameters per token, making it computationally environment friendly whereas sustaining…
