Uni-MoE-2.0-Omni: An Open Qwen2.5-7B Based Omnimodal MoE for Text, Image, Audio and Video Understanding
How do you construct one open mannequin that may reliably perceive textual content, photographs, audio and video whereas nonetheless working effectively? A group of researchers from Harbin Institute of Technology, Shenzhen launched Uni-MoE-2.0-Omni, a totally open omnimodal massive mannequin that pushes Lychee’s Uni-MoE line towards language centric multimodal reasoning. The system is skilled from scratch…
