MiniMax-01 is Now Open-Source: Scaling Lightning Attention for the AI Agent Era

MiniMax-01 is Now Open-Source: Scaling Lightning Attention for the AI Agent Era

Today, MiniMax has released and open-sourced the all-new MiniMax-01 series of models, which includes two models: the foundational language model MiniMax-Text-01 and the visual multi-modal model MiniMax-VL-01.

Report link: filecdn.minimax.chat/_Arxiv_MiniMax_01_Repo..

Innovative Lightning Attention Architecture, with Top-tier Model Performance

In the MiniMax-01 series, we have made a bold innovation: for the first time on a large scale, we have implemented a novel Lightning Attention mechanism, offering a new alternative to the traditional Transformer architecture. This model boasts a staggering 456 billion parameters, with 45.9 billion parameters activated per inference. Its comprehensive performance is on par with the leading models globally, while it efficiently handles the world's longest context length of up to 4 million tokens—20 to 32 times more than other leading models.