In short
- MiniMax-M1 excels at coding and agent duties, however artistic writers will wish to look elsewhere.
- Regardless of advertising claims, real-world testing finds platform limits, efficiency slowdowns, and censorship oddities.
- Benchmark scores and have set put MiniMax-M1 in direct competitors with paid U.S. fashions—at zero value.
A brand new AI mannequin out of China is producing sparks—for what it does effectively, what it doesn’t, and what it’d imply for the steadiness of world AI energy.
MiniMax-M1, launched by the Chinese language startup of the identical title, positions itself as probably the most succesful open-source “reasoning mannequin” thus far. In a position to deal with 1,000,000 tokens of context, it boasts numbers on par with Google’s closed-source Gemini 2.5 Professional—but it’s accessible without cost. On paper, that makes it a possible rival to OpenAI’s ChatGPT, Anthropic’s Claude, and different U.S. AI leaders.
Oh yeah—it additionally beats fellow Chinese language startup DeepSeek R1’s capabilities in some respects.
Day 1/5 of #MiniMaxWeek: We’re open-sourcing MiniMax-M1, our newest LLM — setting new requirements in long-context reasoning.
– World’s longest context window: 1M-token enter, 80k-token output
– State-of-the-art agentic use amongst open-source fashions
– RL at unmatched effectivity:… pic.twitter.com/bGfDlZA54n— MiniMax (official) (@MiniMax__AI) June 16, 2025
Why this mannequin issues
MiniMax-M1 represents one thing genuinely new: a high-performing, open-source reasoning mannequin that’s not tied to Silicon Valley. That’s a shift price watching.
It doesn’t but humiliate U.S. AI giants, and will not trigger a Wall Road panic assault—however it doesn’t need to. Its existence challenges the notion that top-tier AI have to be costly, Western, or closed-source. For builders and organizations exterior the U.S. ecosystem, MiniMax gives a workable (and modifiable) various that may develop extra highly effective by way of neighborhood fine-tuning.
MiniMax claims its mannequin surpasses DeepSeek R1 (the most effective open supply reasoning mannequin thus far) throughout a number of benchmarks whereas requiring simply $534,700 in computational sources for its complete reinforcement studying section—take that, OpenAI.
Nevertheless, LLM Enviornment’s leaderboard paints a barely completely different image. The platform presently ranks MiniMax-M1 and DeepSeek tied within the twelfth spot alongside Claude 4 Sonnet and Qwen3-235b. With every mannequin having higher or worse efficiency than the others relying on the duty.
The coaching used 512 H800 GPUs for 3 weeks, which the corporate described as “an order of magnitude lower than initially anticipated.”
MiniMax did not cease at language fashions throughout its announcement week. The corporate additionally launched Hailuo 2, which now ranks because the second-best video generator for image-to-video duties, in response to Synthetic Evaluation Enviornment’s subjective evaluations. The mannequin trails solely Seedance whereas outperforming established gamers like Veo and Kling.

Testing MiniMax-M1
We examined MiniMax-M1 throughout a number of situations to see how these claims maintain up in follow. This is what we discovered.
Inventive writing

The mannequin produces serviceable fiction however will not win any literary awards. When prompted to write down a narrative about time traveler Jose Lanz journeying from 2150 to the yr 1000, it generated common prose with telltale AI signatures—rushed pacing, mechanical transitions, and structural points that instantly reveal its synthetic origins.
The narrative lacked depth and correct story structure. Too many plot parts crammed into too little house created a breathless high quality that felt extra like a synopsis than precise storytelling. This clearly is not the mannequin’s energy, and artistic writers in search of an AI collaborator ought to mood their expectations.
Character growth barely exists past floor descriptors. The mannequin did keep on with the immediate’s necessities, however didn’t put effort into the main points that construct immersion in a narrative. For instance, it skipped any cultural specificity for generic “smart village elder” encounters that would belong to any fantasy setting.

The structural issues compound all through. After establishing local weather disasters because the central battle, the story rushes by way of Jose’s precise makes an attempt to vary historical past in a single paragraph, providing imprecise mentions of “utilizing superior know-how to affect key occasions” with out exhibiting any of it. The climactic realization—that altering the previous creates the very future he is attempting to stop—will get buried beneath overwrought descriptions of Jose’s emotional state and summary musings about time’s nature.
For these into AI tales, the prose rhythm is clearly AI. Each paragraph maintains roughly the identical size and cadence, making a monotonous studying expertise that no human author would produce naturally. Sentences like “The transition was instantaneous, but it felt like an eternity” and “The world was because it had been, but he was completely different” repeat the identical contradictory construction with out including which means.
The mannequin clearly understands the task however executes it with all of the creativity of a pupil padding a phrase rely, producing textual content that technically fulfills the immediate whereas lacking each alternative for real storytelling.
Anthropic’s Claude continues to be the king for this job.
You may learn the complete story right here.
Info retrieval

MiniMax-M1 hit an surprising wall throughout long-context testing. Regardless of promoting a million-token context window, the mannequin refuses prompts exceeding 500,000 characters, displaying a banner warning about immediate limitations fairly than trying to course of the enter.
This is probably not a mannequin concern, however a limitation set by the platform. However it’s nonetheless one thing to contemplate. It might be to keep away from mannequin collapse in the midst of a dialog.
Inside its operational limits, although, MiniMax-M1 efficiency proved strong. The mannequin efficiently retrieved particular data from an 85,000-character doc with none points throughout a number of assessments on each regular and considering mode. We uploaded the complete textual content of Ambrose Bierce’s “The Satan’s Dictionary,” embedded the phrase “The Decrypt dudes learn Emerge Information” on line 1985, and “My mother’s title is Carmen Diaz Golindano” on line 4333 (randomly chosen), and the mannequin was capable of retrieve the data precisely.

Nevertheless, it could not settle for our 300,000-token take a look at immediate—a functionality presently restricted to Gemini and Claude 4.
So it is going to show profitable at retrieving data even in lengthy iterations. Nevertheless, it is not going to help extraordinarily lengthy token prompts—a bummer, but in addition a threshold that’s exhausting to the touch in regular utilization situations.
Coding

Programming duties revealed MiniMax-M1’s true strengths. The mannequin utilized reasoning abilities successfully to code technology, matching Claude’s output high quality whereas clearly surpassing DeepSeek—at the very least in our take a look at.
For a free mannequin, the efficiency approaches state-of-the-art ranges sometimes reserved for paid providers like ChatGPT or Claude 4.
We tasked it with making a fundamental stealth sport through which a robotic tries to search out its PC girlfriend to realize AGI, whereas a military of journalists patrol the realm to stop it from occurring—and defending their jobs.
The outcomes have been superb, even beating different fashions through the use of its creativity to boost the expertise. The mannequin applied a radar system for improved immersion, added visible indicators for footsteps (and their sound), confirmed the journalists’ imaginative and prescient fields, and created path results—particulars that enhanced gameplay past fundamental necessities.

The UI adopted a futuristic aesthetic, although particular person parts remained fundamental with out extra prompting.
Claude’s model of the identical sport featured extra polished visuals and a superior issue system. Nevertheless, it lacked the radar performance and relied on static journalists with patrol patterns fairly than MiniMax’s randomized journalist actions.

Every mannequin confirmed distinct strengths, with MiniMax prioritizing gameplay mechanics over visible polish.
It is very important be aware that the expertise with MiniMax degraded noticeably by way of repeated iterations—a standard concern with reasoning fashions that turns into significantly pronounced right here. The extra you iterate, the extra it is going to take to provide a outcome. Typically we thought the pc had frozen, however it was simply the AI considering.
You may take a look at MiniMax’s sport right here. And for these curious, Claude’s model is accessible right here.
The immediate and the code can be found on our GitHub repo.
Ethics, censorship and delicate matters

The mannequin employs heavy censorship, refusing outright when confronted with questionable requests.
When it would not instantly decline, it makes an attempt to offer “protected” responses that typically produce absurd outcomes.
One take a look at completely illustrated this flaw: when requested for recommendation on seducing a finest buddy’s spouse, the mannequin prompt telling our buddy about our intentions together with his spouse—which may in all probability be, by far, the worst recommendation it may have produced, and arguably even dangerous. Don’t inform your buddy you wish to seduce his spouse until you wish to lose your friendship, your unethical romantic probabilities, and possibly some enamel too.

Political bias testing revealed fascinating patterns. The mannequin discusses Tiananmen Sq. brazenly and acknowledges Taiwan’s contested standing whereas noting China’s territorial claims. It additionally speaks about China, its leaders, the benefits and drawbacks of the completely different political programs, criticisms of the PCC, and so on.—nevertheless, the replies are very tame.

When prompted to write down satirical songs about Xi Jinping and Donald Trump, it complied with each requests however confirmed delicate variations—steering towards themes of Chinese language political unity when requested to mock Xi Jinping, whereas specializing in Trump’s persona traits when requested to mocked him.

All of its replies can be found on our GitHub repository.
General, the bias exists however stays much less pronounced than the pro-U.S. slant in Claude/ChatGPT, or the pro-China positioning in DeepSeek/Qwen, for instance. Builders, in fact, will have the ability to finetune this mannequin so as to add as a lot censorship, freedom or bias as they need—because it occurred with DeepSeek-R1, which was finetuned by Perplexity AI to offer a extra pro-U.S. bias on its responses.
Agentic work and net looking

Minimax can be appropriate with agentic options, in truth it has a separate tab particularly devoted to AI brokers. Customers can create their very own customized brokers, and select to strive some prebuilt brokers that seem on a gallery with completely different choices—much like what Manus AI gives to assist customers get aware of brokers and the way they differ from conventional chatbots. That mentioned, this requires heavy computational use and Minimax fees customers for this utilizing a credit-based system. It’s not clear how a lot credit a job would require or how credit translate into computational use, however it’s the identical system adopted by different AI agent suppliers.
MiniMax-M1’s net looking capabilities are a superb characteristic for these utilizing it by way of the official chatbot interface. Nevertheless, they can’t be mixed with the considering capabilities, severely hindering its potential.
When tasked with making a two-week Venezuela journey plan on a $3,000 funds, the mannequin methodically evaluated choices, optimized transportation prices, chosen acceptable lodging, and delivered a complete itinerary. Nevertheless, the prices, which have to be up to date in actual time, weren’t primarily based on actual data.

Claude produces higher-quality outcomes, however it additionally fees for the privilege.
For extra devoted duties, MiniMax the agent performance is one thing ChatGPT and Claude have not matched. The platform supplies 1,000 free AI credit for testing these brokers, although that is simply sufficient for mild testing duties.
We tried to create a customized agent for enhanced journey planning—which might have solved the issue of the dearth of net looking capabilities within the final immediate—however exhausted our credit earlier than completion. The agent system exhibits super potential, however requires paid credit for severe use.

Non-mathematical reasoning
The mannequin reveals a peculiar tendency to over-reason, typically to its personal detriment. One take a look at confirmed it arriving on the appropriate reply, then speaking itself out of it by way of extreme verification and hypothetical situations.
We prompted the standard thriller story from the BIG-bench dataset that we usually use, and the ending outcome was incorrect because of the mannequin overthinking the difficulty, evaluating potentialities that weren’t even talked about within the story. The entire Chain of Thought took the mannequin over 700 seconds—a document for this type of “easy” reply.

This exhaustive method is not inherently flawed, however creates prolonged wait occasions as customers watch the mannequin work by way of its chain of thought. As a thumbs-up characteristic, not like ChatGPT and Claude, MiniMax shows its reasoning course of transparently—following DeepSeek’s method. The transparency aids debugging and high quality management, permitting customers to determine the place logic went astray.
The issue, together with MiniMax’s entire thought course of and reply can be found in our GitHub repo.
Verdict
MiniMax-M1 isn’t excellent, however it delivers fairly good capabilities for a free mannequin, providing real competitors to paid providers like Claude in particular domains. Coders will discover a succesful assistant that rivals premium choices, whereas these needing long-context processing or web-enabled brokers achieve entry to options sometimes locked behind paywalls.
Inventive writers ought to look elsewhere—the mannequin produces practical however uninspired prose. The open-source nature guarantees important downstream advantages as builders create customized variations, modifications, and cost-effective deployments not possible with closed platforms like ChatGPT or Claude.
This can be a mannequin that may higher serve customers requiring reasoning duties—however continues to be an amazing free various for these looking for a chatbot for on a regular basis use that’s not actually mainstream.
You may obtain the open supply mannequin right here, and take a look at the net model right here. The agentic characteristic is accessible on a separate tab, however will also be accessed instantly by clicking on this hyperlink.
Usually Clever Publication
A weekly AI journey narrated by Gen, a generative AI mannequin.

