Close Menu

    Subscribe to Updates

    Get the latest creative news from FooBar about art, design and business.

    What's Hot

    Gen Z And Millennials Are Racing To Upskill In AI

    December 6, 2025

    AI deepfakes of real doctors spreading health misinformation on social media | Health

    December 6, 2025

    AI labs like Meta, Deepseek, and Xai earned worst grades possible on an existential safety index

    December 5, 2025
    Facebook X (Twitter) Instagram
    ailogicnews.aiailogicnews.ai
    • Home
    ailogicnews.aiailogicnews.ai
    Home»Deepseek»MiniMax M1 model claims Chinese LLM crown from DeepSeek • The Register
    Deepseek

    MiniMax M1 model claims Chinese LLM crown from DeepSeek • The Register

    AI Logic NewsBy AI Logic NewsJune 17, 2025No Comments3 Mins Read
    Share Facebook Twitter Pinterest LinkedIn Tumblr Reddit Telegram Email
    Share
    Facebook Twitter LinkedIn Pinterest Email

    MiniMax, an AI firm based in Shanghai, has released an open-source reasoning model that challenges Chinese rival DeepSeek and US-based Anthropic, OpenAI, and Google in terms of performance and cost.

    MiniMax-M1 was released Monday under an Apache software license, and thus is actually open source, unlike Meta’s Llama family, offered under a community license that’s not open source, and DeepSeek, which is only partially under an open-source license.

    “In complex, productivity-oriented scenarios, M1’s capabilities are top-tier among open-source models, surpassing domestic closed-source models and approaching the leading overseas models, all while offering the industry’s best cost-effectiveness,” MiniMax boasts in a blog post.

    According to the blog post, M1 is competitive with OpenAI o3, Gemini 2.5 Pro, Claude 4 Opus, DeepSeek R1, DeepSeek R1-0528, and Qwen3-235B on various benchmarks (AIME 2024, LiveCodeBench, SWE-bench Verified, Tau-bench, and MRCR), coming in behind some models and ahead of others to varying degrees. As always, take vendor-supplied benchmark results with a grain of salt, but the source code is available on GitHub should you wish to confirm its performance independently.

    But MiniMax makes clear that it’s trying to supplant DeepSeek as the leading industry disruptor by noting that its context window (the amount of input it can handle) is one million tokens, which rivals Google Gemini 2.5 Pro and is eight times the capacity of DeepSeek R1.

    Day 1/5 of #MiniMaxWeek: We’re open-sourcing MiniMax-M1, our latest LLM — setting new standards in long-context reasoning.- World’s longest context window: 1M-token input, 80k-token output- State-of-the-art agentic use among open-source models- RL at unmatched efficiency:… pic.twitter.com/bGfDlZA54n

    — MiniMax (official) (@MiniMax__AI) June 16, 2025

    In terms of output, the model can manage 80,000 tokens, better than DeepSeek’s 64,000 token capacity but shy of OpenAI’s o3, which can spit out 100,000 tokens in response to a prompt.

    Backed by Alibaba Group, Tencent, and IDG Capital, MiniMax claims its Lightning Attention mechanism, a way to calculate attention matrices that improves both training and inference efficiency, gives its M1 model an advantage when computing long context inputs and when trying to reason.

    “For example, when performing deep reasoning with 80,000 tokens, it requires only about 30 percent of the computing power of DeepSeek R1,” the company claims. “This feature gives us a substantial computational efficiency advantage in both training and inference.”

    This more efficient computation method, in conjunction with an improved reinforcement learning algorithm called CISPO (detailed in M1’s technical report [PDF]), translates to lower computing costs.

    “The entire reinforcement learning phase used only 512 [Nvidia] H800s for three weeks, with a rental cost of just $537,400,” MiniMax claims. “This is an order of magnitude less than initially anticipated.” ®



    Source link

    Share. Facebook Twitter Pinterest LinkedIn Tumblr Email
    Previous ArticleWhy AI Complements (But Doesn’t Rep
    Next Article AI Will Reduce Our Corporate Workforce
    AI Logic News

    Related Posts

    Deepseek

    AI labs like Meta, Deepseek, and Xai earned worst grades possible on an existential safety index

    December 5, 2025
    Deepseek

    ByteDance and DeepSeek Are Placing Very Different AI Bets

    December 5, 2025
    Deepseek

    Nvidia claims 10x speed boost on new server for DeepSeek-style AI models

    December 4, 2025
    Demo
    Top Posts

    FTC’s Holyoak Has Her Eyes On DeepSeek

    February 22, 20256 Views

    OpenAI Rejects Elon Musks Bid Further Escalating The Feud

    February 17, 20253 Views

    Optimize Inventory Management with AI for Small Online Retailers

    February 17, 20253 Views
    Latest Reviews
    ailogicnews.ai
    © 2025 Lee Enterprises

    Type above and press Enter to search. Press Esc to cancel.