• bitcoinBitcoin(BTC)$70,703.004.78%
  • ethereumEthereum(ETH)$2,060.273.81%
  • tetherTether(USDT)$1.000.02%
  • binancecoinBNB(BNB)$645.673.67%
  • rippleXRP(XRP)$1.393.64%
  • usd-coinUSDC(USDC)$1.000.00%
  • solanaSolana(SOL)$86.854.47%
  • tronTRON(TRX)$0.286244-0.66%
  • Figure HelocFigure Heloc(FIGR_HELOC)$1.04-0.72%
  • dogecoinDogecoin(DOGE)$0.0924342.72%
  • whitebitWhiteBIT Coin(WBT)$55.993.37%
  • USDSUSDS(USDS)$1.000.00%
  • cardanoCardano(ADA)$0.2609792.36%
  • bitcoin-cashBitcoin Cash(BCH)$446.53-0.16%
  • leo-tokenLEO Token(LEO)$9.120.68%
  • HyperliquidHyperliquid(HYPE)$34.8013.07%
  • chainlinkChainlink(LINK)$9.104.15%
  • moneroMonero(XMR)$344.51-0.60%
  • Ethena USDeEthena USDe(USDE)$1.000.06%
  • CantonCanton(CC)$0.148538-0.91%
  • stellarStellar(XLM)$0.1563584.85%
  • USD1USD1(USD1)$1.00-0.01%
  • RainRain(RAIN)$0.0089741.69%
  • daiDai(DAI)$1.000.00%
  • litecoinLitecoin(LTC)$54.151.49%
  • hedera-hashgraphHedera(HBAR)$0.0954060.09%
  • paypal-usdPayPal USD(PYUSD)$1.00-0.03%
  • avalanche-2Avalanche(AVAX)$9.423.65%
  • suiSui(SUI)$0.965.84%
  • zcashZcash(ZEC)$220.737.99%
  • shiba-inuShiba Inu(SHIB)$0.0000064.84%
  • the-open-networkToncoin(TON)$1.341.18%
  • crypto-com-chainCronos(CRO)$0.0763531.95%
  • tether-goldTether Gold(XAUT)$5,140.571.34%
  • World Liberty FinancialWorld Liberty Financial(WLFI)$0.1015701.44%
  • MemeCoreMemeCore(M)$1.48-3.59%
  • pax-goldPAX Gold(PAXG)$5,178.661.27%
  • polkadotPolkadot(DOT)$1.522.28%
  • uniswapUniswap(UNI)$3.964.55%
  • mantleMantle(MNT)$0.683.84%
  • Pi NetworkPi Network(PI)$0.2193691.15%
  • okbOKB(OKB)$98.480.45%
  • Circle USYCCircle USYC(USYC)$1.120.00%
  • BlackRock USD Institutional Digital Liquidity FundBlackRock USD Institutional Digital Liquidity Fund(BUIDL)$1.000.00%
  • BittensorBittensor(TAO)$199.541.69%
  • SkySky(SKY)$0.0768044.77%
  • Falcon USDFalcon USD(USDF)$1.00-0.02%
  • AsterAster(ASTER)$0.700.67%
  • Global DollarGlobal Dollar(USDG)$1.000.01%
  • aaveAave(AAVE)$111.214.25%
TradePoint.io
  • Main
  • AI & Technology
  • Stock Charts
  • Market & News
  • Business
  • Finance Tips
  • Trade Tube
  • Blog
  • Shop
No Result
View All Result
TradePoint.io
No Result
View All Result

Baidu Releases ERNIE-4.5-21B-A3B-Thinking: A Compact MoE Model for Deep Reasoning

September 10, 2025
in AI & Technology
Reading Time: 6 mins read
A A
Baidu Releases ERNIE-4.5-21B-A3B-Thinking: A Compact MoE Model for Deep Reasoning
ShareShareShareShareShare

Baidu AI Research team has just released ERNIE-4.5-21B-A3B-Thinking, a new reasoning-focused large language model designed around efficiency, long-context reasoning, and tool integration. Being part of the ERNIE-4.5 family, this model is a Mixture-of-Experts (MoE) architecture with 21B total parameters but only 3B active parameters per token, making it computationally efficient while maintaining competitive reasoning capability. Released under the Apache-2.0 license, it is accessible for both research and commercial deployment via Hugging Face.

What is the architectural design of ERNIE-4.5-21B-A3B-Thinking?

ERNIE-4.5-21B-A3B-Thinking is built on a Mixture-of-Experts backbone. Instead of activating all 21B parameters, the router selects a subset of experts, resulting in 3B active parameters per token. This structure reduces computation without compromising the specialization of different experts. The research team applies router orthogonalization loss and token-balanced loss to encourage diverse expert activation and stable training.

YOU MAY ALSO LIKE

Andrej Karpathy’s new open source ‘autoresearch’ lets you run hundreds of AI experiments a night — with revolutionary implications

Rode’s Rodecaster Video Core makes livestreaming even cheaper

This design provides a middle ground between small dense models and ultra-large systems. The research team’s assumptions include a theory that ~3B active parameters per token may represent a practical sweet spot for reasoning performance versus deployment efficiency.

How does the model handle long-context reasoning?

A defining capability of ERNIE-4.5-21B-A3B-Thinking is its 128K context length. This allows the model to process very long documents, perform extended multi-step reasoning, and integrate structured data sources such as academic papers or multi-file codebases.

The research team achieves this through progressive scaling of Rotary Position Embeddings (RoPE)—gradually increasing the frequency base from 10K up to 500K during training. Additional optimizations, including FlashMask attention and memory-efficient scheduling, make these long-context operations computationally feasible.

What training strategy supports its reasoning?

The model follows the multi-stage recipe defined across the ERNIE-4.5 family:

  1. Stage I – Text-only pretraining builds the core language backbone, starting with 8K context and expanding to 128K.
  2. Stage II – Vision training is skipped for this text-only variant.
  3. Stage III – Joint multimodal training is not used here, as A3B-Thinking is purely textual.

Post-training focuses on reasoning tasks. The research team employs Supervised Fine-Tuning (SFT) across mathematics, logic, coding, and science, followed by Progressive Reinforcement Learning (PRL). Reinforcement stages begin with logic, then extend to mathematics and programming, and finally to broader reasoning tasks. This is enhanced by Unified Preference Optimization (UPO), which integrates preference learning with PPO to stabilize alignment and reduce reward hacking.

What role does tool usage play in this model?

ERNIE-4.5-21B-A3B-Thinking supports structured tool and function calling, making it useful for scenarios where external computation or retrieval is required. Developers can integrate it with vLLM, Transformers 4.54+, and FastDeploy. This tool-use capability is particularly suited for program synthesis, symbolic reasoning, and multi-agent workflows.

Built-in function calling allows the model to reason over long contexts while dynamically invoking external APIs, a key requirement for applied reasoning in enterprise systems.

How does ERNIE-4.5-21B-A3B-Thinking perform on reasoning benchmarks?

It show strong performance improvements across logical reasoning, mathematics, scientific QA, and programming tasks. In evaluations, the model demonstrates:

  • Enhanced accuracy in multi-step reasoning datasets, where long chains of thought are required.
  • Competitiveness with larger dense models on STEM reasoning tasks.
  • Stable text generation and academic synthesis performance, benefiting from extended context training.

These results suggest that the MoE structure amplifies reasoning specialization, making it efficient without requiring trillion-scale dense parameters.

https://huggingface.co/baidu/ERNIE-4.5-21B-A3B-Thinking

How does it compare to other reasoning-focused LLMs?

This release gets into the landscape that includes OpenAI’s o3, Anthropic’s Claude 4, DeepSeek-R1, and Qwen-3. Many of these competitors rely on dense architectures or larger active parameter counts. Baidu research team’s choice of a compact MoE with 3B active parameters offers a different balance:

  • Scalability: Sparse activation reduces compute overhead while scaling expert capacity.
  • Long-context readiness: 128K context is directly trained, not retrofitted.
  • Commercial openness: Apache-2.0 license lowers adoption friction for enterprises.

Summary

ERNIE-4.5-21B-A3B-Thinking explains how deep reasoning can be achieved without massive dense parameter counts. By combining efficient MoE routing, 128K context training, and tool integration, Baidu’s research team offers a model that balances research-grade reasoning with deployment feasibility.


Check out the Model on Hugging Face and PAPER. Feel free to check out our GitHub Page for Tutorials, Codes and Notebooks. Also, feel free to follow us on Twitter and don’t forget to join our 100k+ ML SubReddit and Subscribe to our Newsletter.


Asif Razzaq is the CEO of Marktechpost Media Inc.. As a visionary entrepreneur and engineer, Asif is committed to harnessing the potential of Artificial Intelligence for social good. His most recent endeavor is the launch of an Artificial Intelligence Media Platform, Marktechpost, which stands out for its in-depth coverage of machine learning and deep learning news that is both technically sound and easily understandable by a wide audience. The platform boasts of over 2 million monthly views, illustrating its popularity among audiences.

Credit: Source link

ShareTweetSendSharePin

Related Posts

Andrej Karpathy’s new open source ‘autoresearch’ lets you run hundreds of AI experiments a night — with revolutionary implications
AI & Technology

Andrej Karpathy’s new open source ‘autoresearch’ lets you run hundreds of AI experiments a night — with revolutionary implications

March 10, 2026
Rode’s Rodecaster Video Core makes livestreaming even cheaper
AI & Technology

Rode’s Rodecaster Video Core makes livestreaming even cheaper

March 9, 2026
You can (sort of) block Grok from editing your uploaded photos
AI & Technology

You can (sort of) block Grok from editing your uploaded photos

March 9, 2026
How to make your e-commerce product visible to AI agents? Use this new system trusted by L’Oréal, Unilever, Mars & Beiersdorf
AI & Technology

How to make your e-commerce product visible to AI agents? Use this new system trusted by L’Oréal, Unilever, Mars & Beiersdorf

March 9, 2026
Next Post
NATO member Poland's military says it shot down Russian drones in its airspace – Axios

NATO member Poland's military says it shot down Russian drones in its airspace - Axios

Leave a Reply Cancel reply

Your email address will not be published. Required fields are marked *

Search

No Result
View All Result
AI Storage Need Fuels Hard Drive Demand

AI Storage Need Fuels Hard Drive Demand

March 9, 2026
This 1 Condition Is Impacting More Women In Their Early 30s And 40s, Here's What You Need To Know – BuzzFeed

This 1 Condition Is Impacting More Women In Their Early 30s And 40s, Here's What You Need To Know – BuzzFeed

March 8, 2026
NetEase is reportedly pulling funding for Yakuza creator’s studio

NetEase is reportedly pulling funding for Yakuza creator’s studio

March 8, 2026

About

Learn more

Our Services

Legal

Privacy Policy

Terms of Use

Bloggers

Learn more

Article Links

Contact

Advertise

Ask us anything

©2020- TradePoint.io - All rights reserved!

Tradepoint.io, being just a publishing and technology platform, is not a registered broker-dealer or investment adviser. So we do not provide investment advice. Rather, brokerage services are provided to clients of Tradepoint.io by independent SEC-registered broker-dealers and members of FINRA/SIPC. Every form of investing carries some risk and past performance is not a guarantee of future results. “Tradepoint.io“, “Instant Investing” and “My Trading Tools” are registered trademarks of Apperbuild, LLC.

This website is operated by Apperbuild, LLC. We have no link to any brokerage firm and we do not provide investment advice. Every information and resource we provide is solely for the education of our readers. © 2020 Apperbuild, LLC. All rights reserved.

No Result
View All Result
  • Main
  • AI & Technology
  • Stock Charts
  • Market & News
  • Business
  • Finance Tips
  • Trade Tube
  • Blog
  • Shop

© 2023 - TradePoint.io - All Rights Reserved!