• Kinza Babylon Staked BTCKinza Babylon Staked BTC(KBTC)$83,270.000.00%
  • Steakhouse EURCV Morpho VaultSteakhouse EURCV Morpho Vault(STEAKEURCV)$0.000000-100.00%
  • Stride Staked InjectiveStride Staked Injective(STINJ)$16.51-4.18%
  • Vested XORVested XOR(VXOR)$3,404.231,000.00%
  • FibSwap DEXFibSwap DEX(FIBO)$0.0084659.90%
  • ICPanda DAOICPanda DAO(PANDA)$0.003106-39.39%
  • TruFin Staked APTTruFin Staked APT(TRUAPT)$8.020.00%
  • bitcoinBitcoin(BTC)$106,006.000.48%
  • ethereumEthereum(ETH)$2,517.88-0.08%
  • VNST StablecoinVNST Stablecoin(VNST)$0.0000400.67%
  • tetherTether(USDT)$1.000.00%
  • rippleXRP(XRP)$2.284.57%
  • binancecoinBNB(BNB)$651.830.29%
  • Wrapped SOLWrapped SOL(SOL)$143.66-2.32%
  • solanaSolana(SOL)$150.83-0.39%
  • usd-coinUSDC(USDC)$1.000.00%
  • dogecoinDogecoin(DOGE)$0.184663-1.36%
  • tronTRON(TRX)$0.2857122.23%
  • cardanoCardano(ADA)$0.670.37%
  • staked-etherLido Staked Ether(STETH)$2,516.870.08%
  • wrapped-bitcoinWrapped Bitcoin(WBTC)$105,943.000.38%
  • Gaj FinanceGaj Finance(GAJ)$0.0059271.46%
  • Content BitcoinContent Bitcoin(CTB)$24.482.55%
  • USD OneUSD One(USD1)$1.000.11%
  • HyperliquidHyperliquid(HYPE)$35.211.08%
  • SuiSui(SUI)$3.21-1.87%
  • Wrapped stETHWrapped stETH(WSTETH)$3,035.980.20%
  • UGOLD Inc.UGOLD Inc.(UGOLD)$3,042.460.08%
  • ParkcoinParkcoin(KPK)$1.101.76%
  • chainlinkChainlink(LINK)$13.79-1.34%
  • avalanche-2Avalanche(AVAX)$20.78-0.72%
  • stellarStellar(XLM)$0.2714841.67%
  • leo-tokenLEO Token(LEO)$9.13-0.35%
  • bitcoin-cashBitcoin Cash(BCH)$417.101.86%
  • ToncoinToncoin(TON)$3.17-0.42%
  • shiba-inuShiba Inu(SHIB)$0.000013-1.60%
  • hedera-hashgraphHedera(HBAR)$0.1710010.76%
  • USDSUSDS(USDS)$1.000.00%
  • Yay StakeStone EtherYay StakeStone Ether(YAYSTONE)$2,671.07-2.84%
  • litecoinLitecoin(LTC)$88.05-0.79%
  • wethWETH(WETH)$2,517.23-0.15%
  • Wrapped eETHWrapped eETH(WEETH)$2,686.02-0.32%
  • polkadotPolkadot(DOT)$4.090.88%
  • moneroMonero(XMR)$328.930.54%
  • Pundi AIFXPundi AIFX(PUNDIAI)$16.000.00%
  • Binance Bridged USDT (BNB Smart Chain)Binance Bridged USDT (BNB Smart Chain)(BSC-USD)$1.00-0.12%
  • PengPeng(PENG)$0.60-13.59%
  • Ethena USDeEthena USDe(USDE)$1.000.01%
  • Bitget TokenBitget Token(BGB)$4.66-0.21%
  • MurasakiMurasaki(MURA)$4.32-12.46%
TradePoint.io
  • Main
  • AI & Technology
  • Stock Charts
  • Market & News
  • Business
  • Finance Tips
  • Trade Tube
  • Blog
  • Shop
No Result
View All Result
TradePoint.io
No Result
View All Result

Qwen AI Releases Qwen2.5-7B-Instruct-1M and Qwen2.5-14B-Instruct-1M: Allowing Deployment with Context Length up to 1M Tokens

January 27, 2025
in AI & Technology
Reading Time: 6 mins read
A A
Qwen AI Releases Qwen2.5-7B-Instruct-1M and Qwen2.5-14B-Instruct-1M: Allowing Deployment with Context Length up to 1M Tokens
ShareShareShareShareShare

YOU MAY ALSO LIKE

AI Liability Insurance: The Next Step in Safeguarding Businesses from AI Failures

Mini Motorways is getting a creative mode

The advancements in large language models (LLMs) have significantly enhanced natural language processing (NLP), enabling capabilities like contextual understanding, code generation, and reasoning. However, a key limitation persists: the restricted context window size. Most LLMs can only process a fixed amount of text, typically up to 128K tokens, which limits their ability to handle tasks requiring extensive context, such as analyzing lengthy documents or debugging large codebases. These constraints often necessitate workarounds like text chunking, increasing computational complexity. Overcoming these challenges requires models that can extend context lengths efficiently without compromising performance.

Qwen AI’s Latest Release

Qwen AI has introduced two new models, Qwen2.5-7B-Instruct-1M and Qwen2.5-14B-Instruct-1M, designed to support context lengths of up to 1 million tokens. Developed by the Qwen team at Alibaba Group, these models also come with an open-sourced inference framework optimized for handling long contexts. This advancement enables developers and researchers to work with larger datasets in a single pass, offering a practical solution for applications that demand extended context processing. Additionally, the models feature improvements in sparse attention mechanisms and kernel optimization, resulting in faster processing times for extended inputs.

Technical Details and Benefits

The Qwen2.5-1M series retains a Transformer-based architecture, incorporating features like Grouped Query Attention (GQA), Rotary Positional Embeddings (RoPE), and RMSNorm for stability over long contexts. Training involved both natural and synthetic datasets, with tasks like Fill-in-the-Middle (FIM), paragraph reordering, and position-based retrieval enhancing the model’s ability to handle long-range dependencies. Sparse attention methods such as Dual Chunk Attention (DCA) allow for efficient inference by dividing sequences into manageable chunks. Progressive pre-training strategies, which gradually scale context lengths from 4K to 1M tokens, optimize efficiency while controlling computational demands. The models are fully compatible with vLLM’s open-source inference framework, simplifying integration for developers.

Results and Insights

Benchmark results demonstrate the capabilities of the Qwen2.5-1M models. In the Passkey Retrieval Test, the 7B and 14B variants successfully retrieved hidden information from 1 million tokens, showcasing their effectiveness in long-context scenarios. In other benchmarks, including RULER and Needle in a Haystack (NIAH), the 14B model outperformed alternatives like GPT-4o-mini and Llama-3. Sparse attention techniques contributed to reduced inference times, achieving speedups of up to 6.7x on Nvidia H20 GPUs. These results highlight the models’ ability to combine efficiency with high performance, making them suitable for real-world applications requiring extensive context.

Conclusion

The Qwen2.5-1M series addresses critical limitations in NLP by significantly extending context lengths while maintaining efficiency and accessibility. By overcoming constraints that have long hindered LLMs, these models open new possibilities for applications ranging from analyzing large datasets to processing entire code repositories. With innovations in sparse attention, kernel optimization, and long-context pre-training, Qwen2.5-1M offers a practical and effective tool for tackling complex, context-heavy tasks.


Check out the Paper, Models on Hugging Face and Technical Details. All credit for this research goes to the researchers of this project. Also, don’t forget to follow us on Twitter and join our Telegram Channel and LinkedIn Group. Don’t Forget to join our 70k+ ML SubReddit.

🚨 [Recommended Read] Nebius AI Studio expands with vision models, new language models, embeddings and LoRA (Promoted)


Asif Razzaq is the CEO of Marktechpost Media Inc.. As a visionary entrepreneur and engineer, Asif is committed to harnessing the potential of Artificial Intelligence for social good. His most recent endeavor is the launch of an Artificial Intelligence Media Platform, Marktechpost, which stands out for its in-depth coverage of machine learning and deep learning news that is both technically sound and easily understandable by a wide audience. The platform boasts of over 2 million monthly views, illustrating its popularity among audiences.

📄 Meet ‘Height’:The only autonomous project management tool (Sponsored)

Credit: Source link

ShareTweetSendSharePin

Related Posts

AI Liability Insurance: The Next Step in Safeguarding Businesses from AI Failures
AI & Technology

AI Liability Insurance: The Next Step in Safeguarding Businesses from AI Failures

June 8, 2025
Mini Motorways is getting a creative mode
AI & Technology

Mini Motorways is getting a creative mode

June 7, 2025
Agent-based computing is outgrowing the web as we know it
AI & Technology

Agent-based computing is outgrowing the web as we know it

June 7, 2025
New Tales and Emeteria unveil Fading Echo action-adventure game
AI & Technology

New Tales and Emeteria unveil Fading Echo action-adventure game

June 7, 2025
Next Post
Low water supply causing ‘a real challenge’ in battling wildfires, says L.A. Fire Dept. Official

Low water supply causing ‘a real challenge’ in battling wildfires, says L.A. Fire Dept. Official

Leave a Reply Cancel reply

Your email address will not be published. Required fields are marked *

Search

No Result
View All Result
Urgent water rescues amid severe storm threat in the Midwest

Urgent water rescues amid severe storm threat in the Midwest

June 7, 2025
IO Interactive marries Hitman with 007 and MindsEye

IO Interactive marries Hitman with 007 and MindsEye

June 6, 2025
Cuban inmates protest deportation at Miami detention center

Cuban inmates protest deportation at Miami detention center

June 6, 2025

About

Learn more

Our Services

Legal

Privacy Policy

Terms of Use

Bloggers

Learn more

Article Links

Contact

Advertise

Ask us anything

©2020- TradePoint.io - All rights reserved!

Tradepoint.io, being just a publishing and technology platform, is not a registered broker-dealer or investment adviser. So we do not provide investment advice. Rather, brokerage services are provided to clients of Tradepoint.io by independent SEC-registered broker-dealers and members of FINRA/SIPC. Every form of investing carries some risk and past performance is not a guarantee of future results. “Tradepoint.io“, “Instant Investing” and “My Trading Tools” are registered trademarks of Apperbuild, LLC.

This website is operated by Apperbuild, LLC. We have no link to any brokerage firm and we do not provide investment advice. Every information and resource we provide is solely for the education of our readers. © 2020 Apperbuild, LLC. All rights reserved.

No Result
View All Result
  • Main
  • AI & Technology
  • Stock Charts
  • Market & News
  • Business
  • Finance Tips
  • Trade Tube
  • Blog
  • Shop

© 2023 - TradePoint.io - All Rights Reserved!