• bitcoinBitcoin(BTC)$70,104.001.29%
  • ethereumEthereum(ETH)$2,036.950.89%
  • tetherTether(USDT)$1.000.00%
  • binancecoinBNB(BNB)$643.800.83%
  • rippleXRP(XRP)$1.391.30%
  • usd-coinUSDC(USDC)$1.00-0.01%
  • solanaSolana(SOL)$86.120.44%
  • tronTRON(TRX)$0.2859550.07%
  • Figure HelocFigure Heloc(FIGR_HELOC)$1.04-0.53%
  • dogecoinDogecoin(DOGE)$0.0935962.86%
  • whitebitWhiteBIT Coin(WBT)$55.470.73%
  • USDSUSDS(USDS)$1.000.00%
  • cardanoCardano(ADA)$0.2609051.09%
  • bitcoin-cashBitcoin Cash(BCH)$448.370.07%
  • leo-tokenLEO Token(LEO)$9.210.61%
  • HyperliquidHyperliquid(HYPE)$34.570.96%
  • moneroMonero(XMR)$361.905.62%
  • chainlinkChainlink(LINK)$9.000.70%
  • Ethena USDeEthena USDe(USDE)$1.000.03%
  • CantonCanton(CC)$0.1489301.67%
  • stellarStellar(XLM)$0.1576883.63%
  • USD1USD1(USD1)$1.00-0.05%
  • RainRain(RAIN)$0.009005-0.47%
  • daiDai(DAI)$1.000.02%
  • litecoinLitecoin(LTC)$53.99-0.25%
  • avalanche-2Avalanche(AVAX)$9.552.12%
  • hedera-hashgraphHedera(HBAR)$0.094859-0.54%
  • paypal-usdPayPal USD(PYUSD)$1.000.01%
  • suiSui(SUI)$0.961.68%
  • zcashZcash(ZEC)$219.520.66%
  • shiba-inuShiba Inu(SHIB)$0.0000062.77%
  • the-open-networkToncoin(TON)$1.30-1.97%
  • crypto-com-chainCronos(CRO)$0.075411-0.02%
  • tether-goldTether Gold(XAUT)$5,165.880.48%
  • World Liberty FinancialWorld Liberty Financial(WLFI)$0.1016320.93%
  • pax-goldPAX Gold(PAXG)$5,207.040.52%
  • MemeCoreMemeCore(M)$1.43-4.56%
  • polkadotPolkadot(DOT)$1.48-1.58%
  • uniswapUniswap(UNI)$3.87-1.04%
  • mantleMantle(MNT)$0.703.91%
  • Pi NetworkPi Network(PI)$0.2336919.17%
  • okbOKB(OKB)$98.120.08%
  • Circle USYCCircle USYC(USYC)$1.120.00%
  • BlackRock USD Institutional Digital Liquidity FundBlackRock USD Institutional Digital Liquidity Fund(BUIDL)$1.000.00%
  • BittensorBittensor(TAO)$199.351.56%
  • SkySky(SKY)$0.0793903.95%
  • Falcon USDFalcon USD(USDF)$1.000.00%
  • AsterAster(ASTER)$0.70-0.05%
  • Global DollarGlobal Dollar(USDG)$1.000.00%
  • aaveAave(AAVE)$111.133.17%
TradePoint.io
  • Main
  • AI & Technology
  • Stock Charts
  • Market & News
  • Business
  • Finance Tips
  • Trade Tube
  • Blog
  • Shop
No Result
View All Result
TradePoint.io
No Result
View All Result

NVIDIA AI Releases Nemotron-Terminal: A Systematic Data Engineering Pipeline for Scaling LLM Terminal Agents

March 10, 2026
in AI & Technology
Reading Time: 5 mins read
A A
NVIDIA AI Releases Nemotron-Terminal: A Systematic Data Engineering Pipeline for Scaling LLM Terminal Agents
ShareShareShareShareShare

The race to build autonomous AI agents has hit a massive bottleneck: data. While frontier models like Claude Code and Codex CLI have demonstrated impressive proficiency in terminal environments, the training strategies and data mixtures behind them have remained closely guarded secrets. This lack of transparency has forced researchers and devs into a costly cycle of trial and error.

NVIDIA is now breaking that silence by unveiling a comprehensive framework for building high-performance terminal agents. By introducing Terminal-Task-Gen and the Terminal-Corpus dataset, NVIDIA is essentially giving the developer community the blueprints to build agents that don’t just ‘chat’ about code, but actually execute it with surgical precision.

YOU MAY ALSO LIKE

Social Security watchdog investigating claims that DOGE engineer copied its databases

Epic is increasing the price of Fortnite’s V-Bucks currency

https://arxiv.org/pdf/2602.21193

The Data Scarcity Problem

The challenge of training an agent for the command line is two-fold. First, there is a scarcity of foundational resources—specifically, diverse task prompts and the complex dependency files needed to create realistic environments. Second, capturing ‘trajectories’ (the step-by-step terminal interactions) is logistically painful. Human interactions are slow to record, and synthetic generation via LLM agents is prohibitively expensive because it requires fresh Docker environment instantiation for every single turn.

Terminal-Task-Gen: A Two-Pronged Strategy

NVIDIA’s solution is a ‘coarse-to-fine’ data generation pipeline called Terminal-Task-Gen. It utilizes two distinct strategies to scale training data without breaking the bank.

1. Dataset Adaptation (The Coarse Layer)

Instead of starting from scratch, the team leverages high-quality existing Supervised Fine-Tuning (SFT) datasets from math, code, and software engineering (SWE) domains. They transform these static prompts into interactive terminal tasks.

  • Math and Code: Using 163K math prompts and 35K code prompts, they wrap these challenges in a terminal scaffold.
  • SWE: They pull 32K unique prompts from repositories like SWE-bench and SWE-reBench. The clever part? This process doesn’t require an LLM “in the loop” for the initial adaptation, making it incredibly efficient to scale volume.

2. Synthetic Task Generation (The Fine Layer)

To bridge the gap between general reasoning and the specific rigors of terminal agency, NVIDIA team uses Terminal-Task-Gen to create novel, executable tasks.

  • Seed-based Generation: The LLM uses existing scientific computing or algorithmic problems as “inspiration” to synthesize new tasks. The agent is forced to install packages, read input files, and write results—mirroring a real-world developer workflow.
  • Skill-based Generation: This is where it gets technical. NVIDIA curated a taxonomy of “primitive terminal skills” across nine domains, including Security, Data Science, and System Administration. The LLM is then instructed to combine 3–5 of these primitives (like graph traversal + network configuration + file I/O) into a single, complex task.

Solving the Infrastructure Overhead

One of the most significant engineering breakthroughs in this research is the move to Pre-Built Docker Images. Previous frameworks often generated a unique Dockerfile for every single task, leading to massive build-time overhead and frequent failures. NVIDIA team instead maintains nine shared base images pre-configured with essential libraries (like pandas for data science or cryptography tools for security). This ‘single-pass’ creation method allows for massive parallelization and a significantly smaller resource footprint.

Performance: When 32B Beats 480B

The results of this data-centric approach are staggering. NVIDIA team used this pipeline to train the Nemotron-Terminal family of models, initialized from Qwen3.

On the Terminal-Bench 2.0 benchmark, which tests agents on end-to-end workflows like training machine learning models or debugging system environments, the improvements were vertical:

  • Nemotron-Terminal-8B: Jumped from a 2.5% success rate to 13.0%.
  • Nemotron-Terminal-32B: Achieved a 27.4% accuracy.

To put that in perspective, the 32B model outperformed the 480B Qwen3-Coder (23.9%) and rivaled the performance of closed-source giants like Grok 4 (23.1%) and GPT-5-Mini (24.0%). This proves that for terminal agents, high-quality, diverse trajectory data is a more powerful lever than sheer parameter scale.

Critical Insights

NVIDIA’s research also debunks several common myths in data engineering:

  • Don’t Filter Out Errors: The research team found that keeping ‘unsuccessful’ trajectories in the training data actually improved performance (12.4% vs 5.06% for success-only filtering). Exposing models to realistic error states and recovery patterns makes them more robust.
  • Skip the Curriculum: They experimented with ‘curriculum learning’ (training on easy data before hard data) but found that simple mixed training was just as effective, if not better.
  • Context Length Limits: While terminal trajectories can be long, most high-quality supervision fits within a standard 32,768-token window. Extending the context length slightly hurt performance, likely because long-tail trajectories tend to be noisier.

Check out Paper and HF Project Page. Also, feel free to follow us on Twitter and don’t forget to join our 120k+ ML SubReddit and Subscribe to our Newsletter. Wait! are you on telegram? now you can join us on telegram as well.

The post NVIDIA AI Releases Nemotron-Terminal: A Systematic Data Engineering Pipeline for Scaling LLM Terminal Agents appeared first on MarkTechPost.

Credit: Source link

ShareTweetSendSharePin

Related Posts

Social Security watchdog investigating claims that DOGE engineer copied its databases
AI & Technology

Social Security watchdog investigating claims that DOGE engineer copied its databases

March 10, 2026
Epic is increasing the price of Fortnite’s V-Bucks currency
AI & Technology

Epic is increasing the price of Fortnite’s V-Bucks currency

March 10, 2026
OpenAI upgrades ChatGPT with interactive learning tools as lawsuits and Pentagon backlash mount
AI & Technology

OpenAI upgrades ChatGPT with interactive learning tools as lawsuits and Pentagon backlash mount

March 10, 2026
Anthropic and OpenAI just exposed SAST’s structural blind spot with free tools
AI & Technology

Anthropic and OpenAI just exposed SAST’s structural blind spot with free tools

March 10, 2026
Next Post
Pentagon gives Anthropic an ultimatum amid fight over military AI guardrails

Pentagon gives Anthropic an ultimatum amid fight over military AI guardrails

Leave a Reply Cancel reply

Your email address will not be published. Required fields are marked *

Search

No Result
View All Result
Stay Tuned NOW Streaming Behind The Scenes! – March 02

Stay Tuned NOW Streaming Behind The Scenes! – March 02

March 7, 2026
I’m 0,000 In Debt, Should I File Bankruptcy?

I’m $250,000 In Debt, Should I File Bankruptcy?

March 6, 2026
LIVE: 2026 primary election results coverage for Texas, North Carolina, Arkansas | NBC News

LIVE: 2026 primary election results coverage for Texas, North Carolina, Arkansas | NBC News

March 6, 2026

About

Learn more

Our Services

Legal

Privacy Policy

Terms of Use

Bloggers

Learn more

Article Links

Contact

Advertise

Ask us anything

©2020- TradePoint.io - All rights reserved!

Tradepoint.io, being just a publishing and technology platform, is not a registered broker-dealer or investment adviser. So we do not provide investment advice. Rather, brokerage services are provided to clients of Tradepoint.io by independent SEC-registered broker-dealers and members of FINRA/SIPC. Every form of investing carries some risk and past performance is not a guarantee of future results. “Tradepoint.io“, “Instant Investing” and “My Trading Tools” are registered trademarks of Apperbuild, LLC.

This website is operated by Apperbuild, LLC. We have no link to any brokerage firm and we do not provide investment advice. Every information and resource we provide is solely for the education of our readers. © 2020 Apperbuild, LLC. All rights reserved.

No Result
View All Result
  • Main
  • AI & Technology
  • Stock Charts
  • Market & News
  • Business
  • Finance Tips
  • Trade Tube
  • Blog
  • Shop

© 2023 - TradePoint.io - All Rights Reserved!