• bitcoinBitcoin(BTC)$75,754.00-1.10%
  • ethereumEthereum(ETH)$2,245.96-1.98%
  • tetherTether(USDT)$1.00-0.03%
  • rippleXRP(XRP)$1.37-0.61%
  • binancecoinBNB(BNB)$617.14-1.18%
  • usd-coinUSDC(USDC)$1.00-0.01%
  • solanaSolana(SOL)$82.94-1.38%
  • tronTRON(TRX)$0.3237080.55%
  • Figure HelocFigure Heloc(FIGR_HELOC)$1.03-0.62%
  • dogecoinDogecoin(DOGE)$0.1057015.75%
  • whitebitWhiteBIT Coin(WBT)$53.89-0.48%
  • USDSUSDS(USDS)$1.00-0.03%
  • leo-tokenLEO Token(LEO)$10.370.14%
  • HyperliquidHyperliquid(HYPE)$39.74-1.03%
  • cardanoCardano(ADA)$0.246090-0.59%
  • bitcoin-cashBitcoin Cash(BCH)$445.93-1.58%
  • moneroMonero(XMR)$377.31-1.05%
  • chainlinkChainlink(LINK)$9.09-1.78%
  • CantonCanton(CC)$0.1523352.18%
  • zcashZcash(ZEC)$325.94-2.91%
  • stellarStellar(XLM)$0.159530-1.42%
  • USD1USD1(USD1)$1.00-0.07%
  • daiDai(DAI)$1.000.00%
  • MemeCoreMemeCore(M)$3.39-1.73%
  • litecoinLitecoin(LTC)$55.24-0.62%
  • avalanche-2Avalanche(AVAX)$9.12-0.67%
  • hedera-hashgraphHedera(HBAR)$0.088474-0.94%
  • Ethena USDeEthena USDe(USDE)$1.000.01%
  • RainRain(RAIN)$0.0078035.06%
  • shiba-inuShiba Inu(SHIB)$0.0000060.73%
  • suiSui(SUI)$0.91-1.79%
  • paypal-usdPayPal USD(PYUSD)$1.00-0.04%
  • the-open-networkToncoin(TON)$1.31-0.34%
  • crypto-com-chainCronos(CRO)$0.068312-0.59%
  • Circle USYCCircle USYC(USYC)$1.120.01%
  • tether-goldTether Gold(XAUT)$4,563.73-0.57%
  • Global DollarGlobal Dollar(USDG)$1.000.00%
  • BittensorBittensor(TAO)$250.66-4.30%
  • BlackRock USD Institutional Digital Liquidity FundBlackRock USD Institutional Digital Liquidity Fund(BUIDL)$1.000.00%
  • pax-goldPAX Gold(PAXG)$4,559.42-0.63%
  • mantleMantle(MNT)$0.62-1.49%
  • polkadotPolkadot(DOT)$1.21-1.66%
  • uniswapUniswap(UNI)$3.17-2.17%
  • World Liberty FinancialWorld Liberty Financial(WLFI)$0.063139-13.59%
  • SkySky(SKY)$0.081086-5.71%
  • Pi NetworkPi Network(PI)$0.174330-11.58%
  • Falcon USDFalcon USD(USDF)$1.00-0.03%
  • okbOKB(OKB)$82.18-0.84%
  • nearNEAR Protocol(NEAR)$1.32-2.23%
  • AsterAster(ASTER)$0.66-0.08%
TradePoint.io
  • Main
  • AI & Technology
  • Stock Charts
  • Market & News
  • Business
  • Finance Tips
  • Trade Tube
  • Blog
  • Shop
No Result
View All Result
TradePoint.io
No Result
View All Result

Stanford Researchers Release OpenJarvis: A Local-First Framework for Building On-Device Personal AI Agents with Tools, Memory, and Learning

March 12, 2026
in AI & Technology
Reading Time: 6 mins read
A A
Stanford Researchers Release OpenJarvis: A Local-First Framework for Building On-Device Personal AI Agents with Tools, Memory, and Learning
ShareShareShareShareShare

Stanford researchers have introduced OpenJarvis, an open-source framework for building personal AI agents that run entirely on-device. The project comes from Stanford’s Scaling Intelligence Lab and is presented as both a research platform and deployment-ready infrastructure for local-first AI systems. Its focus is not only model execution, but also the broader software stack required to make on-device agents usable, measurable, and adaptable over time.

Why OpenJarvis?

According to the Stanford research team, most current personal AI projects still keep the local component relatively thin while routing core reasoning through external cloud APIs. That design introduces latency, recurring cost, and data exposure concerns, especially for assistants/agents that operate over personal files, messages, and persistent user context. OpenJarvis is designed to shift that balance by making local execution the default and cloud usage optional.

YOU MAY ALSO LIKE

Mark Zuckerberg Says Meta Is Working On AI Agents For Personal And Business Use

The retrieval rebuild: Why hybrid retrieval intent tripled as enterprise RAG programs hit the scale wall

The research team ties this release to its earlier Intelligence Per Watt research. In that work, they report that local language models and local accelerators can accurately serve 88.7% of single-turn chat and reasoning queries at interactive latencies, while intelligence efficiency improved 5.3× from 2023 to 2025. OpenJarvis is positioned as the software layer that follows from that result: if models and consumer hardware are becoming practical for more local workloads, then developers need a standard stack for building and evaluating those systems.

https://scalingintelligence.stanford.edu/blogs/openjarvis/

The Five-Primitives Architecture

At the architectural level, OpenJarvis is organized around five primitives: Intelligence, Engine, Agents, Tools & Memory, and Learning. The research team describes these as composable abstractions that can be benchmarked, substituted, and optimized independently or used together as an integrated system. This matters because local AI projects often mix inference, orchestration, tools, retrieval, and adaptation logic into a single hard-to-reproduce application. OpenJarvis instead tries to give each layer a more explicit role.

Intelligence: The Model Layer

The Intelligence primitive is the model layer. It sits above a changing set of local model families and provides a unified model catalog so developers do not have to manually track parameter counts, hardware fit, or memory tradeoffs for every release. The goal is to make model choice easier to study separately from other parts of the system, such as the inference backend or agent logic.

Engine: The Inference Runtime

The Engine primitive is the inference runtime. It is a common interface over backends such as Ollama, vLLM, SGLang, llama.cpp, and cloud APIs. The engine layer is framed more broadly as hardware-aware execution, where commands such as jarvis init detect available hardware and recommend a suitable engine and model configuration, while jarvis doctor helps maintain that setup. For developers, this is one of the more practical parts of the design: the framework does not assume a single runtime, but treats inference as a pluggable layer.

Agents: The Behavior Layer

The Agents primitive is the behavior layer. Stanford describes it as the part that turns model capability into structured action under real device constraints such as bounded context windows, limited working memory, and efficiency limits. Rather than relying on one general-purpose agent, OpenJarvis supports composable roles. The Stanford article specifically mentions roles such as the Orchestrator, which breaks complex tasks into subtasks, and the Operative, which is intended as a lightweight executor for recurring personal workflows. The docs also describe the agent harness as handling the system prompt, tools, context, retry logic, and exit logic.

Tools & Memory: Grounding the Agent

The Tools & Memory primitive is the grounding layer. This primitive includes support for MCP (Model Context Protocol) for standardized tool use, Google A2A for agent-to-agent communication, and semantic indexing for local retrieval over notes, documents, and papers. It also support for messaging platforms, webchat, and webhooks. It also covers a narrower tools view that includes web search, calculator access, file I/O, code interpretation, retrieval, and external MCP servers. OpenJarvis is not just a local chat interface; it is intended to connect local models to tools and persistent personal context while keeping storage and control local by default.

Learning: Closed-Loop Improvement

The fifth primitive, Learning, is what gives the framework a closed-loop improvement path. Stanford researchers describe it as a layer that uses local interaction traces to synthesize training data, refine agent behavior, and improve model selection over time. OpenJarvis supports optimization across four layers of the stack: model weights, LM prompts, agentic logic, and the inference engine. Examples listed by the research team include SFT, GRPO, DPO, prompt optimization with DSPy, agent optimization with GEPA, and engine-level tuning such as quantization selection and batch scheduling.

Efficiency as a First-Class Metric

A major technical point in OpenJarvis is its emphasis on efficiency-aware evaluation. The framework treats energy, FLOPs, latency, and dollar cost as first-class constraints alongside task quality. It also emphasizes on a hardware-agnostic telemetry system for profiling energy on NVIDIA GPUs via NVML, AMD GPUs, and Apple Silicon via powermetrics, with 50 ms sampling intervals. The jarvis bench command is meant to standardize benchmarking for latency, throughput, and energy per query. This is important because local deployment is not only about whether a model can answer a question, but whether it can do so within real limits on power, memory, and response time.

Developer Interfaces and Deployment Options

From a developer perspective, OpenJarvis exposes several entry points. The official docs show a browser app, a desktop app, a Python SDK, and a CLI. The browser-based interface can be launched with ./scripts/quickstart.sh, which installs dependencies, starts Ollama and a local model, launches the backend and frontend, and opens the local UI. The desktop app is available for macOS, Windows, and Linux, with the backend still running on the user’s machine. The Python SDK exposes a Jarvis() object and methods such as ask() and ask_full(), while the CLI includes commands like jarvis ask, jarvis serve, jarvis memory index, and jarvis memory search.

The docs also state that all core functionality works without a network connection, while cloud APIs are optional. For dev teams building local applications, another practical feature is jarvis serve, which starts a FastAPI server with SSE streaming and is described as a drop-in replacement for OpenAI clients. That lowers the migration cost for developers who want to prototype against an API-shaped interface while still keeping inference local.


Check out Repo, Docs and Technical details. Also, feel free to follow us on Twitter and don’t forget to join our 120k+ ML SubReddit and Subscribe to our Newsletter. Wait! are you on telegram? now you can join us on telegram as well.

The post Stanford Researchers Release OpenJarvis: A Local-First Framework for Building On-Device Personal AI Agents with Tools, Memory, and Learning appeared first on MarkTechPost.

Credit: Source link

ShareTweetSendSharePin

Related Posts

Mark Zuckerberg Says Meta Is Working On AI Agents For Personal And Business Use
AI & Technology

Mark Zuckerberg Says Meta Is Working On AI Agents For Personal And Business Use

April 29, 2026
The retrieval rebuild: Why hybrid retrieval intent tripled as enterprise RAG programs hit the scale wall
AI & Technology

The retrieval rebuild: Why hybrid retrieval intent tripled as enterprise RAG programs hit the scale wall

April 29, 2026
Qwen Team Releases FlashQLA: a High-Performance Linear Attention Kernel Library That Achieves Up to 3× Speedup on NVIDIA Hopper GPUs
AI & Technology

Qwen Team Releases FlashQLA: a High-Performance Linear Attention Kernel Library That Achieves Up to 3× Speedup on NVIDIA Hopper GPUs

April 29, 2026
Definity embeds agents inside Spark pipelines to catch failures before they reach agentic AI systems
AI & Technology

Definity embeds agents inside Spark pipelines to catch failures before they reach agentic AI systems

April 29, 2026
Next Post
Trump praises Republicans for helping pass the One Big Beautiful Bill

Trump praises Republicans for helping pass the One Big Beautiful Bill

Leave a Reply Cancel reply

Your email address will not be published. Required fields are marked *

Search

No Result
View All Result
2026 NFL Draft live updates: Fernando Mendoza awaits call from Raiders at No. 1; will Jeremiyah Love go top 5? – Yahoo Sports

2026 NFL Draft live updates: Fernando Mendoza awaits call from Raiders at No. 1; will Jeremiyah Love go top 5? – Yahoo Sports

April 23, 2026
Musk and OpenAI’s lawyer share testy exchange

Musk and OpenAI’s lawyer share testy exchange

April 29, 2026
‘I want to find Lynette’: Brian Hooker speaks after release from police custody

‘I want to find Lynette’: Brian Hooker speaks after release from police custody

April 25, 2026

About

Learn more

Our Services

Legal

Privacy Policy

Terms of Use

Bloggers

Learn more

Article Links

Contact

Advertise

Ask us anything

©2020- TradePoint.io - All rights reserved!

Tradepoint.io, being just a publishing and technology platform, is not a registered broker-dealer or investment adviser. So we do not provide investment advice. Rather, brokerage services are provided to clients of Tradepoint.io by independent SEC-registered broker-dealers and members of FINRA/SIPC. Every form of investing carries some risk and past performance is not a guarantee of future results. “Tradepoint.io“, “Instant Investing” and “My Trading Tools” are registered trademarks of Apperbuild, LLC.

This website is operated by Apperbuild, LLC. We have no link to any brokerage firm and we do not provide investment advice. Every information and resource we provide is solely for the education of our readers. © 2020 Apperbuild, LLC. All rights reserved.

No Result
View All Result
  • Main
  • AI & Technology
  • Stock Charts
  • Market & News
  • Business
  • Finance Tips
  • Trade Tube
  • Blog
  • Shop

© 2023 - TradePoint.io - All Rights Reserved!