• bitcoinBitcoin(BTC)$80,242.00-0.53%
  • ethereumEthereum(ETH)$2,283.61-0.08%
  • tetherTether(USDT)$1.000.00%
  • binancecoinBNB(BNB)$673.702.01%
  • rippleXRP(XRP)$1.44-0.81%
  • usd-coinUSDC(USDC)$1.00-0.10%
  • solanaSolana(SOL)$93.29-1.82%
  • tronTRON(TRX)$0.3506080.67%
  • Figure HelocFigure Heloc(FIGR_HELOC)$1.040.71%
  • dogecoinDogecoin(DOGE)$0.1122552.85%
  • whitebitWhiteBIT Coin(WBT)$59.02-0.36%
  • USDSUSDS(USDS)$1.00-0.02%
  • cardanoCardano(ADA)$0.268328-2.10%
  • HyperliquidHyperliquid(HYPE)$39.51-3.18%
  • leo-tokenLEO Token(LEO)$10.01-1.56%
  • zcashZcash(ZEC)$549.84-0.47%
  • bitcoin-cashBitcoin Cash(BCH)$437.21-0.60%
  • moneroMonero(XMR)$407.921.40%
  • chainlinkChainlink(LINK)$10.350.45%
  • CantonCanton(CC)$0.153102-4.54%
  • the-open-networkToncoin(TON)$2.17-7.70%
  • stellarStellar(XLM)$0.161655-1.24%
  • suiSui(SUI)$1.23-3.34%
  • litecoinLitecoin(LTC)$57.870.33%
  • USD1USD1(USD1)$1.000.05%
  • daiDai(DAI)$1.000.03%
  • avalanche-2Avalanche(AVAX)$9.890.24%
  • MemeCoreMemeCore(M)$3.20-1.45%
  • hedera-hashgraphHedera(HBAR)$0.092723-2.13%
  • Ethena USDeEthena USDe(USDE)$1.000.00%
  • shiba-inuShiba Inu(SHIB)$0.000006-1.14%
  • RainRain(RAIN)$0.0075300.10%
  • Global DollarGlobal Dollar(USDG)$1.00-0.01%
  • paypal-usdPayPal USD(PYUSD)$1.00-0.01%
  • crypto-com-chainCronos(CRO)$0.077773-3.69%
  • Circle USYCCircle USYC(USYC)$1.120.00%
  • BittensorBittensor(TAO)$301.10-3.88%
  • tether-goldTether Gold(XAUT)$4,684.36-0.02%
  • BlackRock USD Institutional Digital Liquidity FundBlackRock USD Institutional Digital Liquidity Fund(BUIDL)$1.000.00%
  • uniswapUniswap(UNI)$3.69-2.06%
  • polkadotPolkadot(DOT)$1.372.71%
  • pax-goldPAX Gold(PAXG)$4,683.66-0.02%
  • mantleMantle(MNT)$0.66-2.21%
  • World Liberty FinancialWorld Liberty Financial(WLFI)$0.0670940.04%
  • nearNEAR Protocol(NEAR)$1.592.84%
  • OndoOndo(ONDO)$0.390331-4.34%
  • Ondo US Dollar YieldOndo US Dollar Yield(USDY)$1.13-0.47%
  • okbOKB(OKB)$84.88-0.97%
  • Pi NetworkPi Network(PI)$0.170947-0.63%
  • Falcon USDFalcon USD(USDF)$1.000.10%
TradePoint.io
  • Main
  • AI & Technology
  • Stock Charts
  • Market & News
  • Business
  • Finance Tips
  • Trade Tube
  • Blog
  • Shop
No Result
View All Result
TradePoint.io
No Result
View All Result

Google DeepMind Introduces an AI-Enabled Mouse Pointer Powered by Gemini That Captures Visual and Semantic Context Around the Cursor

May 13, 2026
in AI & Technology
Reading Time: 6 mins read
A A
Google DeepMind Introduces an AI-Enabled Mouse Pointer Powered by Gemini That Captures Visual and Semantic Context Around the Cursor
ShareShareShareShareShare

The mouse pointer has sat at the center of personal computing for more than half a century. It tracks cursor position. It registers clicks. Beyond that, it does almost nothing. Google DeepMind researchers outlined a set of experimental principles and demos for an AI-enabled pointer that goes considerably further: one that understands not just where you are pointing, but what you are pointing at and why it matters.

The system is powered by Gemini and is currently in the experimental stage. Two demos are live in Google AI Studio today: one for editing an image and one for finding places on a map, both operable by pointing and speaking. A deeper integration called Magic Pointer is also rolling out inside Chrome, and a further integration is planned for Googlebook, Google’s new line of Gemini-powered laptops announced this week.

https://deepmind.google/blog/ai-pointer/

What DeepMind is Targeting

The frustration DeepMind researchers are addressing is a familiar one for anyone who has tried to use an AI assistant while already in the middle of work. Because a typical AI tool lives in its own window, users need to drag their world into it. The research team wants the opposite — intuitive AI that meets users across all the tools they use, without interrupting their flow.

In practice, today’s AI workflow often looks like this: you are working inside a document or a browser tab, you spot something you want to ask about, you switch to a chat interface, you re-describe what you were looking at, you run the query, and you paste the result back. This maps to a concrete technical gap: current LLM interfaces are largely text-in, text-out. They have no awareness of the screen state around them. The AI-enabled pointer is an attempt to close that gap by giving the model real-time visual and semantic context derived from cursor position and hover state — without requiring users to manually serialize that context into a written prompt.

YOU MAY ALSO LIKE

Sony’s Xperia 1 VIII Has Bigger Camera Sensors And A New Look

Mira Murati’s Thinking Machines Lab Introduces Interaction Models: A Native Multimodal Architecture for Real-Time Human-AI Collaboration

Four interaction principles

DeepMind researchers have developed four principles that together shift the hard work of conveying context and intent from the user to the computer, replacing text-heavy prompts with simpler, more intuitive interactions.

The first is Maintain the flow. AI capabilities should work across all apps, not force users into ‘AI detours’ between them. The prototype AI-enabled pointer is available wherever the user is working. For example, they could point at a PDF and request a bullet-point summary to paste directly into an email, hover over a table of statistics and request a pie chart version, or highlight a recipe and ask for all the ingredients doubled. This is a direct architectural stance: instead of building AI assistance as a sidecar application, the capability lives at the pointer level and is present in whichever tool the user is already working in.

The second is Show and tell. Current AI models demand precise instructions. To get a good response, a user has to write a detailed prompt. An AI-enabled pointer would streamline this process by smoothly capturing the visual and semantic context around the pointer, letting the computer ‘see’ and understand what’s important to the user. In the experimental system, just point, and the AI knows exactly which word, paragraph, part of an image, or code block the user needs help with. From a technical standpoint, this means the system treats cursor hover state and the surrounding UI content as structured model inputs — comparable to how multimodal models process image and text together, except here the visual region is dynamically cropped and contextualized in real time around a moving cursor.

The third is Embrace the power of ‘This’ and ‘That‘. In everyday interactions with each other, humans rarely speak in long, detailed paragraphs. We might say, ‘Fix this’, ‘Move that here’, or ‘What does this mean?’ — while relying on physical gestures and our shared context to fill in any gaps in understanding. An AI system that understands this combination of context, pointing and speech would allow users to make complex requests in natural shorthand, no fiddly prompting required. The name of the principle is deliberate: deictic language (words like ‘this’ and ‘that’ that depend on physical reference to carry meaning) is how humans naturally communicate when they can point at something. The AI-enabled pointer is designed to handle exactly that class of instruction without needing the user to spell out what “this” refers to.

The fourth is Turn pixels into actionable entities. For decades, computers have only tracked where we are pointing. AI can now also understand what the user is pointing at. This transforms pixels into structured entities, such as places, dates, and objects, that users can interact with instantly. A photo of a scribbled note becomes an interactive to-do list; a paused frame in a travel video becomes a booking link for that cool-looking restaurant. For ML engineers, this is the most technically substantive of the four principles. It describes an entity extraction step that happens at inference time on whatever visual content is under the cursor — converting raw pixel regions into typed, actionable objects rather than leaving them as unstructured screen content.

Where it is going

Google DeepMind is now integrating these principles to reimagine pointing in Chrome and the new Googlebook laptop experience. Starting now, instead of writing a complex prompt, users can use their pointer to ask Gemini in Chrome about the part of the webpage they care about. For example, selecting a few products on a page and asking to compare them, or pointing to where they want to visualize a new couch in their living room.

Key Takeaways

  • Google DeepMind introduces experimental demos of an AI-enabled mouse pointer powered by Gemini that captures visual and semantic context around the cursor — no manual prompting required.
  • The system is built on four principles: Maintain the flow, Show and tell, Embrace the power of “This” and “That”, and Turn pixels into actionable entities.
  • “Turn pixels into actionable entities” is the key technical idea — the pointer converts on-screen content into structured entities like places, dates, and objects that users can act on instantly.
  • Two live demos are available now in Google AI Studio (image editing and map search); Gemini in Chrome is rolling out today, with Magic Pointer for Googlebook coming later this year.
  • The core design shift: instead of users dragging context into an AI window, the AI follows the cursor across every app the user is already working in.

Check out the Technical details. Also, feel free to follow us on Twitter and don’t forget to join our 150k+ ML SubReddit and Subscribe to our Newsletter. Wait! are you on telegram? now you can join us on telegram as well.

Need to partner with us for promoting your GitHub Repo OR Hugging Face Page OR Product Release OR Webinar etc.? Connect with us


Michal Sutter is a data science professional with a Master of Science in Data Science from the University of Padova. With a solid foundation in statistical analysis, machine learning, and data engineering, Michal excels at transforming complex datasets into actionable insights.

Credit: Source link

ShareTweetSendSharePin

Related Posts

Sony’s Xperia 1 VIII Has Bigger Camera Sensors And A New Look
AI & Technology

Sony’s Xperia 1 VIII Has Bigger Camera Sensors And A New Look

May 13, 2026
Mira Murati’s Thinking Machines Lab Introduces Interaction Models: A Native Multimodal Architecture for Real-Time Human-AI Collaboration
AI & Technology

Mira Murati’s Thinking Machines Lab Introduces Interaction Models: A Native Multimodal Architecture for Real-Time Human-AI Collaboration

May 13, 2026
Family Sues OpenAI, Alleging ChatGPT Advice Led To Accidental Overdose
AI & Technology

Family Sues OpenAI, Alleging ChatGPT Advice Led To Accidental Overdose

May 13, 2026
Arm Warns of Phone Market Weakness | Bloomberg Tech 5/7/2026
AI & Technology

Arm Warns of Phone Market Weakness | Bloomberg Tech 5/7/2026

May 13, 2026
Next Post
King Charles on WHCD shooting: ‘Such acts of violence will never succeed’  

King Charles on WHCD shooting: ‘Such acts of violence will never succeed’  

Leave a Reply Cancel reply

Your email address will not be published. Required fields are marked *

Search

No Result
View All Result
King Charles and Queen Camilla depart from White House

King Charles and Queen Camilla depart from White House

May 11, 2026
U.S. economy adds 115,000 jobs, a strong gain for an uncertain labor market – The Washington Post

U.S. economy adds 115,000 jobs, a strong gain for an uncertain labor market – The Washington Post

May 8, 2026
King Charles and Queen Camilla arrive at the White House ahead of address to Congress

King Charles and Queen Camilla arrive at the White House ahead of address to Congress

May 13, 2026

About

Learn more

Our Services

Legal

Privacy Policy

Terms of Use

Bloggers

Learn more

Article Links

Contact

Advertise

Ask us anything

©2020- TradePoint.io - All rights reserved!

Tradepoint.io, being just a publishing and technology platform, is not a registered broker-dealer or investment adviser. So we do not provide investment advice. Rather, brokerage services are provided to clients of Tradepoint.io by independent SEC-registered broker-dealers and members of FINRA/SIPC. Every form of investing carries some risk and past performance is not a guarantee of future results. “Tradepoint.io“, “Instant Investing” and “My Trading Tools” are registered trademarks of Apperbuild, LLC.

This website is operated by Apperbuild, LLC. We have no link to any brokerage firm and we do not provide investment advice. Every information and resource we provide is solely for the education of our readers. © 2020 Apperbuild, LLC. All rights reserved.

No Result
View All Result
  • Main
  • AI & Technology
  • Stock Charts
  • Market & News
  • Business
  • Finance Tips
  • Trade Tube
  • Blog
  • Shop

© 2023 - TradePoint.io - All Rights Reserved!