Need help?
<- Back

Comments (18)

  • avaer
    There's also the Prompt API, currently in Origin Trial, which supports this api surface for sites:https://developer.chrome.com/docs/ai/prompt-apiI just checked the stats: Model Name: v3Nano Version: 2025.06.30.1229 Backend Type: GPU (highest quality) Folder size: 4,072.13 MiB Different use case but a similar approach.I expect that at some point this will become a native web feature, but not anytime soon, since the model download is many multiples the size of the browser itself. Maybe at some point these APIs could use LLMs built into the OS, like we do for graphics drivers.
  • veunes
    It’s a neat idea, but giving a 2B model full JS execution privileges on a live page is a bit sketchy from a security standpoint. Plus, why tie inference to the browser lifecycle at all? If Chrome crashes or the tab gets discarded, your agent's state is just gone. A local background daemon with a "dumb" extension client seems way more predictable and robust fwiw
  • emregucerr
    I would love to see someone build it as some kind of an SDK. App builders could use it as a local LLM plugin when dealing with data involving sensitive information.It's usually too much when an app asks someone to setup a local LLM but this I believe could solve that problem?
  • dabrez
    I have this written a a project I will attempt to do in the future, I also call it "weapons grade unemployment" in the notes I was proposing to use granite but the principle still stands. You beat me to it.
  • montroser
    Not sure if I actually want this (pretty sure I don't) -- but very cool that such a thing is now possible...
  • eric_khun
    it would be awesome if a local model would be directly embeded to chrome and developer could query them.Anyone know if this is somehow possible without going through an extension?
  • Morpheus_Matrix
    [flagged]