Amplifier Browser Runtime

AI as a Document

What happens when you put an AI agent framework
inside a single HTML file

February 2026 Active

Every AI product today requires...

๐Ÿ”‘

An Account

Sign up, verify email, set up a profile

๐Ÿ—๏ธ

An API Key

Generate, store, rotate, protect

๐ŸŒ

Internet

Always-on connection to a remote server

๐Ÿ’ณ

A Subscription

Monthly fees, usage limits, billing anxiety

๐Ÿคž

Trust

That your data stays private on someone else's server

What if none of that was necessary?

One file. Everything inside.

file:///ai-tutor.html
Explain how photosynthesis works
Photosynthesis is the process by which plants convert sunlight, water, and COโ‚‚ into glucose and oxygen. It occurs in two stages: the light-dependent reactions in the thylakoid membranes...
Amplifier kernel via Pyodide/WebAssembly
Local LLM via WebLLM/WebGPU
Specialized bundle with domain knowledge
Complete chat interface

Open it in Chrome. No server. No API key. No internet after first model download.

Five layers, one file

Your Browser
โ†’
Pyodide (WASM)
โ†’
amplifier-core
โ†’
WebLLM Provider
โ†’
WebGPU ยท Your GPU

~6 lines to initialize

const amp = new AmplifierBrowser(); await amp.init({ model: "Phi-3.5-mini-instruct-q4f16_1-MLC" }); const reply = await amp.chat("Explain quantum entanglement");

What happens under the hood

The AmplifierBrowser facade boots Python inside WebAssembly via Pyodide. The same amplifier-core kernel that powers terminal sessions initializes in your browser tab.

WebLLM loads a quantized model directly onto your GPU through WebGPU. No network calls. No intermediary. Your prompt goes straight to local silicon.

Running on YOUR GPU

General Purpose

Phi-3.5-mini

Microsoft's compact powerhouse. Strong reasoning for its size, ideal for general Q&A, tutoring, and code assistance.

~4 GB VRAM ยท 3.8B parameters
Fast Responses

Llama-3.2-3B

Meta's speed-optimized model. Quick responses for conversational use, summaries, and lightweight tasks.

~4 GB VRAM ยท 3B parameters
Multilingual ยท Quality

Qwen2.5-7B

Alibaba's multilingual model. Higher quality output with support for many languages. Needs more VRAM but delivers more capable results.

~8 GB VRAM ยท 7B parameters
๐Ÿ”’ Your data never leaves your device. Fully offline after first model download.

When AI is just a file...

๐Ÿ“š

Education

A teacher sends students an HTML file. Inside is a specialized tutor that knows the curriculum, answers questions, and works offline in any classroom.

๐Ÿข

Enterprise

A company distributes a portable knowledge base. No server infrastructure. No ongoing costs. Every employee gets AI on their laptop.

๐Ÿ”’

Hostile Environments

A journalist in a sensitive region has AI assistance without any network traffic. No server logs. No API calls to intercept. Complete operational security.

๐Ÿ’ป

Prototyping

A developer prototypes agent behavior in a browser tab. No deployment pipeline. No cloud account. Open a file, iterate, ship.

Same kernel.
Different mineral.

โฌ›

Terminal

macOS, Linux, Windows โ€” the classic CLI experience

amplifier-core
๐Ÿ“ฆ

Docker

Containerized workloads with full isolation

amplifier-core
๐Ÿ”—

SSH

Remote machines and build servers

amplifier-core
๐ŸŒ

Browser

WebAssembly โ€” the new frontier

amplifier-core
~2,600
Lines of Kernel
5
Protocol Contracts
4
Runtimes

Same protocols. Same composition model. Same bundle format. The platform() function in the kernel has 'wasm' as a first-class return value โ€” the browser was a design target from day one, not an afterthought.

Built and tested by an agent

The webruntime-developer agent doesn't just build browser AI apps โ€” it tests them autonomously with Playwright and WebGPU flags. No untested code reaches the user.

๐Ÿ”จ
Build HTML
๐Ÿงช
Test
๐Ÿ”ง
Fix
โœ“
Re-test
๐Ÿ“ฆ
Deliver

Headless WebGPU testing

Playwright launches Chromium with WebGPU flags enabled. The agent generates tests, runs them headless, reads console output on failure, and iterates until everything works.

Agent-driven iteration

When a test fails, the agent reads the console logs, diagnoses the issue, fixes the code, and re-runs. This loop continues until the app works correctly โ€” no human debugging required.

AI as a document
changes everything

๐Ÿ“ง

Email it

Attach an AI assistant to a message. The recipient opens a file and has a working AI.

๐Ÿ’พ

USB drive

Carry an AI in your pocket. Works anywhere with a modern browser and a GPU.

๐ŸŒ

Static website

Host AI on a $0/month static site. No backend. No database. No ops team.

๐Ÿ“–

Embed anywhere

Drop an AI into documentation, a wiki, an internal portal โ€” it just works.

No accounts. No infrastructure. No ongoing costs.
The file IS the application.

Research Methodology

Data as of: February 26, 2026

Feature status: Active โ€” code present in production repositories with active development

Research performed:

  • Source code analysis of amplifier_webruntime.py โ€” confirmed imports from amplifier_core.interfaces (Provider, ContextManager, Tool, Orchestrator)
  • Protocol analysis of amplifier_env_common/protocol.py โ€” confirmed 'wasm' as first-class platform() return value
  • WebLLM bundle analysis โ€” confirmed WebGPU integration, model catalog, and provider implementation
  • Agent definition review โ€” webruntime-developer in amplifier-foundation confirmed with Playwright + WebGPU testing workflow
  • Kernel line count: ~2,600 lines with 5 protocol contracts across 4 runtimes

Repositories: amplifier-core (kernel), amplifier-foundation (webruntime-developer agent), webllm bundle (browser LLM provider)

Model VRAM requirements: From WebLLM model catalog documentation. Actual VRAM usage may vary by quantization and browser.

Primary contributors: Not determined โ€” research did not extract individual contributor attribution from git history for this deck. The browser runtime spans multiple repos and contributors.

Gaps: No independent VRAM benchmarks were run. Model parameter counts are from published model cards. Exact initialization line count (~6 lines) is from the AmplifierBrowser facade API surface, not a formal LOC metric. Individual contributor commit shares were not analyzed.

Every other AI product requires infrastructure.
Amplifier's browser runtime requires a browser.

That's not a limitation โ€” it's liberation.

Explore Amplifier on GitHub โ†’
More Amplifier Stories