Hackathon Build

VoiceAgents

Voice-native computing. Local. Private. Instant.
OpenClaw-grade ambition, built for accessibility: the browser and the DAW, commanded by voice, on your machine.

Chromium: \"open reddit and scroll\" LMMS: \"play from bar 8\" System: Local. Private. Fast.
Scroll

Two agents. One thesis.

Chromium

Cursor for the web. Tabs, search, scroll, media, real sites - spoken into motion. The browser stops being an obstacle course.

LMMS

Cursor for the session. Transport, tracks, plugins - voice and text through a real AgentControl boundary. The DAW stops fighting you.

Line for the room: We took “AI is the new cursor” and put it where creators actually live - the open web and the open session. Same pattern. Two industries. One voice.

Chromium

Cursor for the web.

Own the room: Speech becomes a first-class input to the browser - not dictation dumped into a search box.

LMMS Agent

Cursor for music production.

Own the room: Not “chat for musicians” - the moment the DAW takes direction like an instrument, not a fight.

More than a demo.

Closer: Agentic, but accountable.

When everything is voice-native.

Final beat: A world where your computer understands your job - without pretending it owns your life.

Architecture north star.

Full technical plan: local-only inference, gaze as a ranking signal, policy engine for risk tiers, accessibility-first UX - documented for implementers.

voice_agent_full_plan_v4.md

Command schemas, world model, fusion, safety, metrics - the blueprint behind the story.