Speak a command, and Nova AI sees your screen, reasons about what to do, and executes actions autonomously. Search Amazon, fill forms, navigate sites — all by voice. Powered by Amazon Nova 2 Lite multimodal reasoning.
No tab switching, no copy-pasting, no context loss. Just hold, ask, and understand.
Press and hold the shortcut key anywhere on any webpage. A waveform appears — you're live.
"Add Quest protein bars to my Amazon cart." "Search YouTube for coding tutorials." "Scroll down and find the cheapest option."
Nova AI sees your screen, reasons about what to do, and clicks, types, navigates — autonomously completing your task step by step.
Watch how to install, configure, and start talking to your screen.
Powerful when you need it, invisible when you don't.
Multimodal AI sees your screenshot + DOM structure. Understands page context, identifies elements, and plans actions with chain-of-thought reasoning.
Hold backtick, speak naturally. Amazon Transcribe converts speech to text, then Nova decides and acts on your behalf.
After each action, Nova re-captures the screen, re-scrapes the DOM, and decides the next step. Loops up to 25 iterations until your task is complete.
Failed selectors? Nova retries with fresh DOM. Navigation mid-task? Waits for reload and continues. Element highlighting shows exactly what's being clicked.
Ask "What is an API?" and watch the answer adapt to who's asking.
Install ScreenSense Voice for free. Speak a command on any website — Nova AI handles the rest.
Clone from GitHubClone the repo, run npm install && npm run build, then load the dist/ folder into chrome://extensions with Developer Mode enabled. Follow the README setup guide for detailed steps.