Give your agents eyes.

macOS automation that sees the screen and does the clicks.

Peekaboo brings high‑fidelity screen capture, AI analysis, and end‑to‑end GUI automation to macOS — as a native CLI and an MCP server.

macOS 15+ Swift 6.2 Node 22+ (MCP) OpenAI · Claude · Grok · Gemini · Ollama

Install

Two ways: native app/CLI via Homebrew, or MCP server via npm. Both expose the same toolset.

brew install steipete/tap/peekaboo
Requires macOS Screen Recording + Accessibility permissions. Grant guide.
One-liner
peekaboo "Open Safari, go to github.com, and search for Peekaboo"

Features

Tools you can trust: deterministic outputs, typed JSON, and composable automation.

Pixel‑accurate capture

Screens, windows, menu bar. Optional Retina 2× scaling. Fast enough for feedback loops.

CLI: peekaboo image, peekaboo see

Click by intent

Resolve UI elements and act on them: click, type, scroll, drag, hotkeys, menus, Spaces.

CLI: peekaboo click, peekaboo type, peekaboo menu

Agent‑native

Natural language that chains tools — plus an MCP server so Claude Desktop/Cursor can drive it.

CLI: peekaboo agent, peekaboo mcp

How it works

A tight loop: capture → interpret → act. Repeat until the task is done.

  1. 01

    See

    Capture a screen/window and get a structured UI map with stable IDs.

  2. 02

    Decide

    Pick a target: a button label, a menu path, a specific window, a Space.

  3. 03

    Do

    Click, type, scroll, drag, or run a full agent plan — with receipts you can log.