I'm visually impaired, made a thing to drive Wayland around with agents

Posted by Amonwilde@reddit | linux | View on Reddit | 35 comments

Hey folks,

I'm a visually impaired Linux user who typically uses x11 and I wanted to learn about Wayland and port some of my accessibility tools over to Wayland/GNOME but hadn't taken the plunge. A A little while back Anthropic released some desktop driver stuff, of course not on Linux, and I was kind of jealous. I thought putting together something that would let agents control a Wayland desktop would let me learn about the Wayland SPI and device stack plus be a cool project.

Tine is a Python CLI plus a small GNOME Shell extension that combines AT-SPI2 accessibility reads, vision fallback via a labeled coordinate grid, and kernel-level /dev/uinput input. It lets an AI coding agent (Claude Code, Codex, anything that can run a shell command) actually use a GNOME Wayland desktop — click buttons, fill forms, read what's on screen — without the Screencast portal throwing a consent dialog on every action.

Repo: https://github.com/smythp/tine

Caveat: use at your own risk. Agents are nondeterministic, etc. With that said, I just put Arch on an old laptop and let agents control it over ssh.

Let me know what you think.