OpenAI recently released GPT-5.4 with computer use support and the results are really impressive - 75.0% on OSWorld, which is above human-level for OS control tasks. I’ve been building a computer-use agent for a while now and plugging in the new model was a great test for the architecture.
The agent is provider-agnostic - right now it supports both OpenAI GPT-5.4 and Anthropic Claude. Adding a new provider is just one adapter file, the rest of the codebase stays untouched. Cross-platform too - same agent code runs on macOS, Windows, Linux, web, and even on a server through abstract ports (Mouse, Keyboard, Screen) with platform-specific drivers underneath.
In the video it draws the sun and geometric shapes from a text prompt - no scripted actions, just the model deciding where to click and drag in real time.
Currently working on:
Would love to hear how others are approaching computer-use agents. Is anyone else experimenting with the new GPT-5.4 computer use?