Skip to content

cloneisme/open-world-agents

 
 

Repository files navigation

Open World Agents

🚀 Open World Agents

Everything you need to build state-of-the-art foundation multimodal desktop agent, end-to-end.

Documentation License: MIT Python 3.11+ GitHub stars

🚀 Quick Start: Record → Train in 3 Steps

# 1. Record desktop interaction
$ ocap my-session.mcap

# 2. Process to training format
$ python scripts/01_raw_events_to_event_dataset.py --train-dir ./

# 3. Train your model
$ python train.py --dataset ./event-dataset

📖 Detailed Guide: Complete Quick Start Tutorial - Step-by-step walkthrough with examples and troubleshooting

Overview

Open World Agents is a comprehensive framework for building AI agents that interact with desktop applications through vision, keyboard, and mouse control. Complete toolkit from data capture to model training and evaluation:

  • 🌍 Environment Framework: "USB-C of desktop agents" - universal interface for native desktop automation with pre-built plugins for desktop control, high-performance screen capture (6x faster), and zero-configuration plugin system
  • 📊 Data Infrastructure: Complete desktop agent data pipeline from recording to training with OWAMcap format - a universal standard powered by mcap
  • 🛠️ CLI Tools: Command-line utilities (owl) for recording, analyzing, and managing agent data
  • 🤖 Examples: Complete implementations and training pipelines for multimodal agents

Why OWA?

Fragmented tools make desktop AI development painful. Most solutions force you to:

  • Stitch together incompatible recording tools
  • Build custom data pipelines from scratch
  • Handle real-time performance issues yourself
  • Start agent development with no examples

OWA solves this with a unified framework: record with ocap, train with standardized datasets, deploy with real-time environment components, and learn from community examples.

What Can You Build?

Anything that runs on desktop. If a human can do it on a computer, you can build an AI agent to automate it.

🤖 Desktop Automation: Navigate applications, automate workflows, interact with any software
🎮 Game AI: Master complex games through visual understanding and real-time decision making
📊 Training Datasets: Capture high-quality human-computer interaction data for foundation models
🤗 Community Datasets: Access and contribute to growing OWAMcap datasets on HuggingFace
📈 Benchmarks: Create and evaluate desktop agent performance across diverse tasks

Project Structure

The repository is organized as a monorepo with multiple sub-repositories under the projects/ directory. Each sub-repository is a self-contained Python package installable via pip or uv and follows namespace packaging conventions.

open-world-agents/
├── projects/
│   ├── mcap-owa-support/     # OWAMcap format support
│   ├── owa-core/             # Core framework and registry system
│   ├── owa-msgs/             # Core message definitions with automatic discovery
│   ├── owa-cli/              # Command-line tools (ocap, owl)
│   ├── owa-env-desktop/      # Desktop environment plugin
│   ├── owa-env-example/      # Example environment implementations
│   ├── owa-env-gst/          # GStreamer-based screen capture
│   └── [your-plugin]/        # Contribute your own plugins!
├── docs/                     # Documentation
└── README.md

Core Packages

owa owa

The easiest way to get started is to install the owa meta-package, which includes all core components and environment plugins:

pip install owa

All OWA packages use namespace packaging and are installed in the owa namespace (e.g., owa.core, owa.cli, owa.env.desktop). For more detail, see Packaging namespace packages. We recommend using uv as the package manager.

Name Release in PyPI Conda Description
owa.core owa-core owa-core Framework foundation with registry system
owa.msgs owa-msgs owa-msgs Core message definitions with automatic discovery
owa.cli owa-cli owa-cli Command-line tools (owl) for data analysis
mcap-owa-support mcap-owa-support mcap-owa-support OWAMcap format support and utilities
ocap 🎥 ocap ocap Desktop recorder for multimodal data capture
owa.env.desktop owa-env-desktop owa-env-desktop Mouse, keyboard, window event handling
owa.env.gst 🎥 owa-env-gst owa-env-gst GStreamer-powered screen capture (6x faster)
owa.env.example - - Reference implementations for learning

🎥 Video Processing Packages: Packages marked with 🎥 require GStreamer dependencies. Install conda install open-world-agents::gstreamer-bundle first for full functionality.

📦 Lockstep Versioning: All first-party OWA packages follow lockstep versioning, meaning they share the same version number to ensure compatibility and simplify dependency management.

💡 Extensible Design: Built for the community! Easily create custom plugins like owa-env-minecraft or owa-env-web to extend functionality.

Community Packages

Help us grow the ecosystem! 🌱 Community-contributed environment plugins extend OWA's capabilities to specialized domains.

Example plugin ideas from the community:

Example Name Description
owa.env.minecraft Minecraft automation & bot framework
owa.env.web Browser automation via WebDriver
owa.env.mobile Android/iOS device control
owa.env.cad CAD software automation (AutoCAD, SolidWorks)
owa.env.trading Financial trading platform integration

💡 Want to contribute? Check our Plugin Development Guide to create your own owa.env.* package!

💭 These are just examples! The community decides what plugins to build. Propose your own ideas or create plugins for any domain you're passionate about.

Desktop Recording with ocap

ocap (Omnimodal CAPture) is a high-performance desktop recorder that captures screen video, audio, keyboard/mouse events, and window events in synchronized formats. Built with Windows APIs and GStreamer for hardware-accelerated recording with H265/HEVC encoding.

  • Complete recording: Video + audio + keyboard/mouse + window events
  • High performance: Hardware-accelerated, ~100MB/min for 1080p
  • Simple usage: ocap my-recording (stop with Ctrl+C)
  • Modern formats: OWAMcap with flexible MediaRef system (supports MKV, images, URLs, embedded data)

📖 Detailed Documentation: See Desktop Recording Guide for complete setup, usage examples, and troubleshooting.

Quick Start

Environment Usage: Three Types of Components

OWA's Environment provides three types of components for real-time agent interaction:

Callables - Direct function calls for immediate actions

from owa.core import CALLABLES
# Components automatically available - zero configuration!

# Get current time, capture screen, click mouse
current_time = CALLABLES["std/time_ns"]()
screen = CALLABLES["desktop/screen.capture"]()
CALLABLES["desktop/mouse.click"]("left", 2)  # Double-click

Listeners - Event monitoring with user-defined callbacks

from owa.core import LISTENERS
import time

# Monitor keyboard events
def on_key(event):
    print(f"Key pressed: {event.vk}")

listener = LISTENERS["desktop/keyboard"]().configure(callback=on_key)
with listener.session:
    input("Press Enter to stop...")

# Periodic tasks
def on_tick():
    print(f"Tick: {CALLABLES['std/time_ns']()}")

with LISTENERS["std/tick"]().configure(callback=on_tick, interval=1).session:
    time.sleep(3)  # Prints every second for 3 seconds

Runnables - Background processes that can be started/stopped

from owa.core import RUNNABLES

# Periodic screen capture
capture = RUNNABLES["gst/screen_capture"]().configure(fps=60)
with capture.session:
    frame = capture.grab()

Message Types - Access structured message definitions

from owa.core import MESSAGES

# Message types automatically available
KeyboardEvent = MESSAGES["desktop/KeyboardEvent"]
ScreenCaptured = MESSAGES["desktop/ScreenCaptured"]

High-Performance Screen Capture

import time
from owa.core import CALLABLES, LISTENERS, MESSAGES

# Components and messages automatically available - no activation needed!

def on_screen_update(frame, metrics):
    print(f"📸 New frame: {frame.frame_arr.shape}")
    print(f"⚡ Latency: {metrics.latency*1000:.1f}ms")

    # Access screen message type from registry
    ScreenCaptured = MESSAGES['desktop/ScreenCaptured']
    print(f"Frame message type: {ScreenCaptured}")

# Start real-time screen capture
screen = LISTENERS["gst/screen"]().configure(
    callback=on_screen_update, fps=60, show_cursor=True
)

with screen.session:
    print("🎯 Agent is watching your screen...")
    time.sleep(5)

Powered by the powerful Gstreamer and Windows API, our implementation is 6x faster than comparatives.

Library Avg. Time per Frame Relative Speed
owa.env.gst 5.7 ms 1× (Fastest)
pyscreenshot 33 ms 🚶‍♂️ 5.8× slower
PIL 34 ms 🚶‍♂️ 6.0× slower
MSS 37 ms 🚶‍♂️ 6.5× slower
PyQt5 137 ms 🐢 24× slower

📌 Tested on: Intel i5-11400, GTX 1650

Not only does owa.env.gst achieve higher FPS, but it also maintains lower CPU/GPU usage, making it the ideal choice for screen recording. Same applies for ocap, since it internally imports owa.env.gst.

📊 See detailed benchmarks and methodology →

Desktop Recording & Dataset Sharing

Record your desktop usage data and share with the community:

# Install GStreamer dependencies (for video recording) and ocap
conda install open-world-agents::gstreamer-bundle && pip install ocap

# Record desktop activity (includes video, audio, events)
ocap my-session

# Upload to HuggingFace, browse community datasets!
# Visit: https://huggingface.co/datasets?other=OWA

🤗 Community Datasets: Democratizing Desktop Agent Data

Browse Available Datasets: 🤗 datasets?other=OWA

  • Growing Collection: Hundreds of community-contributed datasets
  • Standardized Format: All use OWAMcap for seamless integration
  • Interactive Preview: Hugging Face Spaces Visualizer
  • Easy Sharing: Upload recordings directly with one command

🚀 Impact: OWA has democratized desktop agent data, growing from zero to hundreds of public datasets in the unified OWAMcap format.

Access Community Datasets:

# Load datasets from HuggingFace
from owa.data import load_dataset

# Browse available OWAMcap datasets
datasets = load_dataset.list_available(format="OWA")

# Load a specific dataset
data = load_dataset("open-world-agents/example_dataset")

Data Format Preview

OWAMcap combines the robustness of MCAP with specialized desktop interaction schemas. Perfect synchronization of screen captures, input events, and window context with nanosecond precision.

Key Features:

  • 🔄 Universal Standard: Unlike fragmented formats, enables seamless dataset combination for large-scale foundation models (OWAMcap)
  • 🎯 High-Performance Multimodal Storage: Lightweight MCAP container with nanosecond precision for synchronized data streams (MCAP)
  • 🔗 Flexible MediaRef: Smart references to both external and embedded media (file paths, URLs, data URIs, video frames) with lazy loading - keeps metadata files small while supporting rich media (OWAMcap)Learn more
  • 🤗 Training Pipeline Ready: Native HuggingFace integration, seamless dataset loading, and direct compatibility with ML frameworks (Ecosystem)Browse datasets | Data pipeline

📖 Learn More: Why OWAMcap? | Complete Format Guide | vs Other Formats

$ owl mcap info example.mcap
library:   mcap-owa-support 0.5.1; mcap 1.3.0
profile:   owa
messages:  864
duration:  10.3574349s
start:     2025-06-27T18:49:52.129876+09:00 (1751017792.129876000)
end:       2025-06-27T18:50:02.4873109+09:00 (1751017802.487310900)
compression:
        zstd: [1/1 chunks] [116.46 KiB/16.61 KiB (85.74%)] [1.60 KiB/sec]
channels:
        (1) window           11 msgs (1.06 Hz)    : desktop/WindowInfo [jsonschema]
        (2) keyboard/state   11 msgs (1.06 Hz)    : desktop/KeyboardState [jsonschema]
        (3) mouse/state      11 msgs (1.06 Hz)    : desktop/MouseState [jsonschema]
        (4) screen          590 msgs (56.96 Hz)   : desktop/ScreenCaptured [jsonschema]
        (5) mouse           209 msgs (20.18 Hz)   : desktop/MouseEvent [jsonschema]
        (6) keyboard         32 msgs (3.09 Hz)    : desktop/KeyboardEvent [jsonschema]
channels: 6
attachments: 0
metadata: 0

🛠️ CLI Tools (owl)

# Data analysis
owl mcap info session.mcap              # File overview & statistics
owl mcap cat session.mcap --n 10        # View messages
owl video probe session.mkv             # Video analysis

# Environment management
owl env list                            # List plugins
owl env search "mouse.*click"           # Search components
owl messages show desktop/KeyboardEvent # View schemas

💡 Complete CLI Reference: For detailed information about all CLI commands and options, see the CLI Tools documentation.

Installation

Quick Start

# Install all OWA packages
pip install owa

# For video recording/processing, install GStreamer dependencies first:
conda install open-world-agents::gstreamer-bundle
pip install owa

💡 When do you need GStreamer?

  • Video recording with ocap desktop recorder
  • Real-time screen capture with owa.env.gst
  • Video processing capabilities

Skip GStreamer if you only need:

  • Data processing and analysis
  • ML training on existing datasets
  • Headless server environments

Editable Install (Development)

For development or contributing to the project, you can install packages in editable mode. For detailed development setup instructions, see the Installation Guide.

Features

🌍 Environment Framework: "USB-C of Desktop Agents"

  • ⚡ Real-time Performance: Optimized for responsive agent interactions (GStreamer components achieve <30ms latency)
  • 🔌 Zero-Configuration: Automatic plugin discovery via Python Entry Points
  • 🌐 Event-Driven: Asynchronous processing that mirrors real-world dynamics
  • 🧩 Extensible: Community-driven plugin ecosystem

→ View Environment Framework Guide

📊 Data Infrastructure: Complete Pipeline

  • 🔄 Universal Standard: Unlike fragmented formats, enables seamless dataset combination for large-scale foundation models (OWAMcap)
  • 🎯 High-Performance Multimodal Storage: Lightweight MCAP container with nanosecond precision for synchronized data streams (MCAP)
  • 🔗 Flexible MediaRef: Smart references to both external and embedded media (file paths, URLs, data URIs, video frames) with lazy loading - keeps metadata files small while supporting rich media (OWAMcap)Learn more
  • 🤗 Training Pipeline Ready: Native HuggingFace integration, seamless dataset loading, and direct compatibility with ML frameworks (Ecosystem)Browse datasets | Data pipeline

→ View Data Infrastructure Guide

🤗 Community & Ecosystem

  • 🌱 Growing Ecosystem: Hundreds of community datasets in unified OWAMcap format
  • 🤗 HuggingFace Integration: Native dataset loading, sharing, and interactive preview tools
  • 🧩 Extensible Architecture: Modular design for custom environments, plugins, and message types
  • 💡 Community-Driven: Plugin ecosystem spanning gaming, web automation, mobile control, and specialized domains

→ View Community Datasets

Documentation

Contributing

We welcome contributions! Whether you're:

  • Building new environment plugins
  • Improving performance
  • Adding documentation
  • Reporting bugs

Please see our Contributing Guide for details.

License

This project is released under the MIT License. See the LICENSE file for details.

About

Everything you need to build state-of-the-art foundation multimodal desktop agent, end-to-end.

Resources

License

Contributing

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • Python 80.6%
  • Jupyter Notebook 14.8%
  • JavaScript 2.2%
  • Shell 0.9%
  • CSS 0.5%
  • HTML 0.5%
  • Other 0.5%