We Give Robots Vision

Hardware-agnostic edge robot vision framework with world models for predictive intelligence. Runs on Jetson, Pi, Intel, Hailo & Qualcomm.

Features

Object Detection

80+ object classes in real-time

Depth Estimation

Measure distance to everything

Face Detection

Identify and track faces

Gesture Recognition

Stop, wave, point commands

Pose Estimation

Detect body positions

Object Tracking

Follow objects across frames

Visual SLAM

Build maps, navigate autonomously

VLA Models

Vision-Language-Action

🔦

LIDAR Processing

Point cloud obstacle detection

🔀

Sensor Fusion

Camera + Depth + LIDAR

📸

Multi-Camera

Multiple camera support

Action Chunking

10-30 Hz real-time control

🧠

Diffusion Policy

Robot manipulation AI

🛡️

Safety Controller

E-STOP, velocity limits

❤️

Health Monitor

24/7 with auto-recovery

⬆️

OTA Updates

Model updates + rollback

🧠

World Models

Predictive planning at 200 Hz

🔮

V-JEPA 2

Spatiotemporal perception

📐

Depth Anything V3

35.7% better than MiDaS

🏭

Industry Templates

Warehouse, QA, Agriculture, Retail

🚀

Fleet Management

Multi-device deployment

🔧

Hardware Abstraction

TensorRT, OpenVINO, TVM, Hailo, QNN

Performance

Configuration FPS Use Case
Detection only (TensorRT) 35-40 YOLO11n FP16
Full pipeline (default) 4-6 All models enabled
Full pipeline + turbo 8-12 Aggressive frame skipping
Minimal 15-20 Detection + depth + tracking
World model (LeWM 15M) 100-200 Hz Planning only, <10ms

Hardware

Jetson Orin Nano

40 TOPS, 5-15W

Raspberry Pi 5

+ AI HAT 26 TOPS

Intel Core Ultra

48 TOPS NPU

Hailo-8

26 TOPS, 3.5W

Qualcomm RB5

15-30 TOPS

Camera

CSI, USB, RealSense

Quick Start

bash
git clone https://github.com/mandarwagh9/openeyes.git
cd openeyes
pip install -r requirements.txt
python src/main.py --debug

World Models & Predictive Intelligence (v2.5.0)

bash
# World model with predictive tracking
python src/main.py --world-model lewm --follow

# With safety evaluation
python src/main.py --world-model lewm --safety-predict

# Industry template (warehouse)
python src/main.py --template warehouse --debug

# Turbo mode for maximum FPS
python src/main.py --turbo --world-model lewm

# Process video files
python src/main.py --video input.mp4 --output output.mp4

New in v0.7.0 - v1.0.0

bash
# Multi-Modal Sensing (v0.7.0)
python src/main.py --lidar --lidar-topic /scan
python src/main.py --realsense
python src/main.py --multi-camera

# VLA & Performance (v0.8.0)
python src/main.py --int8 --dla
python src/main.py --diffusion-policy
python src/main.py --action-chunking --control-freq 20

# Safety & Reliability (v1.0.0)
python src/main.py --safety --max-velocity 1.0 --min-distance 0.3
python src/main.py --health-monitor
python src/main.py --ota-update

Models

Model Type Size
YOLO11n/12n/26nDetection2.6-5.4MB
MiDaS + Depth Anything V3Depth350MB
MediaPipeFace/Gesture/Pose~20MB
LeWMWorld Model15M params
V-JEPA 2Perception80-600M params
SmolVLAVLA450M params
OpenVLAVLA7B params