See Everything. Capture Everything. Train Everything.

The Data Infrastructure for Physical AI

The world's leading AI researchers are building world models and spatial intelligence. They need high-fidelity 3D training data from real environments. DreamVu's omnidirectional capture platform delivers it at scale.

1,000+
Hours of 3D Data
500+
Distinct Skills
360°
Spatial Capture
360° Omnidirectional Capture
The Race to Build World Models
Starts with Real-World Data
The most influential minds in AI have converged on one conclusion: the next frontier isn't larger language models — it's machines that understand and interact with the physical world.

Spatial Intelligence Is the Next Frontier

Fei-Fei Li — the Stanford professor who created ImageNet and catalyzed the deep learning revolution — has made spatial intelligence the focus of her latest company, World Labs. Her thesis: AI must learn to perceive, reason about, and act in three-dimensional space. Not from text. Not from flat images. From spatially rich, real-world data.

"Spatial intelligence is the next major capability AI needs to develop. It's how humans and animals make sense of the world — and it's what's missing from today's AI systems."

— Fei-Fei Li, Stanford HAI & World Labs

Yann LeCun — Meta's Chief AI Scientist and Turing Award winner — has been equally direct. He argues that the path to truly intelligent machines runs through world models: internal representations of how the physical world works, learned from observation, not text.

"A system trained on text will never understand the physical world. You need world models — learned from video and sensory data — that can predict what happens next."

— Yann LeCun, Meta AI & NYU

Both visions share a common prerequisite: massive amounts of high-fidelity, spatially aware, real-world 3D data. And that's exactly what doesn't exist today — at least, not at the scale or quality these models demand.

🌍

World Models Need Real Worlds

VLA (Vision-Language-Action) models can't learn physics, spatial relationships, or manipulation skills from 2D images and text. They need dense 3D captures of real environments with real people performing real tasks.

⚠️

The Data Bottleneck Is Critical

Billions have been poured into model architectures — GR00T, RT-2, Octo, π₀ — but the training data barely exists. Open-source robotics datasets are small, narrow-FOV, and lack the 3D spatial richness these models require.

🎯

DreamVu Fills the Gap

Our patented omnidirectional 3D capture technology produces exactly the data that world models and spatial AI systems need — 360° depth + RGB, 3D occupancy maps, semantic labels, and skill segmentation at scale.

Why Training Physical AI Is So Hard
Humanoid robots don't just navigate — they manipulate objects, coordinate limbs, understand context, and learn from watching others. Current data falls short.
👁️

The Perspective Problem

Humanoids need egocentric views (what they see) and exocentric views (how they appear to others). Traditional capture misses half the picture. DreamVu's 360° capture gives you both — simultaneously.

🧩

The Multimodal Gap

Physical AI models need vision + language + action data together. Most datasets provide vision only — leaving teams to stitch together incomplete signals. DreamVu delivers all three, synchronized.

🔄

The Sim-to-Real Gap

Humanoids trained in simulation fail when deployed in real environments. DreamVu captures the real world in formats that translate directly into Isaac Sim and back — closing the sim-to-real loop.

From Real World to Training Pipeline
Our end-to-end platform transforms real-world captures into VLA-ready training datasets — delivered in Isaac Sim, LeRobot, and Open X-Embodiment formats.
1

Synchronized Capture

Hardware-synced egocentric + 360° exocentric cameras with full RGB + depth in real environments

2

Multimodal Annotation

AI-assisted (SAM2, Grounding DINO) + human QA delivers vision, language, and action labels — 10× faster than traditional 3D annotation

3

3D Reconstruction

3D Gaussian Splatting creates photorealistic scenes with all annotations preserved — ready for simulation conversion

4

Simulation Conversion

Automated USD export with physics properties for NVIDIA Isaac Sim and Unreal Engine 5.3+

5

Synthetic Generation

1,000+ frames/hour with domain randomization — all modalities and skill transfer demos preserved

6

Real-World Validation

Continuous verification: sim-to-real transfer rates, manipulation success, and skill transfer effectiveness

Vision Data

  • Synchronized ego + 360° exo video with depth
  • Object segmentation with instance IDs
  • 6DOF object poses
  • Manipulation affordances (grip types, approach vectors)
  • Human and robot demonstrations in 360° view

Language Data

  • QA pairs describing objects, actions, and scene elements
  • Action summaries for every sequence
  • Spatial relations between objects and actors
  • Context descriptions for scene understanding
  • Ready for VLA instruction following

Action Data

  • Full trajectories for every actor in 360° scene
  • Movement paths with timestamps
  • Interaction sequences showing manipulation
  • Demonstration labels for skill transfer
  • Kinematic data where available
🟢 NVIDIA Isaac Sim Native USD
🤗 Hugging Face LeRobot RLDS
📦 Open X-Embodiment
🎮 Unreal Engine 5.3+ Physics
The Alia 360° Camera
The world's only omnidirectional vision system with high-resolution 3D depth perception at the edge — protected by 32+ patents.

Alia Specs

Full 360° coverage with long-range, high-resolution 3D depth in a single compact unit. No stitching, no blind spots, no multi-sensor calibration.
360°
Horizontal FOV
6912×3072
Stereo Resolution
120m
Detection Range
30fps
Frame Rate
Image & Depth
Vertical FOV57.2°
Depth Range0cm – 20m (no blind spot)
Depth Accuracy7mm at 5 meters
OutputReal-time RGB + Depth
Edge AI & Durability
ProcessingEmbedded edge AI
Human Detection120m range
Facial Recognition40m range
Ingress RatingIP67
CompatibilityUbuntu, ROS, OpenCV

Purpose-Built for Physical AI Data

Developed from breakthrough research at IIIT Hyderabad (published at CVPR 2016) and refined over 8 years of production deployment in autonomous mobile robots, UV disinfection systems, and industrial applications worldwide.

Alia is the only camera that combines full 360° coverage with long-range, high-resolution 3D depth sensing in a single compact unit — with on-board edge AI processing. This proprietary technology creates a defensible moat: we capture spatial data that no other company can replicate.

🔵

Complete Scene Coverage

Traditional cameras see 60–90°. In a warehouse or retail environment, most action happens outside that cone. Alia captures everything in all directions simultaneously.

👥

Skill Transfer at Scale

When multiple humans and robots demonstrate tasks throughout an environment, one Alia captures all demonstrations happening anywhere in the space — no repositioning required.

🔷

3D Gaussian Splatting

The 360° coverage provides ideal input for photorealistic 3D reconstruction. All multimodal annotations propagate automatically from 2D frames to the 3D scene.

🛡️

32+ Patents

Protected omnidirectional 3D vision technology with 8+ years of production deployment. A defensible competitive advantage that ensures unique data capture capabilities.

Grocery Store Dataset
1,000 hours of enriched 3D video captured in partnership with a leading grocery store chain — the most comprehensive real-world manipulation dataset ever created for Physical AI.

Grocery Retail

1,000
Hours of Enriched 3D Video
Capture Locations 5 Operational Stores
Distinct Skills 500+
Capture Technology Alia 360° Depth + RGB
Annotation Layers Occupancy, Semantic, Skills, Interactions
Formats Isaac Sim · LeRobot · Open X

Why Grocery?

A grocery store contains more distinct manipulation tasks per square foot than almost any other environment — making it the ideal proving ground for Physical AI.

Unmatched Skill Density

Picking, placing, stacking, scanning, bagging, mopping, organizing — 500+ distinct skills captured across customer, staff, and logistics operations.

Massive Market Pull

Autonomous restocking and checkout are among the highest-demand use cases for humanoid robots, targeting the $22B machine vision in retail market.

Transferable Complexity

If a VLA model can handle a cluttered grocery aisle with customers, carts, and staff in motion, it transfers to warehouses, fulfillment centers, and retail at large.

Open Teaser on Hugging Face

A curated 20–30 hour subset available in LeRobot format — try before you buy, benchmark against your existing training data.

Simple Per-Hour Pricing
Every hour includes the full annotation stack: 360° 3D capture, occupancy maps, semantic labels, skill segmentation, and Isaac Sim-native delivery.

Showcase License

$500/hr
Non-exclusive annual license to the grocery store dataset. Less than the loaded cost of one robotics engineer — for data that shaves months off training.
  • Full 1,000-hour dataset
  • All annotation layers included
  • Isaac Sim + LeRobot + Open X formats
  • Annual license, non-exclusive
Get Started

Exclusive Access

base rate
Time-limited exclusivity on any dataset — showcase or custom. Gain a competitive advantage with data your competitors can't access.
  • $1,500/hr for showcase exclusivity
  • $3,000/hr for custom exclusivity
  • Time-limited exclusive license
  • Priority capture scheduling
Contact Sales
Built for the Platforms You Use
DreamVu data integrates natively with the leading robotics AI platforms and training pipelines.

Isaac Sim-native USD scenes, GR00T training pipeline integration, Isaac Lab compatibility, and Omniverse support

Open teaser dataset in LeRobot RLDS format — discoverable by the global research community

Full Open X-Embodiment compatibility — seamless integration with existing VLA training pipelines

Built for the Teams Building Physical AI

Humanoid Robotics

Foundation model teams at NVIDIA, Figure AI, 1X Technologies, and Agility Robotics building VLA models that need diverse, high-quality 3D training data.

Frontier AI Labs

Embodiment research programs at Google DeepMind, Meta FAIR, and OpenAI pushing the boundaries of spatial intelligence and world models.

Retail & Logistics

Large retailers and logistics operators — Amazon Robotics, Walmart, Ocado, DHL — investing in automation and needing environment-specific training data.

From Breakthrough Research to
Physical AI Infrastructure
DreamVu began with breakthrough research in computational imaging at IIIT Hyderabad — a new optical design for capturing 360° stereoscopic video in a single shot, published at CVPR. Eight years of production deployment later, we're now building the data infrastructure the humanoid robotics industry needs.
SR

Sashi Reddi

Chairman

Managing Partner at SRI Capital. Founder & former CEO of AppLabs (acquired by CSC). PhD Wharton, MS NYU, BTech IIT Delhi.

RA

Rajat Aggarwal

Chief Executive Officer

BS & MS in computer vision from IIIT Hyderabad. His CVPR'16 paper on computational cameras became the seed for DreamVu.

AN

Dr. Anoop Namboodiri

Chief Science Officer

Professor at IIIT Hyderabad. 75+ published papers. Built systems currently deployed at massive scale.

PS

Parikshit Sakurikar

VP Imaging & AI

PhD in Computational Photography from IIT Hyderabad. Eight years focused on ML, high-performance computing for CV.

SRI Capital
Ben Franklin Technology Partners
Broad Street Angels
Philadelphia, PA
US Headquarters
Hyderabad, India
R&D Center

Ready to Train Physical AI
That Actually Works?

See how DreamVu's omnidirectional 3D data can accelerate your world model and spatial AI programs.

Request a Demo Talk to Us