Small but important quality of life update by @rerun.io: Customizable frustum colors. Finally we can distinguish estimates and ground truth in the same 3D view ;)
Small but important quality of life update by @rerun.io: Customizable frustum colors. Finally we can distinguish estimates and ground truth in the same 3D view ;)
Have a sim vehicle in jolt-rust working with stacked hinges for wheel rotation and steering on two of four wheels, no suspension, visualized in @rerun.io:
β¨ Massive Pipeline Refactor β One Framework for Egoβ―+β―Exo Datasets, Visualized with @rerun.io π
After a refactoring, my entire egocentric/exocentric pipeline is now modular. One codebase handles different sensor layouts and outputs a unified, multimodal timeseries file that you can open in Rerun.
MVP of Multiview Video β Camera parameters + 3D keypoints. Visualized with @rerun.io
Trying to wrap my head around fwd/bwd kinematics for imitation learning, so I built a fullyβdifferentiable kinematic hand skeleton in JAX and visualized it with @rerun.io new callback system in a Jupyter Notebook. This shows each joint angle and how it impacts the kinematic skeleton.
Exciting news for egui: there is a draft branch for switching the text handling to Parley, which will bring support for color emojis, right-to-left text, access to system fonts, and much more! github.com/emilk/egui/p...
@rerun.io v0.23 is finally out! π Iβve extended my @gradio-hf.bsky.social annotation pipeline to support multiview videos using the callback system introduced in 0.23.
Visualized with @rerun.io, Iβve integrated videoβbased depth estimation into my robotβtraining pipeline to make data collection as accessible as possibleβwithout requiring specialized hardware.
I extended my previous @rerun.io and @gradio-hf.bsky.social annotation pipeline for multiple views. You can see how powerful this is when using Meta's Segment Anything and multi-view geometry. Only annotating 2 views, I can triangulate the other 6 views and get masks extremely quickly!
Hereβs a sneak peek using @rerun.io and @gradio-hf.bsky.social for data annotation. It uses Video Depth Anything and Segment Anything 2 under the hood to generate segmentation masks and depth maps/point clouds. More to share next week.
Using @rerun.io , I established a baseline from the HoCAP dataset and conducted a qualitative comparison among the ground-truth calibrated cameras, Dust3r, and VGGTβall within rerun. The improvements are evident in both the camera parameters and the multi-view depth map/point cloud.
Rerun keeps growing! Iβm so happy we started this thing, and that weβre building in the open, in Rust π€
11/ Want early access or excited about building the future of Physical AI? Reach out here π 5li7zhj98k8.typeform.com/to/a5XDpBkZ or check out open roles: rerun.io/careers. Let's build!
10/ It comes with visualization built-in for fast data observability over both online and offline systems. The query engine enables you to combine vector search and full dataframe queries, over both raw logs and structured datasets, to support robotics-aware data science and dataset curation.
9/ We are now building a new database and cloud data platform for Physical AI. The database is built around the same data model as the open-source project.
8/ This data model is core to our popular open source framework for logging and visualizing multimodal data, which companies like Meta, Google, @hf.co's LeRobot, and Unitree Robotics have adopted it in their own open-source projects.
7/ We spent the first 2 years of the company iterating on a data model for Physical AI that works for both messy online logs and efficient columnar storage of offline pipeline data.
6/ Rerun is building a unified, multimodal data stackβone data model and platform supporting both online and offline workflows seamlessly, retaining semantic richness throughout.
5/ Until now, no single tool has spanned the entire stack from online data capture, to offline dataset management.
4/ Hand-written online code is being replaced with ML, requiring advanced offline pipelines to collect, manage, label, and curate massive datasets.
3/ Physical AI systems rely on two key components:
Online systems: Run live on robots, processing data and interacting with the real world in real-time;
Offline systems: Run in data centers, analyzing data and improving online systems through training and simulations.
2/ Physical AIβrobotics, drones, autonomous vehiclesβis rapidly evolving, powered by advances in machine learning. But today's data stacks aren't built for this new era.
1/ We just raised $17M to build the multimodal data stack for Physical AI! π
Lead: pointnine.com
With: costanoa.vc, Sunflower Capital,
@seedcamp.com
Angels including: @rauchg.blue, Eric Jang, Oliver Cameron, @wesmckinney.com , Nicolas Dessaigne, Arnav Bimbhet
Thesis: rerun.io/blog/physica...
More progress towards building a straightforward method to collect first-person (ego) and third-person (exo) data for robotic training in @rerun.io. Iβve been using the HO-cap dataset to establish a baseline, and here are some updates Iβve made (code at the end)
Finally finished porting mast3r-slam to @rerun.io and adding a @gradio-hf.bsky.social interface. Really cool to see it running on any video I throw at it, I've included the code at the end
I'm working towards an easy method to collect a combined third-person and first-person pose dataset starting Assembly101 from Meta, with near real-time performance via @rerun.io visualization. The end goal is robot imitation learning with Hugging Face LeRobot
Following up on my prompt depth anything post, I'm starting a bit of a miniseries where I'm going through the tutorials of
Lerobot to understand better how I can get a real robot to work on my custom dataset. Using @rerun.io to visualize
code: github.com/rerun-io/pi0...
The blueprint and stream panels support range selection using shift-click, which speeds up bulk operations like adding entities to views.