This example showcases how to use Rerun with gradio to generate an annotation app. It consists of two different modes both of which leverage Segment Anything 2.
The first mode focuses on tracking an object in a monocular video stream. In addition to segmentation masks, it generates real-time depth maps and point clouds to provide full 3D spatial context, enabling users to visualize, inspect, and annotate the tracked object directly in three-dimensional space.
The second mode uses a multiview RGB-D video dataset. By obtaining segmentation masks from two synchronized and calibrated RGB-D views, the app triangulates these 2D masks to reconstruct a precise 3D mask of the chosen object. It then propagates this 3D mask across all camera views and through each frame of the videos, resulting in a fully tracked 3D object trajectory over time.
This is an external example. Check the repository for more information on how to run the code.
TLDR: make sure you have the Pixi package manager installed and run
git clone https://github.com/rerun-io/annotation-example
cd annotation-example
pixi run app
this will run the single view (monocular) app
pixi run multiview-app
will run the multiview rgb-d app