A Nimbo Plugin Concept for Community-Led Heritage Conservation
Combining satellite time-series, feedforward 4D reconstruction, neural radiance fields, and Gaussian splatting into a single navigable temporal depth—so communities can see, argue for, and protect the living patterns of their places.
Core Insight
No single technology captures what heritage is. Nimbo sees the landscape changing from orbit, month by month. D4RT reconstructs the living street from a grandmother's phone video. NeRF interpolates between the moments no one recorded. And 3DGS makes all of it navigable on any device, by anyone.
Combined, they produce something none can produce alone: a continuous, multi-scale, community-owned temporal record of place—from the orbital to the ornamental, from decades to instants.
Smartphone footage captured by residents during daily life, ceremonies, seasonal events
Feedforward 4D reconstruction. Encoder → Global Scene Representation → Independent point queries at 200+ FPS
NeRF interpolates temporal gaps. 3DGS accumulates Gaussians for real-time web rendering on consumer hardware
Geo-registered into monthly satellite basemaps. NDVI, radar, and infrared layers reveal landscape-scale change
Navigable 4D model. Community-owned. Exportable as evidence for planning, grants, UNESCO documentation
Architecture
Chronotope operates across three nested spatial-temporal scales, each powered by different technology, unified through a shared coordinate system.
Workflow
Technology Integration
Each technology solves a problem the others cannot. Their combination produces capabilities none possess independently.
4D Reconstruction Engine
What It Does
Feedforward transformer encodes video into Global Scene Representation. Lightweight cross-attention decoder independently queries 3D position of any point in space and time.
Why It's Essential
200+ FPS inference. Handles dynamic scenes. Single unified decoder for depth, point clouds, tracking, and camera parameters. No iterative optimization needed.
Zhang et al., Google DeepMind, 2025
Real-Time Rendering Layer
What It Does
Explicit point-based scene representation using anisotropic Gaussians with learned opacity, color, and covariance. Rasterization-based rendering at 100+ FPS.
Why It's Essential
Web-deployable via WebGL/WebGPU. Consumer hardware compatible. Editable—communities can annotate and modify. Compact storage for archival.
Kerbl et al., 2023; heritage applications at CIPA 2025
Temporal Interpolation Fabric
What It Does
Continuous volumetric representation that maps 5D coordinates (x,y,z,θ,φ) to color and density. Extended here to 6D with temporal dimension.
Why It's Essential
Fills gaps between discrete captures with plausible geometry. Enables smooth temporal navigation. Provides the continuous spatial representation that regularizes 3DGS artifacts.
Mildenhall et al., 2020; Fang et al., ICCV 2025
Orbital-Scale Temporal Canvas
What It Does
Cloud-free monthly satellite basemaps from Sentinel 1/2 via Copernicus. 2.5m resolution with super-resolution. NDVI, infrared, radar layers. API via TMS/WMTS.
Why It's Essential
Only platform providing monthly global updates. Already has QGIS plugin architecture. European sovereignty (GAFA-free). Built for change detection.
Kermap, Rennes, France
The D4RT Advantage
Previous approaches (COLMAP, test-time optimization) require minutes to hours per scene. D4RT's feedforward encoder processes video in a single pass. A community member's 30-second walking video becomes a 4D point cloud before they put their phone away.
Heritage isn't static. Markets pulse, water flows, ceremonies unfold. Unlike MegaSaM or π³ which fail on dynamic content, D4RT's unified decoder tracks all pixels through time—the only method that captures the living dynamics of place.
One decoder for depth, point clouds, camera parameters, and temporal correspondence. No ensemble of task-specific models. The query (u,v,t_src,t_tgt,t_cam) can probe any point in space and time independently—enabling the flexible multi-scale architecture Chronotope requires.
Each D4RT query is decoded independently—no inter-query dependencies. This means the occupancy-grid algorithm can efficiently track all pixels in a video with 5–15× adaptive speedup, making dense 4D reconstruction feasible on edge devices.
Design Philosophy
What Chronotope Resists
What Chronotope Enables