At Gestalt, we believe art is meant to be understood, not just observed. We're building the first accessibility platform with zero marginal cost per venue — connecting physical galleries with AI-powered audio descriptions, contributor-sourced data, and ADA-compliant experiences that require no hardware installation.
We're dedicated to removing the friction between curiosity and understanding. By running CLIP visual recognition directly on visitors' phones, we allow museums to offer instantly accessible, deeply contextual audio descriptions — with no QR codes, no beacons, and no hardware contracts. Accessibility is not a fallback. It's the primary interface.
A world where every physical object can instantly share its origins, meaning, and significance. Where 50,000+ public murals are indexed and accessible. Where blind and low-vision visitors experience art as richly as anyone else.
Every visitor who captures and describes an artwork enriches the index for the next person. CLIP matching and algorithmic trust scoring filter contributions automatically — quality scales without a moderation team. 1,700+ artworks indexed with no licensing fees and no data acquisition budget.
CLIP ViT-B/32 runs on the visitor's phone via ONNX WASM. No server round-trips, no images leaving the device. The entire recognition system has zero marginal cost per venue — a structural advantage no hardware-dependent competitor can match.