Research overview
A research program built around the product, not around a slide deck.
Pic2Nav studies photo geolocation as a real systems problem: messy evidence, route-level validation, retrieval memory, and feedback loops that improve the stack over time.
Current snapshot
Measured from the live system on April 1, 2026
399
stored recognitions in the live system
88.73%
positive rate across feedback-bearing records
99
vectors in the deployed baseline index
14
unique places in the current place-grouped held-out split
Research tracks
Three things we are actually working on.
The research page should describe the current system honestly: what is strong, what is experimental, and where the product and model stack currently meet.
Hybrid inference routing
The product is designed around evidence hierarchy: direct signals first, broader reasoning only when the image actually needs it.
Feedback-driven retrieval memory
Recognitions, confirmations, and corrections are persisted so the system can improve as an operational memory, not just a static model.
Backbone experiments
We test stronger vision backbones against the production CLIP baseline using deterministic place-grouped held-out evaluation.
System program
The stack is organized around evidence, not hype.
Direct evidence
EXIF GPS and visible-address shortcuts
The route checks deterministic signals first so the stack can return precise results without unnecessary model escalation.
Retrieval and priors
NaviSense V3
The ML service combines image embeddings, retrieval memory, and geospatial priors to narrow candidates before route-side validation.
Scene reasoning
Claude and Google Vision
When evidence is partial or messy, OCR, landmark hints, phone numbers, and scene-country reasoning become part of the decision path.
Backbone evaluation
StreetCLIP is the first experiment that clearly moved the retrieval path.
We now have a deterministic place-grouped held-out evaluation from 38 canonical records spanning 14 unique places. That split is still small, but it is much more honest than the earlier tiny diagnostic slice.
The strongest current production path is still hybrid orchestration rather than a solved end-to-end geolocation regressor.
Claude-assisted address resolution currently outperforms direct ML acceptance in user-confirmed feedback-bearing production cases.
The held-out comparison is more honest than the earlier tiny diagnostic slice, but it is still directional rather than benchmark-grade proof.
Production baseline
CLIP ViT-B/32
The current deployed baseline is still the safer production default, but the direct geolocation head remains weak in absolute terms.
Experiment service
StreetCLIP
The first controlled backbone comparison shows a clear gain on the retrieval-driven path, even though the canonical corpus is still small.
Outputs
Publications, datasets, and the product surface stay connected.
Working paper
A systems paper framing Pic2Nav as a deployable hybrid photo geolocation stack.
Read publication notesDatasets and corpora
The project is moving toward stronger canonical evaluation and better place-grouped testing discipline.
Explore datasetsAPI and product surface
The research stack stays close to the actual interface rather than living separately from the product.
View API accessNext step
Test the research stack on real images.
The cleanest way to understand Pic2Nav is still to use it: upload an image, inspect the evidence path, and review how the system handles confidence and failure.
Current stance
Honest about what is solved, and what is not.
The research page now reflects the actual paper and live measurements: strong hybrid orchestration, improving retrieval, and a geolocation head that still needs better data and stronger evaluation.
Pic2Nav is best understood as a deployable geolocation stack for messy real-world photographs, not as a single solved benchmark model.