Oceanir
BLOGABOUTHOW IT WORKSPRICING

[ ABOUT ]

Visual Intelligence Research

Oceanir Incorporated is a research group specializing in advanced visual positioning systems and geolocation intelligence.

Learn more about us

[ QUICK LINKS ]

How It WorksPricingNewsroomContact

[ CAREERS ]

Join Our Team

We're looking for exceptional talent to help build the future of visual intelligence.

View Open Roles

contact@oceanir.ai

More
PricingBlogAbout
View pricing
[ PLATFORM ]

Geolocation Intelligence

Advanced visual positioning technology for intelligence professionals, researchers, and organizations requiring precision location analysis.

Explore the platform
[ USE CASES ]
01Insurance02Investigative Journalism03Security Analysis04Enterprise05Law Enforcement
[ GET STARTED ]

Try It Free

Start analyzing images with our visual intelligence platform. No credit card required.

Launch App

Trusted by researchers worldwide

More
PricingBlogAbout
View pricing
⌘K
↑↓Navigate↵SelectescClose
[ 12 suggestions ]
BLOGABOUTHOW IT WORKSPRICINGGET STARTED
Archive IndexBRIEF_OCU_001
PRODUCTJan 29, 20268 min read
Status: Released

Introducing Oculus, Search, and Memory

A standalone hybrid-reasoning VLM, semantic search across your analyses, and persistent memory that learns.

Oceanir Team

Author

Oceanir Team

Product Engineering

Oculus - Hybrid Reasoning VLMBRIEF_OCU_001

Contents

01Three Releases02Oculus: The VLM03Architecture04Oculus Capabilities05oceanir-search06Oceanir-Memory07Getting Started

Document Info

Classification:
Public
Department:
Product Engineering
Model:
Oculus-0.1
Architecture:
Oceanir-Oculus OO1
[ Section 00 - Introduction ]

Today we're releasing three things: Oculus, a standalone hybrid-reasoning vision-language model that outperforms systems 10x its size; oceanir-search, semantic search across your analysis history; and Oceanir-Memory, persistent context that learns from every query.

For months, we've been working on a fundamental question: how do you build a vision model that's small enough to run anywhere, smart enough to reason through complex visual tasks, and integrated enough to remember what it's learned?

The answer is Oculus—and the ecosystem we're building around it.

[ Section 01 - Three Releases ]

What We're Shipping

01 — Model

Oculus-0.1

Hybrid-reasoning VLM built on the OO1 Architecture. Small model, large-model performance.

02 — Feature

oceanir-search

Semantic search across your entire analysis history using natural language.

03 — Feature

Oceanir-Memory

Persistent memory that learns from every analysis. Cross-session context.

[ Section 02 - Oculus: The VLM ]

A Small Model That Thinks Big

Oculus is a hybrid-reasoning vision-language model built on the Oceanir-Oculus OO1 Architecture. It's designed to outperform systems 10x larger on visual reasoning and perception tasks, with optimization for running on commodity GPUs or edge devices.

The key insight: instead of scaling parameters, we scale reasoning. Oculus uses structured thinking traces to work through complex visual tasks step-by-step, and perceptive tool calling to zoom and crop on relevant regions automatically.

"The best vision model isn't the largest one—it's the one that knows where to look and how to think about what it sees."

- OO1 Architecture Paper, 2025

Python
from oceanir import Oculus

model = Oculus.from_pretrained("OceanirAI/Oculus-0.1")

# Basic VQA
answer = model.ask("image.jpg", "What is this?")

# With reasoning traces
answer = model.ask("scene.jpg", "Count the people", think=True)

# With focus/zoom for fine details
answer = model.ask("document.jpg", "Read the fine print", focus=True)

# Structured JSON output
result = model.generate(image, prompt="Describe objects", mode="json")
[ Section 03 - Architecture ]

How Oculus Works

Oculus introduces two core mechanisms that enable small-model performance at large-model quality:

01

Thinking Traces

When think=True, Oculus generates structured reasoning wrapped in <think>...</think> tags. This multi-step reasoning dramatically improves performance on counting, spatial relationships, and ambiguous scenes.

02

Perceptive Focus

When focus=True, Oculus automatically identifies regions of interest and zooms/crops to examine them in detail. This enables fine-grained perception—reading small text, detecting tiny objects, analyzing dense scenes—without increasing model size.

Output Modes
Text
model.ask(image, question)
Natural language answer
Reasoning
model.ask(..., think=True)
Answer with <think> trace
JSON
model.generate(..., mode="json")
Structured JSON
Boxes
model.detect(image)
Bounding boxes + labels
Polygons
model.segment(image)
Segmentation masks
OCR
model.ocr(image)
Extracted text + locations
UI
model.detect_ui(image)
UI elements + types
[ Section 04 - Oculus Capabilities ]

What Oculus Can Do

Oculus brings six core capabilities to visual understanding:

Reasoning via Thinking Traces

01

Structured reasoning traces for multi-step decisions. Improves small-object understanding and ambiguous spatial tasks.

Perceptive Focus (Zoom & Crop)

02

Automatic zooming and cropping to focus on relevant regions. Dramatically improves fine-grained perception.

Structured Outputs

03

Reliable JSON generation for predictable downstream integration. Points, boxes, polygons, and more.

Complex OCR

04

Enhanced text recognition across cluttered, low-resolution, and distorted regions. Works on documents, diagrams, and dense scenes.

Desktop & UI Understanding

05

Better performance on UI navigation and everyday workflows. Detects UI elements with types and bounding boxes.

Edge-Ready Architecture

06

Optimized for commodity GPUs and edge devices. Small model footprint with large-model performance.

[ Section 05 - oceanir-search ]

Find What You've Seen

Traditional search requires exact matches—dates, filenames, coordinates. oceanir-search works differently. It understands what you're describing and retrieves matching analyses using semantic similarity.

Search your analysis history using natural language:

  • "That street with the blue and white tiles in Portugal"
  • "The intersection near that pink Art Deco building"
  • "Photos from last week that showed rooftop terraces"
  • "All analyses where we detected Japanese text"

oceanir-search indexes every analysis automatically—extracted features, detected objects, OCR text, reasoning traces—and makes it all searchable through a single query interface.

[ Section 06 - Oceanir-Memory ]

Context That Persists

Every geolocation tool we've used has the same limitation: no memory. Analyze an image of a street in Barcelona, and later when you see a similar street, the system processes it like it's the first time. This is wasteful.

Oceanir-Memory changes this. Every analysis you run contributes to a persistent knowledge store that improves future queries:

01Instant Recognition

Previously analyzed locations are recognized immediately. No redundant processing.

02Pattern Learning

Architectural patterns, signage styles, and regional signatures are indexed automatically.

03Cross-Session Context

Insights from one session inform the next. Your knowledge graph grows with every analysis.

04Privacy-First Storage

Memory is encrypted per-user. You can delete everything instantly—no backups, no holds.

[ Section 07 - Getting Started ]

Try It Now

Oculus-0.1 is available now for research and non-commercial use. Install the Python SDK and start experimenting:

Terminal
pip install oceanir

oceanir-search and Oceanir-Memory are available to all Pro and Enterprise users on the Oceanir platform. Your analysis history is automatically indexed—just start searching.

Action Required

Experience the Full Stack

Oculus, oceanir-search, and Oceanir-Memory are available now. Run Oculus locally or use the full platform with search and memory.

Launch OceanirView Plans
End of DocumentBRIEF_OCU_001

Related Documents

Jan 4, 2026product

Orca 1.3: How We Expanded Internationally

Dec 28, 2025product

Hyper-Local Miami

Dec 20, 2025updates

Update 1.53: New Pricing, New Settings

Return to Archive Index
Ready when you are

Ready when you are.

Let's Talk

We locate images with
precision, privacy, and purpose.

Company

  • About
  • Brand
  • Careers
  • Contact

Product

  • Analyze Image
  • How It Works
  • Pricing
  • Pro

Resources

  • Blog
  • Docs
  • Privacy
  • Terms

Social

  • X
  • GitHub
  • HuggingFace
  • TikTok
  • Discord
Oceanir
TermsPrivacyCookies|Copyright © 2026 Oceanir