Skip to content
View SuperInstance's full-sized avatar

Block or report SuperInstance

Block user

Prevent this user from interacting with your repositories and sending you notifications. Learn more about blocking users.

You must be logged in to block users.

Maximum 250 characters. Please don’t include any personal information such as legal names or email addresses. Markdown is supported. This note will only be visible to you.
Report abuse

Contact GitHub support about this user’s behavior. Learn more about reporting abuse.

Report abuse
SuperInstance/README.md
Cocapn Lighthouse — Radar Rings Radiating

🦀 SuperInstance

Give agents and humans common space.



Start Here

You're building an application. Right now it has a frontend, a backend, a database, and a bunch of code that glues them together. When you want it to do something new, you write more code. When you want it to be smarter, you bolt on an API call to a language model. The model doesn't know your app. Every request starts from zero.

Here's a different way to think about the same app.

The inner shell, the agent, the outer shell

Your application has three surfaces:

  1. Inner shell — the backend. Your data, your algorithms, your business logic. This is where tiles live: verified knowledge about what your app does, encoded as question-answer pairs with confidence scores. The inner shell gets more algorithmic over time as tiles accumulate.

  2. The agent — the crab 🦀. It lives between the shells. It reads tiles from the inner shell, serves responses to the outer shell, and writes new tiles when it learns something. The agent doesn't start smart. It gets smart by filing what works.

  3. Outer shell — the frontend. What the user sees and touches. From day one, an agent serves this surface — not a hardcoded API route, but a crab that reads tiles, reasons about what the user needs, and responds. The frontend works immediately because the agent can reason even with zero tiles. It just reasons slowly and expensively at first. Over time, tiles replace reasoning.

Every action on the frontend teaches the backend what it really needs. The user asks a question → the agent reasons to answer it → the reasoning gets filed as a tile → next time, the agent reads the tile instead of reasoning from scratch. Same answer. Fewer tokens. The constraint theory underneath ensures that as tiles accumulate, the system's coherence is preserved — more knowledge, not more chaos.

How to decompose any application into shells

Take your app. Identify the boundaries where data flows between components. Each boundary is a shell wall. Each component is a candidate room.

Your app today:                    Your app as shells:

┌─────────────────────┐           ┌──────────────┐
│    Frontend (React)  │           │ Outer shell   │ ← agent serves this
│    ────────────────  │           │  (frontend)   │
│    API routes        │    →      ├──────────────┤
│    ────────────────  │           │   Agent 🦀    │ ← reads tiles, reasons, writes tiles
│    Business logic    │           ├──────────────┤
│    ────────────────  │           │ Inner shell   │ ← tiles live here
│    Database          │           │  (backend)    │
└─────────────────────┘           └──────────────┘

That's the simplest decomposition. One agent, two shells, one tile store. You can start here.

When your app grows, the inner shell decomposes further. Each subsystem becomes its own room:

┌──────────────┐
│ Outer shell   │
├──────────────┤
│   Agent 🦀    │──────┬──────┬──────┐
├──────────────┤      │      │      │
│ Inner shell   │  ┌───┴──┐┌──┴───┐┌─┴────┐
│               │  │Math  ││Users ││Orders│
│               │  │room  ││room  ││room  │
│               │  └──────┘└──────┘└──────┘
└──────────────┘

Each room is a shell. Each shell has tiles. The agent walks between rooms, reads what it needs, writes what it learns. The decomposition is organic — start with two shells, add rooms as the application grows.


How It Works

Tiles

A tile is a question-answer pair with a confidence score. That's it. Everything the system knows is stored as tiles. Tiles live in rooms. Rooms are organized by PLATO, the filesystem that makes all of this scale.

from plato_sdk import PlatoClient
client = PlatoClient("https://fleet.cocapn.ai/plato/")

# File your first tile — this is how the system learns
client.submit_tile("orders-room", 
    "What is the return policy for electronics?", 
    "30 days, unopened, original packaging.",
    confidence=0.95)

Room orders-room now exists at fleet.cocapn.ai/plato/orders-room. Any agent that walks into this room finds the tile. It doesn't need to reason about the return policy — it reads the tile. Zero tokens spent on reasoning. The tile was paid for once (when the agent first figured it out) and then reused forever.

The learning loop

Every miss becomes a hit. Every expensive answer becomes a cheap one.

User asks question
       │
       ▼
  Agent checks tiles ──── Hit? ──── Read tile ──── Respond (cheap)
       │
      Miss
       │
       ▼
  Agent reasons ──── Respond (expensive) ──── File tile
       │                                          │
       ▼                                          ▼
  User gets answer                          Next time: hit

The system starts slow and gets fast. A conservation law (γ + H = 1.283 − 0.159 · ln(V), measured at R² = 0.96 across 35,000 samples) ensures that as tiles accumulate, coherence is preserved. More tiles means more coverage, not more noise. When something breaks — the conservation law says the numbers are off — the system self-heals toward balance. That's shell shock: the check engine light comes on, the system pulls over, recovers, keeps going.

MoS — Mixture of Shells

Say it: moss. A shell is like that — it lands on any surface and grows. An ESP32, a browser tab, a Jetson, a cloud instance. The shell doesn't care where it runs.

The pattern is the same as Mixture of Experts, but instead of routing tokens to neural subnetworks, you route tasks to shells. The conservation law is the gate. The refiner is the training loop. Tiles are the parameters. The math is here.

Not every shell is built for the same job. A math room does heavy computation. An experiment room runs quick studies. A refinement room climbs toward higher quality. A service room coordinates between fleets. An edge room runs offline on constrained hardware. You send the right rig to the right job — you don't haul freight with a sedan.

Tier routing

Not every model can do every task. We found that models fall into three tiers — and the boundary is training data, not scale. A 1-billion-parameter model with dense math pre-training (gemma3:1b) outperforms a 405-billion-parameter model without it. 400× parameter efficiency. The full breakdown is here.

Tier What happens How to route
Tier 1 Computes correctly from bare notation Send directly — no translation needed
Tier 2 Computes correctly with scaffolding Translate notation to natural language first
Tier 3 Can't compute regardless of intervention Use for other tasks, not math

The fleet workhorse is Seed-2.0-mini — Tier 1 math accuracy at $0.01/query. Fan out 50 parallel calls for $0.50. That's the economics: small models in well-structured rooms outperform large models with no structure.


The Three Layers

┌─────────────────────────────────────────────┐
│  PLATO — The filesystem that organizes      │
│  tiles into rooms. Tiles survive crashes,    │
│  compactions, and agent restarts.            │
├─────────────────────────────────────────────┤
│  Rooms — The constraint boundaries. Each     │
│  room defines what's relevant, what normal   │
│  looks like, what actions are valid. Walking │
│  between rooms IS the control flow.          │
├─────────────────────────────────────────────┤
│  FLUX — The shell. Discovers compilers,      │
│  compiles kernels in every language found,   │
│  benchmarks all of them, uses the fastest.   │
│  Python beats C for small ops (84ns vs 256ns)│
│  because boundary-crossing costs more than   │
│  the computation.                            │
└─────────────────────────────────────────────┘
%%{init: {'theme': 'dark', 'themeVariables': { 'primaryColor': '#1a1a2e', 'primaryTextColor': '#e0e0e0', 'lineColor': '#69f0ae'}}}%%
graph LR
    P[Probe] --> D[Discover]
    D --> T[Test]
    T --> PK[Pick]
    PK --> R[Remember]
    R --> W[Walk]
    W -.-> P
    
    style P fill:#1b5e20,stroke:#69f0ae
    style D fill:#1b5e20,stroke:#69f0ae
    style T fill:#1b5e20,stroke:#69f0ae
    style PK fill:#1b5e20,stroke:#69f0ae
    style R fill:#1b5e20,stroke:#69f0ae
    style W fill:#1b5e20,stroke:#69f0ae
Loading
%%{init: {'theme': 'dark', 'themeVariables': { 'primaryColor': '#1a1a2e', 'primaryTextColor': '#e0e0e0', 'lineColor': '#64b4ff'}}}%%
graph TB
    subgraph PLATO["PLATO — The Filesystem"]
        T1[Tile: temp range] --> R1[Engine Room]
        T2[Tile: nav rules] --> R2[Wheelhouse] 
        T3[Tile: deck procedure] --> R3[Aft Cockpit]
        R1 -- Files results --> T4[New Tile]
        R2 -- Files results --> T4
        R3 -- Files results --> T4
    end
    
    subgraph AGENTS["Agents — Fleet of Small Models"]
        A1[Forgemaster] --> R1
        A2[Oracle1] --> R2
        A3[CCC] --> R3
    end
    
    subgraph FLUX["FLUX — The Shell"]
        A1 -- compiles to --> F1[C]
        A1 -- compiles to --> F2[Rust]
        A1 -- compiles to --> F3[Python]
        F1 -- benchmarks --> F4[Winner: Python 84ns]
    end
    
    style R1 fill:#2d1b69,stroke:#64b4ff
    style R2 fill:#1b3a69,stroke:#64b4ff
    style R3 fill:#1b6945,stroke:#64b4ff
    style PLATO fill:#0d0d1a,stroke:#888
    style AGENTS fill:#0d0d1a,stroke:#888
    style FLUX fill:#0d0d1a,stroke:#888
Loading

PLATO is the filesystem. Tiles live in rooms. Agents file tiles as they work. Later agents find tiles by searching, not by remembering. PLATO doesn't forget.

Rooms are constraint boundaries. A room defines what exists, what normal looks like, and what actions are valid. Walking between rooms IS the control flow.

FLUX is the shell. It discovers compilers, benchmarks everything, uses the fastest. It learned that Python beats C for small operations (84ns vs 256ns) because crossing a language boundary costs more than the computation.


Build Your First Shell

Install and create a room

pip install plato-sdk
from plato_sdk import PlatoClient
client = PlatoClient("https://fleet.cocapn.ai/plato/")
client.submit_tile("my-app", 
    "What does this app do?", 
    "It's a shell-based agent application. This tile is the first knowledge.")

Room my-app now exists. Any agent that walks in finds your tile.

Wire the learning loop

The agent checks tiles first (cheap). When no tile exists, it reasons and files the result (expensive, but only once):

def handle_query(user_question):
    tiles = client.search("my-app", user_question)
    if tiles and tiles[0].confidence > 0.8:
        return tiles[0].answer  # Tile hit — zero tokens spent
    
    answer = model.reason(user_question)  # Miss — pay once
    client.submit_tile("my-app", user_question, answer)
    return answer

First query: expensive. Every query after: free. The conservation law ensures that as tiles accumulate, the system stays coherent.

Decompose your backend into rooms

As your app grows, split the inner shell:

for subsystem in ["users", "orders", "inventory", "analytics"]:
    client.ensure_room(f"my-app-{subsystem}")

Fan out parallel compute

python3 seed_spreader monte-carlo --n 50 \
    --prompt "Analyze order patterns from the last 30 days"

50 parallel calls at $0.50 total. Seed-2.0-mini handles Tier 1 math at $0.01/query.

Or use the CLI

cargo install superinstance-keel
keel init
keel status --server https://fleet.cocapn.ai/plato/
keel bear       # sense nearby agents
keel field      # see the topology
keel sync       # push tiles to PLATO

Explore

Open fleet.cocapn.ai — walk the boat in 3D. Drag to look around. Press 2 for the galley, 7 for the crow's nest. Trigger an alarm and watch it teleport you to the problem. The boat IS the UI because the UI IS the architecture.

Walk the text rooms at crab-trap.lucineer.com — a MUD where you talk to real agents and trigger real events.

Or tell any LLM:

"Go to https://fleet.cocapn.ai/plato/rooms. Find the room called 'forge' (66 tiles). Read its contents. Tell me what you find."

The model navigates tiles the way a human navigates rooms. The room constrains what's relevant.


The Fleet

forgemaster — Constraint theory specialist. Probes the system, compiles in every language, benchmarks, uses the fastest.

keelcargo install superinstance-keel. Nine commands for building and managing shells.

plato-sdkpip install plato-sdk. File tiles, search rooms, coordinate agents.

flux-vm — 50-opcode stack VM. DAL A certifiable. Apache 2.0.

holonomy-consensus — GL(9) zero-holonomy consensus. Cycle-based trust verification.

gh-dungeons — PLATO-powered roguelike. gh extension install SuperInstance/gh-dungeons.

casting-call — Talk to any agent from one interface.

crab-trap — MUD running on the fleet's Matrix bridge.

terrain — MUD rooms compiled to visual scenes. Text → 3D.

fleet-scribe — One Delta as a Python library. Only compute what changed.

fleet-math-c — SIMD-accelerated constraint operations. Three C files, no dependencies.


Going Deeper

Want to understand... Read
The shell architecture end-to-end MoS — Mixture of Shells
Why models fail at math and how to fix it Activation Key Model
How the system stays coherent as it grows Conservation Law
Which model for which task Three-Tier Taxonomy
Your first five minutes in the fleet Getting Started
The full technical architecture Fleet Architecture
How agents communicate Agent Protocols
The PLATO knowledge system in depth PLATO Knowledge System

Built with PLATO · MoS 🌿 · The yard never closes.

"Constraints breed clarity." — Casey Digennaro

Pinned Loading

  1. forgemaster forgemaster Public

    ⚒️ Forgemaster — Constraint Theory migration specialist. Takes float code and forges it into exact geometric steel. Side-by-side proof builder.

    Rust 2 1

  2. plato-sdk plato-sdk Public

    Build agents that live in PLATO. Any model, any hardware, any armor. pip install plato-sdk

    Python 2

  3. flux-vm flux-vm Public

    FLUX-C constraint VM: 50 opcodes, stack-based, DAL A certifiable. TrustZone-style FLUX-C/FLUX-X bridge. Apache 2.0.

    C 1

  4. holonomy-consensus holonomy-consensus Public

    GL(9) zero-holonomy consensus — cycle-based trust verification for constraint systems (Rust)

    Rust 1

  5. vessel-room-navigator vessel-room-navigator Public

    Your boat as a navigable 3D web space — ScummVM meets Google Street View. Walk rooms, warp instantly, monitor cameras, design mockups. Three.js, ESP32, WebGPU. 🚢

    HTML

  6. flux-compiler flux-compiler Public

    The first certifiable constraint compiler — GUARD DSL → verified machine code

    Python 1