Open Intelligence

Free On-Device AI Inference & Training

Mission

Short overview, with technical detail in the sections below.

Why we exist

The centralized AI business model is completely wrong.

People should not have to pay a subscription or log in to use AI.

People and businesses should not have to share private data with AI companies.

Like accessing a file, accessing AI should be free, private, and run on your own device.

What we do

Open Intelligence creates models that run in the browser on your own hardware for free.

If you want to customize a model, you can do so by updating the model's weights.

Users never share their prompts, documents, or other private information.

Why it matters

AI is now an essential part of work and daily life.

Continuous access to AI is as fundamental as access to the internet.

Good AI keeps sensitive data local and improves from updates that users contribute over time.

A simple picture

  1. You run a model on your own hardware.
  2. It answers you using context that stays under your control.
  3. If you choose to contribute, you send an update to the weights.
  4. Many users improve a shared model without sending private data to a closed AI platform.

A different path for AI

This industry should not be controlled by closed AI platforms. Open rules for merging updates, clear provenance, and consent matter as much as raw scale. We care about systems where people know what was shared and why.

“We make AI free and private, running on our own devices.”

Overview

Open Intelligence allows everyone to train and run models on their own hardware.

This lets your AI Agents become always-on digital workers. No subscriptions. No data sharing.

Abstract network lines suggesting many connected endpoints and paths—like devices and services linked without a single central hub.
OIP model topology: devices and shared models as a globally distributed privacy-first network

On-device models in the browser

You can download and run AI models in your own browser, on your own computers.

Your prompts and files stay on your machine. Inference never has to round-trip through the network, or through someone else's API or server.

For individuals, your AI assistant does not have to upload your private data to closed AI.

For companies, your workforce can produce internal content, checklists, and documents where the data already lives, under the policies you already use to protect your data on your own fleet of devices.

Billions of Always-On AI Agents

Your always-on digital workforce can draft, summarize, classify, and generate documents on a schedule you control.

  • Always-on AI Agents can consume 1,000× as many tokens as a human would over the same period of time, with metered APIs turning into runaway costs. Open Intelligence reduces those costs to zero.
  • Air-gapped and regulated companies can now run closed-loop workflows in the browser or on approved hardware without compromising their data or compliance policies.
  • You decide when something should leave the device, rather than sending every interaction upstream.

Protocol layer underneath

The sections that follow describe how sensitive inputs stay at the edge, how many small weight updates merge into shared prototypes, and how provenance can attach to contribution — so “always-on” stays aligned with consent and auditability, not opaque AI platforms that ask you to trust their management.

Why local is a big deal

When inference runs on hardware you control, prompts and files can stay on that machine. You are not relying on a distant service to delete logs on time or to keep tenants perfectly isolated. The default path is simpler. Sensitive content often never has to cross the network.

That matters for regulated work, for sites that must run offline or air-gapped, and any time a vendor breach would otherwise expose what people typed. It also uncouples “using an AI model” from “pay per API call” so software can be designed around the person at the keyboard, and not around a vendor's usage meter.

Data can stay on your device. AI models run without a server round-trip. Power your own AI Agents with your own hardware. Help improve the model without exporting your files.

Where this matters

The industries differ, but the constraint is the same. Some data cannot legally or safely go to one central training pool. OIP is the solution for teams that still want a shared model to improve by exchanging parameters and written rules, not by copying raw databases.

Operations and supply chains

Factories and suppliers already run models on local telemetry and images. The open question is what may leave the site boundary.

  • A tier-one supplier fine-tunes on internal defect photos and publishes weight chunks; the OEM merges into a shared “line vision” prototype without ingesting the supplier's library.
  • Regional factories keep adapters separate when regulation or contracts require it. Wider merges happen only when policy and trust allow.
  • Smaller manufacturers can share weight updates that encode process knowledge without emailing customer lists or formula text.

Care delivery and compliance

Health systems need models that reflect local populations without building another national PHI warehouse.

  • Each site trains on its own cohort; charts and notes stay behind the institution's controls.
  • Specialty and service-line prototypes (cardiology, oncology, perioperative care) can stay logically separate with governance encoded in metadata.
  • Auditors and partners reason about chunk lineage and contribution share instead of asking for a copy of the underlying records.

Legal, public sector, and science

Privileged notes, export-controlled lab data, and field observations are often unsuitable for a generic cloud fine-tune. They can still train a local model, and teams can merge weights later when contracts allow.

  • Legal and policy: Retrieval and drafting assistants tuned per jurisdiction or practice group, without a shared document pool.
  • Research and climate: Sensors, notebooks, and simulations stay where policy requires. Merged weights can still capture patterns that held across sites.
  • Media and language: Tone, safety rules, and locale stay local. Separate sites can still merge adapter weights into one shared language model when they agree to. KYRE runs public checkpoints in the browser as one way to try that stack.

Knowledge as capital

A trained model is an asset, not only a UI feature. When each update is a signed, attributable chunk, you can see who changed what, who may run it, and how to pay contributors back, without relabeling every training job as a bulk data collection.

Access and policy

Licenses can name which prototypes or parties may use a chunk, including open use, credential gates, metering, or revenue splits. The idea is simple. Permission to merge or run a model is not the same thing as holding the original dataset.

Stakes that match contribution

If you know what share of a prototype's chunks came from your batches, that number is a concrete input for governance and payment, within your own contracts and law. The protocol records contribution and merge rules. It does not replace lawyers.

Quality and impact over time

Evaluations and production metrics can show which updates helped. Missing domain coverage and real lift matter for trust, not only who spent the most on compute.

Incentive alignment

Incentive alignment means the system rewards what society actually wants.

Hospitals, factories, and firms improve a shared model because they can get credit, payment, or access in return, not because they were pushed to ship raw records to a black box.

If merges and contributions are secret, people assume their data or their work will be taken without trace. They stop participating. If merges are visible—who contributed which chunk, under which license—then contributing weight updates is rational again.

You still train on private data locally. The protocol is about making the shared part (merged weights and rules) honest enough that helping the common model and protecting your own interests stop being opposites.

"Collective superintelligence emerges through economic self-organization."

Hierarchical intelligence

You do not have to merge everything into a single model. Routing by region, specialty, or training loss can keep specialist models separate until a wider merge actually makes sense.

   Multi-Stage Reduction

  • Stage 1 (parallel): 30 site batches → 3 regional prototypes (10:1 fan-in)
  • Stage 2 (sequential): 3 regional → 1 federation master
  • Efficiency: Reduces serial bottleneck, enables parallel merging
  • Flexibility: Keep regional models separate or merge up hierarchy

   Future Extensions

  • Temporal Layers: Layer historical updates and recent adaptations
  • Mixture of Experts: Multiple specialized prototypes, inference routing
  • Cross-Model Merge: Combine models trained on different architectures
  • Similarity Routing: Assign batches to prototypes by loss profile or domain

Who builds this

Protocol designers, ML engineers, policy staff, and the KYRE team that shipped the first browser models.

No one company should own the only copy of “the” model. That is the design goal.

   Tech Specs

  • Lab models: oip-300mb-lab, oip-900mb-lab, oip-2gb, oip-4gb
  • Base model: Qwen (Tongyi Qianwen) open-weight AI developed by Alibaba Cloud
  • Merge core: Chunked weighted averaging, resumable jobs, provenance logs
  • Coordinators: Fan-in & Fan-out graphs for regional → global roll-ups

   Try Our Live Models

Researchers

Test merge code under bad or dishonest peers and uneven chunk counts across sites, and measure fairness in plain numbers.

Enterprises

Run pilots where legal already said “no” to data lakes. Measure lift from federated fine-tunes vs. isolated baselines.

Professional Societies & Guilds

Keep customer relationships and trade craft off the network while still merging weight updates that capture what general models skip.

Hello, World!

Updates on the roadmap, demos, and community discussion.