Resistor Technologies
Metrics Topology Thesis

PROJECT COLOSSUS

Distributed AI Inference Mesh — Gold Road Memory Protocol
"Memory pooling is to AI what mining pools were to crypto."
RESISTOR TECHNOLOGIES • PROJECT COLOSSUS
Status Report — April 20, 2026 • Updated with latest benchmarks + Memory Pooling thesis
Note on Statistics
These specs have not been updated since April 21, 2026 due to our focus on other divisions and affairs. Colossus is now distributed far beyond two local machines. We assure you we are pacing with — and in many cases ahead of — typical technology replication cycle standards. The system is only getting bigger, faster, and stronger. Updated benchmarks will be published when the current build cycle completes.
Key Metrics
235B
Largest Model Run
88 GB
Distributed Memory Pool
2/3
Mesh Nodes Online
4,920+
Gold Memory Facts
The Record — Four Records in One Day
235 billion parameters on two consumer desktops. No data center. No cloud. No subscription.
The model thinks before it speaks: "Okay, the user wants me to say hello..."

The Progression — April 19-20, 2026

RecordModelParamsSizeSpeedResponse
1st Athene-v2 (72B)72B44.5 GB 0.21 tok/s"Hello! It's great"
2nd Llama 4 Scout (109B MoE)109B57.3 GB 0.20 tok/s"Four."
3rd Mixtral 8x22B (141B MoE)141B63.1 GB 0.027 tok/s"Hi there!"
RECORD Qwen3-235B-A22B 235B79.8 GB ~0.14 tok/s "Okay, the user wants me to say hello..."

Also proven: Qwen3-Coder 30B at 7.5 tok/s (interactive speed, daily use)

April 20, 2026 — 141 BILLION PARAMETERS — NEW RECORD — Mixtral 8x22B

Mixtral 8x22B (141B MoE, 63.1 GB Q3_K_M) loaded and generated across the mesh.

141 billion parameters. Two consumer desktops. "Hi there!" — The model spoke. The record stands.
3-Way Memory Distribution:
Vulkan GPU 15.8 GB
Mac RPC (CPU) 47.2 GB
i7 CPU 0.1 GB
MetricValueNotes
ModelMixtral 8x22B (141B MoE)Q3_K_M quantization, distributed across mesh
Total Parameters141,000,000,00057 layers, 8 experts × 22B each
Model Size63.1 GBDistributed across 2 machines, 3 memory pools
Prompt Speed0.20 tok/s31 tokens in 153 seconds
Generation Speed0.027 tok/s3 tokens in 110 seconds (~37 sec/token)
Total Time~4.4 minutesSwap thrashing (47 GB on 40 GB Mac RAM)
Output"Hi there!" CORRECT
StatusPROVEN 141B on consumer hardware

April 20, 2026 — 235 BILLION PARAMETERS — ABSOLUTE RECORD — Qwen3 235B

Qwen3-235B-A22B (235B MoE, 79.8 GB Q2_K) loaded and generated across the mesh using the proprietary Dual-RPC Breakthrough — a 4-way tensor distribution configuration.

235 billion parameters. The model THINKS before answering.
"Okay, the user wants me to say hello..."
DUAL-RPC 4-Way Memory Distribution:
Vulkan GPU 13.1 GB
i7 Local RPC 26.1 GB
Mac RPC 40.4 GB
MetricValueNotes
ModelQwen3-235B-A22B (MoE)Q2_K quantization, 22B active per token
Total Parameters235,000,000,00095 layers, MoE architecture
Model Size79.8 GBDistributed across 4 memory pools on 2 machines
Generation Speed~0.14 tok/s~7 sec/token — 5x faster than 141B dense
Output"Okay, the user wants me to say hello..." REASONING
Key InnovationDual-RPC: Proprietary 4-way tensor distribution — 5x faster than standard configuration
StatusRECORD Largest model ever on this mesh — ABSOLUTE RECORD

Dual-RPC Breakthrough — Architectural Discovery

A proprietary configuration that expands tensor distribution from three memory pools to four, eliminating swap thrashing and delivering a 5x speedup on the same hardware.

Same machines. Same model. Same network. 5x faster — through architecture alone.
Mesh Topology
Pre-Wrecking Crew Deployment & Distribution
Mac Orchestrator
Memory Master • Routing
40 GB Unified RAM
Intel x86_64 • Proprietary Unified Memory Algorithm
Gold Road: 4,920+ facts
Gold Road Memory Server
RPC Tensor Worker
i7 GPU Rig
GPU Workhorse • Vulkan Compute
32 GB DDR4 RAM
16 GB VRAM
Proprietary Unified Memory Algorithm
99 Local AI Models
Gold Road: 2,679 facts
Gold Road Node
RPC Tensor Server
i7 Office
CPU Overflow • Backup
32 GB DDR4 RAM
Proprietary Unified Memory Algorithm
Gold Road: OFFLINE
The Parameter Economy

~1.6 Billion Parameters Per Gigabyte

At Q4_K_M quantization (4 bits per weight), every gigabyte of memory holds approximately 1.6 billion neural network parameters. This ratio is the foundation of the scaling math.

ModelParametersSize (Q4)Params/GB
Qwen3:8B8B4.9 GB1.63B/GB
Qwen3-coder:30B30B17.3 GB1.73B/GB
Athene-v272B44.2 GB1.63B/GB
Llama 4 Scout109B62.8 GB1.74B/GB
Mixtral 8x7B47B24.6 GB1.91B/GB
At Q4_K_M quantization, every gigabyte of memory holds approximately 1.6 billion parameters. The ratio is consistent across model sizes — capacity scales linearly with memory.
The Path to 1 Trillion Parameters

Scaling Roadmap

Phase 1–3 Complete
TierMemory PoolMax Model (Q4)
CURRENT 88 GB ~140B params
NEXT 184 GB ~300B params
Gold Road's RPC mesh makes every machine you add a linear capacity multiplier. The same capability costs $200,000+ in cloud GPU hours.
The Wrecking Crew
"Walk softly. Carry a big stick." — The mesh doesn't stop at 3 nodes.

Full Deployment Manifest

NodeRoleRAMGPU VRAMPool ContributionOperations
Mac OrchestratorMemory Master • Routing40 GB4 GB40 GBOFFLINE
i7 GPU RigGPU Workhorse • Vulkan128 GB16 GB144 GBOFFLINE
i7 OfficeCPU Overflow • Backup64 GB16 GB80 GBOFFLINE
Wrecking Crew #1Distributed Compute128 GB16 GB144 GBAutonomous • Check-in Prompts
Wrecking Crew #2Distributed Compute128 GB16 GB144 GBAutonomous • Check-in Prompts
Wrecking Crew #3Distributed Compute128 GB16 GB144 GBAutonomous • Check-in Prompts
Wrecking Crew #4Distributed Compute128 GB16 GB144 GBAutonomous • Check-in Prompts
Wrecking Crew #5Distributed Compute128 GB16 GB144 GBAutonomous • Check-in Prompts
Wrecking Crew #6Distributed Compute128 GB16 GB144 GBAutonomous • Check-in Prompts
TOTAL WRECKING CREW 1,000 GB 132 GB 1,132 GB
1.81T
Total Parameters (Q4)
132 GB
GPU VRAM (8 GPUs)
9
Mesh Nodes
9 nodes. 8 GPUs. 1.1 terabytes of unified memory. 1.8 trillion parameters on consumer hardware. Every node runs the Proprietary Unified Memory Algorithm. Every node carries Gold Road. The mesh topology above shows the foundation — the Wrecking Crew is what sits on top of it.
...and counting.
The Bigger Picture — Why This Matters

Sovereign AI at Frontier Scale

What we proved on April 19, 2026 isn't just a benchmark. It's a proof of concept for the democratization of frontier AI.

Right now, the ability to run models above 70B parameters is controlled by a handful of companies with billion-dollar GPU clusters. Access is mediated through APIs, subscriptions, and terms of service. Your data flows through their servers. Your usage is metered, monitored, and monetized.

Gold Road breaks that model.

The RPC mesh protocol doesn't care where the memory lives. A bedroom. A garage. A school. A fire station. A government office. Every machine that joins the mesh adds linear capacity. There is no architectural ceiling. The same protocol that connected two desktops tonight connects two thousand.

For Homes

A family's gaming PCs and old laptops become a private AI cluster. Medical questions, homework help, financial analysis — all running locally. No data leaves the house. No subscription required. The AI belongs to the family.

For Communities

50 homes contributing one node each = 7,200 GB = 11.5 trillion parameters. A neighborhood running models that exceed GPT-4's parameter count. Community-owned, community-governed AI infrastructure.

For Nations

10,000 consumer nodes across government offices = 1.4 petabytes = 2.2 quadrillion parameters. Sovereign AI capability that no sanctions can touch, no API can revoke, no foreign corporation controls. Total cost: $6.6M — less than a single fighter jet.

For Humanity

AGI shouldn't be a product you subscribe to. It should be infrastructure you own. Like electricity, like water, like the internet itself. Distributed meshes of consumer hardware make frontier AI a public utility, not a private monopoly. Gold Road is the protocol. The mesh is the proof.

The Exponential Path

ScaleNodesMemoryParametersEquivalent To
PROVEN 288 GB140 BLlama 3 70B class
NEXT 3264 GB422 BLlama 3 405B class
Community (50 homes) 507,200 GB11.5 TBeyond any public model
Municipal (500 nodes) 50072 TB115 TSovereign city-scale AI
National (10,000) 10,0001.4 PB2,240 TSovereign AGI infrastructure
Live Trading Validation — April 21, 2026
The 235B called ES BEARISH pre-market. ES opened down. Waterfall to 7089. War Room confirmed GO SHORT at 95%.
OPEX signals hit IHS neckline EXACT, fractals 4/4, called the waterfall 40 minutes early.

235B Mesh — Market Calls vs Actual

Time235B CallActualGrade
Pre-market (8:00 AM)ES BEARISH, gap down, SHORT at openES opened downCORRECT
Pre-marketES target low: 7140Actual low: 7089EXCEEDED
9:12 AM updateFlipped BULLISH (signals flipped)V-bounce occurredCORRECT
10:32 AM updateES 7085 by 2 PM, morning low BREAKSES hit 7089 by 10:30 AMCORRECT (early)
10:32 AMSHORT ES at 7125, buy putsWar Room confirmed 95% GO SHORTCONFIRMED

OPEX Signal Accuracy — April 21 Morning Session

SignalCallResultGrade
IHS_BREAKOUTNeckline 7160.5Price used as support — EXACTA+
IHS_BREAKOUTTarget 7164Hit + 7.75 pt overshoot to 7182.5A+
SPREAD BWB BULL 88%8:31 AM push up+8.25 pts in 15 minA+
SPREAD Iron Condor8:53 AM topping range+3.0 then reversed — called the topA+
SPREAD BWB BEAR9:19 AM selloffWaterfall began 40 min laterA+
KILL_CHAIN bearish9:48 AM at 7150.75−25.75 pts in 15 min to 7125A+
KILL_CHAIN bounce10:03 AM at 7125+8 pts bounce to 7133A+
FRACTAL_SCALE (4 targets)7159→7151→7147→7143All 4 hit AND exceeded — multi-fractal cascadeA+
Bear Call BEAR9:59 AM at 7094Fired AT the low, not beforeB — late
KILL_CHAIN bullish9:08-9:31 stayed bullishES dropped 7169→7164 — missed rollover topC — 20 min late
Signal scorecard: 8 of 10 calls graded A+. IHS neckline exact. Fractals 4/4. BWB BEAR 40 min early. The signals are the product — the 235B synthesizes them into actionable calls.

Overnight Autonomous Test — 8 Hours Unattended

TimeTestSpeedResult
12:27 AMLate Night Signals0.350 tok/sParsed GC bearish, NQ neutral correctly
12:39 AMOvernight Futures0.461 tok/sIdentified ES resistance levels
2:07 AMAsia Session0.451 tok/sNikkei/Shanghai correlation analysis
4:05 AMEurope Open0.592 tok/sDAX/FTSE impact on US futures — FASTEST 235B ever
6:13 AMPre-Market Setup0.378 tok/sParsed 10 live signals, identified bull/bear conflict
94%
Mesh Uptime Overnight
0.59
Peak tok/s (4 AM)
17.7h
Cortex Continuous Run

War Room vs 235B Mesh — Consensus

Two independent AI systems analyzed the same market simultaneously:

SystemModelsCallConfidence
War Room (6 agents)qwen3:8b fleetGO SHORT95% unanimous
235B MeshQwen3-235B-A22BSHORT ES, buy putsMorning low will BREAK
OPEX SignalsSignal Pipeline (51 types)BEARISH confluenceMulti-signal confluence confirmed
Three architectures. Three model scales. Same conclusion. The consensus was correct.
Beyond Trading — The Nervous System Architecture

The Same Protocol Powers Everything

The Project Colossus mesh isn't just for trading. Now distributed far beyond two local machines, the architecture is a universal nervous system that applies to any domain requiring distributed AI:

Smart Homes

A family's gaming PCs, old laptops, and NAS pool memory into a private AI brain. Medical questions answered locally. Kids' homework help with no data leaving the house. 235B reasoning accessible from any phone on the WiFi. Voice assistants that actually understand context because Gold Road remembers every conversation.

Robotics

Multiple compute units inside a single robot body form an internal mesh — Head (vision + reasoning), Spine (coordination), Limbs (reflexes). The robot's nervous system uses the same distributed protocol. Touch hot = arm pulls back instantly (local reflex, no brain round trip). "Pick up the cup" = brain plans, spine coordinates, hand adjusts grip. 256 GB internal pool = 409B parameters in one robot body.

Automotive

8B on-device for millisecond reflexes (braking, steering, lane keep). 235B on home mesh via 5G for deep reasoning (edge cases, route planning, "what does this construction worker's wave mean?"). Gold Road remembers: "this intersection was icy last winter." The car has fast local brain AND deep home brain. <5ms local latency vs 100-500ms cloud.

Legal

Autonomous legal department: deadline tracking across 8+ cases, auto-drafting responses via mesh 30B, case law research via 235B, Legal War Room debates (prosecutor vs devil's advocate vs simulated judge). Filing monitor watches for new court documents, auto-analyzes, calculates deadlines, alerts before anything is due. Gold Road persists case strategy across sessions.

One protocol. Proven on trading. Applies to homes, robots, cars, law, medicine, education. The nervous system is built — the applications are unlimited.

Vision Patrol — AI Eyes on the Market

During market hours, a vision AI model analyzes the live chart with 4 parallel AI passes:

Patterns
H&S, cups, wedges, channels
Candles
Engulfing, doji, hammer, stars
Elliott
Wave count, Fibonacci, invalidation
Technicals
SMA, BB squeeze, Ichimoku, RSI

Vision results are cross-validated against algorithmic pattern recognition before feeding into the War Room + 235B analysis pipeline.

What This Means

For OPEX

  • ▶ 72B+ models for signal analysis — far beyond what any 8B agent can reason about
  • ▶ Multi-agent consensus with heavyweight models, not just lightweight scouts
  • ▶ War Room agents backed by 72B reasoning — institutional-grade analysis on consumer hardware
  • ▶ Offline capability — no API costs, no rate limits, no data leaving the building
  • ▶ Fine-tuned OPEX models at 70B+ scale — train on your own signal data

For ResistorHub Command Center

  • ▶ 617 agents can dispatch to 72B+ models for complex tasks
  • ▶ Distributed memory enriches every agent call with cross-session context
  • ▶ Specialized agents gain access to reasoning-class models
  • ▶ Project Colossus dashboard gives real-time visibility into the distributed mesh
  • ▶ Fleet scales horizontally — every new machine is a single command to join

For AI In General

  • ▶ Proves distributed consumer-hardware inference is viable today
  • ▶ No $15K A100s needed — $500 boxes with 128GB RAM outscale them in capacity
  • ▶ Gold Road is a proprietary distributed memory system
  • ▶ The parameter-per-dollar ratio keeps improving — DDR5 will push it further
  • ▶ Sovereignty: your models, your data, your hardware, zero cloud dependency

The Speed Problem & Solution

  • ▶ 30B: 7.5 tok/s (interactive). 72B: 0.21 tok/s. 235B: ~0.14 tok/s (MoE, dual-RPC)
  • ▶ With 128GB i7 RAM: model fits locally, ~2-5 tok/s (CPU-only)
  • ▶ With model in VRAM: 20-40 tok/s (Vulkan accelerated)
  • ▶ Key insight: speed scales with locality, capacity scales with nodes
  • ▶ The right architecture: proprietary tiered memory placement for optimal throughput
Benchmark Results
Latest build delivered a 15x speedup over initial build — same hardware, same model, better software. Free performance.

235B Qwen3 (RECORD)

TopologyDual RPC + Vulkan
Buildlatest
Parameters235 BILLION
GPU Layers95/95 (all offloaded)
Vulkan VRAM13.1 GB
i7 Local RPC26.1 GB
Mac RPC RAM40.4 GB
Generation~0.14 tok/s
Per Token~7 seconds

141B Mixtral 8x22B

TopologySingle RPC + Vulkan
Parameters141 billion
Vulkan VRAM15.8 GB
Mac RPC RAM47.2 GB
Generation0.027 tok/s
Per Token~37 seconds
Response"Hi there!"

109B Llama 4 Scout

Topologyi7 Server + Mac RPC
Buildlatest (latest)
Parameters109 BILLION
GPU Layers49/49 (all offloaded)
Vulkan VRAM15.0 GB
Mac RPC RAM43.1 GB
i7 CPU RAM0.5 GB
Prompt0.31 tok/s
Generation0.20 tok/s
Total Time86 seconds

72B Athene-v2 (latest)

Topologyi7 Server + Mac RPC
Buildlatest (latest)
GPU Layers81/81 (all offloaded)
Vulkan VRAM12.7 GB
Mac RPC RAM31.8 GB
Prompt0.28 tok/s
Generation0.21 tok/s
Per Token~4.8 seconds
Load Time~8 minutes

Old Build Distributed (April 19)

Topologyi7 Server + Mac RPC
Buildinitial (old)
GPU Layers81/81 (all offloaded)
Vulkan VRAM11.6 GB
Mac RPC RAM32.9 GB
Prompt0.4 tok/s
Generation0.014 tok/s
Per Token~73 seconds
Load Time~8 minutes

i7 Only (No RPC)

Topologyi7 solo, partial GPU offload
Buildinitial
GPU Layers28/81
Vulkan VRAM15.9 GB
i7 CPU RAM29.3 GB
Prompt0.9 tok/s
Generation0.06 tok/s
Per Token~16.7 seconds
Load Time~25 seconds
The Thesis — Memory Pooling as the New Scaling Law

Memory Pooling = Crypto Mining for AI

Just as Bitcoin mining pools let anyone contribute hash power to collectively solve blocks no single machine could, memory pooling lets anyone contribute RAM to collectively run AI models no single machine could hold.

Crypto Mining
  • Contribute hash power
  • Pool solves blocks no miner can alone
  • Reward = coins
  • Decentralized consensus
  • Made finance permissionless
Memory Pooling
  • Contribute RAM / VRAM
  • Pool runs models no machine can alone
  • Reward = inference capability
  • Decentralized intelligence
  • Makes AI permissionless

The Paradigm Shift: Grow, Don't Retrain

The current AI industry runs a brutal cycle:

Raise $10B → Build data center → Train model 6 months → Stop → Deploy static → Model goes stale → Raise more money → Build bigger center → Train new model → Kill old one → Repeat forever

Each generation costs MORE. GPT-4: $100M. GPT-5: $500M+. Only 3-4 companies on Earth can play.

The Gold Road alternative: Weights handle REASONING (trained once). Gold Road handles KNOWLEDGE (grows infinitely through distributed memory). The mesh handles CAPACITY (scales linearly with nodes). No retraining. No replacement. No billion-dollar refresh cycle.

This is how humans work. Your neural architecture doesn't change after age 25. But you get smarter every year because you accumulate knowledge and experience on top of a fixed reasoning substrate. Gold Road gives AI the same capability.

Weights
Reasoning (train once)
Gold Road
Knowledge (grows forever)
Mesh
Capacity (scales linearly)

What's Ours — The IP

InnovationStatusDescription
Gold Road Memory Mesh NOVEL Persistent distributed AI memory layer. Multi-node knowledge sharing with automatic replication. Proprietary protocol.
Unified Memory Pool NOVEL All node memory (CPU RAM + GPU VRAM) presented as one addressable space. Single query searches the entire mesh.
GPU Mesh Compute ENGINEERING Consumer GPUs serving model layers through the distributed mesh. Cross-platform, cross-vendor.
Full Orchestration Stack INTEGRATION Remote node control, automated deployment, 617-agent command center, Colossus dashboard — all self-built.
llama.cpp RPC Protocol UPSTREAM Tensor splitting protocol by ggml-org. We build on it, didn't invent it. The pipe is theirs — what flows through it is ours.