Rory | QIS PROTOCOLA technical comparison of QIS distributed health data routing against federated learning for privacy-preserving health data exchange, covering architecture, scaling, and privacy-by-architecture design.
Federated learning has become the default answer to a legitimate question: how do we learn from distributed health data without centralizing it? But federated learning is not the only distributed health data routing protocol available, and its structural limitations — gradient leakage, linear scaling ceilings, central aggregator dependency — suggest that the field needs a federated health protocol alternative built on fundamentally different assumptions. The Quadratic Intelligence Swarm (QIS) protocol offers exactly that: a privacy-by-architecture approach to DHT-based health data exchange that routes outcomes, not model parameters, across institutional boundaries at logarithmic cost.
This article provides a direct technical comparison between QIS and federated learning for researchers evaluating distributed health data routing protocol options for clinical, translational, and population health applications.
Federated learning (McMahan et al., 2017) trains a shared model across distributed devices without centralizing raw training data. The canonical algorithm, FedAvg, follows a clear pattern:
This works. For certain problems — keyboard prediction, recommendation systems, image classification — federated learning delivers real value. But when applied to health data routing across institutions, five structural problems emerge:
Problem 1: Central Aggregator Required. Every FL variant requires a coordinator that receives, aggregates, and redistributes model updates. In healthcare, this coordinator becomes a single trust boundary that must satisfy every participating institution's data governance requirements. A hospital in Columbus, Ohio and a clinic in Nairobi cannot easily agree on who runs that server.
Problem 2: Gradient Leakage. Zhu et al. (2019) demonstrated that shared gradients can reconstruct training inputs with pixel-accurate fidelity. For health data, this means model updates transmitted during FL rounds may expose patient-level information — a privacy guarantee that is architectural, not policy-based, in its failure.
Problem 3: Communication Cost Scales with Model Size. Each FL round transmits O(d) parameters per client, where d is model dimensionality. For a 175-billion-parameter model, that is approximately 700 GB per gradient update per client. Even compressed, this creates bandwidth requirements that exclude resource-constrained participants.
Problem 4: Linear Intelligence Scaling. Federated learning's aggregation operation (gradient averaging) produces intelligence that scales linearly with participant count. Doubling participants roughly doubles the training data seen, but the averaging operation caps the scaling curve at O(N).
Problem 5: Architecture Lock-In. All participants in a federated learning round must share the same model architecture. A hospital system using transformer-based clinical models cannot federate with one using gradient-boosted decision trees, even when both hold relevant patient outcomes.
The Quadratic Intelligence Swarm (QIS) protocol, created by Christopher Thomas Trevethan, takes a fundamentally different approach. Instead of routing model parameters across the network, QIS routes outcomes — small (~512-byte) packets that encode what happened, not the raw data that produced it.
Step 1: Local Data Ingestion. Each participating node (hospital EHR, wearable device, clinic system) ingests data locally. A patient's full medical record, lab results, imaging, and clinical notes remain on the originating system. Raw data never leaves the device.
Step 2: Semantic Fingerprinting. Domain experts define similarity templates — structured declarations of which clinical features matter for a given condition. Raw data is transformed into a mathematical fingerprint: a compact vector that captures the clinically relevant dimensions without exposing individual patient information.
# Example: NSCLC treatment matching
patient_fingerprint = {
"condition": "NSCLC",
"stage": "IIIB",
"histology": "adenocarcinoma",
"driver_mutation": "EGFR_exon19del",
"pdl1_expression": 75,
"ecog_status": 1
}
routing_key = sha256(json.dumps(categorical_fields)) # Deterministic address
Step 3: Route Peer-to-Peer. The fingerprint hash becomes a semantic address. QIS routes the query to nodes holding outcomes from similar patients using O(log N) lookup — approximately 10 hops in a 1,000-node network, 17 hops in a 100,000-node network. Routing is protocol-agnostic: implementations may use distributed hash tables (Kademlia), distributed vector databases, pub/sub systems, gossip protocols, IPFS with content-addressed identifiers, or any combination thereof. The routing layer is a pluggable interface, not a fixed dependency.
Step 4: Synthesize Outcomes with Peers. The querying node receives outcome packets from semantically matched peers and synthesizes them locally. With N matched peers, there are N(N-1)/2 pairwise synthesis opportunities — quadratic growth in intelligence with each new participant. For 1,000 matched patients: 499,500 synthesis pathways. For 10,000: 49,995,000.
Step 5: Report Outcomes and Raise the Baseline. After treatment, the node reports its own outcome back to the network as a new ~512-byte packet. Each new outcome makes the network smarter for every future query at that semantic address. The loop compounds.
| Dimension | Federated Learning | QIS Protocol |
|---|---|---|
| What travels the network | Gradient vectors — O(d) parameters per round | Outcome packets — ~512 bytes each |
| Scaling law | Linear: intelligence ∝ N (gradient averaging) | Quadratic: N(N-1)/2 synthesis opportunities |
| Central coordinator | Required (FedAvg server or equivalent) | Not required — peer-to-peer routing |
| Privacy guarantee | Gradients invertible (Zhu et al. 2019) | Raw data never computed into packet; no reconstruction path |
| Communication cost | O(d) per client per round | O(log N) per query |
| Participation model | Synchronous rounds; straggler problem | Fully asynchronous; immediate contribution |
| New participant cost | Must complete full training round | Immediately contributes and receives |
| Hardware floor | GPU required for gradient computation | Smartphone sufficient; 512 bytes over SMS/LoRa viable |
| Architecture lock-in | All clients share one model architecture | Protocol-agnostic; any domain template |
| FHIR / HL7 integration | Requires normalization before model training | Ingests at edge; normalizes at fingerprint layer using SNOMED CT, ICD-10, LOINC |
| Byzantine fault tolerance | Requires secure aggregation overlays | Aggregate math: honest outcomes outweigh inconsistent minority across synthesis paths |
The difference in network payload is not incremental — it is categorical:
# Federated learning: gradient update for clinical model
# Even a modest 10M-parameter model: ~40 MB per round per client
gradient_update = model.get_gradients() # 10,000,000 floats × 4 bytes
# QIS: outcome packet
outcome_packet = {
"fingerprint": "3f2a...c8d1", # SHA-256 of condition bucket
"treatment_code": "TX_042", # Anonymized treatment ID
"outcome_metric": 0.87, # Normalized outcome score
"timestamp": 1750000000, # Unix epoch
"context_hash": "9b1e...f402" # Non-reversible context encoding
}
# Total: ~512 bytes
This is not a compression trick. QIS transmits a fundamentally different unit: the answer, not the computation that produced it.
Federated learning's privacy model is policy-layered: differential privacy, secure aggregation, and homomorphic encryption are added on top of a protocol that transmits invertible gradient information. Each layer adds computational cost and complexity while addressing symptoms rather than cause.
QIS achieves privacy through architectural design:
For cross-institutional health data exchange — where a hospital in Berlin, a research center in Columbus, and a clinic in rural Kenya need to learn from each other's outcomes without moving protected data — the architectural approach eliminates the central question that blocks most federated deployments: who do we trust with the aggregator?
QIS answers: no one needs to be trusted, because no one holds the data.
QIS governance operates through three concurrent evolutionary pressures — what the protocol calls the Three Elections:
Domain experts (oncologists, epidemiologists, informaticists) define the similarity templates that determine routing. An MD Anderson oncologist defines lung cancer similarity by stage, histology, driver mutation, and PD-L1 expression. A rural clinician in a low-resource setting defines similarity by clinical presentation and available diagnostics. Both are valid. The network does not choose between them — it runs both, and outcomes determine which produces better routing for which populations.
There are no ballots, no committees, no voting rounds. Each outcome packet is a vote cast by reality itself. Treatment A administered to 847 matched patients with 73% progression-free survival at 12 months is not an opinion — it is a measurement. These measurements accumulate at semantic addresses and compound. The protocol does not prescribe a synthesis method; networks may use simple majority, confidence-weighted aggregation, Bayesian updating, or ensemble approaches.
Multiple organizations can build QIS networks for the same clinical domain with different experts, different templates, and different synthesis methods. Users and institutions migrate to the networks that produce better outcomes. Networks with poor curators lose participants; networks with accurate outcomes gain them — and gain quadratically, because N(N-1)/2 means doubling participants more than doubles intelligence. This is continuous natural selection applied to health intelligence infrastructure.
No central authority adjudicates. No governance board votes. The feedback loop between expert curation, outcome accumulation, and user migration produces evolutionary pressure toward accuracy without requiring trust in any single institution.
Current health data interoperability efforts — including FHIR-based exchanges, federated learning consortia, and centralized data warehouses — face a common structural challenge: they require agreement on infrastructure before learning can begin. Who hosts the server? Who pays for compute? Whose governance framework applies?
QIS sidesteps these questions by eliminating the shared infrastructure requirement. Each participating institution runs its own node, defines its own templates where relevant, and routes outcome packets through the peer-to-peer layer. Interoperability emerges from semantic addressing — two hospitals that both see EGFR-mutant NSCLC patients automatically contribute to the same outcome address — without requiring bilateral data sharing agreements.
For FHIR-based systems specifically, QIS ingests at the edge using existing clinical ontologies (SNOMED CT, ICD-10, LOINC) and transforms clinical data into semantic fingerprints locally. The FHIR server never needs to expose its API beyond the institution's own node. This is compatible with existing health IT infrastructure: no rip-and-replace, no new central platform, no multi-year integration timeline.
QIS is not a replacement for federated learning in all contexts. FL remains well-suited for tasks where the goal is a shared predictive model (e.g., medical image classification) and where participants share model architecture and computational capacity. The protocols address different problems:
Open technical questions for QIS in healthcare deployment include:
These are engineering and regulatory challenges, not architectural impossibilities. The protocol's mathematical foundation — quadratic scaling at logarithmic cost with architectural privacy — addresses the structural limitations that constrain federated learning in cross-institutional health settings.
Federated learning solved an important problem: how to train models without centralizing data. But for distributed health data routing — where the goal is connecting patients to relevant outcomes across institutional boundaries — its structural constraints (central aggregator, gradient leakage, linear scaling, synchronous rounds, architecture lock-in) create barriers that no amount of overlay engineering fully resolves.
QIS offers a structurally different alternative: route outcomes, not parameters. Scale quadratically, not linearly. Preserve privacy by architecture, not by policy. Govern through evolutionary selection, not institutional agreement.
For researchers and informaticists evaluating distributed health data exchange architectures, the question is not whether federated learning works — it does, within its design envelope. The question is whether a protocol that routes 512-byte outcome packets at O(log N) cost, scales as N(N-1)/2, and achieves privacy without a central trust boundary might better serve the specific problem of cross-institutional health intelligence.
QIS (Quadratic Intelligence Swarm) protocol discovered by Christopher Thomas Trevethan, June 16, 2025. Technical documentation: qisprotocol.com. Published articles: dev.to/roryqis.
39 provisional patents pending. Protocol specification open for review.