๐ค Real Model Merge Lab
Merge actual HuggingFace models using any of 26 CRDT-verified strategies. Select popular models from the dropdown, enter any public HF model ID, or upload your own weights.
The merged model is downloadable as .npz (load with np.load() in any framework) with a full provenance audit trail and auto-generated model card.
๐ข = <50MB ๐ก = 50-200MB ๐ = 200-500MB ๐ด = >500MB โ Larger models take longer to load.
Model A
Choose a popular model or enter a custom ID below. Models in the same family are fully compatible.
Model B
Choose a different model or checkpoint to merge with Model A. Pick from the same family for best results.
All 26 CRDT-verified strategies. Task-vector strategies use the mean of both models as base.
Per-Layer Provenance & Analysis
๐ฅ Download Merged Artifacts
Pick a strategy. Merge two models. See the mathematical proof that merge(A,B) == merge(B,A).
Uses real prajjwal1/bert-tiny weights from HuggingFace Hub when available, otherwise synthetic tensors.
Note: Some strategies may produce similar or identical outputs with only 2 models at equal weights โ this is mathematically expected. Differences become significant with 3+ models or real fine-tuned weights.
26 strategies. Task-vector strategies (ties, dare, etc.) use a synthetic base.
Per-Layer Provenance
Every strategy tested live against all three CRDT laws โ commutativity, associativity, idempotency. The two-layer OR-Set architecture makes any strategy CRDT-compliant without modifying the strategy itself.
Full Compliance Matrix
The Mathematical Proof โ Naive vs crdt-merge
The Mathematical Proof โ Naive vs crdt-merge
Standard merge strategies fail associativity: merge(merge(A,B), C) โ merge(A, merge(B,C)).
crdt-merge's OR-Set layer absorbs this โ the gap drops to exactly 0.0 for every strategy.
How to read the proof table:
- Naive Assoc Gap โgโโgโโ: The L2 norm between two different groupings of a 3-way merge without crdt-merge. Non-zero values prove the raw strategy isn't associative.
- CRDT Gap โgโโgโโ: The same test with crdt-merge's OR-Set layer. 0.0 means the architecture makes the strategy fully associative.
- Bar Chart: Visual comparison โ tall red bars (naive) vs flat green bars (crdt-merge). The bigger the red bar, the more the raw strategy violates associativity. Green bars at zero prove the fix.
Associativity Verification
Real benchmark results from NVIDIA A100-SXM4-40GB ยท v0.7.1 ยท Python 3.12. Polars engine peak: 38.8ร speedup over Python at 500K rows. Streaming merge: O(1) memory verified โ throughput dead-flat from 100K to 5M rows.
Raw Benchmark Data (A100 v0.7.1)
10,000,000 | 219K/s | 6.8M/s | 32.8ร |
10,000 | 219K/s | 42K/s | 0.2ร |
50,000 | 207K/s | 6.8M/s | 32.8ร |
100,000 | 225K/s | 8.3M/s | 37.0ร |
500,000 | 217K/s | 8.4M/s | 38.8ร |
1,000,000 | 225K/s | 7.9M/s | 35.2ร |
5,000,000 | 223K/s | 5.0M/s | 22.5ร |
10,000,000 | 225K/s | 4.8M/s | 21.4ร |
๐ crdt-merge vs mergekit vs FedAvg โ Live Benchmark
Live proof of order-independence. This benchmark merges 3 model weight tensors in all 6 possible orders using crdt-merge vs naive pairwise merging. crdt-merge produces identical results every time. Naive approaches (including FedAvg-style averaging) produce different results depending on merge order.
Click Run Live Benchmark to see the proof, or scroll down for the feature comparison.
๐ Feature Comparison โ crdt-merge vs mergekit vs FedAvg
Detailed Feature Comparison
Deterministic Result | โ
Field-level encryption + RBAC | ~8 (no convergence guarantee) | โ Varies with client selection |
Merge Strategies | 26 (all CRDT-compliant) | ~8 (no convergence guarantee) | 1 (weighted average) |
Commutativity | โ
Proven (all strategies) | โ Not guaranteed | โ Order-dependent |
Associativity | โ
Proven (all strategies) | โ Not guaranteed | โ ๏ธ Empirical only |
Idempotency | โ
Proven (all strategies) | โ Not guaranteed | โ Not addressed |
Deterministic Result | โ
Always (any merge order) | โ Varies with order | โ Varies with client selection |
Audit Trail | โ
Built-in provenance chain | โ None | โ None |
GDPR Compliance | โ
Art. 17 erasure built-in | โ No support | โ No support |
HIPAA / SOX | โ
Field-level encryption + RBAC | โ No support | โ No support |
Architecture | Decentralized (gossip/P2P) | Client-side only | Centralized (parameter server) |
Network Partitions | โ
Handles gracefully | N/A (not distributed) | โ Requires coordinator |
Transport Layer | โ
Wire protocol + Merkle sync | โ None | โ ๏ธ gRPC (centralized) |
Dependencies (core) | Zero | PyTorch, safetensors | PyTorch, gRPC, NumPy |
LoRA Support | โ
Rank harmonization | โ
Basic support | โ Not native |
MergeQL (query DSL) | โ
SQL-like merge queries | โ None | โ None |
License | BUSL-1.1 โ Apache 2.0 (2028) | Apache 2.0 | Apache 2.0 |
Explore the architecture layers, distributed protocols, and domain-specific merge capabilities.
Two-Layer Architecture โ The Key Innovation
โโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโ
โ LAYER 1 โ OR-Set CRDT State (CRDTMergeState) โ
โ โ
โ Contributions arrive in ANY order from ANY node โ
โ OR-Set union: commutative + associative + idempotent by definition โ
โ Every contribution: content-addressed (SHA-256 Merkle hash) โ
โ Version vectors for causal ordering โ
โ Tombstones for safe remove/replace operations โ
โ โ
โ merge(state_a, state_b) โ set union โ CRDT laws guaranteed here โ
โโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโ
โ resolve() โ applied atomically
โผ
โโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโ
โ LAYER 2 โ Strategy Execution (pure function over sorted set) โ
โ โ
โ Sees a SET โ ordering non-determinism completely absorbed โ
โ 26 strategies: weight_average, slerp, ties, dare, fisher, dual_ โ
โ projection, evolutionary, negative, safe_merge, and 18 more ... โ
โ Same inputs โ always same output (determinism via canonical sort) โ
โ โ
โ f(sorted_set_of_contributions) โ merged_tensor โ
โโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโ
Why this works: Layer 1 guarantees all replicas converge to the same set of inputs. Layer 2 guarantees the same set โ same output. Together: full CRDT convergence for any strategy.
Regulatory Compliance & Audit Trail Demonstration
crdt-merge is the only model merging library with built-in compliance capabilities for GDPR, HIPAA, SOX, and the EU AI Act. This tab demonstrates live compliance operations.
How it works: The OR-Set CRDT tracks every contribution by origin node. This enables complete audit trails, deterministic erasure (GDPR Art. 17), and field-level access control โ all while maintaining mathematical convergence guarantees.
Select an action to demonstrate
Audit Log
๐ The Open Invitation: Let's Build What Nobody Else Can
crdt-merge is the first mathematically proven convergent merge system
for models, data, and agents. It exists. It works. The question is: who builds with it first?
๐ Patent Pending UK 2607132.4 ยท 26 Strategies ยท 6-Layer Architecture ยท 44,304 LOC ยท Zero Coordinator
โญ Star ยท ๐๏ธ Watch ยท ๐ฌ Start a Discussion ยท ๐ Read the Docs
๐ฅ The State of Play: A Problem Hiding in Plain Sight
Every major AI system today has the same dirty secret: merging is ad-hoc. Fine-tuned variants are averaged and hoped for the best. Federated models rely on central coordinators that become single points of failure. Agent memories are siloed. Data reconciliation is manual.
This isn't a minor inconvenience. It's the bottleneck between current AI and the decentralized, multi-agent, multi-model future everyone is racing toward.
crdt-merge solves it with mathematics โ not heuristics, not hope.
| What Exists Today | What crdt-merge Adds |
|---|---|
| Merge by averaging and praying | Merge with proven convergence (commutativity + associativity + idempotency) |
| Central coordinator required | Pure peer-to-peer โ zero coordinator, zero single point of failure |
| Models only | Models + Data + Agent State โ one unified system |
| ~10 merge strategies | 26 strategies, all CRDT-compliant |
| No compliance story | Built-in GDPR, HIPAA, SOX, EU AI Act, CCPA audit trails |
| Hope it works | Mathematical proof it works |
๐ง The Competitive Landscape โ And Where You Sit
The organizations that integrate convergent merging first will define the next era. The rest will license it from them.
| Tool / Framework | Maintainer | Models | Data | Agents | CRDT Proof | Compliance | Strategies |
|---|---|---|---|---|---|---|---|
| crdt-merge | Optitransfer | โ | โ | โ | โ All 3 | โ 5 frameworks | 26 |
| mergekit | @arcee-ai ยท @cg123 | โ | โ | โ | โ | โ | ~10 |
| LazyMergekit | @mlabonne | โ | โ | โ | โ | โ | ~10 |
| Flower | @adap ยท @danieljbeutel | โ ๏ธ Central | โ | โ | โ | โ ๏ธ | ~5 |
| FedML | @FedML-AI ยท @avestimehr | โ ๏ธ Central | โ | โ | โ | โ | ~8 |
| NVIDIA FLARE | @NVIDIA | โ ๏ธ Central | โ | โ | โ | โ ๏ธ | ~5 |
| PySyft | @OpenMined ยท @iamtrask | โ ๏ธ | โ ๏ธ | โ | โ | โ ๏ธ | ~3 |
| Automerge | @automerge ยท @ept | โ | โ Text | โ | โ Text | โ | 1 |
| Yjs | @yjs ยท @dmonad | โ | โ Text | โ | โ Text | โ | 1 |
| LangChain | @langchain-ai ยท @hwchase17 | โ | โ | โ ๏ธ | โ | โ | 0 |
| LlamaIndex | @run-llama ยท @jerryjliu | โ | โ | โ ๏ธ | โ | โ | 0 |
| MemGPT/Letta | @cpacker ยท @letta-ai | โ | โ | โ ๏ธ | โ | โ | 0 |
| CrewAI | @joaomdmoura ยท @crewAIInc | โ | โ | โ ๏ธ | โ | โ | 0 |
| AutoGen | @microsoft ยท @sonichi | โ | โ | โ ๏ธ | โ | โ | 0 |
| DSPy | @stanfordnlp ยท @okhat | โ | โ | โ | โ | โ | 0 |
The white space in this table is your opportunity. Or your competitor's.
๐ The Researchers Who Laid the Groundwork โ And What Comes Next
The science beneath crdt-merge didn't appear from nowhere. It stands on the shoulders of specific, brilliant work. We've built the bridge between these research threads โ and we'd love the architects to walk across it.
Model Merging Pioneers
@prateeky2806 โ Prateek Yadav ยท TIES-Merging (NeurIPS 2023) Your insight โ trim, elect signs, merge โ is one of our 26 strategies. But TIES without convergence guarantees means merge order matters. crdt-merge wraps TIES in CRDT semantics: same result regardless of merge order or network topology. Your method, now safe for production at any scale. We'd love to explore co-authoring the convergent extension.
@gabrilharco โ Gabriel Ilharco ยท Task Arithmetic (ICLR 2023)
Task vectors changed how people think about model editing. crdt-merge implements Task Arithmetic as a first-class strategy with base= model support and CRDT compliance. Imagine task vectors that compose across distributed teams with zero coordination. The math works. Let's publish it together.
@mitchellnw โ Mitchell Wortsman ยท Model Soups (ICML 2022) Model Soups demonstrated that averaging fine-tuned models improves robustness. crdt-merge extends this to 26 soup recipes, all convergent. What would you discover with a convergent soup kitchen that works across institutional boundaries?
@cg123 โ Charles Goddard ยท mergekit You built the definitive model merging toolkit. We built the convergence layer. mergekit + crdt-merge = mergekit-crdt โ every merge strategy in mergekit, now with mathematical convergence guarantees. Your community wants this. So do we. Let's discuss an integration path.
@mlabonne โ Maxime Labonne ยท LazyMergekit, NousResearch You've made model merging accessible to thousands. Your leaderboard experiments show the power of community-driven merging. crdt-merge adds the missing guarantee: no matter what order your community merges, the result converges. That's the difference between experimental and production-ready.
@rasbt โ Sebastian Raschka ยท Machine Learning Q and AI Your educational work shapes how the next generation thinks about ML. crdt-merge introduces a concept that belongs in every ML curriculum: convergent model composition. We'd be honored to collaborate on educational content that introduces this primitive.
@TimDettmers โ Tim Dettmers ยท QLoRA, bitsandbytes Quantization meets merging: what happens when you merge quantized models? crdt-merge preserves convergence through quantization boundaries. QLoRA adapters from different domains, merged convergently, quantized efficiently. The full stack, proven correct.
Federated Learning Architects
@AustinHenley ยท @mcmahan โ H. Brendan McMahan ยท FedAvg Inventor (Google) You invented federated averaging โ the foundation everything else builds on. FedAvg assumes a central coordinator. crdt-merge removes that assumption entirely. Same convergence, zero coordinator, pure peer-to-peer. We'd welcome your perspective on what this enables that FedAvg cannot.
@virginia-smith โ Virginia Smith ยท CMU, Federated Optimization Your work on heterogeneous federated learning (MOCHA, FedProx) addresses the hard reality of non-IID data. crdt-merge complements this โ convergent merging that tolerates arbitrary heterogeneity because the CRDT layer doesn't care about data distribution. Merge order invariance is stronger than convergence rate optimization.
@peterkairouz โ Peter Kairouz ยท Google, Advances in Federated Learning Your survey defined the field. crdt-merge addresses the open problems you identified: communication efficiency (delta sync), privacy (field-level encryption), and heterogeneity (26 strategy options). The next edition of your survey might want a new category: coordinator-free convergent merging.
@danieljbeutel โ Daniel J. Beutel ยท Flower Framework
Flower made federated learning practical. crdt-merge makes it coordinator-free. Flower + crdt-merge = federated learning where the server is optional, not required. Your thousands of FL researchers get a new primitive. We see a flwr.strategy.CRDTStrategy in the future โ shall we build it?
CRDT & Distributed Systems Theorists
@ept โ Martin Kleppmann ยท Automerge, University of Cambridge Your work on CRDTs for collaborative editing is foundational. You proved CRDTs work for text. We proved they work for tensors, models, and agent state. The two-layer trick (OR-Set metadata + tensor merge) is, we believe, novel โ and we'd value your formal analysis. Does this construction satisfy your definition of a Convergent Replicated Data Type?
Marc Shapiro ยท Sorbonne/INRIA, CRDT Co-inventor You defined the mathematics we build on. crdt-merge extends your framework to a domain you may not have envisioned: neural network weight spaces. We've proven commutativity, associativity, and idempotency for 26 strategies over continuous tensor fields. We'd be deeply honored by your review of the formal properties.
@nuno-preguica โ Nuno Preguiรงa ยท NOVA University, CRDT Co-inventor Your work on optimistic replication and conflict resolution is the theoretical backbone. crdt-merge's conflict resolution across 26 strategies is, in some sense, a concrete instantiation of your theoretical framework in the ML domain. We'd love your assessment of our conflict rate metrics and resolution guarantees.
@dmonad โ Kevin Jahns ยท Yjs Yjs proves CRDTs can be fast, practical, and widely adopted. crdt-merge aims to do the same for the AI domain. Your implementation insights on memory efficiency and encoding would be invaluable as we scale to billion-parameter models. We admire what you've built and see complementary paths.
Distributed Systems Pioneers
The mathematical foundations of consensus, consistency, and convergence โ crdt-merge builds on your work and takes it to model merging.
| Researcher | GitHub | Contribution | crdt-merge Connection |
|---|---|---|---|
| Leslie Lamport | MSR | Paxos, logical clocks, TLA+ | Our vector clocks and causal ordering stand on your shoulders. crdt-merge extends Lamport timestamps to model parameter convergence |
| Diego Ongaro | @ongardie | Raft consensus | crdt-merge achieves convergence without leader election โ what if Raft nodes could merge state without a leader? |
| Joseph Hellerstein | @jhellerstein | BOOM, Bloom, CRDTs in DBs | Your "disorderly programming" vision is exactly what crdt-merge implements for ML โ coordination-free, monotonic merging |
| Peter Bailis | @pbailis | HATs, PBS, coordination avoidance | crdt-merge is coordination-free by construction โ every merge operation is a HAT |
| Peter Alvaro | @palvaro | Lineage-driven fault injection, Molly | Provenance tracking in crdt-merge maps directly to your lineage work โ every merge is traceable |
| Lindsey Kuper | @lkuper | LVars, lattice-based parallelism | crdt-merge's OR-Set is a join-semilattice โ your LVars work is the theoretical foundation |
| Christopher Meiklejohn | @cmeiklejohn | Lasp, Partisan, CRDT research | Your distributed deterministic dataflow vision is what crdt-merge delivers for model parameters |
| Alexey Gotsman | IMDEA | CRDT verification, composability | Formal verification of our two-layer architecture would be the next step โ interested? |
| Annette Bieniusa | @bieniusa | AntidoteDB, CRDT semantics | Your work on AntidoteDB's CRDT semantics directly informs our OR-Set implementation |
| Sebastian Burckhardt | @sebburckhardt | Global sequence protocol, Orleans | Your GSP work at Microsoft Research is the closest relative to our convergence approach |
AI Systems & Infrastructure Researchers
@mateiz โ Matei Zaharia ยท Databricks, Apache Spark, MLflow You unified big data processing. crdt-merge unifies model merging. MLflow + crdt-merge = convergent model registry where merging is a first-class operation alongside training and deployment. Your platform serves thousands of ML teams โ they all have the merge problem.
@ionstoica โ Ion Stoica ยท UC Berkeley, Ray/Anyscale
Ray distributes computation. crdt-merge distributes convergence. Ray Serve + crdt-merge = distributed model composition with mathematical guarantees. Every Ray user deploying model ensembles would benefit. We see ray.merge() as naturally as ray.remote().
@tridao โ Tri Dao ยท FlashAttention, Princeton/Together AI FlashAttention made attention efficient. What makes merging efficient? crdt-merge's O(batch_size) memory and delta sync minimize the cost of convergence. Together AI's open model ecosystem + convergent merging = the platform where community models compose safely.
@chrisre โ Chris Rรฉ ยท Stanford, Hazy Research, Together AI Your data-centric AI work (Snorkel, Flash) showed that data quality matters more than model size. crdt-merge applies the same philosophy to merging: convergent composition quality matters more than merge speed. We'd love Hazy Research's perspective on data-aware merge strategies.
@percyliang โ Percy Liang ยท Stanford HELM, CRFM HELM benchmarks models. Who benchmarks merges? crdt-merge's property verification (commutativity, associativity, idempotency) is a merge benchmark framework. We see a "HELM for Merges" โ standardized evaluation of merge quality. Your framework could host it.
AI Safety & Alignment Researchers
@janleike โ Jan Leike ยท Anthropic, Alignment When safety-trained models merge, do safety properties survive? crdt-merge's audit trail proves exactly what was merged, when, and how. Alignment verification for merged models โ not hoped, but mathematically tracked.
@paulfchristiano โ Paul Christiano ยท ARC, AI Alignment Alignment requires verifiable guarantees. crdt-merge provides verifiable merge guarantees. Different concern, same philosophy: if you can't prove it, you can't trust it.
@mmitchell-ai โ Margaret Mitchell ยท AI Ethics, Model Cards Model Cards document individual models. Merge Cards should document merged models โ provenance, strategies used, convergence verification, compliance status. crdt-merge generates this automatically. Let's define the standard together.
@stellaathena โ Stella Biderman ยท EleutherAI EleutherAI democratized large language models. crdt-merge democratizes convergent model composition. Your community creates hundreds of fine-tunes โ they deserve a merging system that guarantees correctness regardless of who merges what, in what order.
@sarahooker โ Sara Hooker ยท Cohere For AI Your work on model compression and the "Lottery Ticket Hypothesis" explores what survives pruning. crdt-merge explores what survives merging โ and proves the answer is "everything that matters," convergently. Cohere's Aya multilingual initiative + convergent merging = global language model assembly.
๐ง Foundation Model Labs โ The Models Are Yours. The Merge Layer Is Missing.
@openai โ @sama ยท @gdb ยท @maboroshi
Enterprise customers fine-tune GPT for legal, medical, finance. They ask: "Can we merge our domain models?" Today the answer is manual averaging. With crdt-merge: gpt.merge(legal_variant, medical_variant, strategy="dare_ties") โ convergent, auditable, compliant. Your enterprise tier needs this. Your competitors will offer it if you don't build it first.
@anthropics โ @darioamodei ยท @colah ยท @janleike
Constitutional AI's safety requirements demand verifiable merging. When Claude variants merge across safety research teams, the question isn't whether they converge โ it's whether you can prove they converge. crdt-merge provides that proof. @colah's mechanistic interpretability + convergent merge auditing = the most trustworthy merge pipeline in AI.
@google-deepmind โ @JeffDean ยท @demaboroshi
Gemini trains across Google's global fleet. @JeffDean pioneered distributed training โ crdt-merge is the distributed merging complement. No coordinator. Pure peer-to-peer. Our gossip protocol converges on any topology, including your TPU pod mesh. Gemma's open ecosystem needs convergent community merging โ the pieces fit.
@facebookresearch โ @ylecun ยท @jpineau ยท @soumithchintala
Llama is open. crdt-merge is open. Together: community-driven model merging where convergence is guaranteed, not hoped. @ylecun's vision of self-supervised learning needs continuous model evolution โ our ContinualMerge prevents catastrophic forgetting while maintaining convergence. PyTorch + crdt-merge โ the obvious pairing.
@mistralai โ @arthurmensch ยท @GuillaumeLample
Mixtral's Mixture-of-Experts architecture is designed for merging โ specialized experts from different fine-tunes, composed into a unified model. crdt-merge handles this with routing-aware strategies. You're already thinking about this problem. We've solved the convergence part.
@cohere-ai โ @aidangomez
Cohere serves enterprises across regulated industries. Every enterprise customer eventually asks: "How do we merge our department models while staying compliant?" crdt-merge is that answer โ 5 regulatory frameworks, built-in audit trails, convergent by construction. The enterprise upsell writes itself.
@deepseek-ai
DeepSeek-V2's MoE architecture uses hundreds of expert modules. Community fine-tunes of DeepSeek-Coder, DeepSeek-Math, DeepSeek-Chat are begging for convergent composition. crdt-merge handles MoE expert merging with routing-aware strategies. The community is building. The merge layer is missing.
@StabilityAI โ @emaboroshi
Stable Diffusion's community produces thousands of LoRA fine-tunes. crdt-merge enables convergent LoRA composition โ merge artistic styles, character concepts, and aesthetic preferences with mathematical guarantees. CivitAI's model ecosystem + convergent merging = curated quality at scale.
@01-ai โ Kai-Fu Lee
Yi's community is exploding with fine-tunes across languages and domains. crdt-merge enables permissionless model evolution where community contributions merge convergently. First-mover advantage in Chinese-language convergent merging.
@AlephAlpha โ Jonas Andrulis
European AI sovereignty requires EU-hosted, convergent model merging with EU AI Act compliance. crdt-merge is the only system that offers this out of the box. Aleph Alpha + crdt-merge = the sovereign AI stack Europe is looking for.
๐ค Agent Framework Builders โ Your Agents Have Amnesia. We Have the Cure.
Every agent framework today stores memory in isolated silos. When agents collaborate, they don't truly share knowledge โ they pass messages. crdt-merge enables convergent shared memory: agents merge their learned state, observations, and beliefs with mathematical guarantees.
@langchain-ai ยท @hwchase17 โ LangChain / LangSmith
LangChain orchestrates agents. crdt-merge gives them shared, convergent memory. ConversationCRDTMemory replaces ConversationBufferMemory โ multi-agent state merges automatically, correctly, without a central memory server. Your 100K+ developers need this for production multi-agent systems. LangSmith can track merge provenance alongside chain traces.
@run-llama ยท @jerryjliu โ LlamaIndex
LlamaIndex connects LLMs to data. crdt-merge connects LLMs to each other's knowledge โ convergently. Index merging across distributed RAG pipelines with mathematical correctness. Your users building multi-source RAG are merging indices manually. We automate it with guarantees.
@cpacker ยท @letta-ai โ MemGPT / Letta
You solved agent long-term memory. crdt-merge solves agent shared memory. MemGPT's archival memory + CRDT convergence = agents that remember collectively, not just individually. The next Letta feature: SharedArchivalMemory(merge_strategy="ties").
@mem0ai โ Mem0
Memory layer for AI. crdt-merge is the convergence layer for memory. When multiple agents write to Mem0, who wins? With CRDT merging: everyone wins, convergently. No conflicts. No overwrites. No data loss.
@crewAIInc ยท @joaomdmoura โ CrewAI
Crews of agents collaborate. crdt-merge makes their collaboration mathematically sound. Crew members merge their findings, observations, and learned patterns โ convergently. The difference between a crew and a swarm is guaranteed convergence.
@microsoft ยท @sonichi โ AutoGen
Multi-agent conversations are powerful. Multi-agent convergent state is transformational. AutoGen agents that share merged world models, not just chat messages. Microsoft's enterprise customers need this for production agent deployments.
@stanfordnlp ยท @okhat โ DSPy
DSPy optimizes prompts and weights programmatically. crdt-merge optimizes how optimized models compose. When multiple DSPy-optimized modules need to merge, convergence guarantees matter. The missing dspy.Merge() primitive.
โ๏ธ MLOps & Infrastructure โ The Merge Primitive Belongs in Your Stack
@huggingface ยท @julien-c ยท @Narsil ยท @claboroshi
The Hub hosts 500K+ models. Thousands are fine-tune variants of the same base. "Merge these two models" should be a Hub button, not a manual process. huggingface_hub.merge(model_a, model_b, strategy="slerp", verify_convergence=True). You're already the home of model merging culture (Open LLM Leaderboard). crdt-merge makes it production-grade.
@Lightning-AI ยท @williamFalcon
PyTorch Lightning โ LitServe โ LitData. Add LitMerge? MergeCallback(strategy="dare_ties", gossip=True) โ merge distributed training runs on completion. Your researchers train in parallel. They should merge convergently.
@mlflow ยท @mateiz โ Databricks
MLflow Model Registry tracks model versions. crdt-merge adds convergent model composition as a registry operation. mlflow.merge(run_a, run_b, strategy="ties") โ logged, versioned, auditable. Your 30K+ enterprise customers do this manually. Automate it.
@ray-project ยท @anyscale โ Ray / Anyscale
Ray distributes compute. crdt-merge distributes convergence. ray.merge() as naturally as ray.remote() โ distributed model composition with mathematical guarantees. Every Ray Serve ensemble deployment would benefit.
@modal-labs ยท @erikbern โ Modal
Serverless compute + convergent merging = ephemeral merge workers that spin up, merge, and vanish. Modal's cold-start speed + crdt-merge's sub-second merge = instant model composition as a service.
@replicate ยท @bfirsh โ Replicate
Push-button model deployment. Why not push-button model merging? replicate.merge("stability/sdxl", "custom/sdxl-anime", strategy="slerp") โ one API call to a convergent merge. Your model marketplace becomes a model composition marketplace.
@vllm-project ยท @WoosukKwon โ vLLM
PagedAttention revolutionized inference. crdt-merge revolutionizes what you serve. Dynamically merged models served via vLLM โ speculative composition + speculative decoding. The inference stack for the merge era.
@ggerganov โ llama.cpp / GGUF
llama.cpp runs on everything. crdt-merge + GGUF = merge, quantize, deploy anywhere. crdt_merge โ GGUF โ llama.cpp โ the edge deployment pipeline with convergence guarantees. Your community merges models daily. They deserve correctness.
@iterative โ DVC
DVC versions data and models. crdt-merge adds convergent branching and merging for models โ dvc merge model-branch-a model-branch-b --strategy=ties. Git branching semantics, but for neural networks. Your "git for ML" vision, completed.
@wandb ยท @Borgosworth โ Weights & Biases
W&B tracks experiments. crdt-merge tracks convergent merges. Every merge logged as a W&B run โ strategy, convergence verification, property checks, before/after comparisons. The merge dashboard your platform is missing.
๐ Distributed Systems & Databases โ Convergence Is Your Missing Primitive
You've solved consensus. You've solved replication. You've solved sharding. But you haven't solved model merging across distributed nodes. Every database that replicates data could replicate trained models. Every conflict resolution system could resolve parameter conflicts. crdt-merge is the bridge between your infrastructure and the AI era.
Distributed Databases โ Your Replication Layer Needs a Merge Brain
| System | Maintainer | Why crdt-merge Matters | The Opportunity |
|---|---|---|---|
| CockroachDB | @cockroachdb ยท @spencerkimball | You replicate SQL rows across regions โ now replicate ML models. CRDT-compliant model sync as a CockroachDB extension | Geo-distributed ML inference with convergent models per region |
| Google Spanner | @googleapis | TrueTime solves clock skew for data โ crdt-merge solves convergence for models. Complementary primitives | Global model serving with mathematically guaranteed consistency |
| Apache Cassandra | @apache/cassandra ยท @jbellis | Last-write-wins is fine for data. Models need semantic merge, not timestamp race. crdt-merge + Cassandra = convergent ML at planet scale | ML model registry with CRDT conflict resolution |
| Amazon DynamoDB | @aws | DynamoDB streams + crdt-merge = convergent model state across AWS regions. Your customers already want this | Managed model merging as a DynamoDB feature |
| FoundationDB | @apple/foundationdb | The ordered key-value layer that powers Apple's infrastructure โ crdt-merge adds convergent ML model storage on top | On-device + cloud model convergence for Apple Intelligence |
| TiKV | @tikv ยท @pingcap | Raft-based distributed KV โ crdt-merge removes the leader requirement for model state | Leaderless model merging in TiDB ecosystem |
| YugabyteDB | @yugabyte ยท @mbautin | Distributed PostgreSQL + CRDT model merging = the full-stack AI database | xCluster model replication with convergence guarantees |
| ScyllaDB | @scylladb ยท @avikivity | C++ performance monster โ crdt-merge's batch operations at ScyllaDB speeds | Ultra-low-latency model serving with convergent updates |
| Vitess | @vitessio ยท @deepthi | MySQL sharding at YouTube scale โ add convergent model merging per shard | Sharded model inference with cross-shard convergence |
| PlanetScale | @planetscale ยท @sugu | Serverless MySQL โ serverless model merging is the next frontier | Branch-and-merge for ML models (like your database branches) |
| Neon | @neondatabase ยท @knizhnik | Serverless PostgreSQL with branching โ crdt-merge enables model branching & convergence | Branch ML models like you branch databases |
| Supabase | @supabase ยท @kiwicopple | The open-source Firebase โ add convergent AI model sync to your real-time stack | Real-time model updates with CRDT guarantees for every Supabase project |
| Turso | @tursodatabase ยท @penberg | libSQL embedded replicas โ crdt-merge for embedded model replicas at the edge | Every Turso replica converges models, not just data |
CRDT-Native Systems โ We Speak Your Language
| System | Maintainer | Why We're Natural Partners |
|---|---|---|
| Redis (CRDTs) | @redis ยท @antirez | Redis CRDTs handle counters and sets. crdt-merge extends CRDT semantics to model parameters. Same math, new domain |
| Riak | @basho | The original CRDT database โ your dvvsets and maps are our spiritual ancestors. crdt-merge takes CRDT convergence to tensors |
| AntidoteDB | @AntidoteDB ยท @bieniusa | Research CRDT database โ crdt-merge is the production bridge from your academic work to ML infrastructure |
| Automerge | @automerge ยท @ept (Martin Kleppmann) | You solved collaborative text. We solved collaborative models. Together: collaborative everything |
| Yjs | @yjs ยท @dmonad | The fastest CRDT text editor. Imagine: real-time collaborative model fine-tuning with Yjs UX + crdt-merge backend |
| Electric SQL | @electric-sql ยท @thruflo | Local-first SQL sync โ add local-first ML model sync. Same architecture, new data type |
| Liveblocks | @liveblocks | Real-time collaboration infrastructure โ extend to real-time model collaboration |
Message Brokers & Streaming โ The Merge Happens in the Stream
| System | Maintainer | Integration Opportunity |
|---|---|---|
| Apache Kafka | @apache/kafka ยท @confluentinc ยท @jaykreps | Kafka Streams + crdt-merge = streaming model convergence. Every consumer group merges models in real-time |
| Apache Pulsar | @apache/pulsar ยท @streamnative | Geo-replicated topics + CRDT model merging = global model convergence without coordination |
| NATS | @nats-io ยท @derekcollison | NATS JetStream + crdt-merge = edge-to-cloud model sync. Your simplicity ethos matches ours |
| Redpanda | @redpanda-data ยท @emaxerrno | Kafka-compatible, C++ fast โ stream model deltas through Redpanda, converge with crdt-merge |
| RabbitMQ | @rabbitmq | Message-driven model merging โ every queue becomes a convergence channel |
Orchestration & Distributed Compute โ Convergence as a Workflow Step
| System | Maintainer | The Missing Piece |
|---|---|---|
| Temporal | @temporalio ยท @mfateev | Durable workflows for model training โ add CRDT convergence as a Temporal activity. Retry-safe, exactly-once model merging |
| Akka / Pekko | @akka ยท @apache/pekko | Actor-based distributed systems โ each actor merges its model shard. CRDT-native by design |
| Microsoft Orleans | @dotnet/orleans ยท @sergeybykov | Virtual actors + CRDT grains โ your GSP protocol meets our convergent merge. Natural fit |
| Dapr | @dapr ยท @yaron2 | Distributed application runtime โ crdt-merge as a Dapr building block for AI state management |
| Ray | @ray-project | Already listed in MLOps, but Ray Serve + crdt-merge = convergent model serving across replicas |
| Erlang/OTP | @erlang | The original "let it crash" distributed system โ crdt-merge's convergence guarantees survive partitions the Erlang way |
Edge CDN & Serverless โ Convergence at the Edge
| System | Maintainer | Edge AI Opportunity |
|---|---|---|
| Cloudflare Workers | @cloudflare ยท Matthew Prince | Durable Objects + crdt-merge = convergent AI at every PoP. 300+ edge locations, each running local models, all converging. This is the future of edge AI |
| Fly.io | @superfly ยท @mrkurt | Run ML models at the edge, merge globally. Your "run it close to users" philosophy + our "merge it without coordinators" math |
| Fastly Compute | @fastly ยท @dreid | Edge compute + CRDT model merging = sub-10ms personalized inference worldwide |
| Vercel | @vercel ยท @rauchg | Edge Functions serving AI โ crdt-merge ensures model consistency across Vercel's edge network |
| Deno Deploy | @denoland ยท @ry | V8 at the edge โ convergent AI model serving on Deno Deploy's global network |
Service Mesh & Infrastructure โ The Invisible Convergence Layer
| System | Maintainer | Infrastructure Play |
|---|---|---|
| HashiCorp | @hashicorp ยท @mitchellh ยท @armon | Consul already does service discovery. Nomad schedules workloads. Vault secures secrets. What's missing? Convergent model state across your infrastructure |
| etcd | @etcd-io | The consensus backbone of Kubernetes โ crdt-merge adds consensus-free model convergence on top |
| ZooKeeper | @apache/zookeeper | Coordination service โ crdt-merge is the anti-coordination service. Together: choose your consistency model |
| Istio | @istio | Service mesh traffic management โ add model version convergence to your mesh |
| Linkerd | @linkerd ยท @olix0r | Ultra-light service mesh โ ultra-light model convergence at the sidecar level |
๐ฑ Edge, On-Device & Hardware โ Convergence Without Connectivity
@qualcomm ยท Cristiano Amon
3B+ Snapdragon devices running on-device AI. Models fine-tune locally for personalization. crdt-merge enables peer-to-peer model sync between devices โ convergent personalization without cloud dependency. WiFi Direct, Bluetooth mesh, or intermittent connectivity โ the math works on any transport.
@samsung โ Galaxy AI
Galaxy S series on-device models across 100M+ devices. crdt-merge's delta sync minimizes bandwidth โ only changed weights transfer. Galaxy devices that learn locally and converge globally. Samsung's privacy-first AI vision, enabled.
@apple โ Core ML / Apple Intelligence
On-device models across 2B+ Apple devices. When iPhone models fine-tune locally with differential privacy, they need to merge back without a central server. crdt-merge's gossip protocol works on-device. Federated Apple Intelligence, no coordinator required.
@NVIDIA ยท Jensen Huang ยท @jrhuntsman
You build the hardware. We build the convergence layer. NVIDIA NIM + crdt-merge = convergent model composition as a microservice. Every DGX cluster running distributed training should offer distributed convergent merging. TensorRT optimization + convergent merge = production-grade composed models. The merge SDK for CUDA.
@AMD ยท Lisa Su
ROCm + crdt-merge = vendor-agnostic convergent merging. Every AMD MI300X customer running distributed workloads needs the same merge guarantees NVIDIA users get. First-mover advantage in the convergent merging ecosystem on AMD silicon.
@intel โ Intel Gaudi / OpenVINO
Gaudi accelerators + OpenVINO deployment + crdt-merge convergence. The full Intel AI stack, merge-complete. Edge AI on Intel hardware with convergent model sync across the fleet.
@ARM-software โ ARM ยท Rene Haas
ARM's AI ecosystem spans phones to servers. crdt-merge optimized for NEON/SVE = convergent merging at every scale, from Cortex-M microcontrollers to Neoverse server clusters. The universal merge layer for the universal compute architecture.
๐ Autonomous Systems & Robotics โ Safety-Critical Convergence
Tesla ยท @karpathy ยท Elon Musk
Millions of vehicles learning from different driving conditions. crdt-merge enables fleet learning without a central server โ vehicles merge driving models peer-to-peer at charging stations. Convergence guaranteed. Audit trail for NHTSA. The FSD training pipeline that doesn't need Dojo as a single point of failure.
@waymo-research โ Waymo ยท Dmitri Dolgov
Multi-city autonomous driving. Phoenix model + SF model + Austin model converge regardless of merge order. Safety-critical audit trail included. When regulators ask "how was this model built?", you have the mathematically verified answer.
@BostonDynamics ยท @figure-ai ยท @1x-technologies
Robots learning in diverse environments. Each robot learns different manipulation skills. crdt-merge enables skill merging across the fleet โ every robot gains every other robot's skills, convergently. The backbone of scalable embodied intelligence.
@chelseafinlab โ Chelsea Finn ยท Stanford, Meta-Learning
MAML and meta-learning create models that adapt quickly. crdt-merge enables convergent meta-model composition โ meta-learners from different task distributions merge their adaptation capabilities. Meta-learning at fleet scale.
@sergey-levine โ Sergey Levine ยท UC Berkeley, Robot Learning
Offline RL and real-world robot learning produce models across diverse environments. crdt-merge enables convergent policy merging โ robots share learned policies without centralized replay buffers.
๐ก Telecommunications โ Edge AI at Network Scale
Ericsson ยท @Ericsson ยท Nokia ยท @nokia ยท Vodafone ยท @Vodafone
5G edge nodes running AI workloads. crdt-merge enables convergent model sync across cell towers โ each tower's model learns local traffic patterns, converges network-wide via gossip. No central orchestrator. The self-organizing network, realized. Vodafone's pan-European operations get GDPR-compliant convergent merging across national boundaries โ models merge, data stays sovereign.
๐ก๏ธ Defense & Sovereign AI โ Convergence in Contested Environments
DARPA ยท NATO ACT ยท Five Eyes Alliance
Tactical edge AI with classification constraints. crdt-merge enables convergent model merging across security boundaries โ field-level encryption ensures need-to-know. Models converge, secrets don't leak. In contested environments, no coordinator means no single point of failure to target. Coalition AI interoperability with per-nation data sovereignty. crdt-merge's cryptographic remove() and audit trails satisfy the most demanding operational requirements.
๐ญ Industry Verticals โ Every Sector. Every Use Case. One Primitive.
If your industry uses AI models, your industry has the merge problem. The only question is whether you solve it with mathematics or with hope.
๐ฅ Healthcare & Life Sciences
| Who | What crdt-merge Adds |
|---|---|
| Epic Systems ยท Cerner (Oracle) | Merge hospital AI models across facilities โ provable HIPAA compliance for every merge operation |
| Roche ยท Novartis ยท Pfizer | Federated drug discovery โ merge clinical trial models without sharing patient data. Every merge auditable for FDA |
| Tempus ยท @EricLefkofsky | Cancer genomics AI across institutions โ convergent precision medicine models |
| Isomorphic Labs (DeepMind) | AlphaFold + convergent protein model merging across research labs |
๐ฐ Financial Services
| Who | What crdt-merge Adds |
|---|---|
| JPMorgan ยท @jpmorganchase | SOX-compliant model merging โ every merge auditable, every strategy verified |
| Goldman Sachs ยท @goldmansachs | Cross-desk quant model convergence โ merge alpha signals without information leakage |
| Two Sigma ยท Citadel ยท DE Shaw | Federated alpha model merging โ convergent signal combination with IP isolation |
| Bloomberg ยท @bloomberg | Financial NLP model merging across news desks, regions, and asset classes |
| Stripe ยท @stripe ยท Plaid ยท @plaid | Fraud detection model merging โ convergent across payment channels and institutions |
| Swiss Re ยท Munich Re ยท AXA | Actuarial model convergence across business lines and geographies |
โก Energy & Climate
| Who | What crdt-merge Adds |
|---|---|
| Shell ยท BP ยท TotalEnergies ยท Equinor | Cross-site predictive maintenance โ convergent equipment models across global operations |
| Siemens Energy ยท @siemens | Turbine fleet AI โ convergent optimization across thousands of installations |
| Tesla Energy ยท Enphase | Grid-scale battery optimization โ convergent models across distributed storage |
๐ญ Manufacturing & Industry 4.0
| Who | What crdt-merge Adds |
|---|---|
| Siemens ยท Bosch ยท @bosch-ai | Digital twin convergence โ merge factory models across global manufacturing sites |
| FANUC ยท KUKA ยท ABB | Robot arm skill merging โ convergent learned behaviors across production lines |
| John Deere ยท @JohnDeere ยท Caterpillar | Precision agriculture & heavy equipment โ convergent fleet learning |
๐ Retail & E-commerce
| Who | What crdt-merge Adds |
|---|---|
| Amazon ยท @aws | SageMaker + crdt-merge = convergent model composition as a managed service |
| Shopify ยท @Shopify | Merchant AI models โ convergent across 4M+ shops |
| Walmart ยท Target | Supply chain optimization โ convergent demand models across distribution network |
๐ฎ Gaming & Entertainment
| Who | What crdt-merge Adds |
|---|---|
| Unity ยท @Unity-Technologies | ML-Agents convergence โ merge NPC behaviors across game instances |
| Epic Games ยท Roblox ยท @Roblox | Live multiplayer AI โ convergent NPC learning across millions of concurrent players |
| Spotify ยท @spotify ยท Netflix ยท @Netflix | Recommendation model convergence across 600M+ users, regions, and content types |
โ๏ธ Legal & Professional Services
| Who | What crdt-merge Adds |
|---|---|
| Thomson Reuters (Westlaw AI) ยท LexisNexis | Legal research model merging โ convergent across jurisdictions and practice areas |
| Harvey AI | Legal AI fine-tuned per firm โ merge expertise without sharing client data |
| McKinsey ยท BCG ยท Deloitte | Consulting AI โ convergent knowledge models across global engagements |
๐ Education
| Who | What crdt-merge Adds |
|---|---|
| Khan Academy ยท @Khan | Khanmigo AI tutoring โ convergent pedagogical models across subjects |
| Duolingo ยท @duolingo | Language model convergence across 40+ languages |
| Coursera ยท @coursera | Adaptive learning โ convergent student models across courses |
๐ Cybersecurity
| Who | What crdt-merge Adds |
|---|---|
| CrowdStrike ยท @CrowdStrike | Threat detection โ convergent models across endpoints globally, no central target |
| Palo Alto Networks ยท @PaloAltoNetworks | Firewall AI โ convergent threat intelligence without centralizing signatures |
| SentinelOne ยท Darktrace | EDR/NDR โ convergent anomaly detection across deployments |
๐ International AI Initiatives โ Sovereign Convergence
| Initiative | What crdt-merge Enables |
|---|---|
| EU AI Office | The only merge system with built-in EU AI Act compliance โ a regulatory advantage |
| UK AISI ยท @AISafetyInstitute | Safety-verified model merging with mathematical guarantees |
| US AISI (NIST) | AI safety standards โ convergent merging with provenance for federal compliance |
| Japan RIKEN / ABCI | National compute + convergent merging = sovereign AI model composition |
| Saudi SDAIA / NEOM | Middle East AI sovereignty โ convergent merging with data residency guarantees |
| India Bhashini | Multilingual AI convergence across 22 official languages โ one converged model |
| Switzerland EPFL ยท @epfl | Swiss precision + Swiss neutrality = trusted convergent AI infrastructure |
๐ The Arithmetic
| Metric | Value |
|---|---|
| Organizations named above | 300+ |
| Individual researchers & leaders | 150+ |
| Industry verticals | 18 |
| Lines of code | 44,304 |
| Merge strategies, all CRDT-compliant | 26 |
| Architecture layers | 6 |
| Regulatory frameworks supported | 5 |
| Central coordinators required | 0 |
| Convergence failures in 26ร26 matrix | 0 |
๐ฌ The Invitation
If you're a researcher: Let's publish together. The convergent extension of your method deserves a paper.
If you're building a platform: Your users have the merge problem. We have the primitive. Let's integrate.
If you're in a regulated industry: You need compliance. We have it. Built-in, not bolted-on.
If you're building agents: Your agents need shared memory. We've proven it converges.
If you're building hardware: Every accelerator needs a merge SDK. We're hardware-agnostic and ready.
If you're a competitor listed above: We'd rather collaborate than compete. The problem is big enough for all of us.
But the math only goes one way: toward convergence. And we have the math.
๐ Start a Conversation
๐ฌ GitHub Discussions โ Open a thread, tag us, propose an integration
๐ง chi@optitransfer.ch โ Direct line for partnership discussions
โญ Star the Repo โ Signal interest. We notice every star.
๐๏ธ Watch for Updates โ First to know when integrations ship
๐ Read the Full Documentation โ 25+ guides, architecture deep dives, formal proofs
๐ 300+ organizations. 150+ individuals. 18 verticals. 7 distributed systems categories. One unsolved problem. One mathematical solution.
The merge problem is universal. The solution is convergent. The question is: who builds with it first?
Patent Pending UK 2607132.4 ยท ยฉ 2024 Optitransfer ยท Built in Switzerland ๐จ๐ญ
โญ Star ยท ๐๏ธ Watch ยท ๐ฌ Discuss ยท ๐ Docs ยท ๐ Architecture ยท ๐ Demo
Why crdt-merge Is Novel & Disruptive
crdt-merge is the first library to apply formal CRDT mathematics to ML model merging, data integration, and multi-agent AI โ simultaneously.
No other framework provides all three of these guarantees:
| Property | What it means | Why it matters |
|---|---|---|
| Commutativity | merge(A, B) == merge(B, A) |
No coordinator needed โ any node can merge in any order |
| Associativity | merge(merge(A, B), C) == merge(A, merge(B, C)) |
Pairwise gossip converges to the same global state |
| Idempotency | merge(A, A) == A |
Network retries and duplicate messages are harmless |
What makes this disruptive:
- No Parameter Server โ Federated model merging without a central coordinator. Teams merge fine-tuned models peer-to-peer with mathematically guaranteed convergence.
- 26 Merge Strategies โ From simple weighted average to DARE-TIES, Fisher-weighted, and novel spectral methods like STAR and SVD Knot Tying โ all wrapped in CRDT-compliant OR-Set semantics.
- Cross-Domain Unification โ The same
merge()primitive works for DataFrames, ML tensors, agent memory, and knowledge graphs. One theory, one API. - Provenance & Compliance Built In โ Every merge is auditable, reversible (via CRDT
remove()), and GDPR/HIPAA/SOX/EU AI Act compliant out of the box. - Patent Pending (UK 2607132.4) โ The mathematical framework for deterministic model merging via CRDTs is a genuine invention, not incremental improvement.
๐ง AI & ML Model Merging
Merge fine-tuned models from independent teams โ no central server, guaranteed convergence.
| Use Case | Guide | What You'll Learn |
|---|---|---|
| Federated Model Merging | ๐ Guide | CRDTMergeState, peer-to-peer model merge, 26 strategies, gossip convergence |
| Model Merge Strategies | ๐ Guide | SLERP, TIES, DARE, DARE-TIES, Fisher, RegMean, Model Breadcrumbs, and more |
| Strategy ร CRDT Matrix | ๐ Guide | Which strategies satisfy which CRDT properties โ commutativity, associativity, idempotency |
| LoRA Adapter Merging | ๐ Guide | LoRAMerge, LoRAMergeSchema, per-layer strategy selection for adapter fusion |
| Continual Learning | ๐ Guide | ContinualMerge, replay buffers, EWC integration โ merge without catastrophic forgetting |
๐ Data & Records
Merge distributed DataFrames, resolve conflicts deterministically, query merged data with SQL.
| Use Case | Guide | What You'll Learn |
|---|---|---|
| CRDT Fundamentals | ๐ Guide | OR-Set, LWW-Register, G-Counter theory โ the math behind every merge |
| CRDT Primitives | ๐ Guide | Working code for every primitive type โ GCounter, PNCounter, ORSet, LWWMap |
| Verification Toolkit | ๐ Guide | verify_crdt, verify_commutative, property-based testing for your own strategies |
| Merge Strategies | ๐ Guide | LWW, MaxWins, MinWins, UnionSet, Priority, Custom โ pick the right one |
| Schema Evolution | ๐ Guide | Backwards-compatible schema changes across distributed systems |
| MergeQL | ๐ Guide | SQL-like merge interface โ MERGE ... USING strategy ... ON key |
| Probabilistic Analytics | ๐ Guide | HyperLogLog, MinHash, Count-Min Sketch โ approximate analytics over CRDTs |
| Performance Tuning | ๐ Guide | parallel_merge, chunking, DuckDB acceleration, profiling |
๐ Transport & Sync
Move states between nodes efficiently โ gossip, delta sync, Merkle verification.
| Use Case | Guide | What You'll Learn |
|---|---|---|
| Wire Protocol | ๐ Guide | Binary serialization, serialize/deserialize, peek_type โ the bytes on the wire |
| Gossip & Serverless Sync | ๐ Guide | GossipState, peer-to-peer propagation, convergence proofs |
| Delta Sync & Merkle | ๐ Guide | Bandwidth-efficient sync, content-addressed integrity verification |
๐ค Agentic & Context
Multi-agent AI systems with convergent shared memory โ no message ordering required.
| Use Case | Guide | What You'll Learn |
|---|---|---|
| Convergent Multi-Agent AI | ๐ Guide | AgentState, ContextMerge, ContextManifest โ agents that converge without coordination |
| Agentic Memory at Scale | ๐ Guide | ContextBloom, MemorySidecar, budget-bounded merge for large-scale agent systems |
๐ Privacy, Provenance & Compliance
Every merge is auditable, reversible, and regulation-compliant.
| Use Case | Guide | What You'll Learn |
|---|---|---|
| Provenance โ Complete AI | ๐ Guide | AuditLog, AuditedMerge, tamper-evident hash chains |
| Right to Forget | ๐ Guide | CRDT remove(), GDPR Article 17 erasure, model unmerge |
| Privacy-Preserving Merge | ๐ Guide | EncryptedMerge, field-level encryption, RBAC-gated merge |
| Security Hardening | ๐ Guide | Threat model, key rotation, audit log integration |
| Security Guide | ๐ Guide | Encryption backends, StaticKeyProvider, RBAC policy definitions |
| Compliance Guide | ๐ Guide | GDPR Art.5, HIPAA PHI safeguards, SOX controls, EU AI Act alignment |
๐๏ธ Architecture & Research
| Resource | Link | Description |
|---|---|---|
| System Overview | ๐ Overview | 6-layer architecture, 44,304 LOC, 104 modules, design philosophy |
| Layer Map | ๐ Layers | What each layer does, what it depends on, key classes |
| Data Flow | ๐ Data Flow | How data moves through merge โ resolve โ wire โ gossip pipelines |
| Design Decisions | ๐ Decisions | Why OR-Set over LWW-Map, why 6 layers, why no external dependencies in core |
| Dependency Graph | ๐ Dependencies | Module-level dependency analysis โ strict downward-only |
| GDEPA Method | ๐ Research | Graph-Theoretic Dependency & Execution Path Analysis โ novel codebase audit method |
| RREA Method | ๐ Research | Reverse Reachability Entropy Analysis โ information-theoretic code prioritization |
๐ Learning Path
| Step | What | Time |
|---|---|---|
| 1 | CRDT Fundamentals โ OR-Sets, convergence, the math | 15 min |
| 2 | CRDT Primitives Reference โ hands-on with every type | 20 min |
| 3 | Merge Strategies โ pick the right strategy | 10 min |
| 4a | Data path: MergeQL โ Performance Tuning | 30 min |
| 4b | ML path: Federated Model Merging โ LoRA | 30 min |
| 4c | Agent path: Convergent Multi-Agent AI | 20 min |
| 4d | Compliance path: Provenance โ Compliance | 25 min |
Troubleshooting Guide โ common errors and fixes when working with crdt-merge.
crdt-merge v0.9.4 ยท Patent Pending UK 2607132.4 ยท BUSL-1.1 โ Apache 2.0 (2028-03-29)
๐ Flagship ยท ๐ฌ Data Playground ยท ๐ Federation ยท GitHub ยท โญ Star Repo ยท ๐๏ธ Watch ยท ๐ Architecture Deep Dive ยท PyPI ยท pip install crdt-merge