Multi-master vector replication with quorum writes, vector clocks, and automatic conflict resolution.
ruvector-replication = "0.1.1"When your vector database runs on more than one node, you need a way to keep data in sync without losing writes or slowing down queries. ruvector-replication handles that: it replicates vectors across nodes, resolves conflicts automatically, and lets you trade off consistency versus speed per-write. It plugs into the RuVector ecosystem alongside Raft consensus and auto-sharding.
| Single-node vector DB | ruvector-replication | |
|---|---|---|
| Availability | One node goes down, everything stops | Replicas serve reads and accept writes |
| Write scaling | One writer | Multi-master -- write to any node |
| Conflict handling | N/A | Vector clocks, last-write-wins, or CRDTs |
| Consistency control | N/A | Per-write: One, Quorum, or All |
| Sync efficiency | N/A | Incremental deltas with compression |
| Recovery | Manual restore from backup | Automatic replica recovery |
use ruvector_replication::{Replicator, ReplicationConfig, ConsistencyLevel};
#[tokio::main]
async fn main() -> Result<(), Box<dyn std::error::Error>> {
let config = ReplicationConfig {
replication_factor: 3,
consistency_level: ConsistencyLevel::Quorum,
sync_interval: Duration::from_millis(100),
batch_size: 1000,
compression: true,
..Default::default()
};
let replicator = Replicator::new(config).await?;
replicator.start().await?;
Ok(())
}| Feature | What It Does | Why It Matters |
|---|---|---|
| Multi-master replication | Write to any node in the cluster | No single point of failure for writes |
| Configurable consistency | Choose One, Quorum, or All per write | Trade latency for safety on a per-operation basis |
| Vector clock conflict resolution | Track causal ordering across nodes | Detect and resolve concurrent writes correctly |
| CRDT support | Conflict-free replicated data types | Guaranteed convergence without coordination |
| Change streams | Real-time replication event stream | Monitor sync status and react to changes |
| Incremental sync with compression | Only send deltas, compressed on the wire | Minimize bandwidth between nodes |
| Automatic recovery | Replicas catch up after failures | No manual intervention on node restart |
| Bandwidth throttling | Cap replication throughput | Protect production traffic from replication storms |
use ruvector_replication::{Replicator, WriteOptions};
// Write with quorum consistency
let options = WriteOptions {
consistency: ConsistencyLevel::Quorum,
timeout: Duration::from_secs(5),
};
replicator.write(vector_entry, options).await?;
// Write with eventual consistency (faster)
let options = WriteOptions {
consistency: ConsistencyLevel::One,
..Default::default()
};
replicator.write(vector_entry, options).await?;// Get replication lag
let lag = replicator.lag().await?;
println!("Replication lag: {:?}", lag);
// Get replica status
for replica in replicator.replicas().await? {
println!("{}: {} (lag: {}ms)",
replica.id,
replica.status,
replica.lag_ms
);
}
// Subscribe to replication events
let mut stream = replicator.events().await?;
while let Some(event) = stream.next().await {
match event {
ReplicationEvent::Synced { node_id, entries } => {
println!("Synced {} entries to {}", entries, node_id);
}
ReplicationEvent::Conflict { key, resolution } => {
println!("Conflict on {}: {:?}", key, resolution);
}
_ => {}
}
}// Replication configuration
pub struct ReplicationConfig {
pub replication_factor: usize,
pub consistency_level: ConsistencyLevel,
pub sync_interval: Duration,
pub batch_size: usize,
pub compression: bool,
pub conflict_resolution: ConflictResolution,
}
// Consistency levels
pub enum ConsistencyLevel {
One, // Write to one replica
Quorum, // Write to majority
All, // Write to all replicas
}
// Conflict resolution strategies
pub enum ConflictResolution {
LastWriteWins,
VectorClock,
Custom(Box<dyn ConflictResolver>),
}
// Replica information
pub struct ReplicaInfo {
pub id: NodeId,
pub status: ReplicaStatus,
pub lag_ms: u64,
pub last_sync: DateTime<Utc>,
}impl Replicator {
pub async fn new(config: ReplicationConfig) -> Result<Self>;
pub async fn start(&self) -> Result<()>;
pub async fn stop(&self) -> Result<()>;
// Write operations
pub async fn write(&self, entry: VectorEntry, options: WriteOptions) -> Result<()>;
pub async fn write_batch(&self, entries: Vec<VectorEntry>, options: WriteOptions) -> Result<()>;
// Monitoring
pub async fn lag(&self) -> Result<Duration>;
pub async fn replicas(&self) -> Result<Vec<ReplicaInfo>>;
pub async fn events(&self) -> Result<impl Stream<Item = ReplicationEvent>>;
// Management
pub async fn add_replica(&self, node_id: NodeId) -> Result<()>;
pub async fn remove_replica(&self, node_id: NodeId) -> Result<()>;
pub async fn force_sync(&self, node_id: NodeId) -> Result<()>;
}┌─────────────────────────────────────────────────────────┐
│ Replication Flow │
│ │
│ Client │
│ │ │
│ ▼ │
│ ┌──────────┐ Quorum Write ┌──────────┐ │
│ │ Primary │────────────────────▶│ Replica 1│ │
│ │ │ │ │ │
│ │ Vectors │────────────────────▶│ Vectors │ │
│ └──────────┘ └──────────┘ │
│ │ │
│ │ Async Replication │
│ └──────────────────────────▶┌──────────┐ │
│ │ Replica 2│ │
│ │ │ │
│ │ Vectors │ │
│ └──────────┘ │
└─────────────────────────────────────────────────────────┘
- ruvector-core - Core vector database engine
- ruvector-cluster - Clustering and sharding
- ruvector-raft - Raft consensus
- Main README - Complete project overview
- API Documentation - Full API reference
- GitHub Repository - Source code
MIT License - see LICENSE for details.