Symptom
Cause: The target still has an old replica (or its snapshots were pruned), so the source can’t send incrementally.
A) Quick checklist (safe for running VMs)
-
Confirm where the VM disk lives (source node)
-
Verify storage mapping (both nodes)
-
Clean the target replica (on TARGET node)
-
Delete the replication job (on SOURCE node)
If it won’t save: check cluster quorum →
pvecm status
(must have quorum to edit).
-
(Optional) Clear stale job state
-
Recreate & seed a new base (SOURCE node, GUI)
-
VM → Replication → Add
-
Target node:
<TARGET_NODE>
-
Target storage:
<STORAGE_ID>
-
Set schedule & Rate limit (helpful during business hours)
-
-
Click Resync (or Run now) → a full initial send runs, then incrementals resume.
B) Verification & logs
C) Notes & gotchas
-
Does this stop the VM? No. The VM keeps running; initial resync just adds I/O & network load.
-
Two defaults gateways / mis-mapped storage IDs can cause weirdness. Ensure the same storage ID is enabled on both nodes and points to an existing ZFS pool on each.
-
Don’t manually delete replication snapshots (those named like
@replicate-<VMID>-0-…
) on either side. -
If the disk bus or storage changes (e.g.,
move_disk
), it’s normal to recreate the replication job.
D) One-copy/paste template (replace ALL placeholders)
E) Prevention tips
-
Keep replication retention ≥ your run interval so common base snapshots stick around.
-
Avoid manual snapshot deletion of
replicate-*
snapshots. -
After storage moves or pool renames, recreate the job.
-
Monitor Datacenter → Tasks and
pvesr status
regularly.
Thank you W3DATA Cloud Tech support Team