Checkpoint sync state on all cancellation types#694
Conversation
…eeded When a Temporal worker shuts down (SIGTERM during rolling deploy), Sync() returned immediately without checkpointing in-memory sync progress. The sync loop only checkpointed on DeadlineExceeded (run duration timeout), silently losing up to 10s of progress on other cancellation causes like worker shutdown. Use a background context with 30s timeout for the checkpoint since the caller's context may already be cancelled, matching the pattern in c1file.Close(). Co-Authored-By: Claude Opus 4.6 <noreply@anthropic.com>
|
No actionable comments were generated in the recent review. 🎉 WalkthroughThe sync package's cancellation handling now ensures checkpoints persist before exiting when context cancellation occurs. A new test validates that syncs can resume from checkpoints after mid-operation cancellation. Changes
Estimated code review effort🎯 3 (Moderate) | ⏱️ ~22 minutes Poem
🚥 Pre-merge checks | ✅ 2 | ❌ 1❌ Failed checks (1 warning)
✅ Passed checks (2 passed)
✏️ Tip: You can configure your own custom pre-merge checks in the settings. ✨ Finishing Touches
🧪 Generate unit tests (beta)
Comment |
…eeded When the sync context is cancelled for reasons other than DeadlineExceeded (e.g. worker shutdown via SIGTERM), the sync loop exited without checkpointing. This lost all progress made since the last periodic checkpoint, forcing a full re-sync of those pages on the next attempt. Now checkpoint on all cancellation types using a background context with a 30s timeout, since the caller's context is already cancelled. Also broaden the cleanup error handler in syncer.go to treat any context cancellation (not just DeadlineExceeded) as ErrSyncNotComplete. This is a rewrite of #694 against the current codebase, which was refactored from a single sync loop into sequentialSync/parallelSync. Co-Authored-By: Claude Opus 4.6 (1M context) <noreply@anthropic.com>
|
Superseded by #738 which rewrites this fix against current main (the sync loop was refactored into sequentialSync/parallelSync since this branch was created). |
Pull request was closed
…eeded When the sync context is cancelled for reasons other than DeadlineExceeded (e.g. worker shutdown via SIGTERM), the sync loop exited without checkpointing. This lost all progress made since the last periodic checkpoint, forcing a full re-sync of those pages on the next attempt. Now checkpoint on all cancellation types using a background context with a 30s timeout, since the caller's context is already cancelled. Also broaden the cleanup error handler in syncer.go to treat any context cancellation (not just DeadlineExceeded) as ErrSyncNotComplete. This is a rewrite of #694 against the current codebase, which was refactored from a single sync loop into sequentialSync/parallelSync. Co-Authored-By: Claude Opus 4.6 (1M context) <noreply@anthropic.com>
Summary
DeadlineExceeded(run duration timeout)context.WithTimeout(context.Background(), 30s)for the checkpoint since the caller's context may already be cancelled — same pattern asc1file.Close()minCheckpointInterval) is silently lost on every rolling deployContext
Combined with the c1-side
WorkerStopTimeoutfix (c1 PR #14204), this ensures the full graceful shutdown chain works:Sync()checkpoints state before returning (this PR)syncer.Close()→ WAL checkpoint + local saveC1zmanager.SaveC1Z()→ uploads to S3 within the 25s grace periodTest plan
TestCancellationCheckpoints— cancels context mid-sync, verifies checkpoint persists, resumes successfullypkg/sync/...tests pass🤖 Generated with Claude Code
Summary by CodeRabbit
Bug Fixes
Tests