-
Notifications
You must be signed in to change notification settings - Fork 39
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
feat: abci state sync #2413
base: v1.8-dev
Are you sure you want to change the base?
feat: abci state sync #2413
Conversation
WalkthroughThe pull request introduces comprehensive state synchronization capabilities to the Dash Drive blockchain platform. It adds new configuration structures, error handling mechanisms, and application traits to support snapshot management. The changes span multiple packages, including Changes
Sequence DiagramsequenceDiagram
participant Client
participant StateSourceApp
participant SnapshotManager
participant GroveDB
Client->>StateSourceApp: list_snapshots()
StateSourceApp->>SnapshotManager: get_snapshots()
SnapshotManager->>GroveDB: Retrieve snapshots
GroveDB-->>SnapshotManager: Return snapshot list
SnapshotManager-->>StateSourceApp: Return snapshots
StateSourceApp-->>Client: Respond with snapshots
Client->>StateSourceApp: offer_snapshot()
StateSourceApp->>SnapshotManager: Create new snapshot session
SnapshotManager->>GroveDB: Prepare for state sync
Client->>StateSourceApp: apply_snapshot_chunk()
StateSourceApp->>SnapshotManager: Apply chunk
SnapshotManager->>GroveDB: Update state
Possibly related PRs
Suggested reviewers
Poem
Thank you for using CodeRabbit. We offer it for free to the OSS community and would appreciate your support in helping us grow. If you find it useful, would you consider giving us a shout-out on your favorite social media? 🪧 TipsChatThere are 3 ways to chat with CodeRabbit:
Note: Be mindful of the bot's finite context window. It's strongly recommended to break down tasks such as reading entire modules into smaller chunks. For a focused discussion, use review comments to chat about specific files and their changes, instead of using the PR comments. CodeRabbit Commands (Invoked using PR comments)
Other keywords and placeholders
CodeRabbit Configuration File (
|
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Actionable comments posted: 12
🧹 Nitpick comments (19)
packages/rs-drive-abci/src/abci/handler/load_snapshot_chunk.rs (2)
27-30
: Include detailed error information when snapshot retrieval failsCurrently, the error handling when retrieving the snapshot at the specified height discards the original error information. Including the underlying error message can provide more context for debugging.
Apply this diff to include the original error details:
- .map_err(|_| AbciError::StateSyncInternalError("load_snapshot_chunk failed".to_string()))? + .map_err(|e| AbciError::StateSyncInternalError(format!("load_snapshot_chunk failed: {}", e)))?
29-30
: Differentiate between not found and other errorsWhen the snapshot is not found, the code returns a generic
StateSyncInternalError
. It would be more appropriate to return an error indicating that the snapshot was not found to distinguish it from other internal errors.Apply this diff to specify a not found error:
.ok_or_else(|| { - AbciError::StateSyncInternalError("load_snapshot_chunk failed".to_string()) + AbciError::StateSyncBadRequest("Snapshot not found at the given height".to_string()) })?;packages/rs-drive-abci/src/abci/app/state_source.rs (1)
86-89
: Include error details in gRPCStatus
for better debuggingWhen mapping errors to
tonic::Status
, consider including more detailed error messages to aid in debugging.Apply this diff to include the error source:
.map_err(|e| tonic::Status::internal(format!("list_snapshots failed: {:?}", e)))Repeat this pattern for other error mappings in the file.
packages/rs-drive-abci/src/abci/handler/offer_snapshot.rs (3)
16-18
: Avoid potential panic when convertingapp_hash
The
try_into()
method can fail if theapp_hash
does not have the expected length. Currently, the error handling uses a generic error message. Include specific details about the expected and actual lengths to improve error clarity.Apply this diff to provide detailed error information:
let request_app_hash: [u8; 32] = request.app_hash.try_into().map_err(|e| { - AbciError::StateSyncBadRequest("offer_snapshot invalid app_hash length".to_string()) + AbciError::StateSyncBadRequest(format!( + "offer_snapshot invalid app_hash length: expected 32 bytes, got {} bytes", + request.app_hash.len() + )) })?;
75-79
: Consolidate duplicate error handling logicThe error handling code for wiping the GroveDB is duplicated in both branches of the conditional. Consider refactoring this into a separate function or closure to reduce code duplication and improve maintainability.
Extract the duplicated code into a helper function:
fn wipe_grovedb(app: &A) -> Result<(), AbciError> { app.platform().drive.grove.wipe().map_err(|e| { AbciError::StateSyncInternalError(format!( "offer_snapshot unable to wipe grovedb: {}", e )) }) }Then, replace the duplicated calls with:
wipe_grovedb(app)?;
69-72
: Clarify error message when a newer snapshot is already being syncedThe error message "offer_snapshot already syncing newest height" might be unclear to users. Consider rephrasing it to more clearly indicate that the offered snapshot is not newer than the current one.
Apply this diff to improve the error message:
return Err(Error::Abci(AbciError::StateSyncBadRequest( - "offer_snapshot already syncing newest height".to_string(), + "Offered snapshot height is not newer than the current syncing snapshot".to_string(), )));packages/rs-drive-abci/src/abci/handler/apply_snapshot_chunk.rs (3)
76-76
: Replaceprintln!
with proper logging macrosUsing
println!
for logging is not recommended in production code. Consider using thetracing
macros for consistent and configurable logging.Apply these diffs to replace
println!
withtracing
macros:For line 76:
-println!("[state_sync] state sync completed. verifying"); +tracing::info!("[state_sync] state sync completed. verifying");For lines 97-100:
-println!( - "[state_sync] incorrect hash in prefix:{:?}", - hex::encode(prefix) -); +tracing::error!( + "[state_sync] incorrect hash in prefix:{:?}", + hex::encode(prefix) +);Also applies to: 97-100
87-91
: Include underlying error details when wrapping errorsCurrently, the error message does not include details from the underlying error, which can make debugging difficult. Including the error message provides more context.
Apply this diff to include the underlying error:
- AbciError::StateSyncInternalError( - "apply_snapshot_chunk unable to verify grovedb".to_string(), - ) + AbciError::StateSyncInternalError(format!( + "apply_snapshot_chunk unable to verify grovedb: {}", + e + ))
71-75
: Include underlying error details when wrapping errorsSimilar to the previous comment, include the error message from
commit_session
to aid in debugging.Apply this diff to include the underlying error:
- AbciError::StateSyncInternalError( - "apply_snapshot_chunk unable to commit session".to_string(), - ) + AbciError::StateSyncInternalError(format!( + "apply_snapshot_chunk unable to commit session: {}", + e + ))packages/rs-drive-abci/src/abci/app/consensus.rs (1)
40-50
: Refactor shared initialization code to avoid duplicationThe
snapshot_manager
initialization code is duplicated in bothConsensusAbciApplication
andFullAbciApplication
. Refactoring this into a common helper function promotes code reuse and maintainability.Consider creating a method in the
Platform
struct or a separate helper function to initialize theSnapshotManager
:impl<C> Platform<C> { fn create_snapshot_manager(&self) -> Result<SnapshotManager, Error> { let checkpoints_path = self .config .state_sync_config .checkpoints_path .to_str() .ok_or_else(|| { Error::InitializationError("Invalid checkpoints path: non-UTF8 characters present".to_string()) })? .to_string(); Ok(SnapshotManager::new( checkpoints_path, self.config.state_sync_config.max_num_snapshots, self.config.state_sync_config.snapshots_frequency, )) } }Then, in
ConsensusAbciApplication::new
:let snapshot_manager = platform.create_snapshot_manager()?;Adjust error handling as needed.
packages/rs-drive-abci/src/abci/app/full.rs (1)
40-50
: Refactor shared initialization code to avoid duplicationThe initialization code for
snapshot_manager
is duplicated across multiple files. Refactoring this code into a shared function improves maintainability.Refer to the suggestion made for
ConsensusAbciApplication
and apply the same refactoring here.packages/rs-drive-abci/tests/strategy_tests/state_sync.rs (1)
255-271
: Fix the elapsed time measurement for per-chunk processingThe
elapsed
variable is calculated usingstart_time.elapsed()
, which measures the total time since the beginning of the test, not the time taken to process the current chunk. To accurately measure the time per chunk, initialize a newInstant
at the start of each loop iteration.Apply this diff to correct the elapsed time measurement:
while let Some(chunk_id) = chunk_queue.pop_front() { + let start_time_chunk = Instant::now(); // ... existing code ... - let elapsed = start_time.elapsed(); + let elapsed = start_time_chunk.elapsed(); // ... existing code ... }packages/rs-drive-abci/src/abci/app/mod.rs (2)
28-32
: Add documentation for the SnapshotManagerApplication trait.The trait would benefit from more detailed documentation explaining:
- Its purpose and responsibilities
- When it should be implemented
- How it relates to state synchronization
/// Platform-based ABCI application pub trait SnapshotManagerApplication { - /// Returns Platform + /// Returns a reference to the SnapshotManager instance + /// + /// This trait provides snapshot management capabilities to ABCI applications, + /// enabling state synchronization features like creating, listing, and managing snapshots. fn snapshot_manager(&self) -> &SnapshotManager; }
52-59
: Add documentation for the SnapshotFetchingApplication trait.The trait would benefit from more detailed documentation explaining:
- Its purpose and responsibilities
- The relationship between snapshot fetching sessions and platform
- Generic parameter constraints
-/// Application that can maintain state sync +/// Application that manages state synchronization through snapshot fetching +/// +/// This trait provides capabilities for maintaining state synchronization by: +/// - Managing snapshot fetching sessions +/// - Accessing the platform instance for state operations +/// +/// Type Parameters: +/// - 'p: Lifetime of the Platform reference +/// - C: Type implementing core RPC functionality pub trait SnapshotFetchingApplication<'p, C> { /// Returns the current snapshot fetching session fn snapshot_fetching_session(&self) -> &RwLock<Option<SnapshotFetchingSession<'p>>>;packages/rs-drive-abci/src/abci/error.rs (1)
57-64
: LGTM! Consider adding documentation examples.The new error variants for state sync are well-structured and appropriately separated between client and server errors. Consider adding documentation examples to illustrate typical error scenarios.
Add documentation examples:
/// Client State sync bad request /// /// # Examples /// ``` /// // Example of a bad request when chunk height is invalid /// StateSyncBadRequest("Invalid chunk height: expected 100, got 50".to_string()) /// ``` #[error("bad request state sync: {0}")] StateSyncBadRequest(String), /// Server State sync bad request /// /// # Examples /// ``` /// // Example of an internal error when snapshot creation fails /// StateSyncInternalError("Failed to create snapshot: IO error".to_string()) /// ``` #[error("internal error state sync: {0}")] StateSyncInternalError(String),packages/rs-drive-abci/src/abci/handler/finalize_block.rs (2)
103-103
: Remove unnecessary clone for primitive type.The
clone()
call is unnecessary forblock_height
as it's likely a primitive type (u64/i64) which implements Copy.- .store(block_height.clone(), Ordering::Relaxed); + .store(block_height, Ordering::Relaxed);
105-114
: Improve snapshot creation error handling and conditional check.The snapshot creation logic can be improved in terms of error handling and Rust idioms.
- if (app.platform().config.state_sync_config.snapshots_enabled) { + if app.platform().config.state_sync_config.snapshots_enabled { app.snapshot_manager() .create_snapshot(&app.platform().drive.grove, block_height as i64) .map_err(|e| { - Error::Execution(ExecutionError::CorruptedDriveResponse(format!( - "Unable to create snapshot:{}", - e - ))) + Error::Execution(ExecutionError::CorruptedDriveResponse( + format!("Unable to create snapshot: {e}") + )) })?; }packages/rs-drive/src/error/drive.rs (1)
191-194
: Enhance snapshot error variant with more specific details.The current
Snapshot
error variant could be more informative by including specific error types or codes.Consider replacing with a more detailed variant:
- /// Error todo - #[error("snapshot error")] - Snapshot(String), + /// Snapshot operation errors + #[error("snapshot error: {kind}: {message}")] + Snapshot { + /// Type of snapshot error + kind: SnapshotErrorKind, + /// Detailed error message + message: String, + },Add the error kind enum:
/// Types of snapshot errors #[derive(Debug, Clone, PartialEq, Eq)] pub enum SnapshotErrorKind { /// Error creating snapshot Creation, /// Error loading snapshot Loading, /// Error applying snapshot chunk ChunkApplication, }packages/rs-drive-abci/.env.local (1)
16-16
: Add a descriptive comment forCHECKPOINTS_PATH
.Consider adding a descriptive comment explaining the purpose of this variable, similar to how
GROVEDB_LATEST_FILE
is documented.+# Path to store ABCI state sync checkpoints CHECKPOINTS_PATH=${DB_PATH}/checkpoints
📜 Review details
Configuration used: CodeRabbit UI
Review profile: CHILL
Plan: Pro
⛔ Files ignored due to path filters (2)
Cargo.lock
is excluded by!**/*.lock
packages/dashmate/templates/platform/drive/tenderdash/config.toml.dot
is excluded by!**/*.dot
📒 Files selected for processing (22)
packages/rs-drive-abci/.env.local
(1 hunks)packages/rs-drive-abci/src/abci/app/consensus.rs
(4 hunks)packages/rs-drive-abci/src/abci/app/full.rs
(4 hunks)packages/rs-drive-abci/src/abci/app/mod.rs
(2 hunks)packages/rs-drive-abci/src/abci/app/state_source.rs
(1 hunks)packages/rs-drive-abci/src/abci/config.rs
(2 hunks)packages/rs-drive-abci/src/abci/error.rs
(1 hunks)packages/rs-drive-abci/src/abci/handler/apply_snapshot_chunk.rs
(1 hunks)packages/rs-drive-abci/src/abci/handler/finalize_block.rs
(3 hunks)packages/rs-drive-abci/src/abci/handler/list_snapshots.rs
(1 hunks)packages/rs-drive-abci/src/abci/handler/load_snapshot_chunk.rs
(1 hunks)packages/rs-drive-abci/src/abci/handler/mod.rs
(1 hunks)packages/rs-drive-abci/src/abci/handler/offer_snapshot.rs
(1 hunks)packages/rs-drive-abci/src/config.rs
(7 hunks)packages/rs-drive-abci/src/platform_types/mod.rs
(1 hunks)packages/rs-drive-abci/src/platform_types/snapshot/mod.rs
(1 hunks)packages/rs-drive-abci/src/server.rs
(2 hunks)packages/rs-drive-abci/tests/strategy_tests/main.rs
(1 hunks)packages/rs-drive-abci/tests/strategy_tests/state_sync.rs
(1 hunks)packages/rs-drive/Cargo.toml
(1 hunks)packages/rs-drive/src/error/drive.rs
(1 hunks)packages/rs-platform-version/Cargo.toml
(1 hunks)
✅ Files skipped from review due to trivial changes (2)
- packages/rs-drive-abci/src/platform_types/mod.rs
- packages/rs-drive-abci/tests/strategy_tests/main.rs
⏰ Context from checks skipped due to timeout of 90000ms (17)
- GitHub Check: Rust packages (dpp) / Unused dependencies
- GitHub Check: Rust packages (dpp) / Tests
- GitHub Check: Rust packages (dpp) / Linting
- GitHub Check: Rust packages (dpp) / Check each feature
- GitHub Check: Rust packages (dash-sdk) / Unused dependencies
- GitHub Check: Rust packages (dash-sdk) / Tests
- GitHub Check: Rust packages (dash-sdk) / Linting
- GitHub Check: Rust packages (dash-sdk) / Check each feature
- GitHub Check: Rust packages (dapi-grpc) / Unused dependencies
- GitHub Check: Rust packages (dapi-grpc) / Tests
- GitHub Check: Rust packages (dapi-grpc) / Linting
- GitHub Check: Rust packages (dapi-grpc) / Check each feature
- GitHub Check: Build Docker images (Drive, drive, drive-abci) / Build Drive image
- GitHub Check: Build Docker images (DAPI, dapi, dapi) / Build DAPI image
- GitHub Check: Build Docker images (Dashmate helper, dashmate-helper, dashmate-helper) / Build Dashmate helper image
- GitHub Check: Build JS packages / Build JS
- GitHub Check: Rust crates security audit
🔇 Additional comments (8)
packages/rs-platform-version/Cargo.toml (1)
14-14
: Consider using a fixed version or commit hash instead of develop branch.Using a git dependency pointing to the
develop
branch can lead to build reproducibility issues and unexpected breaking changes. Consider one of these alternatives:
- Use a specific commit hash for better reproducibility:
grovedb-version = { git = "https://github.com/dashpay/grovedb", rev = "<commit-hash>" }
- Wait for a stable release and use version constraints:
grovedb-version = { version = "^2.2.0" }Let's verify the current state of the grovedb repository:
packages/rs-drive/Cargo.toml (1)
Line range hint
91-116
: Verify feature flags required for ABCI state sync.The PR implements ABCI state sync, but it's unclear which feature flags are required for this functionality. Please:
- Document which feature flags are necessary for state sync
- Consider creating a dedicated feature flag group for state sync functionality
- Ensure all required grovedb features are enabled
Let's analyze the feature usage:
packages/rs-drive-abci/src/abci/handler/load_snapshot_chunk.rs (1)
39-40
: Use consistent GroveDB versioningThe code uses
PlatformVersion::latest().drive.grove_version
to specify the GroveDB version. Ensure that the version used here aligns with the versions used elsewhere in the application to maintain compatibility.packages/rs-drive-abci/tests/strategy_tests/state_sync.rs (1)
71-303
: Test functionrun_state_sync
is well-structured and effectiveThe
run_state_sync
function sets up the environment, runs the chain with the defined strategy, and tests the state synchronization process comprehensively. The use of temporary directories, configuration settings, and performance measurements demonstrates good practices in test implementation.packages/rs-drive-abci/src/abci/app/mod.rs (1)
35-43
: Verify the impact of lifetime parameter change.The lifetime parameter change from 'a to 'p in
TransactionalApplication
trait could affect existing implementations.✅ Verification successful
Lifetime parameter change is safe
The change from 'a to 'p in
TransactionalApplication
trait is a non-breaking change as it only renames the generic lifetime parameter. Existing implementations inConsensusAbciApplication
andFullAbciApplication
already use matching lifetimes, and all trait bounds in handlers are correctly aligned with their function signatures.🏁 Scripts executed
The following scripts were executed for the analysis:
Script:
#!/bin/bash # Search for implementations of TransactionalApplication to verify the impact ast-grep --pattern 'impl$_ TransactionalApplication<$_> for $_'Length of output: 65
Script:
#!/bin/bash # Search for any mentions of TransactionalApplication rg "TransactionalApplication" -A 5 # Also try a more flexible ast-grep pattern ast-grep --pattern 'impl TransactionalApplication' # And check the specific file content cat "packages/rs-drive-abci/src/abci/app/mod.rs"Length of output: 10868
packages/rs-drive-abci/src/abci/handler/mod.rs (1)
38-48
: LGTM! Well-structured module organization.The new snapshot-related modules follow the existing pattern and are properly organized and exported.
Also applies to: 53-62
packages/rs-drive-abci/src/server.rs (1)
35-47
: Consider service registration order impact.While the implementation is correct, consider if the order of service registration matters for the ABCI protocol. Currently, we have:
- CheckTx service
- StateSource service
Verify if this order aligns with the ABCI protocol requirements or if it needs to be documented.
✅ Verification successful
Service registration order appears non-critical
After thorough verification of the codebase, there's no evidence suggesting that the ABCI service registration order is critical. The services appear to operate independently, handling different types of requests. The current implementation is correct as is.
🏁 Scripts executed
The following scripts were executed for the analysis:
Script:
#!/bin/bash # Search for any documentation or comments about ABCI service order requirements rg -i "service.*order" --type md rg -i "abci.*service.*order" --type mdLength of output: 284
Script:
#!/bin/bash # Search for ABCI service implementations and registration patterns ast-grep --pattern 'Server::builder().$$$add_service($$$)' # Look for comments about ABCI services and configuration rg -i "abci.*service" -A 3 -B 3 --type rust --type ts # Search for any configuration or setup related comments rg -i "(abci|service).*configuration" --type rustLength of output: 530
packages/rs-drive-abci/src/abci/config.rs (1)
90-106
: Review environment-specific configurations.The
snapshots_frequency
andmax_num_snapshots
values are identical for local and testnet environments. Consider if these should be different:
- Local environment might benefit from more frequent snapshots for development
- Testnet might need different retention policies
grovedb = { git = "https://github.com/dashpay/grovedb", branch = "develop", optional = true, default-features = false } | ||
grovedb-costs = { git = "https://github.com/dashpay/grovedb", branch = "develop", optional = true } | ||
grovedb-path = { git = "https://github.com/dashpay/grovedb", branch = "develop" } | ||
grovedb-storage = { git = "https://github.com/dashpay/grovedb", branch = "develop", optional = true } | ||
grovedb-version = { git = "https://github.com/dashpay/grovedb", branch = "develop" } | ||
grovedb-epoch-based-storage-flags = { git = "https://github.com/dashpay/grovedb", branch = "develop" } |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
💡 Codebase verification
Pin grovedb dependencies to specific commit for reproducible builds
All grovedb dependencies are using the develop
branch, which can lead to inconsistent versions between builds. To ensure reproducibility and prevent potential compatibility issues:
- Replace all grovedb git dependencies with a specific commit hash:
grovedb = { git = "https://github.com/dashpay/grovedb", rev = "<commit-hash>", optional = true, default-features = false }
grovedb-costs = { git = "https://github.com/dashpay/grovedb", rev = "<commit-hash>", optional = true }
# ... apply same pattern to other grovedb dependencies
- Use the same commit hash across all packages that depend on grovedb to maintain consistency.
🔗 Analysis chain
Ensure consistent versioning across all grovedb dependencies.
All grovedb dependencies are pointing to the develop branch, which could lead to version incompatibilities if different components evolve at different rates. Consider:
- Using the same commit hash across all dependencies:
[dependencies]
grovedb = { git = "https://github.com/dashpay/grovedb", rev = "<commit-hash>", optional = true, default-features = false }
grovedb-costs = { git = "https://github.com/dashpay/grovedb", rev = "<commit-hash>", optional = true }
# ... apply same pattern to other grovedb dependencies
- Using workspace dependencies to ensure consistency:
[workspace.dependencies]
grovedb-common = { git = "https://github.com/dashpay/grovedb", rev = "<commit-hash>" }
[dependencies]
grovedb = { workspace = true, optional = true, default-features = false }
# ... reference workspace for other grovedb dependencies
Let's verify the dependency tree for potential conflicts:
🏁 Scripts executed
The following scripts were executed for the analysis:
Script:
#!/bin/bash
# Check if all grovedb dependencies resolve to the same commit
cargo tree | grep -A 1 'grovedb'
# Look for any dependency conflicts
cargo tree -d | grep 'grovedb'
Length of output: 255
Script:
#!/bin/bash
# Check for Cargo.lock and workspace configurations
fd "Cargo\.(lock|toml)" --type f --exec cat {}
# Look for any workspace configurations
rg "^\[workspace\]" -A 5
# Search for grovedb references in all Cargo.toml files
rg "grovedb.*git.*" --type toml
Length of output: 65647
Script:
#!/bin/bash
# Search for all Cargo.toml files and their content
fd "Cargo.toml" --type f --exec echo "=== {} ===" \; --exec grep -A 10 "grovedb" {} \;
# Look for any workspace dependency declarations
fd "Cargo.toml" --type f --exec grep -l "workspace.dependencies" {} \;
# Search for any commit hashes in git dependencies
fd "Cargo.toml" --type f --exec grep -A 1 "git.*grovedb.*rev\s*=" {} \;
Length of output: 4173
packages/rs-drive-abci/src/abci/handler/apply_snapshot_chunk.rs
Outdated
Show resolved
Hide resolved
/// Configuration for StateSync feature | ||
#[derive(Clone, Debug, Serialize, Deserialize)] | ||
pub struct StateSyncAbciConfig { | ||
/// Enable snapshot | ||
pub snapshots_enabled: bool, | ||
|
||
/// Path to checkpoints | ||
pub checkpoints_path: PathBuf, | ||
|
||
/// Frequency of snapshot creation (in blocks) | ||
pub snapshots_frequency: i64, | ||
|
||
/// Maximum number of snapshots to keep | ||
pub max_num_snapshots: usize, | ||
} |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
🛠️ Refactor suggestion
Add validation for configuration values.
The struct should validate that:
snapshots_frequency
is positivemax_num_snapshots
is greater than zero
+impl StateSyncAbciConfig {
+ pub fn validate(&self) -> Result<(), &'static str> {
+ if self.snapshots_frequency <= 0 {
+ return Err("snapshots_frequency must be positive");
+ }
+ if self.max_num_snapshots == 0 {
+ return Err("max_num_snapshots must be greater than zero");
+ }
+ Ok(())
+ }
+}
📝 Committable suggestion
‼️ IMPORTANT
Carefully review the code before committing. Ensure that it accurately replaces the highlighted code, contains no missing lines, and has no issues with indentation. Thoroughly test & benchmark the code to ensure it meets the requirements.
/// Configuration for StateSync feature | |
#[derive(Clone, Debug, Serialize, Deserialize)] | |
pub struct StateSyncAbciConfig { | |
/// Enable snapshot | |
pub snapshots_enabled: bool, | |
/// Path to checkpoints | |
pub checkpoints_path: PathBuf, | |
/// Frequency of snapshot creation (in blocks) | |
pub snapshots_frequency: i64, | |
/// Maximum number of snapshots to keep | |
pub max_num_snapshots: usize, | |
} | |
/// Configuration for StateSync feature | |
#[derive(Clone, Debug, Serialize, Deserialize)] | |
pub struct StateSyncAbciConfig { | |
/// Enable snapshot | |
pub snapshots_enabled: bool, | |
/// Path to checkpoints | |
pub checkpoints_path: PathBuf, | |
/// Frequency of snapshot creation (in blocks) | |
pub snapshots_frequency: i64, | |
/// Maximum number of snapshots to keep | |
pub max_num_snapshots: usize, | |
} | |
impl StateSyncAbciConfig { | |
pub fn validate(&self) -> Result<(), &'static str> { | |
if self.snapshots_frequency <= 0 { | |
return Err("snapshots_frequency must be positive"); | |
} | |
if self.max_num_snapshots == 0 { | |
return Err("max_num_snapshots must be greater than zero"); | |
} | |
Ok(()) | |
} | |
} |
/// State sync configuration | ||
pub state_sync_config: StateSyncAbciConfig, | ||
|
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Include state_sync_config in deserialization intermediate struct.
The state_sync_config
field is not included in the PlatformConfigIntermediate
struct, which means it won't be deserialized from environment variables or config files.
struct PlatformConfigIntermediate {
// ... other fields ...
+ #[serde(default)]
+ pub state_sync_config: StateSyncAbciConfig,
}
impl<'de> Deserialize<'de> for PlatformConfig {
fn deserialize<D>(deserializer: D) -> Result<PlatformConfig, D::Error>
where
D: Deserializer<'de>,
{
let config = PlatformConfigIntermediate::deserialize(deserializer)?;
Ok(PlatformConfig {
// ... other fields ...
- state_sync_config: StateSyncAbciConfig::default(),
+ state_sync_config: config.state_sync_config,
})
}
}
Also applies to: 281-281
@@ -12,6 +12,12 @@ ABCI_LOG_STDOUT_FORMAT=pretty | |||
ABCI_LOG_STDOUT_COLOR=true | |||
|
|||
DB_PATH=/tmp/db | |||
|
|||
CHECKPOINTS_PATH=${DB_PATH}/checkpoints |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
💡 Codebase verification
The codebase has a proper default production path (/var/lib/dash-platform/data
), but it's being overridden by environment files to use temporary storage:
.env.mainnet
:DB_PATH=/tmp/db
.env.testnet
:DB_PATH=/tmp/db
This configuration will lead to data loss on system reboot. Update the environment files to use persistent storage paths instead of /tmp/db
.
🔗 Analysis chain
Verify database paths for production environments.
Both CHECKPOINTS_PATH
and GROVEDB_LATEST_FILE
inherit from DB_PATH
which is set to /tmp/db
. While this is fine for local development, using /tmp
in production could lead to data loss as it's typically cleared on system reboot.
Run this script to check if there are any production environment files with similar configuration:
Also applies to: 19-19
🏁 Scripts executed
The following scripts were executed for the analysis:
Script:
#!/bin/bash
# Description: Check for production environment files and their DB_PATH configuration
# to ensure they don't use temporary directories.
echo "Searching for environment files:"
fd -t f "^\.env\.(production|prod)" packages/rs-drive-abci/
echo -e "\nChecking DB_PATH in found files:"
rg "^DB_PATH=" packages/rs-drive-abci/.env.*
Length of output: 702
Script:
#!/bin/bash
# Check for deployment documentation and configuration
echo "Searching for deployment-related files and documentation:"
fd -t f "^(deploy|docker|k8s|kubernetes|README)" packages/rs-drive-abci/
echo -e "\nChecking content of found files for database path configuration:"
rg -i "(\bdb[_\s-]path|database[_\s-]path|data[_\s-]dir)" packages/rs-drive-abci/
Length of output: 2638
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Actionable comments posted: 1
♻️ Duplicate comments (1)
packages/rs-drive-abci/src/abci/app/consensus.rs (1)
41-51
:⚠️ Potential issueHandle potential UTF-8 conversion failure
The
to_str()
call on the checkpoints path could fail if the path contains invalid UTF-8 characters.
📜 Review details
Configuration used: CodeRabbit UI
Review profile: CHILL
Plan: Pro
📒 Files selected for processing (2)
packages/rs-drive-abci/src/abci/app/consensus.rs
(3 hunks)packages/rs-drive-abci/src/server.rs
(2 hunks)
⏰ Context from checks skipped due to timeout of 90000ms (4)
- GitHub Check: Build Docker images (Drive, drive, drive-abci) / Build Drive image
- GitHub Check: Build Docker images (DAPI, dapi, dapi) / Build DAPI image
- GitHub Check: Build Docker images (Dashmate helper, dashmate-helper, dashmate-helper) / Build Dashmate helper image
- GitHub Check: Build JS packages / Build JS
🔇 Additional comments (3)
packages/rs-drive-abci/src/abci/app/consensus.rs (3)
24-35
: Good use of lifetime documentation and field organizationThe struct is well-organized with clear documentation for the lifetime parameter. The new fields for snapshot management are logically grouped.
68-82
: Well-structured trait implementationsThe new trait implementations for
SnapshotManagerApplication
andSnapshotFetchingApplication
are clean and follow good separation of concerns.
192-218
: Verify error handling in snapshot operationsThe snapshot-related ABCI methods are properly integrated, but we should verify the error handling in the corresponding handlers.
Let's check the error handling implementation:
✅ Verification successful
Error handling in snapshot operations is robust and comprehensive
All snapshot handlers implement proper error handling with:
- Detailed error messages for debugging
- Proper error propagation and mapping
- Comprehensive coverage of error cases
- Consistent cleanup on errors
🏁 Scripts executed
The following scripts were executed for the analysis:
Script:
#!/bin/bash # Description: Check error handling in snapshot handlers # Expected: Each handler should have proper error handling and mapping # Look for error handling patterns in snapshot handlers rg -A 5 "fn (offer_snapshot|apply_snapshot_chunk|list_snapshots|load_snapshot_chunk)" packages/rs-drive-abci/src/abci/handler/Length of output: 2494
Script:
#!/bin/bash # Check error handling implementation in snapshot handlers # Look for error types and handling patterns # First, check for error type definitions rg "pub.*enum.*Error" packages/rs-drive-abci/src/abci/ # Then check actual implementations in handlers for file in offer_snapshot apply_snapshot_chunk list_snapshots load_snapshot_chunk; do echo "=== $file implementation ===" cat "packages/rs-drive-abci/src/abci/handler/${file}.rs" doneLength of output: 12836
@@ -31,6 +32,7 @@ pub fn start( | |||
|
|||
let check_tx_service = | |||
CheckTxAbciApplication::new(Arc::clone(&platform), Arc::new(check_tx_core_rpc)); | |||
let state_source_service = StateSourceAbciApplication::new(Arc::clone(&platform)); |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Add state source service to the gRPC server
The state_source_service
is instantiated but not added to the gRPC server. This will prevent the state source functionality from being accessible.
Add the service to the gRPC server by modifying the builder chain:
let grpc_server = dapi_grpc::tonic::transport::Server::builder()
.add_service(dapi_grpc::platform::v0::platform_server::PlatformServer::new(query_service))
.add_service(
tenderdash_abci::proto::abci::abci_application_server::AbciApplicationServer::new(
check_tx_service,
),
);
+ .add_service(
+ tenderdash_abci::proto::abci::abci_application_server::AbciApplicationServer::new(
+ state_source_service,
+ ),
+ );
Also applies to: 37-43
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Actionable comments posted: 1
♻️ Duplicate comments (1)
packages/rs-drive-abci/src/abci/app/consensus.rs (1)
41-48
: 🛠️ Refactor suggestionAdd path validation for checkpoints directory
Consider validating that the checkpoints directory exists and is writable before initializing the
SnapshotManager
.
🧹 Nitpick comments (2)
packages/rs-drive-abci/src/abci/app/full.rs (1)
188-214
: Add documentation for snapshot-related ABCI methodsConsider adding documentation comments explaining the purpose and behavior of each snapshot-related method:
offer_snapshot
: When it's called and what it decidesapply_snapshot_chunk
: The chunk application processlist_snapshots
: The listing criteriaload_snapshot_chunk
: The chunk loading process+ /// Handles an incoming snapshot offer from a peer. + /// Returns a decision on whether to accept or reject the snapshot. fn offer_snapshot( &self, request: proto::RequestOfferSnapshot, ) -> Result<proto::ResponseOfferSnapshot, proto::ResponseException> { handler::offer_snapshot(self, request).map_err(error_into_exception) } + /// Applies a chunk of snapshot data to the local state. + /// Returns the result of the chunk application process. fn apply_snapshot_chunk( &self, request: proto::RequestApplySnapshotChunk, ) -> Result<proto::ResponseApplySnapshotChunk, proto::ResponseException> { handler::apply_snapshot_chunk(self, request).map_err(error_into_exception) } + /// Lists available snapshots in the local state. + /// Returns metadata about each available snapshot. fn list_snapshots( &self, request: proto::RequestListSnapshots, ) -> Result<proto::ResponseListSnapshots, proto::ResponseException> { handler::list_snapshots(self, request).map_err(error_into_exception) } + /// Loads a specific chunk of a local snapshot. + /// Returns the requested chunk data for state sync. fn load_snapshot_chunk( &self, request: proto::RequestLoadSnapshotChunk, ) -> Result<proto::ResponseLoadSnapshotChunk, proto::ResponseException> { handler::load_snapshot_chunk(self, request).map_err(error_into_exception) }packages/rs-drive-abci/src/abci/app/state_source.rs (1)
75-97
: Enhance error handling in gRPC method implementations.Currently, all errors in
list_snapshots
andload_snapshot_chunk
are mapped totonic::Status::internal
, which may obscure the nature of the errors. Consider mapping specific error types to more appropriate gRPC status codes to provide clearer feedback to clients.Here is an example of how you might map different error kinds:
async fn list_snapshots( &self, request: tonic::Request<proto::RequestListSnapshots>, ) -> Result<tonic::Response<proto::ResponseListSnapshots>, tonic::Status> { handler::list_snapshots(self, request.into_inner()) .map(tonic::Response::new) - .map_err(|e| tonic::Status::internal(format!("list_snapshots failed: {}", e))) + .map_err(|e| match e { + SomeSpecificErrorType::NotFound => tonic::Status::not_found(e.to_string()), + SomeSpecificErrorType::InvalidInput => tonic::Status::invalid_argument(e.to_string()), + _ => tonic::Status::internal(format!("list_snapshots failed: {}", e)), + }) }Apply similar error handling to
load_snapshot_chunk
:async fn load_snapshot_chunk( &self, request: tonic::Request<proto::RequestLoadSnapshotChunk>, ) -> Result<tonic::Response<proto::ResponseLoadSnapshotChunk>, tonic::Status> { handler::load_snapshot_chunk(self, request.into_inner()) .map(tonic::Response::new) - .map_err(|e| tonic::Status::internal(format!("load_snapshot_chunk failed: {}", e))) + .map_err(|e| match e { + SomeSpecificErrorType::NotFound => tonic::Status::not_found(e.to_string()), + SomeSpecificErrorType::InvalidInput => tonic::Status::invalid_argument(e.to_string()), + _ => tonic::Status::internal(format!("load_snapshot_chunk failed: {}", e)), + }) }This refactoring will help clients better understand the reasons for failures and allow them to handle different error cases appropriately.
📜 Review details
Configuration used: CodeRabbit UI
Review profile: CHILL
Plan: Pro
📒 Files selected for processing (5)
packages/rs-drive-abci/src/abci/app/consensus.rs
(3 hunks)packages/rs-drive-abci/src/abci/app/full.rs
(4 hunks)packages/rs-drive-abci/src/abci/app/state_source.rs
(1 hunks)packages/rs-drive-abci/src/abci/config.rs
(2 hunks)packages/rs-drive-abci/src/platform_types/snapshot/mod.rs
(1 hunks)
🚧 Files skipped from review as they are similar to previous changes (2)
- packages/rs-drive-abci/src/abci/config.rs
- packages/rs-drive-abci/src/platform_types/snapshot/mod.rs
⏰ Context from checks skipped due to timeout of 90000ms (18)
- GitHub Check: Rust packages (drive) / Linting
- GitHub Check: Rust packages (drive) / Formatting
- GitHub Check: Rust packages (dpp) / Unused dependencies
- GitHub Check: Rust packages (dpp) / Tests
- GitHub Check: Rust packages (dpp) / Linting
- GitHub Check: Rust packages (dpp) / Check each feature
- GitHub Check: Rust packages (dash-sdk) / Unused dependencies
- GitHub Check: Rust packages (dash-sdk) / Linting
- GitHub Check: Rust packages (dash-sdk) / Tests
- GitHub Check: Rust packages (dash-sdk) / Check each feature
- GitHub Check: Rust packages (dapi-grpc) / Unused dependencies
- GitHub Check: Rust packages (dapi-grpc) / Tests
- GitHub Check: Rust packages (dapi-grpc) / Linting
- GitHub Check: Rust packages (dapi-grpc) / Check each feature
- GitHub Check: Build Docker images (DAPI, dapi, dapi) / Build DAPI image
- GitHub Check: Build Docker images (Dashmate helper, dashmate-helper, dashmate-helper) / Build Dashmate helper image
- GitHub Check: Build Docker images (Drive, drive, drive-abci) / Build Drive image
- GitHub Check: Build JS packages / Build JS
🔇 Additional comments (7)
packages/rs-drive-abci/src/abci/app/full.rs (2)
Line range hint
1-33
: Clean and well-documented struct definition!The new snapshot-related imports and struct fields are properly organized with clear documentation.
63-77
: Well-implemented trait implementations!The snapshot-related trait implementations are clean and efficient, using references appropriately.
packages/rs-drive-abci/src/abci/app/consensus.rs (3)
24-35
: Great improvements to lifetime parameters!The change from
'a
to'p
makes the code more self-documenting, and the added comment clearly explains the lifetime relationship.
65-79
: Well-structured trait implementations!The snapshot-related trait implementations are clean and consistent with the updated lifetime parameters.
189-215
: Add documentation for snapshot-related ABCI methodsConsider adding documentation comments explaining the purpose and behavior of each snapshot-related method.
packages/rs-drive-abci/src/abci/app/state_source.rs (2)
17-25
: StructStateSourceAbciApplication
is correctly implemented.The struct definition with generic bounds and field declarations is appropriate and follows Rust conventions.
50-63
: Safe initialization ofSnapshotManager
without potential panics.By cloning the
checkpoints_path
and passing it directly toSnapshotManager::new
, the code avoids potential panics that could occur from invalid Unicode conversions. This ensures robust handling of file paths.
let snapshot_manager = SnapshotManager::new( | ||
platform | ||
.config | ||
.state_sync_config | ||
.checkpoints_path.clone(), | ||
platform.config.state_sync_config.max_num_snapshots, | ||
platform.config.state_sync_config.snapshots_frequency, | ||
); |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
🛠️ Refactor suggestion
Add path validation for checkpoints directory
Consider validating that the checkpoints directory exists and is writable before initializing the SnapshotManager
. This would prevent runtime errors during snapshot operations.
let snapshot_manager = SnapshotManager::new(
- platform
- .config
- .state_sync_config
- .checkpoints_path.clone(),
+ {
+ let path = platform.config.state_sync_config.checkpoints_path.clone();
+ std::fs::create_dir_all(&path).map_err(|e| {
+ Error::InitializationError(format!(
+ "Failed to create checkpoints directory: {}",
+ e
+ ))
+ })?;
+ path
+ },
platform.config.state_sync_config.max_num_snapshots,
platform.config.state_sync_config.snapshots_frequency,
);
📝 Committable suggestion
‼️ IMPORTANT
Carefully review the code before committing. Ensure that it accurately replaces the highlighted code, contains no missing lines, and has no issues with indentation. Thoroughly test & benchmark the code to ensure it meets the requirements.
let snapshot_manager = SnapshotManager::new( | |
platform | |
.config | |
.state_sync_config | |
.checkpoints_path.clone(), | |
platform.config.state_sync_config.max_num_snapshots, | |
platform.config.state_sync_config.snapshots_frequency, | |
); | |
let snapshot_manager = SnapshotManager::new( | |
{ | |
let path = platform.config.state_sync_config.checkpoints_path.clone(); | |
std::fs::create_dir_all(&path).map_err(|e| { | |
Error::InitializationError(format!( | |
"Failed to create checkpoints directory: {}", | |
e | |
)) | |
})?; | |
path | |
}, | |
platform.config.state_sync_config.max_num_snapshots, | |
platform.config.state_sync_config.snapshots_frequency, | |
); |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Actionable comments posted: 7
🧹 Nitpick comments (10)
packages/rs-drive-abci/src/abci/handler/apply_snapshot_chunk.rs (5)
38-38
: Extract magic number into a named constant.The hardcoded value
1u16
should be extracted into a named constant with clear documentation of its purpose.+const DEFAULT_SNAPSHOT_CHUNK_VERSION: u16 = 1; + pub fn apply_snapshot_chunk<'a, 'db: 'a, A, C: 'db>( // ... ) { - session.state_sync_info.apply_chunk(..., 1u16, ...) + session.state_sync_info.apply_chunk(..., DEFAULT_SNAPSHOT_CHUNK_VERSION, ...)
76-76
: Use tracing instead of println for consistency.Replace
println!
with the tracing framework for consistent logging throughout the codebase.- println!("[state_sync] state sync completed. verifying"); + tracing::info!("[state_sync] state sync completed. verifying");
92-92
: Use is_empty() instead of comparing len() with zero.For better readability and idiomatic Rust, use
is_empty()
when checking if a collection has no elements.- if incorrect_hashes.len() > 0 { + if !incorrect_hashes.is_empty() {
21-25
: Deduplicate error handling for session locking.The error message for session locking is duplicated. Consider extracting it into a constant or helper function.
const SESSION_LOCK_ERROR: &str = "apply_snapshot_chunk unable to lock session (poisoned)"; fn handle_session_lock_error<T>(result: Result<T, impl std::error::Error>) -> Result<T, Error> { result.map_err(|_| AbciError::StateSyncInternalError(SESSION_LOCK_ERROR.to_string()).into()) }Also applies to: 61-65
77-91
: Improve error context for GroveDB verification.The error message for GroveDB verification could include more context about what failed during verification.
.map_err(|e| { AbciError::StateSyncInternalError( - "apply_snapshot_chunk unable to verify grovedb".to_string(), + format!("apply_snapshot_chunk unable to verify grovedb: {}", e), ) })?;packages/rs-drive-abci/src/platform_types/snapshot/mod.rs (5)
11-13
: Clean up imports and resolve potential duplication.
- Line 11 contains a commented-out import that should be removed
- Both lines 11 and 13 import
PlatformVersion
from different sources, suggesting potential confusion-//use platform_version::version::PlatformVersion;
17-17
: Remove unused constant.The constant
CHUNK_SIZE_16MB
is defined but never used in the code.-const CHUNK_SIZE_16MB: usize = 16 * 1024 * 1024;
36-43
: Remove Default derive and improve documentation.
- The
Default
derive might create invalid instances with empty paths and zero values- The documentation comment is duplicated at line 45
-#[derive(Default, Clone)] +#[derive(Clone)] pub struct SnapshotManager { freq: i64, number_stored_snapshots: usize, checkpoints_path: PathBuf, }
47-55
: Fix copy-pasted documentation comments.Several documentation comments are incorrectly copy-pasted:
- Line 50: "Snapshot accepted" for
app_hash
- Line 53: "Snapshot accepted" for
state_sync_info
/// Snapshot accepted pub snapshot: abci::Snapshot, - /// Snapshot accepted + /// Application hash pub app_hash: Vec<u8>, // sender_metrics: Option<HashMap<String, Metrics>>, - /// Snapshot accepted + /// State synchronization information pub state_sync_info: Pin<Box<MultiStateSyncSession<'db>>>,
205-214
: Add concurrency protection.The
save_snapshots
method should handle concurrent access to prevent race conditions when multiple instances try to update snapshots simultaneously.Consider using a mutex or other synchronization mechanism to protect access to the snapshots data.
📜 Review details
Configuration used: CodeRabbit UI
Review profile: CHILL
Plan: Pro
📒 Files selected for processing (4)
packages/rs-drive-abci/src/abci/config.rs
(2 hunks)packages/rs-drive-abci/src/abci/handler/apply_snapshot_chunk.rs
(1 hunks)packages/rs-drive-abci/src/abci/handler/list_snapshots.rs
(1 hunks)packages/rs-drive-abci/src/platform_types/snapshot/mod.rs
(1 hunks)
🚧 Files skipped from review as they are similar to previous changes (1)
- packages/rs-drive-abci/src/abci/handler/list_snapshots.rs
⏰ Context from checks skipped due to timeout of 90000ms (11)
- GitHub Check: Rust packages (dpp) / Detect immutable structure changes
- GitHub Check: Rust packages (dpp) / Check each feature
- GitHub Check: Rust packages (dash-sdk) / Unused dependencies
- GitHub Check: Rust packages (dash-sdk) / Tests
- GitHub Check: Rust packages (dash-sdk) / Linting
- GitHub Check: Rust packages (dash-sdk) / Check each feature
- GitHub Check: Rust packages (dapi-grpc) / Tests
- GitHub Check: Rust packages (dapi-grpc) / Check each feature
- GitHub Check: Rust packages (dapi-grpc) / Linting
- GitHub Check: Build Docker images (Drive, drive, drive-abci) / Build Drive image
- GitHub Check: Build JS packages / Build JS
🔇 Additional comments (3)
packages/rs-drive-abci/src/abci/config.rs (2)
68-81
: Add validation for configuration values.The struct should validate numeric fields to prevent invalid configurations:
snapshots_frequency
should be positivemax_num_snapshots
should be greater than zeroimpl StateSyncAbciConfig { pub fn validate(&self) -> Result<(), String> { if self.snapshots_frequency <= 0 { return Err("snapshots_frequency must be positive".into()); } if self.max_num_snapshots == 0 { return Err("max_num_snapshots must be greater than zero".into()); } Ok(()) } }
118-122
: 🛠️ Refactor suggestionFollow platform's environment variable naming convention.
The environment variable
CHECKPOINTS_PATH
should follow the platform's naming convention, prefixed withDASH_PLATFORM_
.- std::env::var("CHECKPOINTS_PATH") + std::env::var("DASH_PLATFORM_CHECKPOINTS_PATH")Likely invalid or redundant comment.
packages/rs-drive-abci/src/platform_types/snapshot/mod.rs (1)
21-34
: Well-structured and documented Snapshot entity.The struct has clear field descriptions and appropriate trait derivations.
pub fn default_local() -> Self { | ||
Self { | ||
snapshots_enabled: true, | ||
checkpoints_path: PathBuf::from("/var/lib/dash-platform/data/checkpoints"), | ||
snapshots_frequency: 3, | ||
max_num_snapshots: 10, | ||
} | ||
} | ||
|
||
pub fn default_testnet() -> Self { | ||
Self { | ||
snapshots_enabled: true, | ||
checkpoints_path: PathBuf::from("/var/lib/dash-platform/data/checkpoints"), | ||
snapshots_frequency: 3, | ||
max_num_snapshots: 10, | ||
} | ||
} | ||
|
||
pub fn default_mainnet() -> Self { | ||
Self { | ||
snapshots_enabled: true, | ||
checkpoints_path: PathBuf::from("/var/lib/dash-platform/data/checkpoints"), | ||
snapshots_frequency: 3, | ||
max_num_snapshots: 10, | ||
} | ||
} |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
🛠️ Refactor suggestion
Consolidate duplicate configuration.
The default configurations for local, testnet, and mainnet are identical. Consider:
- Using a single private method to create the base configuration
- Allowing environment-specific overrides only where needed
impl StateSyncAbciConfig {
fn base_config() -> Self {
Self {
snapshots_enabled: true,
checkpoints_path: Self::default_checkpoints_path(),
snapshots_frequency: 3,
max_num_snapshots: 10,
}
}
pub fn default_local() -> Self {
Self::base_config()
}
pub fn default_testnet() -> Self {
Self::base_config()
}
pub fn default_mainnet() -> Self {
Self::base_config()
}
}
refetch_chunks: vec![], // TODO: Check when this is needed | ||
reject_senders: vec![], // TODO: Check when this is needed |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
💡 Codebase verification
Implement chunk refetching and sender rejection for state sync robustness
The empty vectors for refetch_chunks
and reject_senders
should be replaced with proper implementation to:
- Request refetch of corrupted or invalid chunks
- Reject nodes that consistently send invalid chunks
packages/rs-drive-abci/src/abci/handler/apply_snapshot_chunk.rs
: Implement chunk validation and track sender reliability
🔗 Analysis chain
Address TODO comments regarding chunk refetching and sender rejection.
The TODO comments indicate missing implementation details for refetch_chunks
and reject_senders
. These parameters are important for handling corrupted or invalid chunks.
Let's verify if these features are implemented in other state sync implementations:
🏁 Scripts executed
The following scripts were executed for the analysis:
Script:
#!/bin/bash
# Search for refetch_chunks and reject_senders usage in other implementations
rg -A 5 'refetch_chunks|reject_senders' --type rust
Length of output: 1246
Script:
#!/bin/bash
# Find the struct definition and related snapshot handling code
ast-grep --pattern 'struct $_ {
$$$
refetch_chunks: $_,
$$$
}'
# Search for documentation and comments about snapshot handling
rg -B 5 'struct.*Snapshot|fn.*snapshot|//.*snapshot' --type rust
# Look for error handling related to chunks
rg -B 3 'Error.*chunk|chunk.*Error|invalid.*chunk|corrupt.*chunk' --type rust
Length of output: 24974
#[cfg(test)] | ||
mod tests { | ||
use super::*; | ||
use std::fs; | ||
|
||
#[test] | ||
fn test_create_snapshot() { | ||
let test_cases = vec![ | ||
(1000, 1000, vec![1000]), | ||
(1000, 1001, vec![1000, 1001]), | ||
(1000, 1002, vec![1000, 1001, 1002]), | ||
(1000, 1004, vec![1002, 1003, 1004]), | ||
(1000, 1005, vec![1003, 1004, 1005]), | ||
]; | ||
for (start, end, want) in test_cases { | ||
let grove_dir = tempfile::tempdir().unwrap(); | ||
let checkpoints_dir = tempfile::tempdir().unwrap(); | ||
let grove = GroveDb::open(grove_dir.path()).unwrap(); | ||
let manager = | ||
SnapshotManager::new(checkpoints_dir.path().to_str().unwrap().to_string().into(), 3, 1); | ||
for height in start..=end { | ||
manager.create_snapshot(&grove, height).unwrap(); | ||
} | ||
let snapshots = manager.get_snapshots(&grove).unwrap(); | ||
let res: Vec<i64> = snapshots.iter().map(|s| s.height).collect(); | ||
assert_eq!(want, res); | ||
|
||
let paths: Vec<String> = snapshots.iter().map(|s| s.path.to_string()).collect(); | ||
for path in paths { | ||
assert!(Path::new(&path).exists()); | ||
} | ||
fs::remove_dir_all(grove_dir.path()).unwrap(); | ||
} | ||
} | ||
} |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
🛠️ Refactor suggestion
Improve test coverage.
Add tests for:
- Error handling (invalid paths, concurrent access)
- Edge cases (zero/negative freq, zero stored snapshots)
get_snapshot_at_height
functionality
Example test cases to add:
#[test]
fn test_get_snapshot_at_height() {
let grove_dir = tempfile::tempdir().unwrap();
let checkpoints_dir = tempfile::tempdir().unwrap();
let grove = GroveDb::open(grove_dir.path()).unwrap();
let manager = SnapshotManager::new(checkpoints_dir.path().to_str().unwrap().to_string().into(), 3, 1);
// Create some snapshots
for height in 1000..=1002 {
manager.create_snapshot(&grove, height).unwrap();
}
// Test existing height
let snapshot = manager.get_snapshot_at_height(&grove, 1001).unwrap();
assert!(snapshot.is_some());
assert_eq!(snapshot.unwrap().height, 1001);
// Test non-existent height
let snapshot = manager.get_snapshot_at_height(&grove, 999).unwrap();
assert!(snapshot.is_none());
}
#[test]
fn test_invalid_parameters() {
let checkpoints_dir = tempfile::tempdir().unwrap();
let manager = SnapshotManager::new(
checkpoints_dir.path().to_str().unwrap().to_string().into(),
0, // Invalid number_stored_snapshots
-1, // Invalid freq
);
// Assert logs contain error messages
}
let snapshot = Snapshot { | ||
height, | ||
version: SNAPSHOT_VERSION, | ||
path: checkpoint_path.to_str().unwrap().to_string(), | ||
hash: root_hash as [u8; 32], | ||
metadata: vec![], | ||
}; |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Handle non-UTF8 paths safely.
The unwrap()
call on to_str()
could panic with non-UTF8 paths.
let snapshot = Snapshot {
height,
version: SNAPSHOT_VERSION,
- path: checkpoint_path.to_str().unwrap().to_string(),
+ path: checkpoint_path.to_string_lossy().into_owned(),
hash: root_hash as [u8; 32],
metadata: vec![],
};
📝 Committable suggestion
‼️ IMPORTANT
Carefully review the code before committing. Ensure that it accurately replaces the highlighted code, contains no missing lines, and has no issues with indentation. Thoroughly test & benchmark the code to ensure it meets the requirements.
let snapshot = Snapshot { | |
height, | |
version: SNAPSHOT_VERSION, | |
path: checkpoint_path.to_str().unwrap().to_string(), | |
hash: root_hash as [u8; 32], | |
metadata: vec![], | |
}; | |
let snapshot = Snapshot { | |
height, | |
version: SNAPSHOT_VERSION, | |
path: checkpoint_path.to_string_lossy().into_owned(), | |
hash: root_hash as [u8; 32], | |
metadata: vec![], | |
}; |
let root_hash = grove | ||
.root_hash(None, &PlatformVersion::latest().drive.grove_version) | ||
.unwrap() | ||
.map_err(|e| Error::Drive(Drive(DriveError::Snapshot(e.to_string()))))?; | ||
|
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Remove unwrap() to prevent panics.
The unwrap()
call on root_hash
could panic. Handle the error properly.
let root_hash = grove
.root_hash(None, &PlatformVersion::latest().drive.grove_version)
- .unwrap()
.map_err(|e| Error::Drive(Drive(DriveError::Snapshot(e.to_string()))))?;
📝 Committable suggestion
‼️ IMPORTANT
Carefully review the code before committing. Ensure that it accurately replaces the highlighted code, contains no missing lines, and has no issues with indentation. Thoroughly test & benchmark the code to ensure it meets the requirements.
let root_hash = grove | |
.root_hash(None, &PlatformVersion::latest().drive.grove_version) | |
.unwrap() | |
.map_err(|e| Error::Drive(Drive(DriveError::Snapshot(e.to_string()))))?; | |
let root_hash = grove | |
.root_hash(None, &PlatformVersion::latest().drive.grove_version) | |
.map_err(|e| Error::Drive(Drive(DriveError::Snapshot(e.to_string()))))?; | |
pub fn new(checkpoints_path: PathBuf, number_stored_snapshots: usize, freq: i64) -> Self { | ||
if let Err(e) = fs::create_dir_all(checkpoints_path.clone()) { | ||
tracing::error!( | ||
"Failed to create directory {}: {}", | ||
checkpoints_path.display(), | ||
e | ||
); | ||
} | ||
Self { | ||
freq, | ||
number_stored_snapshots, | ||
checkpoints_path, | ||
} | ||
} |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
🛠️ Refactor suggestion
Add validation in constructor.
The new
method should validate its parameters:
freq
should be positivenumber_stored_snapshots
should be greater than zero
pub fn new(checkpoints_path: PathBuf, number_stored_snapshots: usize, freq: i64) -> Self {
+ if freq <= 0 {
+ tracing::error!("Frequency must be positive, got {}", freq);
+ }
+ if number_stored_snapshots == 0 {
+ tracing::error!("Number of stored snapshots must be greater than zero");
+ }
if let Err(e) = fs::create_dir_all(checkpoints_path.clone()) {
📝 Committable suggestion
‼️ IMPORTANT
Carefully review the code before committing. Ensure that it accurately replaces the highlighted code, contains no missing lines, and has no issues with indentation. Thoroughly test & benchmark the code to ensure it meets the requirements.
pub fn new(checkpoints_path: PathBuf, number_stored_snapshots: usize, freq: i64) -> Self { | |
if let Err(e) = fs::create_dir_all(checkpoints_path.clone()) { | |
tracing::error!( | |
"Failed to create directory {}: {}", | |
checkpoints_path.display(), | |
e | |
); | |
} | |
Self { | |
freq, | |
number_stored_snapshots, | |
checkpoints_path, | |
} | |
} | |
pub fn new(checkpoints_path: PathBuf, number_stored_snapshots: usize, freq: i64) -> Self { | |
if freq <= 0 { | |
tracing::error!("Frequency must be positive, got {}", freq); | |
} | |
if number_stored_snapshots == 0 { | |
tracing::error!("Number of stored snapshots must be greater than zero"); | |
} | |
if let Err(e) = fs::create_dir_all(checkpoints_path.clone()) { | |
tracing::error!( | |
"Failed to create directory {}: {}", | |
checkpoints_path.display(), | |
e | |
); | |
} | |
Self { | |
freq, | |
number_stored_snapshots, | |
checkpoints_path, | |
} | |
} |
// TODO: Use Metrics for statistics | ||
struct Metrics { | ||
success: usize, | ||
error: usize, | ||
} | ||
|
||
enum MetricType { | ||
Success, | ||
Error, | ||
} | ||
|
||
impl Metrics { | ||
fn new() -> Self { | ||
Self { | ||
success: 0, | ||
error: 0, | ||
} | ||
} | ||
|
||
fn incr(&mut self, metric: MetricType) { | ||
match metric { | ||
MetricType::Success => self.success += 1, | ||
MetricType::Error => self.error += 1, | ||
} | ||
} | ||
} |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
💡 Codebase verification
Remove unused Metrics implementation
The Metrics struct and its implementation are currently unused in the codebase. Since there's no active metrics tracking, this code should be removed and reintroduced when metrics implementation is actually needed to avoid maintaining unused code.
🔗 Analysis chain
Verify the need for Metrics implementation.
The Metrics implementation is complete but currently unused. Consider either:
- Implementing the metrics tracking now
- Moving this to a separate PR when metrics are needed
🏁 Scripts executed
The following scripts were executed for the analysis:
Script:
#!/bin/bash
# Search for any existing metrics implementations or TODOs
rg -g '!target' -g '*.rs' "TODO.*metrics|metrics.*TODO" .
Length of output: 59
Script:
#!/bin/bash
# Search for Metrics struct usage
ast-grep --pattern 'struct Metrics'
# Search for any metric-related implementations
rg -g '!target' -g '*.rs' -B 2 -A 2 'impl.*Metrics|Metrics::'
# Get more context around the current implementation
rg -g '!target' -g '*.rs' -B 5 -A 5 'struct Metrics'
Length of output: 1572
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Actionable comments posted: 2
🧹 Nitpick comments (3)
packages/rs-drive-abci/src/abci/handler/offer_snapshot.rs (3)
34-39
: Refactor duplicated code to improve maintainabilityThe code for wiping the GroveDB and handling errors is duplicated in both the
if
andelse
blocks. Consider refactoring this repeated code into a helper function to reduce duplication and enhance maintainability.Apply this diff to extract the duplicated code into a helper function:
+fn wipe_grovedb(app: &A) -> Result<(), AbciError> { + app.platform().drive.grove.wipe().map_err(|e| { + AbciError::StateSyncInternalError(format!( + "offer_snapshot unable to wipe grovedb: {}", + e + )) + }) +} // Replace the duplicated code in both blocks with: - app.platform().drive.grove.wipe().map_err(|e| { - AbciError::StateSyncInternalError(format!( - "offer_snapshot unable to wipe grovedb:{}", - e - )) - })?; + wipe_grovedb(app)?;Also applies to: 76-81
40-54
: Refactor duplicated snapshot syncing codeThe code for starting snapshot syncing is repeated in both the
if
andelse
blocks. Refactoring this code into a helper function will reduce duplication and potential errors during future updates.You can create a helper function like this:
+fn start_snapshot_syncing(app: &A, request_app_hash: [u8; 32]) -> Result<_, AbciError> { + app.platform() + .drive + .grove + .start_snapshot_syncing( + request_app_hash, + CURRENT_STATE_SYNC_VERSION, + &PlatformVersion::latest().drive.grove_version, + ) + .map_err(|e| { + AbciError::StateSyncInternalError(format!( + "offer_snapshot unable to start snapshot syncing session: {}", + e + )) + }) +} // Replace the duplicated code in both blocks with: - let state_sync_info = app - .platform() - .drive - .grove - .start_snapshot_syncing( - request_app_hash, - CURRENT_STATE_SYNC_VERSION, - &PlatformVersion::latest().drive.grove_version, - ) - .map_err(|e| { - AbciError::StateSyncInternalError(format!( - "offer_snapshot unable to start snapshot syncing session:{}", - e - )) - })?; + let state_sync_info = start_snapshot_syncing(app, request_app_hash)?;Also applies to: 82-96
17-19
: Enhance error message for invalidapp_hash
lengthThe error message
"offer_snapshot invalid app_hash length"
could be more informative by including the expected and actual lengths, aiding in debugging.Consider updating the error message:
- AbciError::StateSyncBadRequest("offer_snapshot invalid app_hash length".to_string()) + AbciError::StateSyncBadRequest(format!( + "offer_snapshot invalid app_hash length: expected 32 bytes, got {} bytes", + request.app_hash.len() + ))
📜 Review details
Configuration used: CodeRabbit UI
Review profile: CHILL
Plan: Pro
📒 Files selected for processing (1)
packages/rs-drive-abci/src/abci/handler/offer_snapshot.rs
(1 hunks)
⏰ Context from checks skipped due to timeout of 90000ms (18)
- GitHub Check: Rust packages (drive) / Tests
- GitHub Check: Rust packages (drive) / Linting
- GitHub Check: Rust packages (drive) / Formatting
- GitHub Check: Rust packages (dpp) / Unused dependencies
- GitHub Check: Rust packages (dpp) / Tests
- GitHub Check: Rust packages (dpp) / Linting
- GitHub Check: Rust packages (dpp) / Check each feature
- GitHub Check: Rust packages (dash-sdk) / Unused dependencies
- GitHub Check: Rust packages (dash-sdk) / Tests
- GitHub Check: Rust packages (dash-sdk) / Linting
- GitHub Check: Rust packages (dash-sdk) / Check each feature
- GitHub Check: Rust packages (dapi-grpc) / Tests
- GitHub Check: Rust packages (dapi-grpc) / Unused dependencies
- GitHub Check: Rust packages (dapi-grpc) / Check each feature
- GitHub Check: Rust packages (dapi-grpc) / Linting
- GitHub Check: Build Docker images (DAPI, dapi, dapi) / Build DAPI image
- GitHub Check: Build Docker images (Drive, drive, drive-abci) / Build Drive image
- GitHub Check: Build JS packages / Build JS
let mut response = proto::ResponseOfferSnapshot::default(); | ||
response.result = i32::from(response_offer_snapshot::Result::Accept); | ||
Ok(response) |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Ensure consistent setting of response.result
field
In the if
block, response.result
is explicitly set to Accept
:
response.result = i32::from(response_offer_snapshot::Result::Accept);
However, in the else
block, you're returning a default response without setting the result
field. This could lead to inconsistent behavior or incorrect responses to the ABCI client. Ensure that response.result
is set appropriately in both cases.
Apply this diff to set response.result
in the else
block:
let response = proto::ResponseOfferSnapshot::default();
+response.result = i32::from(response_offer_snapshot::Result::Accept);
Ok(response)
Also applies to: 100-101
let session = session_write_guard | ||
.as_mut() | ||
.ok_or(AbciError::StateSyncInternalError( | ||
"offer_snapshot unable to lock session".to_string(), | ||
))?; |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
🛠️ Refactor suggestion
Simplify session retrieval and error handling
Since you're in the else
block where session_write_guard.is_some()
is guaranteed, using ok_or
is unnecessary. You can safely unwrap the Option
or use expect
with a clear message.
Apply this diff to simplify the code:
-let session = session_write_guard
- .as_mut()
- .ok_or(AbciError::StateSyncInternalError(
- "offer_snapshot unable to lock session".to_string(),
- ))?;
+let session = session_write_guard.as_mut().expect(
+ "Session should be present in else block of offer_snapshot"
+);
📝 Committable suggestion
‼️ IMPORTANT
Carefully review the code before committing. Ensure that it accurately replaces the highlighted code, contains no missing lines, and has no issues with indentation. Thoroughly test & benchmark the code to ensure it meets the requirements.
let session = session_write_guard | |
.as_mut() | |
.ok_or(AbciError::StateSyncInternalError( | |
"offer_snapshot unable to lock session".to_string(), | |
))?; | |
let session = session_write_guard.as_mut().expect( | |
"Session should be present in else block of offer_snapshot" | |
); |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Actionable comments posted: 0
🧹 Nitpick comments (4)
packages/rs-drive-abci/src/abci/handler/load_snapshot_chunk.rs (4)
4-4
: Remove commented-out import.The commented-out import
platform_version::version::PlatformVersion
appears to be unused. If it's no longer needed, please remove it to maintain code cleanliness.-//use platform_version::version::PlatformVersion;
19-23
: Consider using a higher logging level for state sync operations.State sync is a critical operation. Consider using
tracing::info!
ortracing::debug!
instead oftrace!
to ensure these important operations are visible in typical logging configurations.- tracing::trace!( + tracing::info!(
24-30
: Improve error messages and fix typo.The error messages could be more descriptive and consistent. Also, there's a typo in "mattched".
- AbciError::StateSyncInternalError("load_snapshot_chunk failed: empty mattched snapshot".to_string()) + AbciError::StateSyncInternalError("load_snapshot_chunk failed: no snapshot found at height".to_string())
44-46
: Add chunk validation before response construction.Consider validating the chunk size and content before constructing the response to prevent potential memory issues or invalid data transmission.
+ // Validate chunk size and content + if chunk.len() > MAX_CHUNK_SIZE { + return Err(AbciError::StateSyncInternalError("Chunk size exceeds maximum allowed size".to_string()).into()); + } let mut response = proto::ResponseLoadSnapshotChunk::default(); response.chunk = chunk;
📜 Review details
Configuration used: CodeRabbit UI
Review profile: CHILL
Plan: Pro
📒 Files selected for processing (1)
packages/rs-drive-abci/src/abci/handler/load_snapshot_chunk.rs
(1 hunks)
⏰ Context from checks skipped due to timeout of 90000ms (20)
- GitHub Check: Rust packages (drive) / Unused dependencies
- GitHub Check: Rust packages (drive) / Tests
- GitHub Check: Rust packages (drive) / Linting
- GitHub Check: Rust packages (drive) / Formatting
- GitHub Check: Rust packages (dpp) / Unused dependencies
- GitHub Check: Rust packages (dpp) / Tests
- GitHub Check: Rust packages (dpp) / Linting
- GitHub Check: Rust packages (dpp) / Check each feature
- GitHub Check: Rust packages (dash-sdk) / Unused dependencies
- GitHub Check: Rust packages (dash-sdk) / Tests
- GitHub Check: Rust packages (dash-sdk) / Linting
- GitHub Check: Rust packages (dash-sdk) / Check each feature
- GitHub Check: Rust packages (dapi-grpc) / Tests
- GitHub Check: Rust packages (dapi-grpc) / Linting
- GitHub Check: Rust packages (dapi-grpc) / Formatting
- GitHub Check: Rust packages (dapi-grpc) / Check each feature
- GitHub Check: Build Docker images (DAPI, dapi, dapi) / Build DAPI image
- GitHub Check: Build Docker images (Drive, drive, drive-abci) / Build Drive image
- GitHub Check: Build Docker images (Dashmate helper, dashmate-helper, dashmate-helper) / Build Dashmate helper image
- GitHub Check: Build JS packages / Build JS
🔇 Additional comments (2)
packages/rs-drive-abci/src/abci/handler/load_snapshot_chunk.rs (2)
11-18
: Well-structured function signature with appropriate generic constraints!The generic design with
SnapshotManagerApplication
andPlatformApplication
trait bounds provides good flexibility while ensuring type safety.
34-43
: Verify version compatibility and conversion safety.Two potential issues to address:
- Converting request.version from i32 to u16 could truncate data.
- Using
PlatformVersion::latest()
might cause compatibility issues with older snapshots.Consider:
- Adding version range validation
- Using the snapshot's original version instead of latest
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Actionable comments posted: 3
♻️ Duplicate comments (2)
packages/rs-drive-abci/src/abci/config.rs (2)
66-81
: 🛠️ Refactor suggestionAdd validation for configuration values.
The struct should validate that:
snapshots_frequency
is positivemax_num_snapshots
is greater than zeroimpl StateSyncAbciConfig { + pub fn validate(&self) -> Result<(), &'static str> { + if self.snapshots_frequency <= 0 { + return Err("snapshots_frequency must be positive"); + } + if self.max_num_snapshots == 0 { + return Err("max_num_snapshots must be greater than zero"); + } + Ok(()) + } }
91-116
: 🛠️ Refactor suggestionConsolidate duplicate configuration.
The default configurations for local, testnet, and mainnet are identical. Consider using a single private method to create the base configuration.
impl StateSyncAbciConfig { fn base_config() -> Self { Self { snapshots_enabled: true, checkpoints_path: Self::default_checkpoints_path(), snapshots_frequency: 10, max_num_snapshots: 100, } } pub fn default_local() -> Self { Self::base_config() } pub fn default_testnet() -> Self { Self::base_config() } pub fn default_mainnet() -> Self { Self::base_config() } }
🧹 Nitpick comments (2)
packages/rs-drive-abci/src/abci/config.rs (1)
118-122
: Use a more specific environment variable name.The environment variable
CHECKPOINTS_PATH
should be prefixed with the application name to avoid conflicts.fn default_checkpoints_path() -> PathBuf { - std::env::var("CHECKPOINTS_PATH") + std::env::var("DASH_PLATFORM_CHECKPOINTS_PATH") .map(PathBuf::from) .unwrap_or_else(|_| PathBuf::from("/var/lib/dash-platform/data/checkpoints")) }packages/rs-drive-abci/tests/strategy_tests/state_sync.rs (1)
156-161
: Document magic numbers in test configuration.The test configuration uses magic numbers without explanation. Consider adding constants with descriptive names or comments explaining these values.
+ // Constants for test configuration + const TEST_SNAPSHOT_FREQUENCY: i64 = 10; // Take snapshot every 10 blocks + const TEST_MAX_SNAPSHOTS: usize = 3; // Keep last 3 snapshots + let local_state_sync_config = StateSyncAbciConfig { snapshots_enabled: true, checkpoints_path: checkpoint_test_directory, - snapshots_frequency: 10, - max_num_snapshots: 3, + snapshots_frequency: TEST_SNAPSHOT_FREQUENCY, + max_num_snapshots: TEST_MAX_SNAPSHOTS, };
📜 Review details
Configuration used: CodeRabbit UI
Review profile: CHILL
Plan: Pro
📒 Files selected for processing (3)
packages/rs-drive-abci/src/abci/config.rs
(2 hunks)packages/rs-drive-abci/src/abci/handler/load_snapshot_chunk.rs
(1 hunks)packages/rs-drive-abci/tests/strategy_tests/state_sync.rs
(1 hunks)
🚧 Files skipped from review as they are similar to previous changes (1)
- packages/rs-drive-abci/src/abci/handler/load_snapshot_chunk.rs
fn get_target_tmp_folder() -> PathBuf { | ||
PathBuf::from("/Users/odysseasg/Downloads/state_sync/").join("tmp") | ||
//get_target_folder().join("tmp") | ||
} |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Remove hardcoded user path.
The function contains a hardcoded user-specific path. Use the commented-out implementation that uses CARGO_MANIFEST_DIR
instead.
fn get_target_tmp_folder() -> PathBuf {
- PathBuf::from("/Users/odysseasg/Downloads/state_sync/").join("tmp")
+ get_target_folder().join("tmp")
}
📝 Committable suggestion
‼️ IMPORTANT
Carefully review the code before committing. Ensure that it accurately replaces the highlighted code, contains no missing lines, and has no issues with indentation. Thoroughly test & benchmark the code to ensure it meets the requirements.
fn get_target_tmp_folder() -> PathBuf { | |
PathBuf::from("/Users/odysseasg/Downloads/state_sync/").join("tmp") | |
//get_target_folder().join("tmp") | |
} | |
fn get_target_tmp_folder() -> PathBuf { | |
get_target_folder().join("tmp") | |
} |
println!( | ||
"source app_hash:{}", | ||
hex::encode( | ||
source_outcome | ||
.abci_app | ||
.platform | ||
.drive | ||
.grove | ||
.root_hash(None, &PlatformVersion::latest().drive.grove_version) | ||
.unwrap() | ||
.unwrap() | ||
) | ||
); | ||
println!( | ||
"target app_hash:{}", | ||
hex::encode( | ||
target_abci_app | ||
.platform() | ||
.drive | ||
.grove | ||
.root_hash(None, &PlatformVersion::latest().drive.grove_version) | ||
.unwrap() | ||
.unwrap() | ||
) | ||
); |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
🛠️ Refactor suggestion
Add explicit assertions for state sync verification.
The test only prints the source and target app hashes without explicitly verifying they match. Add assertions to validate the state sync was successful.
+ let source_hash = source_outcome
+ .abci_app
+ .platform
+ .drive
+ .grove
+ .root_hash(None, &PlatformVersion::latest().drive.grove_version)
+ .unwrap()
+ .unwrap();
+
+ let target_hash = target_abci_app
+ .platform()
+ .drive
+ .grove
+ .root_hash(None, &PlatformVersion::latest().drive.grove_version)
+ .unwrap()
+ .unwrap();
+
+ assert_eq!(
+ source_hash,
+ target_hash,
+ "Source and target app hashes must match after state sync"
+ );
+
println!(
"source app_hash:{}",
- hex::encode(
- source_outcome
- .abci_app
- .platform
- .drive
- .grove
- .root_hash(None, &PlatformVersion::latest().drive.grove_version)
- .unwrap()
- .unwrap()
- )
+ hex::encode(&source_hash)
);
println!(
"target app_hash:{}",
- hex::encode(
- target_abci_app
- .platform()
- .drive
- .grove
- .root_hash(None, &PlatformVersion::latest().drive.grove_version)
- .unwrap()
- .unwrap()
- )
+ hex::encode(&target_hash)
);
📝 Committable suggestion
‼️ IMPORTANT
Carefully review the code before committing. Ensure that it accurately replaces the highlighted code, contains no missing lines, and has no issues with indentation. Thoroughly test & benchmark the code to ensure it meets the requirements.
println!( | |
"source app_hash:{}", | |
hex::encode( | |
source_outcome | |
.abci_app | |
.platform | |
.drive | |
.grove | |
.root_hash(None, &PlatformVersion::latest().drive.grove_version) | |
.unwrap() | |
.unwrap() | |
) | |
); | |
println!( | |
"target app_hash:{}", | |
hex::encode( | |
target_abci_app | |
.platform() | |
.drive | |
.grove | |
.root_hash(None, &PlatformVersion::latest().drive.grove_version) | |
.unwrap() | |
.unwrap() | |
) | |
); | |
let source_hash = source_outcome | |
.abci_app | |
.platform | |
.drive | |
.grove | |
.root_hash(None, &PlatformVersion::latest().drive.grove_version) | |
.unwrap() | |
.unwrap(); | |
let target_hash = target_abci_app | |
.platform() | |
.drive | |
.grove | |
.root_hash(None, &PlatformVersion::latest().drive.grove_version) | |
.unwrap() | |
.unwrap(); | |
assert_eq!( | |
source_hash, | |
target_hash, | |
"Source and target app hashes must match after state sync" | |
); | |
println!( | |
"source app_hash:{}", | |
hex::encode(&source_hash) | |
); | |
println!( | |
"target app_hash:{}", | |
hex::encode(&target_hash) | |
); |
println!( | ||
"best_snapshot height:{} app_hash:{}", | ||
best_snapshot.height, | ||
hex::encode(&best_snapshot.hash) | ||
); | ||
|
||
let target_platform = TestPlatformBuilder::new() | ||
.with_config(config.clone()) | ||
.build_with_mock_rpc(); | ||
let target_abci_app = FullAbciApplication::new(&target_platform); | ||
|
||
let offer_snapshot_request = RequestOfferSnapshot { | ||
snapshot: Some(best_snapshot.clone()), | ||
app_hash: best_snapshot.hash.to_vec(), | ||
}; | ||
|
||
let _ = target_abci_app | ||
.offer_snapshot(offer_snapshot_request) | ||
.expect("should offer_snapshot succeed"); | ||
|
||
let mut chunk_queue: VecDeque<Vec<u8>> = VecDeque::new(); | ||
chunk_queue.push_back(best_snapshot.hash.to_vec()); | ||
|
||
let start_time = Instant::now(); | ||
|
||
let mut duration_sum_fetch: Duration = Duration::ZERO; | ||
let mut duration_sum_apply: Duration = Duration::ZERO; | ||
|
||
let mut chunk_counter = 0; | ||
let mut ops_counter = 0; | ||
while let Some(chunk_id) = chunk_queue.pop_front() { | ||
let request_load_chunk = RequestLoadSnapshotChunk { | ||
height: best_snapshot.height, | ||
version: best_snapshot.version, | ||
chunk_id: chunk_id.clone(), | ||
}; | ||
let start_time_fetch = Instant::now(); | ||
let load_chunk_response = source_outcome | ||
.abci_app | ||
.load_snapshot_chunk(request_load_chunk) | ||
.expect("should fetch chunk"); | ||
duration_sum_fetch += start_time_fetch.elapsed(); | ||
|
||
let request_apply_chunk = RequestApplySnapshotChunk { | ||
chunk_id, | ||
chunk: load_chunk_response.chunk, | ||
..Default::default() | ||
}; | ||
let request_apply_num_ops = request_apply_chunk.chunk.len(); | ||
ops_counter += request_apply_num_ops; | ||
|
||
let elapsed = start_time.elapsed(); | ||
let chunk_id_hex = hex::encode(&request_apply_chunk.chunk_id); | ||
let start_time_apply = Instant::now(); | ||
let apply_chunk_response = target_abci_app | ||
.apply_snapshot_chunk(request_apply_chunk) | ||
.expect("should apply chunk succeed"); | ||
duration_sum_apply += start_time_apply.elapsed(); | ||
println!( | ||
"#{} apply:{} num_ops:{} returned:{} queue:{} {:.2?}", | ||
chunk_counter, | ||
chunk_id_hex, | ||
request_apply_num_ops, | ||
apply_chunk_response.next_chunks.len(), | ||
chunk_queue.len(), | ||
elapsed | ||
); | ||
chunk_queue.extend(apply_chunk_response.next_chunks); | ||
chunk_counter += 1; | ||
} | ||
println!("total chunks:{} ops:{}", chunk_counter, ops_counter); | ||
println!("duration_sum_fetch: {}", duration_sum_fetch.as_secs_f64()); | ||
println!("duration_sum_apply: {}", duration_sum_apply.as_secs_f64()); | ||
|
||
println!( | ||
"source app_hash:{}", | ||
hex::encode( | ||
source_outcome | ||
.abci_app | ||
.platform | ||
.drive | ||
.grove | ||
.root_hash(None, &PlatformVersion::latest().drive.grove_version) | ||
.unwrap() | ||
.unwrap() | ||
) | ||
); | ||
println!( | ||
"target app_hash:{}", | ||
hex::encode( | ||
target_abci_app | ||
.platform() | ||
.drive | ||
.grove | ||
.root_hash(None, &PlatformVersion::latest().drive.grove_version) | ||
.unwrap() | ||
.unwrap() | ||
) | ||
); | ||
} |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
🛠️ Refactor suggestion
Add cleanup after test.
The test creates temporary directories but doesn't clean them up after the test completes. Add cleanup in a Drop
implementation or use a cleanup function.
#[test]
fn run_state_sync() {
+ struct Cleanup {
+ base_dir: PathBuf,
+ }
+
+ impl Drop for Cleanup {
+ fn drop(&mut self) {
+ if let Err(e) = remove_dir(&self.base_dir) {
+ eprintln!("Failed to cleanup test directories: {}", e);
+ }
+ }
+ }
+
let platform_version = PlatformVersion::latest();
// ... rest of the test setup ...
+
+ let _cleanup = Cleanup {
+ base_dir: base_test_directory.clone(),
+ };
// ... rest of the test ...
}
Committable suggestion skipped: line range outside the PR's diff.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Actionable comments posted: 0
🧹 Nitpick comments (5)
packages/rs-drive-abci/src/abci/handler/apply_snapshot_chunk.rs (5)
21-25
: Extract duplicate error message into a constant.The error message "apply_snapshot_chunk unable to lock session (poisoned)" is duplicated. Consider extracting it into a constant for better maintainability.
+const LOCK_SESSION_ERROR: &str = "apply_snapshot_chunk unable to lock session (poisoned)"; + let mut session_write_guard = app.snapshot_fetching_session().write().map_err(|_| { - AbciError::StateSyncInternalError( - "apply_snapshot_chunk unable to lock session (poisoned)".to_string(), - ) + AbciError::StateSyncInternalError(LOCK_SESSION_ERROR.to_string()) })?; // ... later in the code ... let session = session_write_guard .take() .ok_or(AbciError::StateSyncInternalError( - "apply_snapshot_chunk unable to lock session (poisoned)".to_string(), + LOCK_SESSION_ERROR.to_string(), ))?;Also applies to: 63-65
38-38
: Document the purpose of the hardcoded chunk format version.The hardcoded value
1u16
appears to be a chunk format version. Please add a comment explaining its purpose or consider making it a named constant.+// Chunk format version 1 represents the initial version of the snapshot chunk format session .state_sync_info .apply_chunk( &app.platform().drive.grove, &request.chunk_id, request.chunk, - 1u16, + CHUNK_FORMAT_VERSION, &PlatformVersion::latest().drive.grove_version, )
76-76
: Replace println! with tracing macro for consistent logging.Use the tracing macro for consistency with the rest of the codebase.
-println!("[state_sync] state sync completed. verifying"); +tracing::info!("[state_sync] state sync completed. verifying");
92-97
: Simplify the length comparison and improve error message.The length comparison can be simplified, and the error message could include the incorrect hashes for better debugging.
-if incorrect_hashes.len() > 0 { +if !incorrect_hashes.is_empty() { Err(AbciError::StateSyncInternalError(format!( - "apply_snapshot_chunk grovedb verification failed with {} incorrect hashes", - incorrect_hashes.len() + "apply_snapshot_chunk grovedb verification failed with {} incorrect hashes: {:?}", + incorrect_hashes.len(), + incorrect_hashes )))?; }
70-75
: Include original error in StateSyncInternalError.The error handling for
commit_session
discards the original error information. Include it in the error message for better debugging..commit_session(state_sync_info) .map_err(|e| { AbciError::StateSyncInternalError( - "apply_snapshot_chunk unable to commit session".to_string(), + format!("apply_snapshot_chunk unable to commit session: {}", e) ) })?;
📜 Review details
Configuration used: CodeRabbit UI
Review profile: CHILL
Plan: Pro
📒 Files selected for processing (1)
packages/rs-drive-abci/src/abci/handler/apply_snapshot_chunk.rs
(1 hunks)
🔇 Additional comments (2)
packages/rs-drive-abci/src/abci/handler/apply_snapshot_chunk.rs (2)
1-14
: LGTM! Well-structured function signature with proper trait bounds.The function signature is well-designed with appropriate lifetime bounds and trait constraints for the application parameter.
53-54
: Implement chunk refetching and sender rejection for state sync robustness.The empty vectors for
refetch_chunks
andreject_senders
should be replaced with proper implementation to:
- Request refetch of corrupted or invalid chunks
- Reject nodes that consistently send invalid chunks
Issue being fixed or feature implemented
Implementation of state sync abci applications.
What was done?
SnapshotManager
and its related configuration inPlatformConfig
StateSourceAbciApplication
for all the replication reading operations (for the source nodes)ConsensusAbciApplication
FullAbciApplication
for testsHow Has This Been Tested?
New strategy test. Comparing root hash, verifying grovedb after state sync is complete
Breaking Changes
to be determined
Checklist:
For repository code-owners and collaborators only
Summary by CodeRabbit
Summary by CodeRabbit
Based on the comprehensive changes, here are the release notes:
New Features
StateSourceAbciApplication
service to handle additional state source requests.Configuration
Performance
Compatibility
Testing