feat(tests): split large test file up into smaller ones
This commit is contained in:
parent
ab4e5cf38c
commit
0dd0b5f0af
|
|
@ -1,943 +0,0 @@
|
|||
use std::sync::Arc;
|
||||
use readur::{
|
||||
AppState,
|
||||
models::{CreateWebDAVDirectory, User, AuthProvider},
|
||||
services::webdav::{SmartSyncService, SmartSyncStrategy, SmartSyncDecision, WebDAVService, WebDAVConfig},
|
||||
test_utils::{TestContext, TestAuthHelper},
|
||||
};
|
||||
|
||||
/// Mock WebDAV service for testing smart sync scenarios
|
||||
#[derive(Clone)]
|
||||
struct MockWebDAVService {
|
||||
directories: std::collections::HashMap<String, String>, // path -> etag
|
||||
files: Vec<readur::models::FileIngestionInfo>,
|
||||
}
|
||||
|
||||
impl MockWebDAVService {
|
||||
fn new() -> Self {
|
||||
Self {
|
||||
directories: std::collections::HashMap::new(),
|
||||
files: Vec::new(),
|
||||
}
|
||||
}
|
||||
|
||||
fn with_directory_structure(directories: Vec<(String, String)>) -> Self {
|
||||
let mut service = Self::new();
|
||||
for (path, etag) in directories {
|
||||
service.directories.insert(path, etag);
|
||||
}
|
||||
service
|
||||
}
|
||||
|
||||
async fn discover_files_and_directories_mock(
|
||||
&self,
|
||||
_path: &str,
|
||||
_recursive: bool,
|
||||
) -> anyhow::Result<readur::services::webdav::discovery::WebDAVDiscoveryResult> {
|
||||
let directories: Vec<readur::models::FileIngestionInfo> = self.directories
|
||||
.iter()
|
||||
.map(|(path, etag)| readur::models::FileIngestionInfo {
|
||||
path: path.clone(),
|
||||
name: path.split('/').last().unwrap_or("").to_string(),
|
||||
size: 0,
|
||||
mime_type: "".to_string(),
|
||||
last_modified: Some(chrono::Utc::now()),
|
||||
etag: etag.clone(),
|
||||
is_directory: true,
|
||||
created_at: Some(chrono::Utc::now()),
|
||||
permissions: Some(0),
|
||||
owner: None,
|
||||
group: None,
|
||||
metadata: None,
|
||||
})
|
||||
.collect();
|
||||
|
||||
Ok(readur::services::webdav::discovery::WebDAVDiscoveryResult {
|
||||
files: self.files.clone(),
|
||||
directories,
|
||||
})
|
||||
}
|
||||
}
|
||||
|
||||
use tokio::sync::OnceCell;
|
||||
|
||||
static TEST_CONTEXT: OnceCell<TestContext> = OnceCell::const_new();
|
||||
|
||||
/// Helper function to create test database and user using shared TestContext
|
||||
async fn create_test_setup() -> (Arc<AppState>, User) {
|
||||
// Get or create shared test context to avoid multiple database containers
|
||||
let test_context = TEST_CONTEXT.get_or_init(|| async {
|
||||
TestContext::new().await
|
||||
}).await;
|
||||
|
||||
let auth_helper = TestAuthHelper::new(test_context.app().clone());
|
||||
let test_user = auth_helper.create_test_user().await;
|
||||
|
||||
// Convert TestUser to User model for compatibility
|
||||
let user = User {
|
||||
id: test_user.user_response.id,
|
||||
username: test_user.user_response.username,
|
||||
email: test_user.user_response.email,
|
||||
password_hash: Some("hashed_password".to_string()),
|
||||
role: test_user.user_response.role,
|
||||
created_at: chrono::Utc::now(),
|
||||
updated_at: chrono::Utc::now(),
|
||||
oidc_subject: None,
|
||||
oidc_issuer: None,
|
||||
oidc_email: None,
|
||||
auth_provider: AuthProvider::Local,
|
||||
};
|
||||
|
||||
(test_context.state().clone(), user)
|
||||
}
|
||||
|
||||
/// Helper function to create WebDAV service for testing
|
||||
fn create_test_webdav_service() -> WebDAVService {
|
||||
let config = WebDAVConfig {
|
||||
server_url: "https://test.example.com".to_string(),
|
||||
username: "test".to_string(),
|
||||
password: "test".to_string(),
|
||||
watch_folders: vec!["/Documents".to_string()],
|
||||
file_extensions: vec!["pdf".to_string(), "txt".to_string()],
|
||||
timeout_seconds: 30,
|
||||
server_type: Some("generic".to_string()),
|
||||
};
|
||||
|
||||
WebDAVService::new(config).expect("Failed to create WebDAV service")
|
||||
}
|
||||
|
||||
#[tokio::test]
|
||||
async fn test_first_time_sync_full_deep_scan() {
|
||||
// Test Scenario 1: First-time sync with no existing directory ETags
|
||||
// Expected: Should perform full deep scan and establish directory ETag baseline
|
||||
|
||||
let (state, user) = create_test_setup().await;
|
||||
let smart_sync_service = SmartSyncService::new(state.clone());
|
||||
let webdav_service = create_test_webdav_service();
|
||||
|
||||
// Verify no existing directories in database
|
||||
let existing_dirs = state.db.list_webdav_directories(user.id).await
|
||||
.expect("Failed to list directories");
|
||||
assert!(existing_dirs.is_empty(), "Database should start with no tracked directories");
|
||||
|
||||
// Test smart sync evaluation for first-time scenario
|
||||
let decision = smart_sync_service.evaluate_sync_need(user.id, &webdav_service, "/Documents").await
|
||||
.expect("Smart sync evaluation failed");
|
||||
|
||||
match decision {
|
||||
SmartSyncDecision::RequiresSync(SmartSyncStrategy::FullDeepScan) => {
|
||||
// This is expected for first-time sync
|
||||
println!("✅ First-time sync correctly identified need for full deep scan");
|
||||
}
|
||||
other => panic!("Expected FullDeepScan strategy for first-time sync, got: {:?}", other),
|
||||
}
|
||||
|
||||
// Simulate performing the deep scan (this would normally interact with real WebDAV)
|
||||
// For testing, we'll directly save some directory ETags to verify the tracking works
|
||||
let test_directories = vec![
|
||||
("/Documents", "root-etag-123"),
|
||||
("/Documents/Projects", "projects-etag-456"),
|
||||
("/Documents/Archive", "archive-etag-789"),
|
||||
("/Documents/Projects/Current", "current-etag-abc"),
|
||||
];
|
||||
|
||||
for (path, etag) in &test_directories {
|
||||
let dir = CreateWebDAVDirectory {
|
||||
user_id: user.id,
|
||||
directory_path: path.to_string(),
|
||||
directory_etag: etag.to_string(),
|
||||
file_count: 0,
|
||||
total_size_bytes: 0,
|
||||
};
|
||||
|
||||
state.db.create_or_update_webdav_directory(&dir).await
|
||||
.expect("Failed to create directory tracking");
|
||||
}
|
||||
|
||||
// Verify all directories were tracked
|
||||
let tracked_dirs = state.db.list_webdav_directories(user.id).await
|
||||
.expect("Failed to list tracked directories");
|
||||
|
||||
assert_eq!(tracked_dirs.len(), test_directories.len(),
|
||||
"Should track all discovered directories");
|
||||
|
||||
for (expected_path, expected_etag) in &test_directories {
|
||||
let found = tracked_dirs.iter().find(|d| &d.directory_path == expected_path);
|
||||
assert!(found.is_some(), "Directory {} should be tracked", expected_path);
|
||||
|
||||
let dir = found.unwrap();
|
||||
assert_eq!(&dir.directory_etag, expected_etag,
|
||||
"Directory {} should have correct ETag", expected_path);
|
||||
}
|
||||
|
||||
println!("✅ Test passed: First-time sync establishes complete directory ETag baseline");
|
||||
}
|
||||
|
||||
#[tokio::test]
|
||||
async fn test_smart_sync_no_changes_skip() {
|
||||
// Test Scenario 2: Subsequent smart sync with no directory changes
|
||||
// Expected: Should skip sync entirely after ETag comparison
|
||||
|
||||
let (state, user) = create_test_setup().await;
|
||||
let smart_sync_service = SmartSyncService::new(state.clone());
|
||||
|
||||
// Pre-populate database with directory ETags (simulating previous sync)
|
||||
let existing_directories = vec![
|
||||
("/Documents", "root-etag-stable"),
|
||||
("/Documents/Projects", "projects-etag-stable"),
|
||||
("/Documents/Archive", "archive-etag-stable"),
|
||||
];
|
||||
|
||||
for (path, etag) in &existing_directories {
|
||||
let dir = CreateWebDAVDirectory {
|
||||
user_id: user.id,
|
||||
directory_path: path.to_string(),
|
||||
directory_etag: etag.to_string(),
|
||||
file_count: 5,
|
||||
total_size_bytes: 1024000,
|
||||
};
|
||||
|
||||
state.db.create_or_update_webdav_directory(&dir).await
|
||||
.expect("Failed to create existing directory tracking");
|
||||
}
|
||||
|
||||
// Verify directories were created in database
|
||||
let existing_dirs = state.db.list_webdav_directories(user.id).await
|
||||
.expect("Failed to list directories");
|
||||
assert_eq!(existing_dirs.len(), 3, "Should have 3 pre-existing directories");
|
||||
|
||||
// Create mock WebDAV service that returns the same ETags (no changes)
|
||||
let mock_service = MockWebDAVService::with_directory_structure(vec![
|
||||
("/Documents".to_string(), "root-etag-stable".to_string()),
|
||||
("/Documents/Projects".to_string(), "projects-etag-stable".to_string()),
|
||||
("/Documents/Archive".to_string(), "archive-etag-stable".to_string()),
|
||||
]);
|
||||
|
||||
// Test smart sync evaluation - should detect no changes
|
||||
let sync_result = mock_service.discover_files_and_directories_mock("/Documents", false).await
|
||||
.expect("Mock discovery should succeed");
|
||||
|
||||
// Verify mock returns the same ETags
|
||||
assert_eq!(sync_result.directories.len(), 3, "Should discover 3 directories");
|
||||
for directory in &sync_result.directories {
|
||||
let expected_etag = match directory.path.as_str() {
|
||||
"/Documents" => "root-etag-stable",
|
||||
"/Documents/Projects" => "projects-etag-stable",
|
||||
"/Documents/Archive" => "archive-etag-stable",
|
||||
_ => panic!("Unexpected directory: {}", directory.path),
|
||||
};
|
||||
assert_eq!(directory.etag, expected_etag, "Directory {} should have unchanged ETag", directory.path);
|
||||
}
|
||||
|
||||
// Manually test the smart sync logic (since we can't easily mock WebDAVService in evaluate_sync_need)
|
||||
// Get known directories from database
|
||||
let known_dirs: std::collections::HashMap<String, String> = existing_dirs
|
||||
.into_iter()
|
||||
.filter(|dir| dir.directory_path.starts_with("/Documents"))
|
||||
.map(|dir| (dir.directory_path, dir.directory_etag))
|
||||
.collect();
|
||||
|
||||
// Compare with "discovered" directories (same ETags)
|
||||
let mut changed_count = 0;
|
||||
let mut new_count = 0;
|
||||
|
||||
for directory in &sync_result.directories {
|
||||
match known_dirs.get(&directory.path) {
|
||||
Some(known_etag) => {
|
||||
if known_etag != &directory.etag {
|
||||
changed_count += 1;
|
||||
}
|
||||
}
|
||||
None => {
|
||||
new_count += 1;
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
// Verify no changes detected
|
||||
assert_eq!(changed_count, 0, "Should detect no changed directories");
|
||||
assert_eq!(new_count, 0, "Should detect no new directories");
|
||||
|
||||
// This demonstrates the logic that would cause SmartSyncDecision::SkipSync
|
||||
println!("✅ Smart sync no-changes test passed: {} changed, {} new directories detected",
|
||||
changed_count, new_count);
|
||||
println!("✅ In real implementation, this would result in SmartSyncDecision::SkipSync");
|
||||
}
|
||||
|
||||
#[tokio::test]
|
||||
async fn test_deep_scan_resets_directory_etags() {
|
||||
// Test Scenario 5: Manual deep scan should reset all directory ETags at all levels
|
||||
// Expected: All directory ETags should be updated with fresh values from WebDAV
|
||||
|
||||
let (state, user) = create_test_setup().await;
|
||||
let smart_sync_service = SmartSyncService::new(state.clone());
|
||||
let webdav_service = create_test_webdav_service();
|
||||
|
||||
// Pre-populate database with old directory ETags
|
||||
let old_directories = vec![
|
||||
("/Documents", "old-root-etag"),
|
||||
("/Documents/Projects", "old-projects-etag"),
|
||||
("/Documents/Archive", "old-archive-etag"),
|
||||
("/Documents/Projects/Subproject", "old-subproject-etag"),
|
||||
];
|
||||
|
||||
for (path, etag) in &old_directories {
|
||||
let dir = CreateWebDAVDirectory {
|
||||
user_id: user.id,
|
||||
directory_path: path.to_string(),
|
||||
directory_etag: etag.to_string(),
|
||||
file_count: 3,
|
||||
total_size_bytes: 512000,
|
||||
};
|
||||
|
||||
state.db.create_or_update_webdav_directory(&dir).await
|
||||
.expect("Failed to create old directory tracking");
|
||||
}
|
||||
|
||||
// Verify old ETags are in database
|
||||
let pre_scan_dirs = state.db.list_webdav_directories(user.id).await
|
||||
.expect("Failed to list pre-scan directories");
|
||||
assert_eq!(pre_scan_dirs.len(), 4, "Should start with 4 tracked directories");
|
||||
|
||||
for dir in &pre_scan_dirs {
|
||||
assert!(dir.directory_etag.starts_with("old-"),
|
||||
"Directory {} should have old ETag", dir.directory_path);
|
||||
}
|
||||
|
||||
// Create mock WebDAV service that returns new ETags for all directories
|
||||
let mock_service = MockWebDAVService::with_directory_structure(vec![
|
||||
("/Documents".to_string(), "new-root-etag-123".to_string()),
|
||||
("/Documents/Projects".to_string(), "new-projects-etag-456".to_string()),
|
||||
("/Documents/Archive".to_string(), "new-archive-etag-789".to_string()),
|
||||
("/Documents/Projects/Subproject".to_string(), "new-subproject-etag-abc".to_string()),
|
||||
// Additional new directory discovered during deep scan
|
||||
("/Documents/NewlyFound".to_string(), "newly-found-etag-xyz".to_string()),
|
||||
]);
|
||||
|
||||
// Simulate deep scan discovery (this would be called by perform_smart_sync internally)
|
||||
let deep_scan_discovery = mock_service.discover_files_and_directories_mock("/Documents", true).await
|
||||
.expect("Mock deep scan discovery should succeed");
|
||||
|
||||
// Verify deep scan discovers all directories including new ones
|
||||
assert_eq!(deep_scan_discovery.directories.len(), 5, "Deep scan should discover 5 directories");
|
||||
|
||||
// Simulate what perform_smart_sync would do - save all discovered directory ETags
|
||||
for directory_info in &deep_scan_discovery.directories {
|
||||
let webdav_directory = CreateWebDAVDirectory {
|
||||
user_id: user.id,
|
||||
directory_path: directory_info.path.clone(),
|
||||
directory_etag: directory_info.etag.clone(),
|
||||
file_count: 0, // Would be updated by stats
|
||||
total_size_bytes: 0, // Would be updated by stats
|
||||
};
|
||||
|
||||
state.db.create_or_update_webdav_directory(&webdav_directory).await
|
||||
.expect("Failed to update directory ETag during deep scan");
|
||||
}
|
||||
|
||||
// Verify all directory ETags were reset to new values
|
||||
let post_scan_dirs = state.db.list_webdav_directories(user.id).await
|
||||
.expect("Failed to list post-scan directories");
|
||||
|
||||
// Should have one additional directory from deep scan
|
||||
assert_eq!(post_scan_dirs.len(), 5, "Should have 5 directories after deep scan");
|
||||
|
||||
// Verify all ETags are updated
|
||||
for dir in &post_scan_dirs {
|
||||
match dir.directory_path.as_str() {
|
||||
"/Documents" => {
|
||||
assert_eq!(dir.directory_etag, "new-root-etag-123", "Root ETag should be updated");
|
||||
}
|
||||
"/Documents/Projects" => {
|
||||
assert_eq!(dir.directory_etag, "new-projects-etag-456", "Projects ETag should be updated");
|
||||
}
|
||||
"/Documents/Archive" => {
|
||||
assert_eq!(dir.directory_etag, "new-archive-etag-789", "Archive ETag should be updated");
|
||||
}
|
||||
"/Documents/Projects/Subproject" => {
|
||||
assert_eq!(dir.directory_etag, "new-subproject-etag-abc", "Subproject ETag should be updated");
|
||||
}
|
||||
"/Documents/NewlyFound" => {
|
||||
assert_eq!(dir.directory_etag, "newly-found-etag-xyz", "New directory should be tracked");
|
||||
}
|
||||
_ => panic!("Unexpected directory: {}", dir.directory_path),
|
||||
}
|
||||
|
||||
// Verify no old ETags remain
|
||||
assert!(!dir.directory_etag.starts_with("old-"),
|
||||
"Directory {} should not have old ETag: {}", dir.directory_path, dir.directory_etag);
|
||||
}
|
||||
|
||||
println!("✅ Manual deep scan test passed:");
|
||||
println!(" - All {} existing directory ETags were reset", old_directories.len());
|
||||
println!(" - 1 new directory was discovered and tracked");
|
||||
println!(" - Total directories tracked: {}", post_scan_dirs.len());
|
||||
println!(" - Deep scan strategy successfully resets entire ETag baseline");
|
||||
}
|
||||
|
||||
#[tokio::test]
|
||||
async fn test_directory_structure_changes() {
|
||||
// Test Scenario 8: Directory structure changes - new subdirectories should be detected
|
||||
// Expected: New directories get tracked, existing unchanged directories preserved
|
||||
|
||||
let (state, user) = create_test_setup().await;
|
||||
let smart_sync_service = SmartSyncService::new(state.clone());
|
||||
|
||||
// Start with some existing directory tracking
|
||||
let initial_directories = vec![
|
||||
("/Documents", "root-etag-unchanged"),
|
||||
("/Documents/Existing", "existing-etag-unchanged"),
|
||||
];
|
||||
|
||||
for (path, etag) in &initial_directories {
|
||||
let dir = CreateWebDAVDirectory {
|
||||
user_id: user.id,
|
||||
directory_path: path.to_string(),
|
||||
directory_etag: etag.to_string(),
|
||||
file_count: 2,
|
||||
total_size_bytes: 256000,
|
||||
};
|
||||
|
||||
state.db.create_or_update_webdav_directory(&dir).await
|
||||
.expect("Failed to create initial directory tracking");
|
||||
}
|
||||
|
||||
// Simulate discovering new directory structure (this would come from WebDAV)
|
||||
let new_structure = vec![
|
||||
("/Documents", "root-etag-unchanged"), // Unchanged
|
||||
("/Documents/Existing", "existing-etag-unchanged"), // Unchanged
|
||||
("/Documents/NewFolder", "new-folder-etag"), // New directory
|
||||
("/Documents/NewFolder/SubNew", "subnew-etag"), // New subdirectory
|
||||
];
|
||||
|
||||
// In a real scenario, smart sync would detect these changes and track new directories
|
||||
// For testing, we simulate the result of that discovery
|
||||
for (path, etag) in &new_structure {
|
||||
let dir = CreateWebDAVDirectory {
|
||||
user_id: user.id,
|
||||
directory_path: path.to_string(),
|
||||
directory_etag: etag.to_string(),
|
||||
file_count: if path.contains("New") { 0 } else { 2 },
|
||||
total_size_bytes: if path.contains("New") { 0 } else { 256000 },
|
||||
};
|
||||
|
||||
state.db.create_or_update_webdav_directory(&dir).await
|
||||
.expect("Failed to update directory tracking");
|
||||
}
|
||||
|
||||
// Verify all directories are now tracked
|
||||
let final_dirs = state.db.list_webdav_directories(user.id).await
|
||||
.expect("Failed to list final directories");
|
||||
|
||||
assert_eq!(final_dirs.len(), 4, "Should track all 4 directories after structure change");
|
||||
|
||||
// Verify new directories are tracked
|
||||
let new_folder = final_dirs.iter().find(|d| d.directory_path == "/Documents/NewFolder");
|
||||
assert!(new_folder.is_some(), "New folder should be tracked");
|
||||
assert_eq!(new_folder.unwrap().directory_etag, "new-folder-etag");
|
||||
|
||||
let sub_new = final_dirs.iter().find(|d| d.directory_path == "/Documents/NewFolder/SubNew");
|
||||
assert!(sub_new.is_some(), "New subdirectory should be tracked");
|
||||
assert_eq!(sub_new.unwrap().directory_etag, "subnew-etag");
|
||||
|
||||
// Verify unchanged directories preserved
|
||||
let existing = final_dirs.iter().find(|d| d.directory_path == "/Documents/Existing");
|
||||
assert!(existing.is_some(), "Existing directory should be preserved");
|
||||
assert_eq!(existing.unwrap().directory_etag, "existing-etag-unchanged");
|
||||
|
||||
println!("✅ Test passed: Directory structure changes properly tracked");
|
||||
}
|
||||
|
||||
// Additional test stubs for remaining scenarios
|
||||
#[tokio::test]
|
||||
async fn test_smart_sync_targeted_scan() {
|
||||
// Test Scenario 3: Smart sync with single directory changed - should use targeted scan
|
||||
// Expected: Should detect single change and use TargetedScan strategy
|
||||
|
||||
let (state, user) = create_test_setup().await;
|
||||
let smart_sync_service = SmartSyncService::new(state.clone());
|
||||
|
||||
// Pre-populate database with directory ETags (simulating previous sync)
|
||||
let existing_directories = vec![
|
||||
("/Documents", "root-etag-stable"),
|
||||
("/Documents/Projects", "projects-etag-old"), // This one will change
|
||||
("/Documents/Archive", "archive-etag-stable"),
|
||||
("/Documents/Reports", "reports-etag-stable"),
|
||||
];
|
||||
|
||||
for (path, etag) in &existing_directories {
|
||||
let dir = CreateWebDAVDirectory {
|
||||
user_id: user.id,
|
||||
directory_path: path.to_string(),
|
||||
directory_etag: etag.to_string(),
|
||||
file_count: 3,
|
||||
total_size_bytes: 512000,
|
||||
};
|
||||
|
||||
state.db.create_or_update_webdav_directory(&dir).await
|
||||
.expect("Failed to create existing directory tracking");
|
||||
}
|
||||
|
||||
// Verify initial state
|
||||
let existing_dirs = state.db.list_webdav_directories(user.id).await
|
||||
.expect("Failed to list directories");
|
||||
assert_eq!(existing_dirs.len(), 4, "Should have 4 pre-existing directories");
|
||||
|
||||
// Create mock WebDAV service that returns one changed ETag
|
||||
let mock_service = MockWebDAVService::with_directory_structure(vec![
|
||||
("/Documents".to_string(), "root-etag-stable".to_string()),
|
||||
("/Documents/Projects".to_string(), "projects-etag-NEW".to_string()), // Changed!
|
||||
("/Documents/Archive".to_string(), "archive-etag-stable".to_string()),
|
||||
("/Documents/Reports".to_string(), "reports-etag-stable".to_string()),
|
||||
]);
|
||||
|
||||
// Test smart sync evaluation
|
||||
let sync_result = mock_service.discover_files_and_directories_mock("/Documents", false).await
|
||||
.expect("Mock discovery should succeed");
|
||||
|
||||
// Verify mock returns expected ETags
|
||||
assert_eq!(sync_result.directories.len(), 4, "Should discover 4 directories");
|
||||
|
||||
// Get known directories from database for comparison
|
||||
let known_dirs: std::collections::HashMap<String, String> = existing_dirs
|
||||
.into_iter()
|
||||
.filter(|dir| dir.directory_path.starts_with("/Documents"))
|
||||
.map(|dir| (dir.directory_path, dir.directory_etag))
|
||||
.collect();
|
||||
|
||||
// Compare with discovered directories to identify changes
|
||||
let mut changed_directories = Vec::new();
|
||||
let mut new_directories = Vec::new();
|
||||
let mut unchanged_directories = Vec::new();
|
||||
|
||||
for directory in &sync_result.directories {
|
||||
match known_dirs.get(&directory.path) {
|
||||
Some(known_etag) => {
|
||||
if known_etag != &directory.etag {
|
||||
changed_directories.push(directory.path.clone());
|
||||
} else {
|
||||
unchanged_directories.push(directory.path.clone());
|
||||
}
|
||||
}
|
||||
None => {
|
||||
new_directories.push(directory.path.clone());
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
// Verify targeted scan scenario
|
||||
assert_eq!(changed_directories.len(), 1, "Should detect exactly 1 changed directory");
|
||||
assert_eq!(new_directories.len(), 0, "Should detect no new directories");
|
||||
assert_eq!(unchanged_directories.len(), 3, "Should detect 3 unchanged directories");
|
||||
assert_eq!(changed_directories[0], "/Documents/Projects", "Changed directory should be /Documents/Projects");
|
||||
|
||||
// Test strategy selection logic (mirrors SmartSyncService logic)
|
||||
let total_changes = changed_directories.len() + new_directories.len();
|
||||
let total_known = known_dirs.len();
|
||||
let change_ratio = total_changes as f64 / total_known.max(1) as f64;
|
||||
|
||||
// Should use targeted scan (low change ratio, few new directories)
|
||||
let should_use_targeted = change_ratio <= 0.3 && new_directories.len() <= 5;
|
||||
assert!(should_use_targeted, "Should use targeted scan for single directory change");
|
||||
|
||||
println!("✅ Smart sync targeted scan test passed:");
|
||||
println!(" - Changed directories: {:?}", changed_directories);
|
||||
println!(" - New directories: {:?}", new_directories);
|
||||
println!(" - Change ratio: {:.2}%", change_ratio * 100.0);
|
||||
println!(" - Strategy: TargetedScan (as expected)");
|
||||
}
|
||||
|
||||
#[tokio::test]
|
||||
async fn test_smart_sync_fallback_to_deep_scan() {
|
||||
// Test Scenario 4: Smart sync with many directories changed - should fall back to full deep scan
|
||||
// Expected: Should detect many changes and use FullDeepScan strategy
|
||||
|
||||
let (state, user) = create_test_setup().await;
|
||||
let smart_sync_service = SmartSyncService::new(state.clone());
|
||||
|
||||
// Pre-populate database with directory ETags (simulating previous sync)
|
||||
let existing_directories = vec![
|
||||
("/Documents", "root-etag-old"),
|
||||
("/Documents/Projects", "projects-etag-old"),
|
||||
("/Documents/Archive", "archive-etag-old"),
|
||||
("/Documents/Reports", "reports-etag-old"),
|
||||
("/Documents/Images", "images-etag-old"),
|
||||
("/Documents/Videos", "videos-etag-old"),
|
||||
("/Documents/Music", "music-etag-old"),
|
||||
("/Documents/Backup", "backup-etag-old"),
|
||||
];
|
||||
|
||||
for (path, etag) in &existing_directories {
|
||||
let dir = CreateWebDAVDirectory {
|
||||
user_id: user.id,
|
||||
directory_path: path.to_string(),
|
||||
directory_etag: etag.to_string(),
|
||||
file_count: 10,
|
||||
total_size_bytes: 2048000,
|
||||
};
|
||||
|
||||
state.db.create_or_update_webdav_directory(&dir).await
|
||||
.expect("Failed to create existing directory tracking");
|
||||
}
|
||||
|
||||
// Verify initial state
|
||||
let existing_dirs = state.db.list_webdav_directories(user.id).await
|
||||
.expect("Failed to list directories");
|
||||
assert_eq!(existing_dirs.len(), 8, "Should have 8 pre-existing directories");
|
||||
|
||||
// Create mock WebDAV service that returns many changed ETags + new directories
|
||||
let mock_service = MockWebDAVService::with_directory_structure(vec![
|
||||
// Many existing directories with changed ETags
|
||||
("/Documents".to_string(), "root-etag-NEW".to_string()), // Changed
|
||||
("/Documents/Projects".to_string(), "projects-etag-NEW".to_string()), // Changed
|
||||
("/Documents/Archive".to_string(), "archive-etag-NEW".to_string()), // Changed
|
||||
("/Documents/Reports".to_string(), "reports-etag-NEW".to_string()), // Changed
|
||||
("/Documents/Images".to_string(), "images-etag-old".to_string()), // Unchanged
|
||||
("/Documents/Videos".to_string(), "videos-etag-old".to_string()), // Unchanged
|
||||
("/Documents/Music".to_string(), "music-etag-NEW".to_string()), // Changed
|
||||
("/Documents/Backup".to_string(), "backup-etag-old".to_string()), // Unchanged
|
||||
// Many new directories
|
||||
("/Documents/NewProject1".to_string(), "new1-etag".to_string()), // New
|
||||
("/Documents/NewProject2".to_string(), "new2-etag".to_string()), // New
|
||||
("/Documents/NewProject3".to_string(), "new3-etag".to_string()), // New
|
||||
("/Documents/NewProject4".to_string(), "new4-etag".to_string()), // New
|
||||
("/Documents/NewProject5".to_string(), "new5-etag".to_string()), // New
|
||||
("/Documents/NewProject6".to_string(), "new6-etag".to_string()), // New
|
||||
]);
|
||||
|
||||
// Test smart sync evaluation
|
||||
let sync_result = mock_service.discover_files_and_directories_mock("/Documents", false).await
|
||||
.expect("Mock discovery should succeed");
|
||||
|
||||
// Verify mock returns expected ETags
|
||||
assert_eq!(sync_result.directories.len(), 14, "Should discover 14 directories total");
|
||||
|
||||
// Get known directories from database for comparison
|
||||
let known_dirs: std::collections::HashMap<String, String> = existing_dirs
|
||||
.into_iter()
|
||||
.filter(|dir| dir.directory_path.starts_with("/Documents"))
|
||||
.map(|dir| (dir.directory_path, dir.directory_etag))
|
||||
.collect();
|
||||
|
||||
// Compare with discovered directories to identify changes
|
||||
let mut changed_directories = Vec::new();
|
||||
let mut new_directories = Vec::new();
|
||||
let mut unchanged_directories = Vec::new();
|
||||
|
||||
for directory in &sync_result.directories {
|
||||
match known_dirs.get(&directory.path) {
|
||||
Some(known_etag) => {
|
||||
if known_etag != &directory.etag {
|
||||
changed_directories.push(directory.path.clone());
|
||||
} else {
|
||||
unchanged_directories.push(directory.path.clone());
|
||||
}
|
||||
}
|
||||
None => {
|
||||
new_directories.push(directory.path.clone());
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
// Verify fallback to deep scan scenario
|
||||
assert_eq!(changed_directories.len(), 5, "Should detect 5 changed directories");
|
||||
assert_eq!(new_directories.len(), 6, "Should detect 6 new directories");
|
||||
assert_eq!(unchanged_directories.len(), 3, "Should detect 3 unchanged directories");
|
||||
|
||||
// Test strategy selection logic (mirrors SmartSyncService logic)
|
||||
let total_changes = changed_directories.len() + new_directories.len();
|
||||
let total_known = known_dirs.len();
|
||||
let change_ratio = total_changes as f64 / total_known.max(1) as f64;
|
||||
|
||||
// Should fallback to full deep scan (high change ratio OR many new directories)
|
||||
let should_use_full_scan = change_ratio > 0.3 || new_directories.len() > 5;
|
||||
assert!(should_use_full_scan, "Should use full deep scan for many changes");
|
||||
|
||||
// Verify both thresholds are exceeded
|
||||
assert!(change_ratio > 0.3, "Change ratio {:.2}% should exceed 30% threshold", change_ratio * 100.0);
|
||||
assert!(new_directories.len() > 5, "New directories count {} should exceed 5", new_directories.len());
|
||||
|
||||
println!("✅ Smart sync fallback to deep scan test passed:");
|
||||
println!(" - Changed directories: {} ({})", changed_directories.len(), changed_directories.join(", "));
|
||||
println!(" - New directories: {} ({})", new_directories.len(), new_directories.join(", "));
|
||||
println!(" - Unchanged directories: {}", unchanged_directories.len());
|
||||
println!(" - Change ratio: {:.1}% (exceeds 30% threshold)", change_ratio * 100.0);
|
||||
println!(" - New dirs count: {} (exceeds 5 threshold)", new_directories.len());
|
||||
println!(" - Strategy: FullDeepScan (as expected)");
|
||||
}
|
||||
|
||||
#[tokio::test]
|
||||
async fn test_scheduled_deep_scan() {
|
||||
// Test Scenario 6: Scheduled deep scan should reset all directory ETags and track new ones
|
||||
// Expected: Similar to manual deep scan, but triggered by scheduler with different lifecycle
|
||||
|
||||
let (state, user) = create_test_setup().await;
|
||||
let smart_sync_service = SmartSyncService::new(state.clone());
|
||||
|
||||
// Pre-populate database with directory ETags from previous scheduled sync
|
||||
let previous_directories = vec![
|
||||
("/Documents", "scheduled-root-etag-v1"),
|
||||
("/Documents/Quarterly", "scheduled-quarterly-etag-v1"),
|
||||
("/Documents/Monthly", "scheduled-monthly-etag-v1"),
|
||||
("/Documents/Daily", "scheduled-daily-etag-v1"),
|
||||
];
|
||||
|
||||
for (path, etag) in &previous_directories {
|
||||
let dir = CreateWebDAVDirectory {
|
||||
user_id: user.id,
|
||||
directory_path: path.to_string(),
|
||||
directory_etag: etag.to_string(),
|
||||
file_count: 8,
|
||||
total_size_bytes: 1536000,
|
||||
};
|
||||
|
||||
state.db.create_or_update_webdav_directory(&dir).await
|
||||
.expect("Failed to create scheduled directory tracking");
|
||||
}
|
||||
|
||||
// Verify initial scheduled sync state
|
||||
let pre_scheduled_dirs = state.db.list_webdav_directories(user.id).await
|
||||
.expect("Failed to list pre-scheduled directories");
|
||||
assert_eq!(pre_scheduled_dirs.len(), 4, "Should start with 4 scheduled directories");
|
||||
|
||||
for dir in &pre_scheduled_dirs {
|
||||
assert!(dir.directory_etag.contains("v1"),
|
||||
"Directory {} should have v1 ETag", dir.directory_path);
|
||||
}
|
||||
|
||||
// Create mock WebDAV service for scheduled deep scan with updated structure
|
||||
let mock_service = MockWebDAVService::with_directory_structure(vec![
|
||||
// All existing directories get updated ETags
|
||||
("/Documents".to_string(), "scheduled-root-etag-v2".to_string()),
|
||||
("/Documents/Quarterly".to_string(), "scheduled-quarterly-etag-v2".to_string()),
|
||||
("/Documents/Monthly".to_string(), "scheduled-monthly-etag-v2".to_string()),
|
||||
("/Documents/Daily".to_string(), "scheduled-daily-etag-v2".to_string()),
|
||||
// New directories discovered during scheduled scan
|
||||
("/Documents/Weekly".to_string(), "scheduled-weekly-etag-v1".to_string()),
|
||||
("/Documents/Yearly".to_string(), "scheduled-yearly-etag-v1".to_string()),
|
||||
("/Documents/Archives".to_string(), "scheduled-archives-etag-v1".to_string()),
|
||||
]);
|
||||
|
||||
// Simulate scheduled deep scan (this would be triggered by SourceScheduler)
|
||||
let scheduled_discovery = mock_service.discover_files_and_directories_mock("/Documents", true).await
|
||||
.expect("Mock scheduled scan discovery should succeed");
|
||||
|
||||
// Verify scheduled scan discovers expanded directory structure
|
||||
assert_eq!(scheduled_discovery.directories.len(), 7, "Scheduled scan should discover 7 directories");
|
||||
|
||||
// Simulate what scheduled sync would do - perform full deep scan strategy
|
||||
for directory_info in &scheduled_discovery.directories {
|
||||
let webdav_directory = CreateWebDAVDirectory {
|
||||
user_id: user.id,
|
||||
directory_path: directory_info.path.clone(),
|
||||
directory_etag: directory_info.etag.clone(),
|
||||
file_count: 5, // Updated file counts from scan
|
||||
total_size_bytes: 1024000, // Updated sizes from scan
|
||||
};
|
||||
|
||||
state.db.create_or_update_webdav_directory(&webdav_directory).await
|
||||
.expect("Failed to update directory during scheduled scan");
|
||||
}
|
||||
|
||||
// Verify scheduled deep scan results
|
||||
let post_scheduled_dirs = state.db.list_webdav_directories(user.id).await
|
||||
.expect("Failed to list post-scheduled directories");
|
||||
|
||||
// Should have 3 additional directories from scheduled scan
|
||||
assert_eq!(post_scheduled_dirs.len(), 7, "Should have 7 directories after scheduled scan");
|
||||
|
||||
// Verify all existing ETags were updated to v2
|
||||
let mut updated_existing = 0;
|
||||
let mut new_directories = 0;
|
||||
|
||||
for dir in &post_scheduled_dirs {
|
||||
if previous_directories.iter().any(|(path, _)| path == &dir.directory_path) {
|
||||
// Existing directory should be updated
|
||||
assert!(dir.directory_etag.contains("v2"),
|
||||
"Existing directory {} should be updated to v2: {}",
|
||||
dir.directory_path, dir.directory_etag);
|
||||
assert_eq!(dir.file_count, 5, "File count should be updated from scan");
|
||||
assert_eq!(dir.total_size_bytes, 1024000, "Size should be updated from scan");
|
||||
updated_existing += 1;
|
||||
} else {
|
||||
// New directory should be tracked
|
||||
assert!(dir.directory_etag.contains("v1"),
|
||||
"New directory {} should have v1 ETag: {}",
|
||||
dir.directory_path, dir.directory_etag);
|
||||
new_directories += 1;
|
||||
}
|
||||
}
|
||||
|
||||
assert_eq!(updated_existing, 4, "Should update 4 existing directories");
|
||||
assert_eq!(new_directories, 3, "Should discover 3 new directories");
|
||||
|
||||
// Verify no old v1 ETags remain for existing directories
|
||||
for dir in &post_scheduled_dirs {
|
||||
if previous_directories.iter().any(|(path, _)| path == &dir.directory_path) {
|
||||
assert!(!dir.directory_etag.contains("v1"),
|
||||
"Existing directory {} should not have old v1 ETag", dir.directory_path);
|
||||
}
|
||||
}
|
||||
|
||||
println!("✅ Scheduled deep scan test passed:");
|
||||
println!(" - Updated {} existing directories to v2 ETags", updated_existing);
|
||||
println!(" - Discovered and tracked {} new directories", new_directories);
|
||||
println!(" - Total directories tracked: {}", post_scheduled_dirs.len());
|
||||
println!(" - File counts and sizes updated during scan");
|
||||
println!(" - Scheduled deep scan maintains complete directory tracking");
|
||||
}
|
||||
|
||||
#[tokio::test]
|
||||
async fn test_smart_sync_after_deep_scan() {
|
||||
// Test Scenario 7: Smart sync after deep scan should use fresh directory ETags
|
||||
// Expected: After deep scan, smart sync should use the new baseline and detect minimal changes
|
||||
|
||||
let (state, user) = create_test_setup().await;
|
||||
let smart_sync_service = SmartSyncService::new(state.clone());
|
||||
|
||||
// Phase 1: Simulate state after a deep scan has completed
|
||||
let post_deep_scan_directories = vec![
|
||||
("/Documents", "deep-scan-root-fresh"),
|
||||
("/Documents/Active", "deep-scan-active-fresh"),
|
||||
("/Documents/Archive", "deep-scan-archive-fresh"),
|
||||
("/Documents/Processing", "deep-scan-processing-fresh"),
|
||||
];
|
||||
|
||||
for (path, etag) in &post_deep_scan_directories {
|
||||
let dir = CreateWebDAVDirectory {
|
||||
user_id: user.id,
|
||||
directory_path: path.to_string(),
|
||||
directory_etag: etag.to_string(),
|
||||
file_count: 12,
|
||||
total_size_bytes: 2048000,
|
||||
};
|
||||
|
||||
state.db.create_or_update_webdav_directory(&dir).await
|
||||
.expect("Failed to create post-deep-scan directory tracking");
|
||||
}
|
||||
|
||||
// Verify deep scan baseline is established
|
||||
let baseline_dirs = state.db.list_webdav_directories(user.id).await
|
||||
.expect("Failed to list baseline directories");
|
||||
assert_eq!(baseline_dirs.len(), 4, "Should have fresh baseline from deep scan");
|
||||
|
||||
for dir in &baseline_dirs {
|
||||
assert!(dir.directory_etag.contains("fresh"),
|
||||
"Directory {} should have fresh ETag from deep scan", dir.directory_path);
|
||||
}
|
||||
|
||||
// Phase 2: Time passes, then smart sync runs and finds mostly unchanged structure
|
||||
// with just one minor change
|
||||
let mock_service = MockWebDAVService::with_directory_structure(vec![
|
||||
("/Documents".to_string(), "deep-scan-root-fresh".to_string()), // Unchanged from deep scan
|
||||
("/Documents/Active".to_string(), "deep-scan-active-UPDATED".to_string()), // One change!
|
||||
("/Documents/Archive".to_string(), "deep-scan-archive-fresh".to_string()), // Unchanged
|
||||
("/Documents/Processing".to_string(), "deep-scan-processing-fresh".to_string()), // Unchanged
|
||||
]);
|
||||
|
||||
// Phase 3: Smart sync evaluation after deep scan baseline
|
||||
let smart_sync_discovery = mock_service.discover_files_and_directories_mock("/Documents", false).await
|
||||
.expect("Mock smart sync after deep scan should succeed");
|
||||
|
||||
// Verify structure is as expected
|
||||
assert_eq!(smart_sync_discovery.directories.len(), 4, "Should discover same 4 directories");
|
||||
|
||||
// Phase 4: Analyze changes against fresh deep scan baseline
|
||||
let known_dirs: std::collections::HashMap<String, String> = baseline_dirs
|
||||
.into_iter()
|
||||
.filter(|dir| dir.directory_path.starts_with("/Documents"))
|
||||
.map(|dir| (dir.directory_path, dir.directory_etag))
|
||||
.collect();
|
||||
|
||||
let mut changed_dirs_after_deep_scan = Vec::new();
|
||||
let mut unchanged_dirs_after_deep_scan = Vec::new();
|
||||
let mut new_dirs_after_deep_scan = Vec::new();
|
||||
|
||||
for directory in &smart_sync_discovery.directories {
|
||||
match known_dirs.get(&directory.path) {
|
||||
Some(baseline_etag) => {
|
||||
if baseline_etag != &directory.etag {
|
||||
changed_dirs_after_deep_scan.push(directory.path.clone());
|
||||
} else {
|
||||
unchanged_dirs_after_deep_scan.push(directory.path.clone());
|
||||
}
|
||||
}
|
||||
None => {
|
||||
new_dirs_after_deep_scan.push(directory.path.clone());
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
// Phase 5: Verify smart sync detects minimal change against fresh baseline
|
||||
assert_eq!(changed_dirs_after_deep_scan.len(), 1, "Should detect 1 changed directory against fresh baseline");
|
||||
assert_eq!(unchanged_dirs_after_deep_scan.len(), 3, "Should detect 3 unchanged directories against fresh baseline");
|
||||
assert_eq!(new_dirs_after_deep_scan.len(), 0, "Should detect no new directories");
|
||||
|
||||
assert_eq!(changed_dirs_after_deep_scan[0], "/Documents/Active",
|
||||
"Active directory should be the one that changed since deep scan");
|
||||
|
||||
// Phase 6: Verify smart sync strategy selection using fresh baseline
|
||||
let total_changes = changed_dirs_after_deep_scan.len() + new_dirs_after_deep_scan.len();
|
||||
let total_known = known_dirs.len();
|
||||
let change_ratio_vs_baseline = total_changes as f64 / total_known.max(1) as f64;
|
||||
|
||||
// Should use targeted scan (minimal change against fresh baseline)
|
||||
let should_use_targeted = change_ratio_vs_baseline <= 0.3 && new_dirs_after_deep_scan.len() <= 5;
|
||||
assert!(should_use_targeted, "Should use targeted scan for minimal change against fresh baseline");
|
||||
|
||||
// Phase 7: Simulate smart sync updating only the changed directory
|
||||
for dir in &smart_sync_discovery.directories {
|
||||
if changed_dirs_after_deep_scan.contains(&dir.path) {
|
||||
let updated_dir = CreateWebDAVDirectory {
|
||||
user_id: user.id,
|
||||
directory_path: dir.path.clone(),
|
||||
directory_etag: dir.etag.clone(),
|
||||
file_count: 15, // Updated from targeted scan
|
||||
total_size_bytes: 2560000, // Updated from targeted scan
|
||||
};
|
||||
|
||||
state.db.create_or_update_webdav_directory(&updated_dir).await
|
||||
.expect("Failed to update changed directory from smart sync");
|
||||
}
|
||||
}
|
||||
|
||||
// Phase 8: Verify final state maintains fresh baseline with targeted update
|
||||
let final_dirs = state.db.list_webdav_directories(user.id).await
|
||||
.expect("Failed to list final directories");
|
||||
|
||||
assert_eq!(final_dirs.len(), 4, "Should still have 4 directories");
|
||||
|
||||
for dir in &final_dirs {
|
||||
if dir.directory_path == "/Documents/Active" {
|
||||
assert_eq!(dir.directory_etag, "deep-scan-active-UPDATED",
|
||||
"Active directory should have updated ETag");
|
||||
assert_eq!(dir.file_count, 15, "File count should be updated");
|
||||
} else {
|
||||
assert!(dir.directory_etag.contains("fresh"),
|
||||
"Other directories should retain fresh baseline ETags: {}",
|
||||
dir.directory_path);
|
||||
}
|
||||
}
|
||||
|
||||
println!("✅ Smart sync after deep scan test passed:");
|
||||
println!(" - Used fresh deep scan baseline with {} directories", post_deep_scan_directories.len());
|
||||
println!(" - Detected {} changed directory against fresh baseline", changed_dirs_after_deep_scan.len());
|
||||
println!(" - Preserved {} unchanged directories from baseline", unchanged_dirs_after_deep_scan.len());
|
||||
println!(" - Change ratio vs fresh baseline: {:.1}%", change_ratio_vs_baseline * 100.0);
|
||||
println!(" - Strategy: TargetedScan (efficient against fresh baseline)");
|
||||
println!(" - Deep scan provides accurate baseline for subsequent smart syncs");
|
||||
}
|
||||
|
||||
#[tokio::test]
|
||||
async fn test_directory_deletion_handling() {
|
||||
// Test Scenario 9: Directory deletion scenarios should be handled gracefully
|
||||
println!("📝 Test stub: Directory deletion handling");
|
||||
// TODO: Implement directory removal scenarios
|
||||
}
|
||||
|
||||
#[tokio::test]
|
||||
async fn test_webdav_error_fallback() {
|
||||
// Test Scenario 10: WebDAV server errors should fall back to traditional sync
|
||||
println!("📝 Test stub: WebDAV error fallback to traditional sync");
|
||||
// TODO: Implement error scenario testing
|
||||
}
|
||||
|
|
@ -0,0 +1,350 @@
|
|||
use std::sync::Arc;
|
||||
use readur::{
|
||||
AppState,
|
||||
models::{CreateWebDAVDirectory, User, AuthProvider},
|
||||
services::webdav::{SmartSyncService, SmartSyncStrategy, SmartSyncDecision, WebDAVService, WebDAVConfig},
|
||||
test_utils::{TestContext, TestAuthHelper},
|
||||
};
|
||||
|
||||
/// Helper function to create test database and user
|
||||
async fn create_test_setup() -> (Arc<AppState>, User) {
|
||||
let test_context = TestContext::new().await;
|
||||
let auth_helper = TestAuthHelper::new(test_context.app().clone());
|
||||
let test_user = auth_helper.create_test_user().await;
|
||||
|
||||
// Convert TestUser to User model for compatibility
|
||||
let user = User {
|
||||
id: test_user.user_response.id,
|
||||
username: test_user.user_response.username,
|
||||
email: test_user.user_response.email,
|
||||
password_hash: Some("hashed_password".to_string()),
|
||||
role: test_user.user_response.role,
|
||||
created_at: chrono::Utc::now(),
|
||||
updated_at: chrono::Utc::now(),
|
||||
oidc_subject: None,
|
||||
oidc_issuer: None,
|
||||
oidc_email: None,
|
||||
auth_provider: AuthProvider::Local,
|
||||
};
|
||||
|
||||
(test_context.state().clone(), user)
|
||||
}
|
||||
|
||||
/// Helper function to create WebDAV service for testing
|
||||
fn create_test_webdav_service() -> WebDAVService {
|
||||
let config = WebDAVConfig {
|
||||
server_url: "https://test.example.com".to_string(),
|
||||
username: "test".to_string(),
|
||||
password: "test".to_string(),
|
||||
watch_folders: vec!["/Documents".to_string()],
|
||||
file_extensions: vec!["pdf".to_string(), "txt".to_string()],
|
||||
timeout_seconds: 30,
|
||||
server_type: Some("generic".to_string()),
|
||||
};
|
||||
|
||||
WebDAVService::new(config).expect("Failed to create WebDAV service")
|
||||
}
|
||||
|
||||
#[tokio::test]
|
||||
async fn test_deep_scan_resets_directory_etags() {
|
||||
// Integration Test: Manual deep scan should reset all directory ETags at all levels
|
||||
// Expected: Should clear existing ETags and establish fresh baseline
|
||||
|
||||
let (state, user) = create_test_setup().await;
|
||||
|
||||
// Pre-populate database with old directory ETags
|
||||
let old_directories = vec![
|
||||
CreateWebDAVDirectory {
|
||||
user_id: user.id,
|
||||
directory_path: "/Documents".to_string(),
|
||||
directory_etag: "old-root-etag".to_string(),
|
||||
file_count: 5,
|
||||
total_size_bytes: 500000,
|
||||
},
|
||||
CreateWebDAVDirectory {
|
||||
user_id: user.id,
|
||||
directory_path: "/Documents/Projects".to_string(),
|
||||
directory_etag: "old-projects-etag".to_string(),
|
||||
file_count: 10,
|
||||
total_size_bytes: 1000000,
|
||||
},
|
||||
CreateWebDAVDirectory {
|
||||
user_id: user.id,
|
||||
directory_path: "/Documents/Archive".to_string(),
|
||||
directory_etag: "old-archive-etag".to_string(),
|
||||
file_count: 20,
|
||||
total_size_bytes: 2000000,
|
||||
},
|
||||
CreateWebDAVDirectory {
|
||||
user_id: user.id,
|
||||
directory_path: "/Documents/Deep/Nested/Path".to_string(),
|
||||
directory_etag: "old-deep-etag".to_string(),
|
||||
file_count: 3,
|
||||
total_size_bytes: 300000,
|
||||
},
|
||||
];
|
||||
|
||||
for dir in &old_directories {
|
||||
state.db.create_or_update_webdav_directory(dir).await
|
||||
.expect("Failed to create old directory");
|
||||
}
|
||||
|
||||
// Verify old directories were created
|
||||
let before_dirs = state.db.list_webdav_directories(user.id).await.unwrap();
|
||||
assert_eq!(before_dirs.len(), 4, "Should have 4 old directories");
|
||||
|
||||
// Simulate deep scan reset - this would happen during a deep scan operation
|
||||
// For testing, we'll manually clear directories and add new ones
|
||||
|
||||
// Clear existing directories (simulating deep scan reset)
|
||||
for dir in &before_dirs {
|
||||
state.db.delete_webdav_directory(user.id, &dir.directory_path).await
|
||||
.expect("Failed to delete old directory");
|
||||
}
|
||||
|
||||
// Verify directories were cleared
|
||||
let cleared_dirs = state.db.list_webdav_directories(user.id).await.unwrap();
|
||||
assert_eq!(cleared_dirs.len(), 0, "Should have cleared all old directories");
|
||||
|
||||
// Add new directories with fresh ETags (simulating post-deep-scan discovery)
|
||||
let new_directories = vec![
|
||||
CreateWebDAVDirectory {
|
||||
user_id: user.id,
|
||||
directory_path: "/Documents".to_string(),
|
||||
directory_etag: "fresh-root-etag".to_string(),
|
||||
file_count: 8,
|
||||
total_size_bytes: 800000,
|
||||
},
|
||||
CreateWebDAVDirectory {
|
||||
user_id: user.id,
|
||||
directory_path: "/Documents/Projects".to_string(),
|
||||
directory_etag: "fresh-projects-etag".to_string(),
|
||||
file_count: 12,
|
||||
total_size_bytes: 1200000,
|
||||
},
|
||||
CreateWebDAVDirectory {
|
||||
user_id: user.id,
|
||||
directory_path: "/Documents/Archive".to_string(),
|
||||
directory_etag: "fresh-archive-etag".to_string(),
|
||||
file_count: 25,
|
||||
total_size_bytes: 2500000,
|
||||
},
|
||||
CreateWebDAVDirectory {
|
||||
user_id: user.id,
|
||||
directory_path: "/Documents/Deep/Nested/Path".to_string(),
|
||||
directory_etag: "fresh-deep-etag".to_string(),
|
||||
file_count: 5,
|
||||
total_size_bytes: 500000,
|
||||
},
|
||||
CreateWebDAVDirectory {
|
||||
user_id: user.id,
|
||||
directory_path: "/Documents/NewDirectory".to_string(),
|
||||
directory_etag: "brand-new-etag".to_string(),
|
||||
file_count: 2,
|
||||
total_size_bytes: 200000,
|
||||
},
|
||||
];
|
||||
|
||||
for dir in &new_directories {
|
||||
state.db.create_or_update_webdav_directory(dir).await
|
||||
.expect("Failed to create new directory");
|
||||
}
|
||||
|
||||
// Verify fresh directories were created
|
||||
let after_dirs = state.db.list_webdav_directories(user.id).await.unwrap();
|
||||
assert_eq!(after_dirs.len(), 5, "Should have 5 fresh directories after deep scan");
|
||||
|
||||
// Verify ETags are completely different
|
||||
let root_dir = after_dirs.iter().find(|d| d.directory_path == "/Documents").unwrap();
|
||||
assert_eq!(root_dir.directory_etag, "fresh-root-etag");
|
||||
assert_ne!(root_dir.directory_etag, "old-root-etag");
|
||||
|
||||
let projects_dir = after_dirs.iter().find(|d| d.directory_path == "/Documents/Projects").unwrap();
|
||||
assert_eq!(projects_dir.directory_etag, "fresh-projects-etag");
|
||||
assert_ne!(projects_dir.directory_etag, "old-projects-etag");
|
||||
|
||||
let new_dir = after_dirs.iter().find(|d| d.directory_path == "/Documents/NewDirectory").unwrap();
|
||||
assert_eq!(new_dir.directory_etag, "brand-new-etag");
|
||||
|
||||
println!("✅ Deep scan reset test completed successfully");
|
||||
println!(" Cleared {} old directories", old_directories.len());
|
||||
println!(" Created {} fresh directories", new_directories.len());
|
||||
}
|
||||
|
||||
#[tokio::test]
|
||||
async fn test_scheduled_deep_scan() {
|
||||
// Integration Test: Scheduled deep scan should reset all directory ETags and track new ones
|
||||
// This tests the scenario where a scheduled deep scan runs periodically
|
||||
|
||||
let (state, user) = create_test_setup().await;
|
||||
|
||||
// Simulate initial sync state
|
||||
let initial_directories = vec![
|
||||
CreateWebDAVDirectory {
|
||||
user_id: user.id,
|
||||
directory_path: "/Documents".to_string(),
|
||||
directory_etag: "initial-root".to_string(),
|
||||
file_count: 10,
|
||||
total_size_bytes: 1000000,
|
||||
},
|
||||
CreateWebDAVDirectory {
|
||||
user_id: user.id,
|
||||
directory_path: "/Documents/OldProject".to_string(),
|
||||
directory_etag: "initial-old-project".to_string(),
|
||||
file_count: 15,
|
||||
total_size_bytes: 1500000,
|
||||
},
|
||||
];
|
||||
|
||||
for dir in &initial_directories {
|
||||
state.db.create_or_update_webdav_directory(dir).await
|
||||
.expect("Failed to create initial directory");
|
||||
}
|
||||
|
||||
let initial_count = state.db.list_webdav_directories(user.id).await.unwrap().len();
|
||||
assert_eq!(initial_count, 2, "Should start with 2 initial directories");
|
||||
|
||||
// Simulate time passing and directory structure changes
|
||||
// During this time, directories may have been added/removed/changed on the WebDAV server
|
||||
|
||||
// Simulate scheduled deep scan: clear all ETags and rediscover
|
||||
let initial_dirs = state.db.list_webdav_directories(user.id).await.unwrap();
|
||||
for dir in &initial_dirs {
|
||||
state.db.delete_webdav_directory(user.id, &dir.directory_path).await
|
||||
.expect("Failed to delete during deep scan reset");
|
||||
}
|
||||
|
||||
// Simulate fresh discovery after deep scan
|
||||
let post_scan_directories = vec![
|
||||
CreateWebDAVDirectory {
|
||||
user_id: user.id,
|
||||
directory_path: "/Documents".to_string(),
|
||||
directory_etag: "scheduled-root".to_string(), // Changed ETag
|
||||
file_count: 12, // Changed file count
|
||||
total_size_bytes: 1200000, // Changed size
|
||||
},
|
||||
CreateWebDAVDirectory {
|
||||
user_id: user.id,
|
||||
directory_path: "/Documents/NewProject".to_string(), // Different directory
|
||||
directory_etag: "scheduled-new-project".to_string(),
|
||||
file_count: 8,
|
||||
total_size_bytes: 800000,
|
||||
},
|
||||
CreateWebDAVDirectory {
|
||||
user_id: user.id,
|
||||
directory_path: "/Documents/Archive".to_string(), // Completely new directory
|
||||
directory_etag: "scheduled-archive".to_string(),
|
||||
file_count: 30,
|
||||
total_size_bytes: 3000000,
|
||||
},
|
||||
];
|
||||
|
||||
for dir in &post_scan_directories {
|
||||
state.db.create_or_update_webdav_directory(dir).await
|
||||
.expect("Failed to create post-scan directory");
|
||||
}
|
||||
|
||||
// Verify the scheduled deep scan results
|
||||
let final_dirs = state.db.list_webdav_directories(user.id).await.unwrap();
|
||||
assert_eq!(final_dirs.len(), 3, "Should have 3 directories after scheduled deep scan");
|
||||
|
||||
// Verify the directory structure reflects current state
|
||||
let root_dir = final_dirs.iter().find(|d| d.directory_path == "/Documents").unwrap();
|
||||
assert_eq!(root_dir.directory_etag, "scheduled-root");
|
||||
assert_eq!(root_dir.file_count, 12);
|
||||
assert_eq!(root_dir.total_size_bytes, 1200000);
|
||||
|
||||
let new_project = final_dirs.iter().find(|d| d.directory_path == "/Documents/NewProject").unwrap();
|
||||
assert_eq!(new_project.directory_etag, "scheduled-new-project");
|
||||
|
||||
let archive_dir = final_dirs.iter().find(|d| d.directory_path == "/Documents/Archive").unwrap();
|
||||
assert_eq!(archive_dir.directory_etag, "scheduled-archive");
|
||||
|
||||
// Verify old directory is gone
|
||||
assert!(final_dirs.iter().find(|d| d.directory_path == "/Documents/OldProject").is_none(),
|
||||
"Old project directory should be removed after scheduled deep scan");
|
||||
|
||||
println!("✅ Scheduled deep scan test completed successfully");
|
||||
println!(" Initial directories: {}", initial_directories.len());
|
||||
println!(" Final directories: {}", final_dirs.len());
|
||||
println!(" Successfully handled directory structure changes");
|
||||
}
|
||||
|
||||
#[tokio::test]
|
||||
async fn test_deep_scan_performance_with_many_directories() {
|
||||
// Integration Test: Deep scan should perform well even with large numbers of directories
|
||||
// This tests the scalability of the deep scan reset operation
|
||||
|
||||
let (state, user) = create_test_setup().await;
|
||||
|
||||
// Create a large number of old directories
|
||||
let num_old_dirs = 250;
|
||||
let mut old_directories = Vec::new();
|
||||
|
||||
let create_start = std::time::Instant::now();
|
||||
for i in 0..num_old_dirs {
|
||||
let dir = CreateWebDAVDirectory {
|
||||
user_id: user.id,
|
||||
directory_path: format!("/Documents/Old{:03}", i),
|
||||
directory_etag: format!("old-etag-{:03}", i),
|
||||
file_count: i as i32 % 20 + 1, // 1-20 files
|
||||
total_size_bytes: (i as i64 + 1) * 4000, // Varying sizes
|
||||
};
|
||||
|
||||
state.db.create_or_update_webdav_directory(&dir).await
|
||||
.expect("Failed to create old directory");
|
||||
old_directories.push(dir);
|
||||
}
|
||||
let create_duration = create_start.elapsed();
|
||||
|
||||
// Verify old directories were created
|
||||
let before_count = state.db.list_webdav_directories(user.id).await.unwrap().len();
|
||||
assert_eq!(before_count, num_old_dirs, "Should have created {} old directories", num_old_dirs);
|
||||
|
||||
// Simulate deep scan reset - delete all existing
|
||||
let delete_start = std::time::Instant::now();
|
||||
let dirs_to_delete = state.db.list_webdav_directories(user.id).await.unwrap();
|
||||
for dir in &dirs_to_delete {
|
||||
state.db.delete_webdav_directory(user.id, &dir.directory_path).await
|
||||
.expect("Failed to delete directory during deep scan");
|
||||
}
|
||||
let delete_duration = delete_start.elapsed();
|
||||
|
||||
// Verify cleanup
|
||||
let cleared_count = state.db.list_webdav_directories(user.id).await.unwrap().len();
|
||||
assert_eq!(cleared_count, 0, "Should have cleared all directories");
|
||||
|
||||
// Create new directories (simulating rediscovery)
|
||||
let num_new_dirs = 300; // Slightly different number
|
||||
let recreate_start = std::time::Instant::now();
|
||||
for i in 0..num_new_dirs {
|
||||
let dir = CreateWebDAVDirectory {
|
||||
user_id: user.id,
|
||||
directory_path: format!("/Documents/New{:03}", i),
|
||||
directory_etag: format!("new-etag-{:03}", i),
|
||||
file_count: i as i32 % 15 + 1, // 1-15 files
|
||||
total_size_bytes: (i as i64 + 1) * 5000, // Different sizing
|
||||
};
|
||||
|
||||
state.db.create_or_update_webdav_directory(&dir).await
|
||||
.expect("Failed to create new directory");
|
||||
}
|
||||
let recreate_duration = recreate_start.elapsed();
|
||||
|
||||
// Verify final state
|
||||
let final_count = state.db.list_webdav_directories(user.id).await.unwrap().len();
|
||||
assert_eq!(final_count, num_new_dirs, "Should have created {} new directories", num_new_dirs);
|
||||
|
||||
// Performance assertions - should complete within reasonable time
|
||||
assert!(create_duration.as_secs() < 30, "Creating {} directories should take < 30s, took {:?}", num_old_dirs, create_duration);
|
||||
assert!(delete_duration.as_secs() < 15, "Deleting {} directories should take < 15s, took {:?}", num_old_dirs, delete_duration);
|
||||
assert!(recreate_duration.as_secs() < 30, "Recreating {} directories should take < 30s, took {:?}", num_new_dirs, recreate_duration);
|
||||
|
||||
let total_duration = create_duration + delete_duration + recreate_duration;
|
||||
|
||||
println!("✅ Deep scan performance test completed successfully");
|
||||
println!(" Created {} old directories in {:?}", num_old_dirs, create_duration);
|
||||
println!(" Deleted {} directories in {:?}", num_old_dirs, delete_duration);
|
||||
println!(" Created {} new directories in {:?}", num_new_dirs, recreate_duration);
|
||||
println!(" Total deep scan simulation time: {:?}", total_duration);
|
||||
}
|
||||
|
|
@ -0,0 +1,351 @@
|
|||
use std::sync::Arc;
|
||||
use readur::{
|
||||
AppState,
|
||||
models::{CreateWebDAVDirectory, User, AuthProvider},
|
||||
services::webdav::{SmartSyncService, SmartSyncStrategy, SmartSyncDecision, WebDAVService, WebDAVConfig},
|
||||
test_utils::{TestContext, TestAuthHelper},
|
||||
};
|
||||
|
||||
/// Helper function to create test database and user
|
||||
async fn create_test_setup() -> (Arc<AppState>, User) {
|
||||
let test_context = TestContext::new().await;
|
||||
let auth_helper = TestAuthHelper::new(test_context.app().clone());
|
||||
let test_user = auth_helper.create_test_user().await;
|
||||
|
||||
// Convert TestUser to User model for compatibility
|
||||
let user = User {
|
||||
id: test_user.user_response.id,
|
||||
username: test_user.user_response.username,
|
||||
email: test_user.user_response.email,
|
||||
password_hash: Some("hashed_password".to_string()),
|
||||
role: test_user.user_response.role,
|
||||
created_at: chrono::Utc::now(),
|
||||
updated_at: chrono::Utc::now(),
|
||||
oidc_subject: None,
|
||||
oidc_issuer: None,
|
||||
oidc_email: None,
|
||||
auth_provider: AuthProvider::Local,
|
||||
};
|
||||
|
||||
(test_context.state().clone(), user)
|
||||
}
|
||||
|
||||
/// Helper function to create WebDAV service for testing
|
||||
fn create_test_webdav_service() -> WebDAVService {
|
||||
let config = WebDAVConfig {
|
||||
server_url: "https://test.example.com".to_string(),
|
||||
username: "test".to_string(),
|
||||
password: "test".to_string(),
|
||||
watch_folders: vec!["/Documents".to_string()],
|
||||
file_extensions: vec!["pdf".to_string(), "txt".to_string()],
|
||||
timeout_seconds: 30,
|
||||
server_type: Some("generic".to_string()),
|
||||
};
|
||||
|
||||
WebDAVService::new(config).expect("Failed to create WebDAV service")
|
||||
}
|
||||
|
||||
#[tokio::test]
|
||||
async fn test_webdav_error_fallback() {
|
||||
// Integration Test: WebDAV server error scenarios should fall back to traditional sync
|
||||
// Expected: When WebDAV service fails, should gracefully handle errors
|
||||
|
||||
let (state, user) = create_test_setup().await;
|
||||
let smart_sync_service = SmartSyncService::new(state.clone());
|
||||
|
||||
// Create some existing directories to test database robustness
|
||||
let existing_directories = vec![
|
||||
CreateWebDAVDirectory {
|
||||
user_id: user.id,
|
||||
directory_path: "/Documents".to_string(),
|
||||
directory_etag: "existing-root".to_string(),
|
||||
file_count: 5,
|
||||
total_size_bytes: 500000,
|
||||
},
|
||||
CreateWebDAVDirectory {
|
||||
user_id: user.id,
|
||||
directory_path: "/Documents/Projects".to_string(),
|
||||
directory_etag: "existing-projects".to_string(),
|
||||
file_count: 8,
|
||||
total_size_bytes: 800000,
|
||||
},
|
||||
];
|
||||
|
||||
for dir in &existing_directories {
|
||||
state.db.create_or_update_webdav_directory(dir).await
|
||||
.expect("Failed to create existing directory");
|
||||
}
|
||||
|
||||
// Test with a WebDAV service that will fail (invalid URL)
|
||||
let invalid_config = WebDAVConfig {
|
||||
server_url: "https://invalid-server-that-does-not-exist.com".to_string(),
|
||||
username: "invalid".to_string(),
|
||||
password: "invalid".to_string(),
|
||||
watch_folders: vec!["/Documents".to_string()],
|
||||
file_extensions: vec!["pdf".to_string()],
|
||||
timeout_seconds: 1, // Very short timeout to fail quickly
|
||||
server_type: Some("generic".to_string()),
|
||||
};
|
||||
|
||||
let failing_webdav_service = WebDAVService::new(invalid_config)
|
||||
.expect("WebDAV service creation should not fail");
|
||||
|
||||
// Test smart sync evaluation with failing WebDAV service
|
||||
let decision = smart_sync_service.evaluate_sync_need(user.id, &failing_webdav_service, "/Documents").await;
|
||||
|
||||
// The system should handle the WebDAV error gracefully
|
||||
match decision {
|
||||
Ok(SmartSyncDecision::RequiresSync(SmartSyncStrategy::FullDeepScan)) => {
|
||||
println!("✅ WebDAV error correctly falls back to full deep scan");
|
||||
}
|
||||
Err(e) => {
|
||||
println!("✅ WebDAV error handled gracefully: {}", e);
|
||||
// This is acceptable - the system should either fall back or return an error
|
||||
// The important thing is that it doesn't panic or corrupt the database
|
||||
}
|
||||
Ok(other) => {
|
||||
println!("⚠️ WebDAV error resulted in unexpected decision: {:?}", other);
|
||||
// This might be acceptable depending on the implementation
|
||||
}
|
||||
}
|
||||
|
||||
// Verify database state is intact after WebDAV errors
|
||||
let dirs_after_error = state.db.list_webdav_directories(user.id).await.unwrap();
|
||||
assert_eq!(dirs_after_error.len(), 2, "Database should remain intact after WebDAV errors");
|
||||
|
||||
let root_dir = dirs_after_error.iter().find(|d| d.directory_path == "/Documents").unwrap();
|
||||
assert_eq!(root_dir.directory_etag, "existing-root");
|
||||
|
||||
println!("✅ WebDAV error fallback test completed - database remains intact");
|
||||
}
|
||||
|
||||
#[tokio::test]
|
||||
async fn test_database_error_handling() {
|
||||
// Integration Test: Database errors should be handled gracefully
|
||||
// This tests the system's resilience to database connectivity issues
|
||||
|
||||
let (state, user) = create_test_setup().await;
|
||||
let smart_sync_service = SmartSyncService::new(state.clone());
|
||||
|
||||
// Test with invalid user ID (simulates database query errors)
|
||||
let invalid_user_id = uuid::Uuid::new_v4(); // Random UUID that doesn't exist
|
||||
let webdav_service = create_test_webdav_service();
|
||||
|
||||
let decision = smart_sync_service.evaluate_sync_need(invalid_user_id, &webdav_service, "/Documents").await;
|
||||
|
||||
match decision {
|
||||
Ok(SmartSyncDecision::RequiresSync(SmartSyncStrategy::FullDeepScan)) => {
|
||||
println!("✅ Invalid user ID correctly falls back to full deep scan");
|
||||
}
|
||||
Err(e) => {
|
||||
println!("✅ Invalid user ID error handled gracefully: {}", e);
|
||||
// This is the expected behavior - should return an error for invalid user
|
||||
}
|
||||
Ok(other) => {
|
||||
println!("⚠️ Invalid user ID resulted in: {:?}", other);
|
||||
}
|
||||
}
|
||||
|
||||
// Test database connectivity by trying normal operations
|
||||
let test_dir = CreateWebDAVDirectory {
|
||||
user_id: user.id, // Valid user ID
|
||||
directory_path: "/Test".to_string(),
|
||||
directory_etag: "test-etag".to_string(),
|
||||
file_count: 1,
|
||||
total_size_bytes: 100000,
|
||||
};
|
||||
|
||||
// This should work normally
|
||||
state.db.create_or_update_webdav_directory(&test_dir).await
|
||||
.expect("Normal database operations should still work");
|
||||
|
||||
let saved_dirs = state.db.list_webdav_directories(user.id).await.unwrap();
|
||||
assert_eq!(saved_dirs.len(), 1, "Normal database operations should work after error handling");
|
||||
|
||||
println!("✅ Database error handling test completed");
|
||||
}
|
||||
|
||||
#[tokio::test]
|
||||
async fn test_concurrent_smart_sync_operations() {
|
||||
// Integration Test: Concurrent smart sync operations should not interfere with each other
|
||||
// This tests race conditions and database locking
|
||||
|
||||
let (state, user) = create_test_setup().await;
|
||||
|
||||
// Create initial directories
|
||||
let initial_dirs = vec![
|
||||
CreateWebDAVDirectory {
|
||||
user_id: user.id,
|
||||
directory_path: "/Concurrent1".to_string(),
|
||||
directory_etag: "concurrent1-etag".to_string(),
|
||||
file_count: 5,
|
||||
total_size_bytes: 500000,
|
||||
},
|
||||
CreateWebDAVDirectory {
|
||||
user_id: user.id,
|
||||
directory_path: "/Concurrent2".to_string(),
|
||||
directory_etag: "concurrent2-etag".to_string(),
|
||||
file_count: 3,
|
||||
total_size_bytes: 300000,
|
||||
},
|
||||
];
|
||||
|
||||
for dir in &initial_dirs {
|
||||
state.db.create_or_update_webdav_directory(dir).await
|
||||
.expect("Failed to create initial directory");
|
||||
}
|
||||
|
||||
// Run multiple concurrent operations
|
||||
let num_concurrent = 5;
|
||||
let mut handles = Vec::new();
|
||||
|
||||
for i in 0..num_concurrent {
|
||||
let state_clone = state.clone();
|
||||
let user_id = user.id;
|
||||
|
||||
let handle = tokio::spawn(async move {
|
||||
// Each task tries to create/update directories concurrently
|
||||
let dir = CreateWebDAVDirectory {
|
||||
user_id,
|
||||
directory_path: format!("/Concurrent{}", i + 10),
|
||||
directory_etag: format!("concurrent{}-etag", i + 10),
|
||||
file_count: i as i32 + 1,
|
||||
total_size_bytes: (i as i64 + 1) * 100000,
|
||||
};
|
||||
|
||||
// Add some delay to increase chance of race conditions
|
||||
tokio::time::sleep(tokio::time::Duration::from_millis(i as u64 * 10)).await;
|
||||
|
||||
state_clone.db.create_or_update_webdav_directory(&dir).await
|
||||
});
|
||||
|
||||
handles.push(handle);
|
||||
}
|
||||
|
||||
// Wait for all concurrent operations to complete
|
||||
let mut results = Vec::new();
|
||||
for handle in handles {
|
||||
results.push(handle.await.unwrap());
|
||||
}
|
||||
|
||||
// Verify all operations succeeded
|
||||
let mut success_count = 0;
|
||||
let mut error_count = 0;
|
||||
|
||||
for result in results {
|
||||
match result {
|
||||
Ok(_) => success_count += 1,
|
||||
Err(e) => {
|
||||
error_count += 1;
|
||||
println!("Concurrent operation error: {}", e);
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
println!("Concurrent operations: {} successful, {} failed", success_count, error_count);
|
||||
|
||||
// Verify final database state
|
||||
let final_dirs = state.db.list_webdav_directories(user.id).await.unwrap();
|
||||
let expected_total = initial_dirs.len() + success_count;
|
||||
assert_eq!(final_dirs.len(), expected_total,
|
||||
"Should have {} directories after concurrent operations", expected_total);
|
||||
|
||||
// Verify original directories are still intact
|
||||
let concurrent1 = final_dirs.iter().find(|d| d.directory_path == "/Concurrent1").unwrap();
|
||||
assert_eq!(concurrent1.directory_etag, "concurrent1-etag");
|
||||
|
||||
let concurrent2 = final_dirs.iter().find(|d| d.directory_path == "/Concurrent2").unwrap();
|
||||
assert_eq!(concurrent2.directory_etag, "concurrent2-etag");
|
||||
|
||||
println!("✅ Concurrent smart sync operations test completed successfully");
|
||||
println!(" {} initial directories preserved", initial_dirs.len());
|
||||
println!(" {} concurrent operations executed", num_concurrent);
|
||||
println!(" {} operations successful", success_count);
|
||||
}
|
||||
|
||||
#[tokio::test]
|
||||
async fn test_malformed_data_recovery() {
|
||||
// Integration Test: System should handle and recover from malformed data gracefully
|
||||
// This tests robustness against data corruption scenarios
|
||||
|
||||
let (state, user) = create_test_setup().await;
|
||||
|
||||
// Create a directory with normal data first
|
||||
let normal_dir = CreateWebDAVDirectory {
|
||||
user_id: user.id,
|
||||
directory_path: "/Normal".to_string(),
|
||||
directory_etag: "normal-etag".to_string(),
|
||||
file_count: 10,
|
||||
total_size_bytes: 1000000,
|
||||
};
|
||||
|
||||
state.db.create_or_update_webdav_directory(&normal_dir).await
|
||||
.expect("Normal directory creation should work");
|
||||
|
||||
// Test with edge case data
|
||||
let edge_cases = vec![
|
||||
CreateWebDAVDirectory {
|
||||
user_id: user.id,
|
||||
directory_path: "/EmptyPath".to_string(),
|
||||
directory_etag: "".to_string(), // Empty ETag
|
||||
file_count: 0,
|
||||
total_size_bytes: 0,
|
||||
},
|
||||
CreateWebDAVDirectory {
|
||||
user_id: user.id,
|
||||
directory_path: "/SpecialChars".to_string(),
|
||||
directory_etag: "etag-with-special-chars-!@#$%^&*()".to_string(),
|
||||
file_count: -1, // Invalid negative count (should be handled)
|
||||
total_size_bytes: -1000, // Invalid negative size
|
||||
},
|
||||
CreateWebDAVDirectory {
|
||||
user_id: user.id,
|
||||
directory_path: "/VeryLongPath/With/Many/Nested/Directories/That/Goes/On/And/On/For/A/Very/Long/Time/To/Test/Path/Length/Limits".to_string(),
|
||||
directory_etag: "very-long-etag-that-might-exceed-normal-database-field-lengths-and-cause-truncation-issues-if-not-handled-properly".to_string(),
|
||||
file_count: i32::MAX, // Maximum integer value
|
||||
total_size_bytes: i64::MAX, // Maximum long value
|
||||
},
|
||||
];
|
||||
|
||||
let mut successful_edge_cases = 0;
|
||||
let mut failed_edge_cases = 0;
|
||||
|
||||
for edge_case in edge_cases {
|
||||
match state.db.create_or_update_webdav_directory(&edge_case).await {
|
||||
Ok(_) => {
|
||||
successful_edge_cases += 1;
|
||||
println!("✅ Edge case handled successfully: {}", edge_case.directory_path);
|
||||
}
|
||||
Err(e) => {
|
||||
failed_edge_cases += 1;
|
||||
println!("⚠️ Edge case failed as expected: {} - {}", edge_case.directory_path, e);
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
// Verify the normal directory is still accessible
|
||||
let all_dirs = state.db.list_webdav_directories(user.id).await.unwrap();
|
||||
let normal_dir_exists = all_dirs.iter().any(|d| d.directory_path == "/Normal");
|
||||
assert!(normal_dir_exists, "Normal directory should still exist after edge case testing");
|
||||
|
||||
// Verify database is still functional
|
||||
let test_dir = CreateWebDAVDirectory {
|
||||
user_id: user.id,
|
||||
directory_path: "/AfterEdgeCases".to_string(),
|
||||
directory_etag: "after-edge-cases".to_string(),
|
||||
file_count: 5,
|
||||
total_size_bytes: 500000,
|
||||
};
|
||||
|
||||
state.db.create_or_update_webdav_directory(&test_dir).await
|
||||
.expect("Database should still work after edge case testing");
|
||||
|
||||
let final_dirs = state.db.list_webdav_directories(user.id).await.unwrap();
|
||||
let after_edge_case_dir = final_dirs.iter().find(|d| d.directory_path == "/AfterEdgeCases").unwrap();
|
||||
assert_eq!(after_edge_case_dir.directory_etag, "after-edge-cases");
|
||||
|
||||
println!("✅ Malformed data recovery test completed successfully");
|
||||
println!(" {} edge cases handled successfully", successful_edge_cases);
|
||||
println!(" {} edge cases failed as expected", failed_edge_cases);
|
||||
println!(" Database remains functional after edge case testing");
|
||||
}
|
||||
|
|
@ -0,0 +1,138 @@
|
|||
use std::sync::Arc;
|
||||
use readur::{
|
||||
AppState,
|
||||
models::{CreateWebDAVDirectory, User, AuthProvider},
|
||||
services::webdav::{SmartSyncService, SmartSyncStrategy, SmartSyncDecision, WebDAVService, WebDAVConfig},
|
||||
test_utils::{TestContext, TestAuthHelper},
|
||||
};
|
||||
|
||||
/// Helper function to create test database and user
|
||||
async fn create_test_setup() -> (Arc<AppState>, User) {
|
||||
let test_context = TestContext::new().await;
|
||||
let auth_helper = TestAuthHelper::new(test_context.app().clone());
|
||||
let test_user = auth_helper.create_test_user().await;
|
||||
|
||||
// Convert TestUser to User model for compatibility
|
||||
let user = User {
|
||||
id: test_user.user_response.id,
|
||||
username: test_user.user_response.username,
|
||||
email: test_user.user_response.email,
|
||||
password_hash: Some("hashed_password".to_string()),
|
||||
role: test_user.user_response.role,
|
||||
created_at: chrono::Utc::now(),
|
||||
updated_at: chrono::Utc::now(),
|
||||
oidc_subject: None,
|
||||
oidc_issuer: None,
|
||||
oidc_email: None,
|
||||
auth_provider: AuthProvider::Local,
|
||||
};
|
||||
|
||||
(test_context.state().clone(), user)
|
||||
}
|
||||
|
||||
/// Helper function to create WebDAV service for testing
|
||||
fn create_test_webdav_service() -> WebDAVService {
|
||||
let config = WebDAVConfig {
|
||||
server_url: "https://test.example.com".to_string(),
|
||||
username: "test".to_string(),
|
||||
password: "test".to_string(),
|
||||
watch_folders: vec!["/Documents".to_string()],
|
||||
file_extensions: vec!["pdf".to_string(), "txt".to_string()],
|
||||
timeout_seconds: 30,
|
||||
server_type: Some("generic".to_string()),
|
||||
};
|
||||
|
||||
WebDAVService::new(config).expect("Failed to create WebDAV service")
|
||||
}
|
||||
|
||||
#[tokio::test]
|
||||
async fn test_first_time_sync_full_deep_scan() {
|
||||
// Integration Test: First-time sync with no existing directory ETags
|
||||
// Expected: Should perform full deep scan and save all discovered directory ETags
|
||||
|
||||
let (state, user) = create_test_setup().await;
|
||||
let smart_sync_service = SmartSyncService::new(state.clone());
|
||||
|
||||
// Verify no existing directories tracked
|
||||
let existing_dirs = state.db.list_webdav_directories(user.id).await.unwrap();
|
||||
assert_eq!(existing_dirs.len(), 0, "Should start with no tracked directories");
|
||||
|
||||
// Test evaluation for first-time sync
|
||||
let webdav_service = create_test_webdav_service();
|
||||
let decision = smart_sync_service.evaluate_sync_need(user.id, &webdav_service, "/Documents").await;
|
||||
|
||||
match decision {
|
||||
Ok(SmartSyncDecision::RequiresSync(SmartSyncStrategy::FullDeepScan)) => {
|
||||
println!("✅ First-time sync correctly requires FullDeepScan");
|
||||
}
|
||||
Ok(other) => panic!("Expected FullDeepScan for first-time sync, got: {:?}", other),
|
||||
Err(e) => {
|
||||
// WebDAV service will fail in test environment, but the decision logic should still work
|
||||
println!("⚠️ WebDAV service failed as expected in test environment: {}", e);
|
||||
// This is acceptable since we're testing the logic, not the actual WebDAV connection
|
||||
}
|
||||
}
|
||||
|
||||
println!("✅ First-time sync test completed successfully");
|
||||
}
|
||||
|
||||
#[tokio::test]
|
||||
async fn test_first_time_sync_saves_directory_etags() {
|
||||
// Integration Test: First-time sync should save discovered directory ETags to database
|
||||
// This test focuses on the database persistence aspect
|
||||
|
||||
let (state, user) = create_test_setup().await;
|
||||
|
||||
// Manually create directories that would be discovered by WebDAV
|
||||
let discovered_directories = vec![
|
||||
CreateWebDAVDirectory {
|
||||
user_id: user.id,
|
||||
directory_path: "/Documents".to_string(),
|
||||
directory_etag: "root-etag-123".to_string(),
|
||||
file_count: 10,
|
||||
total_size_bytes: 1024000,
|
||||
},
|
||||
CreateWebDAVDirectory {
|
||||
user_id: user.id,
|
||||
directory_path: "/Documents/Projects".to_string(),
|
||||
directory_etag: "projects-etag-456".to_string(),
|
||||
file_count: 5,
|
||||
total_size_bytes: 512000,
|
||||
},
|
||||
CreateWebDAVDirectory {
|
||||
user_id: user.id,
|
||||
directory_path: "/Documents/Archive".to_string(),
|
||||
directory_etag: "archive-etag-789".to_string(),
|
||||
file_count: 20,
|
||||
total_size_bytes: 2048000,
|
||||
},
|
||||
];
|
||||
|
||||
// Save directories (simulating what would happen after WebDAV discovery)
|
||||
for dir in &discovered_directories {
|
||||
state.db.create_or_update_webdav_directory(dir).await
|
||||
.expect("Failed to save directory");
|
||||
}
|
||||
|
||||
// Verify directories were saved
|
||||
let saved_dirs = state.db.list_webdav_directories(user.id).await.unwrap();
|
||||
assert_eq!(saved_dirs.len(), 3, "Should have saved 3 directories");
|
||||
|
||||
// Verify specific directories and their ETags
|
||||
let documents_dir = saved_dirs.iter().find(|d| d.directory_path == "/Documents").unwrap();
|
||||
assert_eq!(documents_dir.directory_etag, "root-etag-123");
|
||||
assert_eq!(documents_dir.file_count, 10);
|
||||
assert_eq!(documents_dir.total_size_bytes, 1024000);
|
||||
|
||||
let projects_dir = saved_dirs.iter().find(|d| d.directory_path == "/Documents/Projects").unwrap();
|
||||
assert_eq!(projects_dir.directory_etag, "projects-etag-456");
|
||||
assert_eq!(projects_dir.file_count, 5);
|
||||
assert_eq!(projects_dir.total_size_bytes, 512000);
|
||||
|
||||
let archive_dir = saved_dirs.iter().find(|d| d.directory_path == "/Documents/Archive").unwrap();
|
||||
assert_eq!(archive_dir.directory_etag, "archive-etag-789");
|
||||
assert_eq!(archive_dir.file_count, 20);
|
||||
assert_eq!(archive_dir.total_size_bytes, 2048000);
|
||||
|
||||
println!("✅ First-time sync directory ETag persistence test completed successfully");
|
||||
}
|
||||
|
|
@ -0,0 +1,172 @@
|
|||
use std::sync::Arc;
|
||||
use readur::{
|
||||
AppState,
|
||||
models::{CreateWebDAVDirectory, User, AuthProvider},
|
||||
services::webdav::{SmartSyncService, SmartSyncStrategy, SmartSyncDecision, WebDAVService, WebDAVConfig},
|
||||
test_utils::{TestContext, TestAuthHelper},
|
||||
};
|
||||
|
||||
/// Helper function to create test database and user
|
||||
async fn create_test_setup() -> (Arc<AppState>, User) {
|
||||
let test_context = TestContext::new().await;
|
||||
let auth_helper = TestAuthHelper::new(test_context.app().clone());
|
||||
let test_user = auth_helper.create_test_user().await;
|
||||
|
||||
// Convert TestUser to User model for compatibility
|
||||
let user = User {
|
||||
id: test_user.user_response.id,
|
||||
username: test_user.user_response.username,
|
||||
email: test_user.user_response.email,
|
||||
password_hash: Some("hashed_password".to_string()),
|
||||
role: test_user.user_response.role,
|
||||
created_at: chrono::Utc::now(),
|
||||
updated_at: chrono::Utc::now(),
|
||||
oidc_subject: None,
|
||||
oidc_issuer: None,
|
||||
oidc_email: None,
|
||||
auth_provider: AuthProvider::Local,
|
||||
};
|
||||
|
||||
(test_context.state().clone(), user)
|
||||
}
|
||||
|
||||
/// Helper function to create WebDAV service for testing
|
||||
fn create_test_webdav_service() -> WebDAVService {
|
||||
let config = WebDAVConfig {
|
||||
server_url: "https://test.example.com".to_string(),
|
||||
username: "test".to_string(),
|
||||
password: "test".to_string(),
|
||||
watch_folders: vec!["/Documents".to_string()],
|
||||
file_extensions: vec!["pdf".to_string(), "txt".to_string()],
|
||||
timeout_seconds: 30,
|
||||
server_type: Some("generic".to_string()),
|
||||
};
|
||||
|
||||
WebDAVService::new(config).expect("Failed to create WebDAV service")
|
||||
}
|
||||
|
||||
#[tokio::test]
|
||||
async fn test_smart_sync_no_changes_skip() {
|
||||
// Integration Test: Smart sync with no directory changes should skip sync entirely
|
||||
// Expected: Should return SkipSync when all directory ETags are unchanged
|
||||
|
||||
let (state, user) = create_test_setup().await;
|
||||
let smart_sync_service = SmartSyncService::new(state.clone());
|
||||
|
||||
// Pre-populate database with known directory ETags
|
||||
let known_directories = vec![
|
||||
CreateWebDAVDirectory {
|
||||
user_id: user.id,
|
||||
directory_path: "/Documents".to_string(),
|
||||
directory_etag: "root-etag-unchanged".to_string(),
|
||||
file_count: 8,
|
||||
total_size_bytes: 800000,
|
||||
},
|
||||
CreateWebDAVDirectory {
|
||||
user_id: user.id,
|
||||
directory_path: "/Documents/Projects".to_string(),
|
||||
directory_etag: "projects-etag-unchanged".to_string(),
|
||||
file_count: 12,
|
||||
total_size_bytes: 1200000,
|
||||
},
|
||||
CreateWebDAVDirectory {
|
||||
user_id: user.id,
|
||||
directory_path: "/Documents/Archive".to_string(),
|
||||
directory_etag: "archive-etag-unchanged".to_string(),
|
||||
file_count: 25,
|
||||
total_size_bytes: 2500000,
|
||||
},
|
||||
];
|
||||
|
||||
for dir in &known_directories {
|
||||
state.db.create_or_update_webdav_directory(dir).await
|
||||
.expect("Failed to create directory tracking");
|
||||
}
|
||||
|
||||
// Verify known directories were created
|
||||
let stored_dirs = state.db.list_webdav_directories(user.id).await.unwrap();
|
||||
assert_eq!(stored_dirs.len(), 3, "Should have 3 known directories");
|
||||
|
||||
// In a real scenario, WebDAV would return the same ETags, indicating no changes
|
||||
// Since we can't mock the WebDAV service easily, we test the database logic
|
||||
|
||||
// Test bulk directory fetching (key performance optimization)
|
||||
let start_time = std::time::Instant::now();
|
||||
let fetched_dirs = state.db.list_webdav_directories(user.id).await.unwrap();
|
||||
let fetch_duration = start_time.elapsed();
|
||||
|
||||
assert!(fetch_duration.as_millis() < 50, "Bulk directory fetch should be fast");
|
||||
assert_eq!(fetched_dirs.len(), 3, "Should fetch all directories efficiently");
|
||||
|
||||
// Verify directory data integrity
|
||||
let docs_dir = fetched_dirs.iter().find(|d| d.directory_path == "/Documents").unwrap();
|
||||
assert_eq!(docs_dir.directory_etag, "root-etag-unchanged");
|
||||
assert_eq!(docs_dir.file_count, 8);
|
||||
|
||||
let projects_dir = fetched_dirs.iter().find(|d| d.directory_path == "/Documents/Projects").unwrap();
|
||||
assert_eq!(projects_dir.directory_etag, "projects-etag-unchanged");
|
||||
assert_eq!(projects_dir.file_count, 12);
|
||||
|
||||
let archive_dir = fetched_dirs.iter().find(|d| d.directory_path == "/Documents/Archive").unwrap();
|
||||
assert_eq!(archive_dir.directory_etag, "archive-etag-unchanged");
|
||||
assert_eq!(archive_dir.file_count, 25);
|
||||
|
||||
println!("✅ No changes sync test completed successfully - bulk fetch in {:?}", fetch_duration);
|
||||
}
|
||||
|
||||
#[tokio::test]
|
||||
async fn test_directory_etag_comparison_efficiency() {
|
||||
// Integration Test: Directory ETag comparison should be efficient for large numbers of directories
|
||||
// This tests the bulk fetching performance optimization
|
||||
|
||||
let (state, user) = create_test_setup().await;
|
||||
|
||||
// Create a larger number of directories to test performance
|
||||
let num_directories = 100;
|
||||
let mut directories = Vec::new();
|
||||
|
||||
for i in 0..num_directories {
|
||||
directories.push(CreateWebDAVDirectory {
|
||||
user_id: user.id,
|
||||
directory_path: format!("/Documents/Folder{:03}", i),
|
||||
directory_etag: format!("etag-folder-{:03}", i),
|
||||
file_count: i as i32 % 10 + 1, // 1-10 files per directory
|
||||
total_size_bytes: (i as i64 + 1) * 10000, // Varying sizes
|
||||
});
|
||||
}
|
||||
|
||||
// Batch insert directories
|
||||
let insert_start = std::time::Instant::now();
|
||||
for dir in &directories {
|
||||
state.db.create_or_update_webdav_directory(dir).await
|
||||
.expect("Failed to create directory");
|
||||
}
|
||||
let insert_duration = insert_start.elapsed();
|
||||
|
||||
// Test bulk fetch performance
|
||||
let fetch_start = std::time::Instant::now();
|
||||
let fetched_dirs = state.db.list_webdav_directories(user.id).await.unwrap();
|
||||
let fetch_duration = fetch_start.elapsed();
|
||||
|
||||
// Verify all directories were created and fetched
|
||||
assert_eq!(fetched_dirs.len(), num_directories, "Should fetch all {} directories", num_directories);
|
||||
|
||||
// Performance assertions
|
||||
assert!(fetch_duration.as_millis() < 200, "Bulk fetch of {} directories should be under 200ms, got {:?}", num_directories, fetch_duration);
|
||||
assert!(insert_duration.as_millis() < 5000, "Bulk insert of {} directories should be under 5s, got {:?}", num_directories, insert_duration);
|
||||
|
||||
// Verify data integrity on a few random directories
|
||||
let dir_50 = fetched_dirs.iter().find(|d| d.directory_path == "/Documents/Folder050").unwrap();
|
||||
assert_eq!(dir_50.directory_etag, "etag-folder-050");
|
||||
assert_eq!(dir_50.file_count, 1); // 50 % 10 + 1 = 1
|
||||
assert_eq!(dir_50.total_size_bytes, 510000); // (50 + 1) * 10000
|
||||
|
||||
let dir_99 = fetched_dirs.iter().find(|d| d.directory_path == "/Documents/Folder099").unwrap();
|
||||
assert_eq!(dir_99.directory_etag, "etag-folder-099");
|
||||
assert_eq!(dir_99.file_count, 10); // 99 % 10 + 1 = 10
|
||||
assert_eq!(dir_99.total_size_bytes, 1000000); // (99 + 1) * 10000
|
||||
|
||||
println!("✅ Directory ETag comparison efficiency test completed successfully");
|
||||
println!(" Created {} directories in {:?}", num_directories, insert_duration);
|
||||
println!(" Fetched {} directories in {:?}", num_directories, fetch_duration);
|
||||
}
|
||||
|
|
@ -0,0 +1,259 @@
|
|||
use std::sync::Arc;
|
||||
use readur::{
|
||||
AppState,
|
||||
models::{CreateWebDAVDirectory, User, AuthProvider},
|
||||
services::webdav::{SmartSyncService, SmartSyncStrategy, SmartSyncDecision, WebDAVService, WebDAVConfig},
|
||||
test_utils::{TestContext, TestAuthHelper},
|
||||
};
|
||||
|
||||
/// Helper function to create test database and user
|
||||
async fn create_test_setup() -> (Arc<AppState>, User) {
|
||||
let test_context = TestContext::new().await;
|
||||
let auth_helper = TestAuthHelper::new(test_context.app().clone());
|
||||
let test_user = auth_helper.create_test_user().await;
|
||||
|
||||
// Convert TestUser to User model for compatibility
|
||||
let user = User {
|
||||
id: test_user.user_response.id,
|
||||
username: test_user.user_response.username,
|
||||
email: test_user.user_response.email,
|
||||
password_hash: Some("hashed_password".to_string()),
|
||||
role: test_user.user_response.role,
|
||||
created_at: chrono::Utc::now(),
|
||||
updated_at: chrono::Utc::now(),
|
||||
oidc_subject: None,
|
||||
oidc_issuer: None,
|
||||
oidc_email: None,
|
||||
auth_provider: AuthProvider::Local,
|
||||
};
|
||||
|
||||
(test_context.state().clone(), user)
|
||||
}
|
||||
|
||||
/// Helper function to create WebDAV service for testing
|
||||
fn create_test_webdav_service() -> WebDAVService {
|
||||
let config = WebDAVConfig {
|
||||
server_url: "https://test.example.com".to_string(),
|
||||
username: "test".to_string(),
|
||||
password: "test".to_string(),
|
||||
watch_folders: vec!["/Documents".to_string()],
|
||||
file_extensions: vec!["pdf".to_string(), "txt".to_string()],
|
||||
timeout_seconds: 30,
|
||||
server_type: Some("generic".to_string()),
|
||||
};
|
||||
|
||||
WebDAVService::new(config).expect("Failed to create WebDAV service")
|
||||
}
|
||||
|
||||
#[tokio::test]
|
||||
async fn test_smart_sync_targeted_scan() {
|
||||
// Integration Test: Smart sync with single directory changed should use targeted scan
|
||||
// Expected: Should return RequiresSync(TargetedScan) when only a few directories have changed
|
||||
|
||||
let (state, user) = create_test_setup().await;
|
||||
|
||||
// Create a scenario with many directories, where only one has changed
|
||||
let unchanged_directories = vec![
|
||||
("/Documents", "root-etag-stable"),
|
||||
("/Documents/Projects", "projects-etag-stable"),
|
||||
("/Documents/Archive", "archive-etag-stable"),
|
||||
("/Documents/Photos", "photos-etag-stable"),
|
||||
("/Documents/Music", "music-etag-stable"),
|
||||
("/Documents/Videos", "videos-etag-stable"),
|
||||
("/Documents/Backup", "backup-etag-stable"),
|
||||
("/Documents/Personal", "personal-etag-stable"),
|
||||
("/Documents/Work", "work-etag-stable"),
|
||||
("/Documents/Temp", "temp-etag-stable"), // 10 directories total
|
||||
];
|
||||
|
||||
// Pre-populate database with known directory ETags
|
||||
for (path, etag) in &unchanged_directories {
|
||||
let dir = CreateWebDAVDirectory {
|
||||
user_id: user.id,
|
||||
directory_path: path.to_string(),
|
||||
directory_etag: etag.to_string(),
|
||||
file_count: 5,
|
||||
total_size_bytes: 500000,
|
||||
};
|
||||
|
||||
state.db.create_or_update_webdav_directory(&dir).await
|
||||
.expect("Failed to create directory tracking");
|
||||
}
|
||||
|
||||
// Verify directories were created
|
||||
let stored_dirs = state.db.list_webdav_directories(user.id).await.unwrap();
|
||||
assert_eq!(stored_dirs.len(), 10, "Should have 10 tracked directories");
|
||||
|
||||
// Test the strategy selection logic for targeted scan
|
||||
// When few directories change (<=30% and <=5 new), should use targeted scan
|
||||
let change_ratio = 1.0 / 10.0; // 1 changed out of 10 = 10%
|
||||
let new_dirs_count = 0; // No new directories
|
||||
|
||||
let should_use_targeted = change_ratio <= 0.3 && new_dirs_count <= 5;
|
||||
assert!(should_use_targeted, "Should use targeted scan for small changes: {:.1}% change ratio", change_ratio * 100.0);
|
||||
|
||||
println!("✅ Targeted scan strategy selection test passed - 10% change triggers targeted scan");
|
||||
}
|
||||
|
||||
#[tokio::test]
|
||||
async fn test_targeted_scan_vs_full_scan_thresholds() {
|
||||
// Integration Test: Test various scenarios for when to use targeted vs full scan
|
||||
// Expected: Strategy should be chosen based on change ratio and new directory count
|
||||
|
||||
let (state, user) = create_test_setup().await;
|
||||
|
||||
// Create base directories for testing different scenarios
|
||||
let base_directories = 20; // Start with 20 directories
|
||||
for i in 0..base_directories {
|
||||
let dir = CreateWebDAVDirectory {
|
||||
user_id: user.id,
|
||||
directory_path: format!("/Documents/Base{:02}", i),
|
||||
directory_etag: format!("base-etag-{:02}", i),
|
||||
file_count: 3,
|
||||
total_size_bytes: 300000,
|
||||
};
|
||||
|
||||
state.db.create_or_update_webdav_directory(&dir).await
|
||||
.expect("Failed to create base directory");
|
||||
}
|
||||
|
||||
// Test Scenario 1: Low change ratio, few new dirs -> Targeted scan
|
||||
let scenario1_changes = 2; // 2 out of 20 = 10%
|
||||
let scenario1_new = 1; // 1 new directory
|
||||
let scenario1_ratio = scenario1_changes as f64 / base_directories as f64;
|
||||
let scenario1_targeted = scenario1_ratio <= 0.3 && scenario1_new <= 5;
|
||||
assert!(scenario1_targeted, "Scenario 1 should use targeted scan: {:.1}% changes, {} new", scenario1_ratio * 100.0, scenario1_new);
|
||||
|
||||
// Test Scenario 2: High change ratio -> Full scan
|
||||
let scenario2_changes = 8; // 8 out of 20 = 40%
|
||||
let scenario2_new = 2; // 2 new directories
|
||||
let scenario2_ratio = scenario2_changes as f64 / base_directories as f64;
|
||||
let scenario2_full_scan = scenario2_ratio > 0.3 || scenario2_new > 5;
|
||||
assert!(scenario2_full_scan, "Scenario 2 should use full scan: {:.1}% changes, {} new", scenario2_ratio * 100.0, scenario2_new);
|
||||
|
||||
// Test Scenario 3: Low change ratio but many new dirs -> Full scan
|
||||
let scenario3_changes = 1; // 1 out of 20 = 5%
|
||||
let scenario3_new = 7; // 7 new directories
|
||||
let scenario3_ratio = scenario3_changes as f64 / base_directories as f64;
|
||||
let scenario3_full_scan = scenario3_ratio > 0.3 || scenario3_new > 5;
|
||||
assert!(scenario3_full_scan, "Scenario 3 should use full scan: {:.1}% changes, {} new", scenario3_ratio * 100.0, scenario3_new);
|
||||
|
||||
// Test Scenario 4: Edge case - exactly at threshold -> Targeted scan
|
||||
let scenario4_changes = 6; // 6 out of 20 = 30% (exactly at threshold)
|
||||
let scenario4_new = 5; // 5 new directories (exactly at threshold)
|
||||
let scenario4_ratio = scenario4_changes as f64 / base_directories as f64;
|
||||
let scenario4_targeted = scenario4_ratio <= 0.3 && scenario4_new <= 5;
|
||||
assert!(scenario4_targeted, "Scenario 4 should use targeted scan: {:.1}% changes, {} new", scenario4_ratio * 100.0, scenario4_new);
|
||||
|
||||
println!("✅ All targeted vs full scan threshold tests passed:");
|
||||
println!(" Scenario 1 (10% changes, 1 new): Targeted scan");
|
||||
println!(" Scenario 2 (40% changes, 2 new): Full scan");
|
||||
println!(" Scenario 3 (5% changes, 7 new): Full scan");
|
||||
println!(" Scenario 4 (30% changes, 5 new): Targeted scan");
|
||||
}
|
||||
|
||||
#[tokio::test]
|
||||
async fn test_directory_change_detection_logic() {
|
||||
// Integration Test: Test the logic for detecting changed, new, and unchanged directories
|
||||
// This is the core of the targeted scan decision making
|
||||
|
||||
let (state, user) = create_test_setup().await;
|
||||
|
||||
// Set up known directories in database
|
||||
let known_dirs = vec![
|
||||
("/Documents", "root-etag-old"),
|
||||
("/Documents/Projects", "projects-etag-stable"),
|
||||
("/Documents/Archive", "archive-etag-old"),
|
||||
("/Documents/ToBeDeleted", "deleted-etag"), // This won't appear in "current"
|
||||
];
|
||||
|
||||
for (path, etag) in &known_dirs {
|
||||
let dir = CreateWebDAVDirectory {
|
||||
user_id: user.id,
|
||||
directory_path: path.to_string(),
|
||||
directory_etag: etag.to_string(),
|
||||
file_count: 3,
|
||||
total_size_bytes: 300000,
|
||||
};
|
||||
|
||||
state.db.create_or_update_webdav_directory(&dir).await
|
||||
.expect("Failed to create known directory");
|
||||
}
|
||||
|
||||
// Simulate current directories from WebDAV (what we'd get from discovery)
|
||||
use std::collections::HashMap;
|
||||
let current_dirs = vec![
|
||||
("/Documents", "root-etag-new"), // Changed
|
||||
("/Documents/Projects", "projects-etag-stable"), // Unchanged
|
||||
("/Documents/Archive", "archive-etag-new"), // Changed
|
||||
("/Documents/NewFolder", "new-folder-etag"), // New
|
||||
];
|
||||
let current_map: HashMap<String, String> = current_dirs.into_iter()
|
||||
.map(|(p, e)| (p.to_string(), e.to_string()))
|
||||
.collect();
|
||||
|
||||
// Get known directories from database
|
||||
let known_map: HashMap<String, String> = state.db.list_webdav_directories(user.id).await
|
||||
.expect("Failed to get known directories")
|
||||
.into_iter()
|
||||
.filter(|dir| dir.directory_path.starts_with("/Documents"))
|
||||
.map(|dir| (dir.directory_path, dir.directory_etag))
|
||||
.collect();
|
||||
|
||||
// Perform comparison logic (mirrors SmartSyncService logic)
|
||||
let mut changed_directories = Vec::new();
|
||||
let mut new_directories = Vec::new();
|
||||
let mut unchanged_directories = Vec::new();
|
||||
|
||||
for (current_path, current_etag) in ¤t_map {
|
||||
match known_map.get(current_path) {
|
||||
Some(known_etag) => {
|
||||
if known_etag != current_etag {
|
||||
changed_directories.push(current_path.clone());
|
||||
} else {
|
||||
unchanged_directories.push(current_path.clone());
|
||||
}
|
||||
}
|
||||
None => {
|
||||
new_directories.push(current_path.clone());
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
// Detect deleted directories (in database but not in current WebDAV response)
|
||||
let mut deleted_directories = Vec::new();
|
||||
for known_path in known_map.keys() {
|
||||
if !current_map.contains_key(known_path) {
|
||||
deleted_directories.push(known_path.clone());
|
||||
}
|
||||
}
|
||||
|
||||
// Verify comparison results
|
||||
assert_eq!(changed_directories.len(), 2, "Should detect 2 changed directories");
|
||||
assert!(changed_directories.contains(&"/Documents".to_string()));
|
||||
assert!(changed_directories.contains(&"/Documents/Archive".to_string()));
|
||||
|
||||
assert_eq!(new_directories.len(), 1, "Should detect 1 new directory");
|
||||
assert!(new_directories.contains(&"/Documents/NewFolder".to_string()));
|
||||
|
||||
assert_eq!(unchanged_directories.len(), 1, "Should detect 1 unchanged directory");
|
||||
assert!(unchanged_directories.contains(&"/Documents/Projects".to_string()));
|
||||
|
||||
assert_eq!(deleted_directories.len(), 1, "Should detect 1 deleted directory");
|
||||
assert!(deleted_directories.contains(&"/Documents/ToBeDeleted".to_string()));
|
||||
|
||||
// Calculate strategy
|
||||
let total_known = known_map.len();
|
||||
let change_ratio = (changed_directories.len() + deleted_directories.len()) as f64 / total_known as f64;
|
||||
let new_dirs_count = new_directories.len();
|
||||
|
||||
let should_use_targeted = change_ratio <= 0.3 && new_dirs_count <= 5;
|
||||
|
||||
println!("✅ Directory change detection logic test completed successfully:");
|
||||
println!(" Changed: {} directories", changed_directories.len());
|
||||
println!(" New: {} directories", new_directories.len());
|
||||
println!(" Unchanged: {} directories", unchanged_directories.len());
|
||||
println!(" Deleted: {} directories", deleted_directories.len());
|
||||
println!(" Change ratio: {:.1}%", change_ratio * 100.0);
|
||||
println!(" Strategy: {}", if should_use_targeted { "Targeted scan" } else { "Full scan" });
|
||||
}
|
||||
Loading…
Reference in New Issue