test(timeline): add integration tests for full seed-expand-collect pipeline
Adds tests/timeline_pipeline_tests.rs with end-to-end integration tests that exercise the complete timeline pipeline against an in-memory SQLite database with realistic data: - pipeline_seed_expand_collect_end_to_end: Full scenario with an issue closed by an MR, state changes, and label events. Verifies that seed finds entities via FTS, expand discovers the closing MR through the entity_references graph, and collect assembles a chronologically sorted event stream containing Created, StateChanged, LabelAdded, and Merged events. - pipeline_empty_query_produces_empty_result: Validates graceful degradation when FTS returns zero matches -- all three stages should produce empty results without errors. - pipeline_since_filter_excludes_old_events: Verifies that the since timestamp filter propagates correctly through collect, excluding events before the cutoff while retaining newer ones. - pipeline_unresolved_refs_have_optional_iid: Tests the Option<i64> target_iid on UnresolvedRef by creating cross-project references both with and without known IIDs. - shared_resolve_entity_ref_scoping: Unit tests for the new shared resolve_entity_ref helper covering project-scoped lookup, unscoped lookup, wrong-project rejection, unknown entity types, and nonexistent entity IDs. Co-Authored-By: Claude Opus 4.6 <noreply@anthropic.com>
This commit is contained in:
354
tests/timeline_pipeline_tests.rs
Normal file
354
tests/timeline_pipeline_tests.rs
Normal file
@@ -0,0 +1,354 @@
|
||||
use lore::core::db::{create_connection, run_migrations};
|
||||
use lore::core::timeline::{TimelineEventType, resolve_entity_ref};
|
||||
use lore::core::timeline_collect::collect_events;
|
||||
use lore::core::timeline_expand::expand_timeline;
|
||||
use lore::core::timeline_seed::seed_timeline;
|
||||
use rusqlite::Connection;
|
||||
use std::path::Path;
|
||||
|
||||
fn setup_db() -> Connection {
|
||||
let conn = create_connection(Path::new(":memory:")).unwrap();
|
||||
run_migrations(&conn).unwrap();
|
||||
conn
|
||||
}
|
||||
|
||||
fn insert_project(conn: &Connection, path: &str) -> i64 {
|
||||
conn.execute(
|
||||
"INSERT INTO projects (gitlab_project_id, path_with_namespace, web_url) VALUES (?1, ?2, ?3)",
|
||||
rusqlite::params![1, path, format!("https://gitlab.com/{path}")],
|
||||
)
|
||||
.unwrap();
|
||||
conn.last_insert_rowid()
|
||||
}
|
||||
|
||||
fn insert_issue(conn: &Connection, project_id: i64, iid: i64, title: &str) -> i64 {
|
||||
conn.execute(
|
||||
"INSERT INTO issues (gitlab_id, project_id, iid, title, state, author_username, created_at, updated_at, last_seen_at, web_url) VALUES (?1, ?2, ?3, ?4, 'opened', 'alice', 1000, 2000, 3000, ?5)",
|
||||
rusqlite::params![iid * 100, project_id, iid, title, format!("https://gitlab.com/g/p/-/issues/{iid}")],
|
||||
)
|
||||
.unwrap();
|
||||
conn.last_insert_rowid()
|
||||
}
|
||||
|
||||
fn insert_mr(
|
||||
conn: &Connection,
|
||||
project_id: i64,
|
||||
iid: i64,
|
||||
title: &str,
|
||||
merged_at: Option<i64>,
|
||||
) -> i64 {
|
||||
conn.execute(
|
||||
"INSERT INTO merge_requests (gitlab_id, project_id, iid, title, state, author_username, created_at, updated_at, last_seen_at, merged_at, merge_user_username, web_url) VALUES (?1, ?2, ?3, ?4, 'merged', 'bob', 1500, 5000, 6000, ?5, 'charlie', ?6)",
|
||||
rusqlite::params![iid * 100, project_id, iid, title, merged_at, format!("https://gitlab.com/g/p/-/merge_requests/{iid}")],
|
||||
)
|
||||
.unwrap();
|
||||
conn.last_insert_rowid()
|
||||
}
|
||||
|
||||
fn insert_document(
|
||||
conn: &Connection,
|
||||
source_type: &str,
|
||||
source_id: i64,
|
||||
project_id: i64,
|
||||
content: &str,
|
||||
) {
|
||||
conn.execute(
|
||||
"INSERT INTO documents (source_type, source_id, project_id, content_text, content_hash) VALUES (?1, ?2, ?3, ?4, ?5)",
|
||||
rusqlite::params![source_type, source_id, project_id, content, format!("hash_{source_id}")],
|
||||
)
|
||||
.unwrap();
|
||||
}
|
||||
|
||||
fn insert_entity_ref(
|
||||
conn: &Connection,
|
||||
project_id: i64,
|
||||
source_type: &str,
|
||||
source_id: i64,
|
||||
target_type: &str,
|
||||
target_id: Option<i64>,
|
||||
ref_type: &str,
|
||||
) {
|
||||
conn.execute(
|
||||
"INSERT INTO entity_references (project_id, source_entity_type, source_entity_id, target_entity_type, target_entity_id, reference_type, source_method, created_at) VALUES (?1, ?2, ?3, ?4, ?5, ?6, 'api', 1000)",
|
||||
rusqlite::params![project_id, source_type, source_id, target_type, target_id, ref_type],
|
||||
)
|
||||
.unwrap();
|
||||
}
|
||||
|
||||
fn insert_state_event(
|
||||
conn: &Connection,
|
||||
project_id: i64,
|
||||
issue_id: Option<i64>,
|
||||
mr_id: Option<i64>,
|
||||
state: &str,
|
||||
created_at: i64,
|
||||
) {
|
||||
let gitlab_id: i64 = rand::random::<u32>().into();
|
||||
conn.execute(
|
||||
"INSERT INTO resource_state_events (gitlab_id, project_id, issue_id, merge_request_id, state, actor_username, created_at) VALUES (?1, ?2, ?3, ?4, ?5, 'alice', ?6)",
|
||||
rusqlite::params![gitlab_id, project_id, issue_id, mr_id, state, created_at],
|
||||
)
|
||||
.unwrap();
|
||||
}
|
||||
|
||||
fn insert_label_event(
|
||||
conn: &Connection,
|
||||
project_id: i64,
|
||||
issue_id: Option<i64>,
|
||||
label: &str,
|
||||
created_at: i64,
|
||||
) {
|
||||
let gitlab_id: i64 = rand::random::<u32>().into();
|
||||
conn.execute(
|
||||
"INSERT INTO resource_label_events (gitlab_id, project_id, issue_id, merge_request_id, action, label_name, actor_username, created_at) VALUES (?1, ?2, ?3, NULL, 'add', ?4, 'alice', ?5)",
|
||||
rusqlite::params![gitlab_id, project_id, issue_id, label, created_at],
|
||||
)
|
||||
.unwrap();
|
||||
}
|
||||
|
||||
/// Full pipeline: seed -> expand -> collect for a scenario with an issue
|
||||
/// that has a closing MR, state changes, and label events.
|
||||
#[test]
|
||||
fn pipeline_seed_expand_collect_end_to_end() {
|
||||
let conn = setup_db();
|
||||
let project_id = insert_project(&conn, "group/project");
|
||||
|
||||
// Issue #5: "authentication error in login"
|
||||
let issue_id = insert_issue(&conn, project_id, 5, "Authentication error in login");
|
||||
insert_document(
|
||||
&conn,
|
||||
"issue",
|
||||
issue_id,
|
||||
project_id,
|
||||
"authentication error in login flow causing 401",
|
||||
);
|
||||
|
||||
// MR !10 closes issue #5
|
||||
let mr_id = insert_mr(&conn, project_id, 10, "Fix auth bug", Some(4000));
|
||||
insert_document(
|
||||
&conn,
|
||||
"merge_request",
|
||||
mr_id,
|
||||
project_id,
|
||||
"fix authentication error in login module",
|
||||
);
|
||||
insert_entity_ref(
|
||||
&conn,
|
||||
project_id,
|
||||
"merge_request",
|
||||
mr_id,
|
||||
"issue",
|
||||
Some(issue_id),
|
||||
"closes",
|
||||
);
|
||||
|
||||
// State changes on issue
|
||||
insert_state_event(&conn, project_id, Some(issue_id), None, "closed", 3000);
|
||||
|
||||
// Label added to issue
|
||||
insert_label_event(&conn, project_id, Some(issue_id), "bug", 1500);
|
||||
|
||||
// SEED: find entities matching "authentication"
|
||||
let seed_result = seed_timeline(&conn, "authentication", None, None, 50, 10).unwrap();
|
||||
assert!(
|
||||
!seed_result.seed_entities.is_empty(),
|
||||
"Seed should find at least one entity"
|
||||
);
|
||||
|
||||
// Verify seeds contain the issue
|
||||
let has_issue = seed_result
|
||||
.seed_entities
|
||||
.iter()
|
||||
.any(|e| e.entity_type == "issue" && e.entity_iid == 5);
|
||||
assert!(has_issue, "Seeds should include issue #5");
|
||||
|
||||
// EXPAND: discover related entities (MR !10 via closes reference)
|
||||
let expand_result = expand_timeline(&conn, &seed_result.seed_entities, 1, false, 100).unwrap();
|
||||
|
||||
// The MR should appear as an expanded entity (or as a seed if it was also matched)
|
||||
let total_entities = seed_result.seed_entities.len() + expand_result.expanded_entities.len();
|
||||
assert!(total_entities >= 2, "Should have at least issue + MR");
|
||||
|
||||
// COLLECT: gather all events
|
||||
let events = collect_events(
|
||||
&conn,
|
||||
&seed_result.seed_entities,
|
||||
&expand_result.expanded_entities,
|
||||
&seed_result.evidence_notes,
|
||||
None,
|
||||
1000,
|
||||
)
|
||||
.unwrap();
|
||||
|
||||
assert!(!events.is_empty(), "Should have events");
|
||||
|
||||
// Verify chronological ordering
|
||||
for window in events.windows(2) {
|
||||
assert!(
|
||||
window[0].timestamp <= window[1].timestamp,
|
||||
"Events must be chronologically sorted: {} > {}",
|
||||
window[0].timestamp,
|
||||
window[1].timestamp
|
||||
);
|
||||
}
|
||||
|
||||
// Verify expected event types are present
|
||||
let has_created = events
|
||||
.iter()
|
||||
.any(|e| matches!(e.event_type, TimelineEventType::Created));
|
||||
let has_state_change = events
|
||||
.iter()
|
||||
.any(|e| matches!(e.event_type, TimelineEventType::StateChanged { .. }));
|
||||
let has_label = events
|
||||
.iter()
|
||||
.any(|e| matches!(e.event_type, TimelineEventType::LabelAdded { .. }));
|
||||
let has_merged = events
|
||||
.iter()
|
||||
.any(|e| matches!(e.event_type, TimelineEventType::Merged));
|
||||
|
||||
assert!(has_created, "Should have Created events");
|
||||
assert!(has_state_change, "Should have StateChanged events");
|
||||
assert!(has_label, "Should have LabelAdded events");
|
||||
assert!(has_merged, "Should have Merged event from MR");
|
||||
}
|
||||
|
||||
/// Verify the pipeline handles an empty FTS result gracefully.
|
||||
#[test]
|
||||
fn pipeline_empty_query_produces_empty_result() {
|
||||
let conn = setup_db();
|
||||
let _project_id = insert_project(&conn, "group/project");
|
||||
|
||||
let seed_result = seed_timeline(&conn, "", None, None, 50, 10).unwrap();
|
||||
assert!(seed_result.seed_entities.is_empty());
|
||||
|
||||
let expand_result = expand_timeline(&conn, &seed_result.seed_entities, 1, false, 100).unwrap();
|
||||
assert!(expand_result.expanded_entities.is_empty());
|
||||
|
||||
let events = collect_events(
|
||||
&conn,
|
||||
&seed_result.seed_entities,
|
||||
&expand_result.expanded_entities,
|
||||
&seed_result.evidence_notes,
|
||||
None,
|
||||
1000,
|
||||
)
|
||||
.unwrap();
|
||||
assert!(events.is_empty());
|
||||
}
|
||||
|
||||
/// Verify since filter propagates through the full pipeline.
|
||||
#[test]
|
||||
fn pipeline_since_filter_excludes_old_events() {
|
||||
let conn = setup_db();
|
||||
let project_id = insert_project(&conn, "group/project");
|
||||
|
||||
let issue_id = insert_issue(&conn, project_id, 1, "Deploy failure");
|
||||
insert_document(
|
||||
&conn,
|
||||
"issue",
|
||||
issue_id,
|
||||
project_id,
|
||||
"deploy failure in staging environment",
|
||||
);
|
||||
|
||||
// Old state change at 2000, recent state change at 8000
|
||||
insert_state_event(&conn, project_id, Some(issue_id), None, "closed", 2000);
|
||||
insert_state_event(&conn, project_id, Some(issue_id), None, "reopened", 8000);
|
||||
|
||||
let seed_result = seed_timeline(&conn, "deploy", None, None, 50, 10).unwrap();
|
||||
let expand_result = expand_timeline(&conn, &seed_result.seed_entities, 0, false, 100).unwrap();
|
||||
|
||||
// Collect with since=5000: should exclude Created(1000) and closed(2000)
|
||||
let events = collect_events(
|
||||
&conn,
|
||||
&seed_result.seed_entities,
|
||||
&expand_result.expanded_entities,
|
||||
&seed_result.evidence_notes,
|
||||
Some(5000),
|
||||
1000,
|
||||
)
|
||||
.unwrap();
|
||||
|
||||
assert_eq!(events.len(), 1, "Only the reopened event should survive");
|
||||
assert_eq!(events[0].timestamp, 8000);
|
||||
}
|
||||
|
||||
/// Verify unresolved references use Option<i64> for target_iid.
|
||||
#[test]
|
||||
fn pipeline_unresolved_refs_have_optional_iid() {
|
||||
let conn = setup_db();
|
||||
let project_id = insert_project(&conn, "group/project");
|
||||
|
||||
let issue_id = insert_issue(&conn, project_id, 1, "Cross-project reference");
|
||||
insert_document(
|
||||
&conn,
|
||||
"issue",
|
||||
issue_id,
|
||||
project_id,
|
||||
"cross project reference test",
|
||||
);
|
||||
|
||||
// Unresolved reference with known iid
|
||||
conn.execute(
|
||||
"INSERT INTO entity_references (project_id, source_entity_type, source_entity_id, target_entity_type, target_entity_id, target_project_path, target_entity_iid, reference_type, source_method, created_at) VALUES (?1, 'issue', ?2, 'issue', NULL, 'other/repo', 42, 'closes', 'description_parse', 1000)",
|
||||
rusqlite::params![project_id, issue_id],
|
||||
)
|
||||
.unwrap();
|
||||
|
||||
// Unresolved reference with NULL iid
|
||||
conn.execute(
|
||||
"INSERT INTO entity_references (project_id, source_entity_type, source_entity_id, target_entity_type, target_entity_id, target_project_path, target_entity_iid, reference_type, source_method, created_at) VALUES (?1, 'issue', ?2, 'merge_request', NULL, 'other/repo', NULL, 'related', 'note_parse', 1000)",
|
||||
rusqlite::params![project_id, issue_id],
|
||||
)
|
||||
.unwrap();
|
||||
|
||||
let seed_result = seed_timeline(&conn, "cross project", None, None, 50, 10).unwrap();
|
||||
let expand_result = expand_timeline(&conn, &seed_result.seed_entities, 1, false, 100).unwrap();
|
||||
|
||||
assert_eq!(expand_result.unresolved_references.len(), 2);
|
||||
|
||||
let with_iid = expand_result
|
||||
.unresolved_references
|
||||
.iter()
|
||||
.find(|r| r.target_type == "issue")
|
||||
.unwrap();
|
||||
assert_eq!(with_iid.target_iid, Some(42));
|
||||
|
||||
let without_iid = expand_result
|
||||
.unresolved_references
|
||||
.iter()
|
||||
.find(|r| r.target_type == "merge_request")
|
||||
.unwrap();
|
||||
assert_eq!(without_iid.target_iid, None);
|
||||
}
|
||||
|
||||
/// Verify the shared resolve_entity_ref works with and without project scoping.
|
||||
#[test]
|
||||
fn shared_resolve_entity_ref_scoping() {
|
||||
let conn = setup_db();
|
||||
let project_id = insert_project(&conn, "group/project");
|
||||
let issue_id = insert_issue(&conn, project_id, 42, "Test issue");
|
||||
|
||||
// Resolve with project filter
|
||||
let result = resolve_entity_ref(&conn, "issue", issue_id, Some(project_id)).unwrap();
|
||||
assert!(result.is_some());
|
||||
let entity = result.unwrap();
|
||||
assert_eq!(entity.entity_iid, 42);
|
||||
assert_eq!(entity.project_path, "group/project");
|
||||
|
||||
// Resolve without project filter
|
||||
let result = resolve_entity_ref(&conn, "issue", issue_id, None).unwrap();
|
||||
assert!(result.is_some());
|
||||
|
||||
// Resolve with wrong project filter
|
||||
let result = resolve_entity_ref(&conn, "issue", issue_id, Some(9999)).unwrap();
|
||||
assert!(result.is_none());
|
||||
|
||||
// Resolve unknown entity type
|
||||
let result = resolve_entity_ref(&conn, "unknown_type", issue_id, None).unwrap();
|
||||
assert!(result.is_none());
|
||||
|
||||
// Resolve nonexistent entity
|
||||
let result = resolve_entity_ref(&conn, "issue", 99999, None).unwrap();
|
||||
assert!(result.is_none());
|
||||
}
|
||||
Reference in New Issue
Block a user