Research synthesis with citation verification¶
Assemble evidence from distributed sources into a grounded synthesis where every material claim is backed by inspectable citations.
Metadata¶
- Pattern id:
research-synthesis-with-citation-verification - Pattern family: Gather / Retrieve / Synthesize
- Problem structure: Context gathering and synthesis (
context-gathering-and-synthesis) - Domains: Research (
research), Compliance (compliance), Engineering (engineering)
Workflow goal¶
Produce a scoped synthesis that answers a question, preserves provenance, and makes citation quality inspectable before downstream use.
Inputs¶
Research question¶
- Description: A scoped question, decision topic, or investigation prompt that defines what evidence should be gathered.
- Kind: request
- Required: Yes
- Examples:
- What changed in the policy landscape for this control?
- What evidence supports this architecture tradeoff?
Source corpus¶
- Description: Searchable documents, records, or knowledge sources that may contain relevant evidence.
- Kind: document-collection
- Required: Yes
- Examples:
- Internal policy library
- RFCs, tickets, and postmortems
- External publications with stable identifiers
Citation policy¶
- Description: Rules defining acceptable sources, recency, attribution requirements, and how uncertainty must be surfaced.
- Kind: policy
- Required: Yes
- Examples:
- Only cite approved internal repositories and primary-source regulations
- Flag unsupported claims instead of inferring citations
Outputs¶
Verified synthesis brief¶
- Description: A concise answer or briefing whose material claims carry explicit citations and uncertainty notes.
- Kind: brief
- Required: Yes
- Examples:
- Compliance obligation summary with source annotations
- Engineering decision memo with linked evidence
Evidence trace¶
- Description: Claim-to-source mapping that lets reviewers inspect provenance and resolve challenges quickly.
- Kind: trace
- Required: Yes
- Examples:
- Paragraph-to-source reference table
- Claim ledger keyed to document excerpts
Open questions¶
- Description: Unresolved gaps, conflicts, or missing evidence that prevented a fully closed synthesis.
- Kind: issue-list
- Required: Yes
- Examples:
- Sources disagree on the effective date
- No approved evidence found for one supporting claim
Environment¶
Operates across document repositories and search systems where evidence coverage, provenance, and reviewer trust matter more than raw summarization speed.
Systems¶
- Document repositories
- Search and retrieval indexes
- Citation or annotation stores
- Review workflow systems
Actors¶
- Requesting analyst or stakeholder
- Reviewing subject-matter owner
- Source custodians
Constraints¶
- Use only sources within the permitted trust boundary for the task.
- Preserve source identifiers and traceability for every consequential claim.
- Surface uncertainty and contradictory evidence instead of flattening it away.
- Do not invent citations when retrieval fails or evidence is weak.
Assumptions¶
- Relevant sources expose stable identifiers or durable references.
- Retrieval access is available for approved corpora.
- A human reviewer can inspect the synthesis before consequential use.
Capability requirements¶
- Retrieval (
retrieval): The pattern depends on finding relevant evidence across scattered repositories before synthesis can begin. - Synthesis (
synthesis): The workflow must compress overlapping findings into a coherent brief without losing important nuance. - Verification (
verification): Claims and citations must be checked against trusted evidence so the output remains grounded. - Memory and state tracking (
memory-and-state-tracking): The system needs durable claim-to-source state so review, revision, and audit do not lose provenance.
Execution architecture¶
- Tool-using single agent (
tool-using-single-agent): A single agent can often manage retrieval, note-taking, and synthesis when the question scope stays bounded. - Human in the loop (
human-in-the-loop): Human review is a normal part of the loop because citation sufficiency and evidence interpretation often require judgment.
Autonomy profile¶
- Level: Human directed (
human-directed) - Reversibility: The generated brief is advisory and can be revised, but downstream consumers should not rely on disputed claims until review is complete.
- Escalation: Escalate whenever source trust is ambiguous, contradictory evidence materially changes the answer, or citation completeness falls below policy thresholds.
Human checkpoints¶
- Confirm the research question and source boundary before broad retrieval starts.
- Review the final synthesis and evidence trace before external sharing or policy use.
- Resolve disputes when sources conflict or citation coverage is incomplete.
Risk and governance¶
- Risk level: Moderate (
moderate) - Failure impact: Unsupported or weakly sourced synthesis can mislead engineering or compliance work, create rework, and weaken audit readiness.
- Auditability: Retain the retrieved source list, claim-to-source mappings, reviewer notes, and unresolved evidence gaps with the final brief.
Approval requirements¶
- Human review is required before the synthesis is used in compliance submissions, external communications, or material decision memos.
- New source classes outside the approved trust boundary require explicit approval.
Privacy¶
- Redact or avoid unnecessary personal or sensitive data in retrieved excerpts.
- Keep access to restricted corpora aligned with source-system permissions.
Security¶
- Preserve source access controls when retrieving or caching evidence.
- Log retrieval and citation-generation actions for later inspection.
Notes: Governance centers on provenance discipline, confidence transparency, and explicit reviewer accountability.
Why agentic¶
- Retrieval paths must adapt as evidence quality and coverage change during the workflow.
- The system has to decide which sources are relevant, duplicative, contradictory, or too weak to cite.
- Maintaining a live evidence trace is stateful work that static summarization pipelines usually handle poorly.
Failure modes¶
Fabricated or unverifiable citation¶
- Impact: Reviewers cannot trust the synthesis, and downstream users may rely on unsupported claims.
- Severity: high
- Detectability: medium
- Mitigations:
- Require source identifiers before a citation can be emitted.
- Block finalization when citation validation fails.
- Preserve reviewer-visible evidence traces for every material claim.
Contradictory evidence is omitted from the summary¶
- Impact: The brief overstates certainty and can bias policy or engineering decisions.
- Severity: medium
- Detectability: medium
- Mitigations:
- Compare retrieved sources for disagreement before synthesis is finalized.
- Include an explicit open-questions or conflicts section.
Stale evidence dominates the synthesis¶
- Impact: The output reflects outdated guidance or obsolete system state.
- Severity: medium
- Detectability: high
- Mitigations:
- Enforce recency checks from citation policy.
- Surface source timestamps in the evidence trace.
Question scope drifts beyond approved source coverage¶
- Impact: The workflow appears complete while leaving key claims outside the validated trust boundary.
- Severity: medium
- Detectability: medium
- Mitigations:
- Confirm scope with a human before broadening retrieval.
- Record unsupported subquestions as open gaps instead of inferring answers.
Evaluation¶
Success metrics¶
- Percentage of material claims with valid traceable citations.
- Reviewer acceptance rate without major provenance corrections.
- Rate of unresolved evidence gaps surfaced before handoff.
Quality criteria¶
- The synthesis clearly separates verified facts, interpretation, and open questions.
- Every consequential claim is inspectable through the evidence trace.
- Source trust boundaries and citation policy exceptions are visible to reviewers.
Robustness checks¶
- Test with partially conflicting sources and ensure disagreement is surfaced.
- Test with revoked or inaccessible sources and verify the workflow blocks unsupported citations.
- Test with sparse evidence to ensure the output degrades into open questions instead of confident hallucination.
Benchmark notes: Evaluate both answer usefulness and provenance integrity; fluent prose without traceable evidence is a failure for this pattern.
Implementation notes¶
Orchestration notes¶
- Separate retrieval, claim extraction, and citation verification steps so unsupported claims can be gated.
- Keep intermediate notes linked to source identifiers rather than freeform memory only.
Integration notes¶
- Common integrations include search indexes, document repositories, and review systems that preserve annotations.
- Avoid implementation assumptions that depend on a specific vendor knowledge base.
Deployment notes¶
- Apply least-privilege access to corpora and reviewer queues.
- Prefer immutable audit records for final claim-to-source mappings.
References¶
Example domains¶
- Research (
research): Produce a literature-backed briefing that ties each key finding to inspectable citations. - Compliance (
compliance): Summarize a control obligation with direct references to policy text and supporting evidence. - Engineering (
engineering): Draft an architecture rationale linked to RFCs, tickets, and operational history.
Related patterns¶
- Incident root cause analysis (provides-context-for)
- Evidence-grounded synthesis often precedes deeper discrepancy investigation when the initial brief surfaces unresolved technical conflicts.
Grounded instances¶
- Vendor data-transfer safeguard obligation synthesis for cross-border review
- Managed Kubernetes node support and hardening obligation synthesis for platform governance review
- Service mesh migration readiness evidence synthesis for architecture review
- Debt covenant obligation synthesis for quarter-close review
- Contingent-worker co-employment and tenure-limit obligation synthesis for program review
- Pay transparency posting obligation synthesis for requisition launch review
- Conveyor safety bulletin synthesis for network readiness review
- Storm-season backup power blackstart evidence synthesis for network assurance review
- Model-serving platform benchmark briefing copilot loop
- Regulatory obligation synthesis for data retention review
- Enterprise support obligation synthesis for severity-one review
- Premium support sovereign logging and evidence-handling obligations synthesis for executive service review
Canonical source¶
data/patterns/gather-retrieve-synthesize/research-synthesis-with-citation-verification.yaml