Granular Storage For Future Workflow Analysis

We wish to capture enough structured history to support traceability, QA, debugging, and future analysis of how radiologists use AI suggestions.

Audit Points

Moment Stored event Future analysis value
Study received Who/what sent it, when, metadata digest, validation status. Integration reliability, missing metadata patterns.
AI run completed Model version, inputs, outputs, confidence distribution, latency. Model monitoring, drift analysis, per-hospital behavior.
Radiologist views study Viewer opened, overlays toggled, AI finding selected. Adoption, trust, attention, workflow friction.
Finding feedback Accepted, rejected, edited, ignored, reclassified, or superseded. Human-AI agreement/disagreement analysis and evaluation.
Legacy report completed Completion timestamp, report reference if available, linked sidecar findings/annotations. Outcome correlation and sidecar workflow evaluation.
Status / handoff recorded Destination or source system, success/failure, retry count, payload digest if a handoff exists. Operational support and hospital onboarding checks without assuming we own final submission.

Analytics Questions We Want To Answer Later

  • Which AI findings are most often accepted without edits?
  • Which finding types are frequently resized, relabeled, or rejected?
  • Do radiologists review AI suggestions before or after reading the raw image stack?
  • How often does the final diagnosis include findings not suggested by AI?
  • Where do turnaround delays occur: ingestion, inference, assignment, sidecar review, or legacy report completion?
  • Do different hospitals have different disagreement patterns after normalizing for modality and model version?
Compliance Boundary Production implementation would require a dedicated security, privacy, and regulatory review.