The Commonplace
Home Dashboard Papers Evidence Digests 🎲
← Papers

AI can raise radiology accuracy and throughput, but benefits are conditional on real-world integration—poorly designed deployments risk automation bias, deskilling and workflow disruption that can erode clinical and economic gains.

Human-AI interaction and collaboration in radiology: from conceptual frameworks to responsible implementation.
B. Koçak, Renato Cuocolo · Fetched March 18, 2026 · Diagnostic and Interventional Radiology
semantic_scholar review_meta low evidence 7/10 relevance DOI Source
AI holds clear potential to improve radiologic accuracy and efficiency, but realized clinical and economic value depends critically on how tools are integrated with radiologists, monitored over time, and governed—evidence on patient outcomes, costs, and long-run workforce effects remains sparse.

Artificial intelligence (AI) is entering routine radiology practice, but most studies evaluate algorithms in isolation rather than their interaction with radiologists in clinical workflows. This narrative review summarizes current knowledge on human-AI interaction in radiology and highlights practical risks and opportunities for clinical teams. First, simple conceptual models of human-AI collaboration are described, such as diagnostic complementarity, which explain when radiologists and AI can achieve synergistic performance exceeding that of either alone. Then, AI tool integration strategies along the imaging pathway are reviewed, from acquisition and triage to interpretation, reporting, and teaching, outlining common interaction models and physician-in-the-loop workflows. Cognitive and professional effects of AI integration are also discussed, including automation bias, algorithmic aversion, deskilling, workload management, and burnout, with specific vulnerabilities for trainees. Furthermore, key elements of responsible implementation are summarized, such as liability and oversight implications, continuous monitoring for performance drift, usable explanations, basic AI literacy, and co-design with radiology teams. Finally, emerging systems are introduced, including vision-language models and adaptive learning loops. This review aims to provide a clear and accessible overview to help the radiology community recognize where human-AI collaboration can add value, where it can cause harm, and which questions future studies must address.

Summary

Main Finding

AI in radiology has clear potential to improve diagnostic performance and workflow efficiency, but real clinical value depends critically on how tools interact with radiologists in practice. Human-AI collaboration can produce synergistic gains (diagnostic complementarity) or generate harms (automation bias, deskilling, workflow disruption). Responsible value capture requires attention to integration design, monitoring, liability, training, and ongoing evaluation — not just algorithmic accuracy measured in isolation.

Key Points

  • Conceptual models of collaboration
    • Diagnostic complementarity: humans and AI can exceed either alone when errors are uncorrelated and tasks are allocated to leverage comparative strengths.
    • Other interaction models include AI as a second reader, triage/prioritization, pre-population of reports, and decision-support with physician-in-the-loop.
  • Integration points across the imaging pathway
    • Acquisition: image quality optimization, protocol selection.
    • Triage: prioritizing critical studies for faster human review.
    • Interpretation/reporting: detection/quantification aids, structured report drafting.
    • Post-interpretation: teaching, quality assurance, and continuous model improvement loops.
  • Cognitive and professional effects
    • Automation bias and algorithmic aversion can respectively increase undue reliance or underuse of tools.
    • Risk of deskilling, especially for trainees who receive less diagnostic practice.
    • Changes in workload composition can reduce routine burden but may shift cognitive load (e.g., follow-up decisions, managing AI outputs).
    • Burnout implications are ambiguous — tools can reduce tedious tasks but also introduce new cognitive, administrative, and liability stresses.
  • Responsible implementation imperatives
    • Legal/liability clarity and oversight pathways.
    • Continuous monitoring for performance drift and distributional shifts.
    • Usable explanations and baseline AI literacy for clinicians.
    • Co-design with frontline radiology teams to fit workflows.
  • Emerging technologies and research needs
    • Vision-language models and adaptive learning loops may expand functionality but raise governance and safety challenges.
    • Gap: few randomized or real-world studies measuring patient outcomes, economic effects, or long-run workforce impacts.

Data & Methods

  • Article type: narrative review synthesizing existing literature on human-AI interaction in radiology.
  • Evidence base: mixture of laboratory evaluation studies, simulation and reader studies, observational deployments, usability/qualitative studies, and a small number of real-world implementation reports.
  • Methods described in the review: conceptual modeling, descriptive summaries of integration strategies, synthesis of cognitive/organizational literature.
  • Limitations of the evidence base highlighted in the review:
    • Heterogeneous study designs and outcomes; many studies evaluate standalone algorithm accuracy rather than clinician-AI joint performance in routine workflows.
    • Limited randomized controlled trials or longitudinal evaluations; few studies measure patient-relevant outcomes or economic impacts.
    • Publication and selection bias toward successful demonstrations; real-world deployment challenges and negative results are underreported.

Implications for AI Economics

  • Value creation and capture
    • Efficiency gains: triage and automation can shorten time-to-diagnosis, increase throughput, and reduce time spent on repetitive tasks, potentially raising productivity per radiologist.
    • Quality improvements: better detection/quantification may reduce downstream costs from missed diagnoses or unnecessary follow-ups, improving cost-effectiveness.
    • Complementarity vs substitution: economic effect depends on whether AI augments radiologists (raising output per worker) or substitutes tasks (reducing demand for certain diagnostic activities). Complementarity can increase total specialist output and value; substitution can compress workforce demand or shift labor to other tasks.
  • Costs and investments
    • Up-front costs: procurement, integration with PACS/EMR, UI/UX development, regulatory compliance, and staff training.
    • Recurring costs: model monitoring, data labeling for drift remediation, software updates, and cybersecurity.
    • Hidden costs: increased liability exposure, workflow redesign burden, and potential productivity loss during transition.
  • Workforce and human capital
    • Short-run productivity may rise, but deskilling risks imply long-run human capital depreciation, especially for trainees; that may raise future labor costs for re-training or reduce the quality premium of experienced radiologists.
    • Task reallocation: some tasks may shift to radiographers, physician extenders, or centralized AI-monitoring teams, affecting wage structures and labor demand across roles.
  • Reimbursement and incentives
    • Current fee-for-service structures may not reward efficiency gains; value-based payment or shared-savings models better align incentives for adoption that improves outcomes and reduces total cost.
    • Clear reimbursement for AI-assisted services (e.g., billing codes) influences adoption and ROI calculations.
  • Liability, regulation, and market adoption
    • Unclear liability frameworks increase perceived and real costs, slowing adoption; insurers and hospitals factor these into procurement decisions.
    • Explainability, trust, and demonstrated real-world effectiveness are key demand-side frictions; small-scale lab gains rarely translate to broad clinical uptake without workflow fit.
  • Research and policy priorities to inform economic decisions
    • Rigorous real-world trials assessing patient outcomes, cost-effectiveness, and labor impacts (including randomized implementation trials where feasible).
    • Comparative studies of integration strategies (e.g., AI-as-triage vs AI-as-second-reader) on throughput, accuracy, and costs.
    • Measurement of long-term effects on workforce skill composition and training pipelines.
    • Development of standard metrics and monitoring frameworks for performance drift and economic returns.
    • Policy levers: clarify liability, create transitional reimbursement for AI-augmented services, subsidize integration/training, and mandate post-deployment monitoring for high-impact systems.

Takeaway: economic benefits from AI in radiology are plausible but contingent on human-AI interaction design, governance, workforce effects, and payment structures. Evaluations that combine clinical outcomes with economic metrics and real-world workflow studies are crucial to determine net value.

Assessment

Paper Typereview_meta Evidence Strengthlow — The review synthesizes largely laboratory evaluations, reader studies, simulations, qualitative/usability reports and a small number of observational deployments; there are few randomized or longitudinal real-world studies measuring patient outcomes or economic impacts, and many studies evaluate standalone algorithm accuracy rather than clinician–AI joint performance, limiting causal inference and external validity. Methods Rigormedium — The article provides a coherent conceptual framework and a broad synthesis of interdisciplinary literature (clinical reader studies, usability, organizational analyses), but it is a narrative review rather than a systematic review or meta-analysis, lacks pre-registered inclusion criteria, quantitative pooling, and is subject to selection and publication biases. SampleEvidence base comprises heterogeneous study types: laboratory algorithm accuracy evaluations across imaging modalities, controlled reader studies and simulations (small-to-moderate sample sizes), observational deployment reports from a few health systems, and qualitative/usability studies with clinicians; data are typically short-term, focus on diagnostic performance or workflow metrics, and rarely include patient-level outcomes, long-run workforce measures, or comprehensive cost data. Themeshuman_ai_collab productivity labor_markets skills_training governance adoption GeneralizabilityPredominance of lab/reader studies limits generalizability to routine clinical workflows, Few randomized or longitudinal real-world evaluations — limited evidence on patient outcomes or long-term economic effects, Evidence concentrated in high-resource health systems and specialty radiology settings — may not generalize to low-resource or non-hospital contexts, Heterogeneity in AI models, tasks, integration designs and regulatory environments reduces transferability across deployments, Small sample sizes and publication/selection bias toward positive demonstrations

Claims (23)

ClaimDirectionConfidenceOutcomeDetails
AI in radiology has clear potential to improve diagnostic performance and workflow efficiency. Decision Quality positive medium diagnostic accuracy (sensitivity/specificity), workflow efficiency (throughput, time-to-diagnosis, time-on-task)
0.07
Real clinical value depends critically on how AI tools interact with radiologists in practice (integration design and human-AI interaction). Decision Quality mixed medium clinician-AI joint diagnostic performance, patient-relevant outcomes, workflow metrics
0.07
Human-AI collaboration can produce synergistic gains (diagnostic complementarity) when errors are uncorrelated and tasks are allocated to leverage comparative strengths. Decision Quality positive medium combined diagnostic accuracy (aggregate sensitivity/specificity), reduction in missed diagnoses
0.07
Human-AI collaboration can also generate harms, including automation bias, deskilling, and workflow disruption. Error Rate negative medium rates of over-reliance on AI, diagnostic error rates attributable to automation bias, measures of clinician skill over time, workflow error/throughput metrics
0.07
Many published studies focus on standalone algorithm accuracy rather than clinician–AI joint performance in routine workflows. Research Productivity negative high proportion of studies reporting standalone algorithm metrics versus those reporting clinician+AI workflow outcomes
0.12
There are limited randomized controlled trials or longitudinal evaluations; few studies measure patient-relevant outcomes or economic impacts. Research Productivity negative high number of RCTs/longitudinal studies, frequency of patient outcome and economic outcome reporting
0.12
Automation bias can increase undue reliance on AI, while algorithmic aversion can drive underuse of helpful tools. Adoption Rate mixed medium rates of clinician acceptance/use of AI recommendations, error rates when following vs. overriding AI
0.07
There is a risk of deskilling, especially for trainees receiving reduced diagnostic practice when AI automates routine tasks. Skill Obsolescence negative low trainee diagnostic performance over time, case exposure counts, measures of retained clinical skill
0.04
Integration points for AI across the imaging pathway include acquisition (image quality/protocol selection), triage (prioritization), interpretation/reporting (detection, quantification, report pre-population), and post-interpretation (teaching, QA, model improvement loops). Organizational Efficiency positive medium site-level implementation metrics by workflow stage (e.g., reduced repeat scans, prioritized read times, report completion time)
0.07
Triage and automation can shorten time-to-diagnosis, increase throughput, and reduce time spent on repetitive tasks. Task Completion Time positive medium time-to-diagnosis, studies-per-hour per radiologist, time spent on repetitive tasks
0.07
Tools that improve detection or quantification may reduce downstream costs from missed diagnoses or unnecessary follow-ups, improving cost-effectiveness in some scenarios. Consumer Welfare positive low downstream healthcare utilization (additional tests, treatments), cost per diagnosis, cost-effectiveness ratios
0.04
Economic outcomes depend on complementarity versus substitution: AI that augments radiologists can raise output per worker; AI that substitutes tasks may reduce demand for certain diagnostic activities. Firm Productivity mixed medium radiologist productivity metrics, employment levels/demand for diagnostic activities
0.07
Up-front implementation costs commonly include procurement, integration with PACS/EMR, UI/UX development, regulatory compliance, and staff training; recurring costs include monitoring, data labeling, software updates, and cybersecurity. Firm Revenue negative medium_high implementation capital expenditures, annual operating expenditures
0.01
Hidden costs can arise from increased liability exposure, workflow redesign burden, and potential productivity loss during transition periods. Organizational Efficiency negative medium measures of productivity during rollout, documented workflow redesign time/costs, liability incidents/concerns
0.07
Changes in workload composition can reduce routine burdens but may shift cognitive load to follow-up decisions and managing AI outputs. Worker Satisfaction mixed medium time allocation across task types, subjective cognitive workload scores, frequency of follow-up decision tasks
0.07
The net effect of AI on clinician burnout is ambiguous: tools can remove tedious tasks but may introduce new cognitive, administrative, and liability stresses. Worker Satisfaction mixed medium burnout survey scores, task satisfaction, administrative burden metrics
0.07
Responsible implementation requires legal/liability clarity, continuous monitoring for performance drift and distributional shifts, usable explanations, baseline AI literacy for clinicians, and co-design with frontline radiology teams. Governance And Regulation positive medium successful deployment metrics, monitoring alerts for drift, clinician comprehension/usability scores
0.07
Emerging technologies such as vision-language models and adaptive learning loops may expand functionality but raise governance and safety challenges. Ai Safety And Ethics mixed low model capability metrics (multimodal performance), incidence of safety/governance incidents
0.04
Explainability, trust, and demonstrated real-world effectiveness are key demand-side frictions; small-scale laboratory gains rarely translate into broad clinical uptake without workflow fit. Adoption Rate negative medium adoption rates, clinician trust/acceptance measures, implementation success rates
0.07
Unclear liability frameworks increase perceived and real costs and can slow adoption by hospitals and insurers. Adoption Rate negative medium_high time-to-adoption, procurement decisions citing liability concerns, insurance/coverage decisions
0.01
Fee-for-service payment structures may not reward efficiency gains from AI; value-based payment or shared-savings models are better aligned to incentivize adoption that reduces total cost and improves outcomes. Governance And Regulation positive medium_high reimbursement levels, adoption under different payment models, cost savings realized
0.01
Research priorities include rigorous real-world trials assessing patient outcomes, cost-effectiveness, and labor impacts; comparative studies of integration strategies; measurement of long-run workforce effects; and development of standard metrics and monitoring frameworks. Research Productivity positive high number and quality of real-world trials, existence of standardized monitoring frameworks, availability of long-term workforce impact studies
0.12
Overall, economic benefits from AI in radiology are plausible but conditional on human-AI interaction design, governance, workforce effects, and payment structures; net value is not determined by algorithmic accuracy alone. Firm Productivity mixed medium net economic value/ROI, clinical outcomes, adoption and sustainability metrics
0.07

Notes