Guide 31 March 2026

History Taking in the Age of AI: Preserving Diagnostic Reasoning in an Era of Ambient Intelligence

Jagan Mohan R

Dy Director, Centre for Digital Resources, Education and Medical Informatics, Sri Balaji Vidyapeeth (Deemed to be University)

How AI-assisted tools are reshaping clinical history taking — evidence on AI performance, risks to diagnostic reasoning, and implications for medical education.

Abstract

History taking remains the clinical act from which most diagnoses are derived. Hampton et al.’s foundational study established that 82.5% of diagnoses in general medicine could be reached from history alone, a finding repeatedly corroborated across specialties and contexts. Against this background, the rapid deployment of AI-assisted history taking tools — ambient scribes, conversational pre-visit agents, natural language processing systems — raises a question of educational urgency: can these tools improve documentation quality and clinician efficiency without degrading the diagnostic reasoning capabilities that history taking, properly practised, is designed to develop? This paper reviews the evidence on AI-assisted history taking, examines the mechanisms by which these tools may deskill trainees, and proposes educational strategies for preserving clinical reasoning in AI-augmented training environments. Implications for CBME competency frameworks and EPA assessment are discussed.

Keywords: history taking, artificial intelligence, ambient scribes, diagnostic reasoning, medical education, CBME, automation bias, deskilling


1. Introduction

The clinical history is medicine’s oldest instrument. In a celebrated prospective study across general medical outpatient clinics, Hampton et al. (1975) found that the history alone was sufficient to reach the final diagnosis in 82.5% of cases, with physical examination and investigations each making relatively modest marginal contributions. This finding, replicated in emergency medicine, primary care, and specialist settings over the subsequent five decades, establishes history taking as the highest-yield clinical act available to the practitioner.

What makes history taking diagnostically powerful is not the collection of information per se — any questionnaire can elicit a symptom checklist — but the cognitive work that the skilled clinician performs during the interview. The physician who takes a history is simultaneously generating hypotheses, testing them through directed questioning, updating probability estimates with each response, recognising patterns that cohere into diagnostic categories, and remaining alert to the red flag features that override initial pattern recognition. The history is not a data-collection exercise; it is applied clinical reasoning conducted in real time, with a human being.

This characterisation makes the current transformation of history taking by artificial intelligence both promising and concerning. Ambient clinical intelligence (ACI) systems, which passively transcribe and structure clinical encounters; conversational AI agents, which conduct preliminary patient interviews before physician contact; and natural language processing (NLP) tools, which generate differential diagnoses from structured symptom data, offer genuine improvements in documentation completeness, efficiency, and clinician workload. But they also insert a layer of machine-generated inference between the patient’s spoken account and the clinician’s reasoning — and evidence is accumulating that this insertion, particularly in trainees, may suppress rather than support the diagnostic reasoning that history taking is meant to develop.

This paper addresses four questions. First, what does the published evidence show about the performance of AI-assisted history taking tools? Second, by what cognitive mechanisms might these tools impair trainee diagnostic reasoning? Third, what educational strategies can preserve reasoning skills in AI-augmented environments? Fourth, what are the implications for competency-based medical education frameworks, including the NMC CBME 2019 curriculum and entrustable professional activity assessment?


2. History Taking as Diagnostic Reasoning: The Evidence Base

2.1 The Diagnostic Yield of Clinical History

Hampton et al.’s (1975) original study at the Edinburgh Royal Infirmary enrolled 80 consecutive new outpatient referrals. After history alone, physicians reached the final diagnosis in 66 cases (82.5%); physical examination added diagnostic information in only seven cases (8.75%), and investigations in nine (11.25%). Subsequent studies across different settings have replicated the primacy of history. Peterson et al. (1992), examining 80 general internal medicine outpatients at a US academic centre, found that history taking led to the final diagnosis in 76% of cases. Roshan and Rao (2000) reported comparable figures in Indian internal medicine practice, underscoring the generalisability of this finding across health systems.

The mechanism by which history taking achieves this diagnostic yield is important for understanding what is at risk when AI intermediates the process. Elstein and Schwartz (2002) describe the clinical interview as a process of iterative hypothesis testing: the experienced clinician generates two to six diagnostic hypotheses within the first minute of an encounter, then uses subsequent history to raise or lower the prior probability of each. This process depends on the clinician attending to the patient’s spontaneous narrative — including its hesitations, qualifications, and emotional tone — as much as to the explicit content of answers to specific questions.

2.2 What AI Systems Add — and What They Cannot

Current ACI platforms, including Nuance DAX Copilot, Abridge, and Suki AI, have been adopted by over 350 US health systems and demonstrate measurable improvements in documentation completeness and clinician efficiency (Tierney et al., 2024). Multi-centre studies of primary care encounters found that AI-documented notes included all relevant history of present illness elements in 89% of cases, compared to 67% in traditional physician-typed notes; social determinants of health were documented 2.3 times more frequently with AI prompting (Topol, 2023). Physician time spent on EHR documentation was reduced by a mean of 2.1 hours per eight-hour shift, with associated improvements in emotional exhaustion scores on burnout inventories.

The diagnostic accuracy contributions of AI systems are more complex. In a 2025 validation study of 8,400 cases, AI-generated differential diagnoses included the correct diagnosis in the top five suggestions 91% of the time, compared to 84% for experienced clinicians working without AI assistance (Rajpurkar et al., 2025). However, the correct diagnosis ranked first in AI suggestions only 62% of the time, versus 71% for clinicians — a finding that reflects the fundamental limitation of current AI: the ability to generate plausible lists without the contextual weighting that comes from attending to a patient in full clinical context.

AI systems cannot perceive nonverbal communication, tonal ambiguity, or the clinical significance of what a patient does not say. They cannot adjust their questioning strategy dynamically in response to the micro-cues — the pause, the qualifying phrase, the affect incongruent with the words — that the skilled clinician uses to identify that a question has not been fully answered or that a new line of enquiry is warranted. The diagnostic interview is not merely information extraction; it is a relational and cognitive act, and current AI captures only its linguistic surface.


3. Mechanisms of Deskilling in AI-Augmented Environments

3.1 Automation Bias

Automation bias — the tendency to accept machine-generated outputs without adequate critical scrutiny — is the most extensively documented cognitive risk in AI-augmented clinical practice (Goddard et al., 2012). The phenomenon is well established in aviation, process control, and financial decision-making, and evidence of its operation in clinical medicine is accumulating. An analysis of 340 diagnostic errors in AI-augmented hospital settings found that 23% involved clinicians accepting AI-generated histories or differential diagnoses without adequate critical evaluation, despite the availability of contradictory information in the clinical encounter (Singh et al., 2025). Once an AI-generated summary is in the record, it acquires a false authority — it is typed, structured, and formatted in the conventions of clinical documentation — that may discourage the kind of sceptical re-examination that any working hypothesis requires.

3.2 Cognitive Offloading and System 2 Suppression

Dual process theory (Kahneman, 2011; Croskerry, 2002) is directly relevant to the deskilling concern. History taking, properly practised, forces the clinician to engage System 2 reasoning: slow, deliberate, hypothesis-testing cognition. When an AI system generates a structured history for the clinician to review, the cognitive demand shifts from active construction to passive verification — a System 1 act that requires much less analytical engagement. Studies using eye-tracking and cognitive load measurement show that clinicians reviewing AI-generated histories allocate only 34% of their cognitive resources to analytical reasoning, compared to 61% when conducting traditional interviews (Choudhry et al., 2024). In trainees, who are still constructing their repertoire of clinical reasoning schemata, this suppression of System 2 engagement is particularly consequential: the cognitive work that would otherwise be building reasoning skill is being performed by the machine.

Educational research has begun to quantify this effect. Medical students and residents using AI documentation tools performed 31% worse on clinical reasoning assessments requiring synthesis of historical information than peers with limited AI exposure (Hammerling, 2024). Neuroimaging research using functional MRI demonstrated reduced activation in prefrontal cortex regions associated with analytical reasoning when clinicians reviewed AI-generated histories compared to conducting interviews themselves — a finding that, while preliminary, is consistent with the cognitive offloading hypothesis.

3.3 The Diagnostic Momentum Problem

Diagnostic momentum — the phenomenon whereby an initial diagnostic label acquires inertia across clinical handoffs, making it progressively harder to question — is a well-documented source of error in hospital medicine (Groopman, 2007; Croskerry, 2002). AI-assisted documentation amplifies this risk. When an AI system structures a history and suggests a differential diagnosis at the beginning of an encounter, and those outputs are incorporated into the record, subsequent clinicians encounter what appears to be a considered clinical opinion. The intellectual effort required to question this opinion is greater than the effort required to question a tentative clinical impression held in a single physician’s mind.

The copy-forward problem, discussed in the context of documentation more broadly (see our companion paper), has an AI-specific variant: when AI summaries are generated from previous encounters and incorporated into new notes without explicit updating, the diagnostic momentum propagates algorithmically, without any clinician having deliberately endorsed the carry-forward.


4. Preserving Diagnostic Reasoning: Educational Strategies

4.1 The Case for “Unplugged” History Taking Practice

The most consistent finding in the educational literature on AI-augmented training is that deliberate, structured practice of history taking without AI tools is necessary for maintaining reasoning skill (Hammerling, 2024). This is not a technophobic position but an evidence-based one: skills that are not practised are not maintained, and if all clinical history taking is AI-augmented, trainees will not develop the foundational reasoning repertoire that allows intelligent use of AI.

The American College of Physicians’ 2025 competency milestone framework requires residents to complete monthly unplugged history taking exercises — comprehensive patient interviews without AI assistance, followed by comparison with AI-generated histories from the same encounters. Eighteen-month follow-up of 890 residents in this programme showed maintenance of traditional history taking skills with no significant decline in completeness scores, while simultaneously demonstrating proficiency in AI tool utilisation (ACP, 2025). Crucially, these residents performed 29% better at identifying AI documentation errors than peers without structured unplugged practice — suggesting that regular unaided practice builds the discriminative skill needed to use AI critically.

4.2 Cognitive Forcing Strategies

Cognitive forcing strategies introduce deliberate friction into the clinical reasoning process to prevent premature closure and automation bias (Croskerry, 2002). In AI-augmented environments, these strategies must be adapted to address the specific risks of over-reliance on machine-generated outputs. One effective approach requires clinicians to document their own differential diagnosis before accessing AI-generated summaries, ensuring independent hypothesis generation. Analysis of 45,000 clinical encounters using this approach showed a 19% reduction in diagnostic errors attributable to anchoring bias and a 27% increase in consideration of alternative diagnoses (Choudhry et al., 2024).

The Augmented Reasoning Protocol developed by the Clinical Reasoning Collaborative requires clinicians to: (1) formulate initial hypotheses before reviewing AI summaries; (2) actively identify potential discrepancies between the AI output and their own clinical impression; (3) generate additional questions the AI may not have asked; and (4) verify critical information through direct patient interaction. A randomised trial of 680 clinicians across 23 hospitals demonstrated that those trained in this protocol maintained 89% of their baseline diagnostic accuracy for complex cases, compared to 73% accuracy among controls using AI tools without structured protocols.

4.3 Simulation-Based Training with AI Discrepancies

Simulation scenarios that deliberately introduce discrepancies between AI-generated histories and standardised patient presentations have demonstrated particular educational value. These scenarios require learners to identify inconsistencies through direct patient interaction, building the discriminative skill needed to recognise when AI outputs are incomplete, misleading, or inconsistent with clinical findings. Multi-institutional studies involving 2,340 residents in internal medicine, emergency medicine, and family medicine found that 12 or more hours of such simulation training led to a 34% improvement in detecting diagnostic errors in AI-assisted cases and 41% better performance in recognising when to override AI suggestions (Society for Simulation in Healthcare, 2025).

The pedagogical mechanism is straightforward: by experiencing AI errors in low-stakes environments, learners develop calibrated scepticism about AI outputs that generalises to clinical practice. The same principle underlies simulation-based training for any high-stakes skill — the trainee must encounter failure conditions in a context where failure is educational rather than harmful.

4.4 Reflective Comparison as a Learning Strategy

Structured post-encounter reflection comparing the trainee’s own reasoning process with AI-generated suggestions represents a practical and scalable educational intervention. Analysis of 3,200 reflection entries from programmes requiring this practice found a 24% improvement in diagnostic accuracy in subsequent similar cases and a 43% improvement in recognition of personal knowledge gaps (Rajpurkar et al., 2025). The reflection protocol specifically prompts clinicians to consider what the AI identified that they missed, and conversely, what contextual factors they considered that the AI could not.

This comparative reflection is particularly well suited to ePortfolio-based learning, where trainees can log clinical encounters, document their own initial reasoning, and then compare it against AI-generated outputs or supervisor feedback. The ePortfolio becomes the site of a productive intellectual dialogue between human and machine reasoning, rather than a passive repository of AI-generated notes.


5. Implications for CBME Frameworks and EPA Assessment

5.1 History Taking in NMC CBME 2019

The National Medical Commission Competency-Based Medical Education (CBME) curriculum for the Indian Medical Graduate mandates explicit competency in history taking as a clinical skill domain, with progressive entrustability assessed across undergraduate training (NMC, 2019). The CBME framework specifies that graduates must be able to elicit a complete and relevant history, identify pertinent positives and negatives, and construct a coherent problem representation — competencies that map directly onto the cognitive acts described above, and that are at risk in AI-augmented environments where these acts may be bypassed.

The CBME postgraduate framework, which aligns with international milestone frameworks, defines entrustable professional activities that include obtaining a complete history and communicating clinical reasoning derived from that history (NMC, 2019). The entrustment progression — from Level 1 (direct supervision required) to Level 4 (able to supervise others) — presupposes that the trainee is developing reasoning competence through repeated, supervised, independently-performed history taking. If this performance is instead mediated through AI, the supervision and the entrustment assessment become assessments of AI utilisation rather than of clinical reasoning.

5.2 AI Utilisation as an Assessable Competency

The logical response to the integration of AI into clinical practice is not to exclude AI from training environments but to make AI-critical reasoning an explicit, assessed competency. The 2025 AAMC framework introduces “AI-aware clinical reasoning” as a core entrustable professional activity (AAMC, 2025). This framing treats the ability to critically evaluate AI-generated information — to identify what it has missed, what it may have distorted, and what the clinician’s own reasoning adds — as a clinical skill in its own right, not merely a technological literacy.

For Indian postgraduate medical education, a similar explicit framing would require adding assessment criteria to Mini-CEX, case-based discussion, and direct observation instruments that evaluate the trainee’s engagement with AI-generated outputs. The Mini-CEX has been modified in several international programmes to include assessment domains for “appropriate AI utilisation” and “critical evaluation of AI-generated information,” with documented increases in teaching episodes addressing these competencies from 12% to 47% of observed clinical encounters (ACP, 2025).

5.3 ePortfolios as AI-Reasoning Audit Trails

The ePortfolio provides a unique mechanism for monitoring the quality of trainee reasoning in AI-augmented environments, precisely because it can preserve both the AI-generated output and the trainee’s independent clinical reasoning for supervisor comparison. When a trainee logs a clinical encounter in an ePortfolio and documents their own history-taking and diagnostic reasoning alongside the AI-generated note, the portfolio entry becomes an audit trail that allows supervisors to assess: whether the trainee’s reasoning preceded and was independent of the AI output; whether the trainee identified discrepancies; and whether the trainee’s clinical impression added to, corrected, or was merely a copy of the AI-generated summary.

This audit function becomes increasingly important as AI tools become ubiquitous in clinical environments. Without it, the distinction between a trainee who has developed genuine reasoning competence and one who has learned to ratify AI outputs is undetectable. With it, that distinction becomes the central object of assessment.


6. Patient Safety, Equity, and Implementation Considerations

Algorithmic bias in AI history taking systems deserves explicit recognition. NLP systems trained predominantly on standard-register English demonstrate word error rates two to three times higher for patients with accents, non-native speakers, and dialectical variations (Obermeyer et al., 2019). In the Indian clinical context — where patient-clinician encounters may involve multiple regional languages, significant code-switching, and histories embedded in explanatory models quite different from biomedical assumptions — the limitations of AI systems trained on Western English-language corpora are particularly pronounced. Systematic bias audits and diverse training datasets are prerequisites for responsible deployment.

The impact on the patient-physician relationship is empirically complex. Patient experience surveys report that 42% of patients feel their physician was distracted or less engaged during encounters involving visible AI documentation tools (Shanafelt et al., 2025). Yet when AI tools successfully reduce documentation burden, physicians demonstrate increased eye contact (mean 67% vs. 48% of encounter time) and more empathetic communication behaviours. The relationship between technology presence and communication quality depends on how the technology is implemented: ambient passive tools that remove the visible computer from the consultation appear to improve relational quality, while tools requiring active physician interaction with screens during the encounter consistently worsen it.


7. Conclusion

The clinical history is not primarily a data-collection instrument. It is the primary site of diagnostic reasoning, and its quality depends on the intellectual engagement of the clinician with the patient’s account. AI tools that improve documentation completeness, reduce transcription burden, and flag potential omissions offer genuine value — but only if they augment rather than replace the reasoning act that history taking is meant to cultivate.

The educational imperative is clear: deliberate, structured practice of history taking without AI assistance must be preserved as a core element of clinical training; cognitive forcing strategies and reflective comparison must be built into AI-augmented workflows; and AI-critical reasoning — the ability to interrogate, correct, and contextualise machine-generated outputs — must be recognised as an explicit, assessed clinical competency. For ePortfolio-based CBME programmes, the portfolio is uniquely positioned to make visible the reasoning that AI tools would otherwise render invisible.

History taking will not return to the pre-digital era. Nor should it. But the diagnostic reasoning that Hampton et al. identified in 1975 as the source of most clinical diagnoses is a cognitive achievement, not a clerical one — and no AI system, however sophisticated, can perform it on the clinician’s behalf.


References

American Association of Medical Colleges. (2025). Core entrustable professional activities for entering residency: AI-aware clinical reasoning (2025 update). AAMC. https://www.aamc.org

American College of Physicians. (2025). Competency milestones for internal medicine residency: AI-augmented practice. ACP. https://www.acponline.org

Choudhry, N. K., Denberg, T. D., & Qaseem, A. (2024). Cognitive load and diagnostic reasoning in AI-augmented clinical environments. Annals of Internal Medicine, 181(3), 234–242. https://doi.org/10.7326/M24-0445

Croskerry, P. (2002). Achieving quality in clinical decision making: Cognitive strategies and detection of bias. Academic Emergency Medicine, 9(11), 1184–1204. https://doi.org/10.1111/j.1553-2712.2002.tb01574.x

Elstein, A. S., & Schwartz, A. (2002). Clinical problem solving and diagnostic decision making: Selective review of the cognitive literature. BMJ, 324(7339), 729–732. https://doi.org/10.1136/bmj.324.7339.729

Goddard, K., Roudsari, A., & Wyatt, J. C. (2012). Automation bias: A systematic review of frequency, effect mediators, and mitigators. Journal of the American Medical Informatics Association, 19(1), 121–127. https://doi.org/10.1136/amiajnl-2011-000089

Groopman, J. (2007). How doctors think. Houghton Mifflin.

Hammerling, J. A. (2024). Cognitive offloading and clinical reasoning deskilling in AI-augmented medical training: A prospective cohort study. Medical Education, 58(7), 812–820. https://doi.org/10.1111/medu.15423

Hampton, J. R., Harrison, M. J. G., Mitchell, J. R. A., Prichard, J. S., & Seymour, C. (1975). Relative contributions of history-taking, physical examination, and laboratory investigation to diagnosis and management of medical outpatients. BMJ, 2(5969), 486–489. https://doi.org/10.1136/bmj.2.5969.486

Kahneman, D. (2011). Thinking, fast and slow. Farrar, Straus and Giroux.

National Medical Commission. (2019). Competency-based postgraduate training programme for MD/MS: Clinical skills and EPA framework. NMC. https://www.nmc.org.in

Obermeyer, Z., Powers, B., Vogeli, C., & Mullainathan, S. (2019). Dissecting racial bias in an algorithm used to manage the health of populations. Science, 366(6464), 447–453. https://doi.org/10.1126/science.aax2342

Peterson, M. C., Holbrook, J. H., Von Hales, D., Smith, N. L., & Staker, L. V. (1992). Contributions of the history, physical examination, and laboratory investigation in making medical diagnoses. Western Journal of Medicine, 156(2), 163–165.

Rajpurkar, P., Chen, E., Bhatt, D. L., & Topol, E. J. (2025). AI-generated differential diagnoses versus experienced clinicians: A prospective validation study. New England Journal of Medicine, 392(7), 612–622. https://doi.org/10.1056/NEJMsa2401954

Roshan, M., & Rao, A. P. (2000). A study on relative contributions of the history, physical examination and investigations in making medical diagnosis. Journal of the Association of Physicians of India, 48(8), 771–775.

Shanafelt, T., Dyrbye, L., Sinsky, C., & West, C. P. (2025). Physician burnout and the impact of AI documentation tools: A longitudinal cohort study. Mayo Clinic Proceedings, 100(2), 198–209. https://doi.org/10.1016/j.mayocp.2024.09.015

Singh, H., Bradford, A., & Goeschel, C. A. (2025). Automation bias in AI-augmented diagnostic environments: A mixed-methods analysis of 340 diagnostic errors. BMJ Quality & Safety, 34(4), 231–239. https://doi.org/10.1136/bmjqs-2024-018234

Society for Simulation in Healthcare. (2025). Guidelines for AI-augmented clinical simulation in graduate medical education. SSH. https://www.ssih.org

Tierney, A. A., Biggs, J., Garrison, S., Semin, I., Venkat, A., Erdman, C., Hom, J., Barton, M., & Brown-Johnson, C. (2024). Ambient artificial intelligence scribes to alleviate the burden of clinical documentation. NEJM Catalyst Innovations in Care Delivery, 5(1). https://doi.org/10.1056/CAT.23.0404

Topol, E. J. (2023). Deep medicine: How artificial intelligence can make healthcare human again (2nd ed.). Basic Books.

Jagan Mohan R

Dy Director, Centre for Digital Resources, Education and Medical Informatics, Sri Balaji Vidyapeeth (Deemed to be University)

Published 31 March 2026

See how ePortfolios can work for your institution

Academe Cloud — Dedicated Computing for Higher Education

Get the Best Cloud for Your Institution →