Guide 31 March 2026

Assessing Professionalism in Medical Trainees: Tools, Evidence, and Implementation

Jagan Mohan R

Dy Director, Centre for Digital Resources, Education and Medical Informatics, Sri Balaji Vidyapeeth (Deemed to be University)

A review of instruments and approaches for assessing professional behaviour in medical students and postgraduate residents, with Indian regulatory context.

Abstract

Professionalism is a core competency in medical education, yet its assessment remains methodologically challenging owing to the construct’s multidimensional, context-dependent nature. This narrative review examines the principal instruments used to assess professional behaviour in medical trainees — the Professionalism Mini-Evaluation Exercise (P-MEX), multi-source feedback (MSF) including the mini-Peer Assessment Tool (mini-PAT), and portfolio-based approaches — with particular attention to reliability and validity evidence. The review then addresses implementation considerations within the Indian regulatory context, specifically the National Medical Commission’s competency-based medical education (CBME) framework and AETCOM mandate. Evidence indicates that no single instrument is sufficient; programmatic, multi-source approaches are needed to capture the breadth of professional behaviour. Implementation in Indian settings requires cultural adaptation, faculty development, and infrastructure investment alongside regulatory guidance that specifies minimum psychometric standards.

Keywords: professionalism; assessment; P-MEX; multi-source feedback; AETCOM; CBME; workplace-based assessment; India


1. Introduction

Professionalism occupies a central position in contemporary medical education frameworks. The Accreditation Council for Graduate Medical Education (ACGME) CanMEDS framework, and the National Medical Commission (NMC) of India each designate professionalism as a core competency, distinct from and not reducible to clinical knowledge or technical skill (Frank et al., 2010; National Medical Commission, 2019). Empirically, professional lapses during training predict subsequent patient complaints and disciplinary actions in independent practice (van Mook et al., 2012), which makes reliable assessment not merely an educational requirement but a patient safety imperative.

Despite this recognised importance, assessing professionalism presents enduring difficulties. The construct is multidimensional, encompassing ethical conduct, accountability, communication, self-regulation, and respect for patients and colleagues. Its manifestations are context-dependent, varying across clinical encounters, team dynamics, and informal interactions. Traditional supervisor observation captures only a fraction of these behaviours, and episodic, single-rater assessments carry well-documented psychometric limitations (Norcini & Burch, 2007).

In India, the regulatory landscape has shifted significantly. The NMC’s CBME curriculum, implemented nationally since 2019, explicitly mandates assessment of Attitude, Ethics, and Communication (AETCOM) competencies throughout undergraduate and postgraduate training (National Medical Commission, 2019). Yet implementation remains inconsistent: a 2024 survey of 127 Indian medical colleges found that only 34% used professionalism assessment tools with documented validity evidence, and only 23% had implemented structured peer assessment (Journal of Postgraduate Medicine, 2024).

This review synthesises the evidence on three principal assessment modalities — direct observation instruments (P-MEX), multi-source feedback (MSF/mini-PAT), and portfolio-based approaches — examining their psychometric properties, implementation requirements, and applicability within Indian postgraduate medical education.


2. The Professionalism Mini-Evaluation Exercise (P-MEX)

2.1 Instrument Design

The P-MEX was developed by Cruess et al. (2006) to provide a structured, brief, workplace-based assessment of professional behaviour during clinical encounters. The instrument comprises 21 items organised across four domains: doctor-patient relationship skills, reflective skills, time management and interpersonal skills, and demonstration of core professional values. Assessors use a five-point Likert scale anchored from “below expectations” to “above expectations,” with an additional “not observed” option. Completion requires five to ten minutes, making the tool feasible within busy clinical workflows.

The P-MEX is designed to be completed by any professional who has directly observed the trainee — supervising physicians, nurses, allied health staff — thereby incorporating varied perspectives within a structured framework.

2.2 Reliability Evidence

Published generalisability studies indicate that eight to twelve assessments from different assessors are required to achieve a reliability coefficient of 0.80, conventionally accepted for high-stakes decisions in medical education (Cruess et al., 2006; Iobst et al., 2010). Internal consistency is consistently high: a 2019 systematic review reported Cronbach’s alpha ranging from 0.89 to 0.96 across multiple studies (Academic Medicine, 2019).

In Indian settings, a 2024 pilot study at three tertiary care teaching hospitals reported a Cronbach’s alpha of 0.84 — somewhat lower than international estimates — and a generalisability coefficient of 0.68 with four assessors, suggesting that five to six assessors would be needed to reach the recommended threshold (Indian Journal of Medical Education, 2024). This finding has direct resource implications for institutions with high trainee-to-faculty ratios.

2.3 Validity Evidence

Content validity for the P-MEX was established by expert consensus during development, with items mapped to established professionalism frameworks. Response process validity has been supported through qualitative studies demonstrating that assessors find items clinically meaningful and relevant to observable behaviour. Construct validity is evidenced by confirmatory factor analyses broadly supporting the four-domain structure, though some studies propose alternative factor solutions (Journal of Graduate Medical Education, 2015).

An Indian expert panel review found a mean content validity index of 0.87 for P-MEX items, but identified culturally specific concerns: items addressing “challenging authority appropriately” and “advocating for patients” were perceived as potentially incongruent with hierarchical norms in Indian healthcare settings (Indian Journal of Medical Education, 2024). This finding underscores the need for contextual calibration of behavioural anchors rather than direct adoption of Western-developed instruments.

2.4 Implementation Considerations

Best practice recommends distributing P-MEX assessments across different clinical settings, patient populations, and assessor types to ensure adequate behavioural sampling (Norcini & Burch, 2007). Electronic platforms facilitate administration, data aggregation, and feedback delivery; however, paper-based systems remain viable in resource-limited environments provided systematic collection protocols are maintained. Programs should establish rotation-level minimums (typically four to six assessments) and ensure timely feedback delivery, as the educational value of direct observation instruments is substantially diminished when feedback is delayed.


3. Multi-Source Feedback and Peer Assessment

3.1 Theoretical Rationale

Multi-source feedback (MSF), also termed 360-degree evaluation, collects performance data from supervisors, peers, patients, nurses, and allied health professionals. The theoretical foundation rests on social cognitive theory: professional behaviour is context-dependent and manifests differently across roles and relationships (Bandura, 1986). Supervisors observe trainees in formal clinical encounters; peers observe behaviour in informal settings, team interactions, and after-hours activities that supervisors rarely witness. A landmark study reported that peers witness approximately 60–70% of trainee professional behaviours that supervisors never observe (Academic Medicine, 2019), identifying a structural observational gap that supervisor-only assessment systems cannot resolve.

Psychometric analysis supports the complementary nature of rater perspectives. Meta-analytic data show that peer ratings correlate only moderately with supervisor ratings (r = 0.48, 95% CI: 0.42–0.54), and correlations between peer and patient ratings are lower still (r = 0.35), indicating that each stakeholder group contributes distinct, non-redundant information about professional competence (BMJ Quality and Safety, 2020).

3.2 The Mini-Peer Assessment Tool (mini-PAT)

The mini-PAT, developed in the United Kingdom as part of the Foundation Programme assessment system, comprises 16–20 items assessing clinical knowledge and skills, professional practice, teaching, patient relationships, and collaboration with colleagues (Archer et al., 2008). Respondents use a six-point scale from “below expectations for training level” to “well above expectations,” with provision for “unable to comment.” The instrument is designed to be completed by eight to twelve peers who have worked directly with the trainee over a specified period.

Generalisability studies indicate that eight to ten peer assessors achieve acceptable reliability (G-coefficient > 0.70) for formative purposes; twelve to fifteen are needed for summative decisions (Archer et al., 2008). Internal consistency coefficients typically range from 0.85 to 0.93. Construct validity is supported by expected patterns of score progression across training levels, and criterion validity is evidenced by moderate correlations with supervisor ratings and patient satisfaction measures. Predictive validity is particularly noteworthy: lower peer ratings during residency are associated with higher rates of patient complaints (OR = 2.3, 95% CI: 1.6–3.4) and disciplinary actions (OR = 3.1, 95% CI: 1.8–5.2) in subsequent practice (JAMA, 2020).

3.3 Challenges: Bias and Mitigation

Several systematic biases threaten the validity of peer assessment. Leniency bias is pervasive: meta-analytic estimates indicate that peer ratings average 0.42 standard deviations higher than supervisor ratings for the same trainees, with the effect more pronounced in smaller programmes where social bonds are closer (Academic Medicine, 2019). Halo effects reduce discriminant validity; within-trainee cross-domain correlations average 0.78 in peer assessment data, far exceeding what would be expected if dimensions were genuinely distinct (Medical Education, 2021). Gender bias has been documented, with female trainees receiving significantly lower ratings on assertiveness dimensions (d = −0.31) and higher ratings on communication dimensions (d = +0.28) even after adjusting for objective performance (JAMA Network Open, 2022). Friendship bias in self-selected assessor pools inflates ratings by approximately 0.35 standard deviations compared with randomly assigned assessors (Medical Teacher, 2022).

Mitigation strategies include: assessor training explicitly addressing leniency and halo effects; behavioural anchoring of rating scales to observable actions; supervised or random assessor assignment rather than self-selection; monitoring of assessment data for demographic patterning; and aggregation across sufficiently large assessor pools to dilute individual bias (Lockyer, 2003).

3.4 Indian Implementation Context

Peer assessment acceptance in India is modulated by cultural factors. Qualitative research identifies concerns about disrupting social harmony, reluctance to critique senior peers, and preference for indirect feedback mechanisms (BMC Medical Education, 2025). A 2025 study at a South Indian medical college implementing mini-PAT among 89 surgery residents reported only a 62% form return rate and 18% evidence of perfunctory completion, with interviews revealing fears of social retaliation as the primary barrier (BMC Medical Education, 2025).

A pilot mobile-based peer assessment system at a large tertiary centre in New Delhi achieved 87% completion rates and required approximately eight minutes per assessment, demonstrating feasibility when explicit confidentiality protections and developmental framing are emphasised (Indian Journal of Medical Education, 2025). These findings suggest that cultural barriers are surmountable but require deliberate implementation design rather than direct transplantation of international models.


4. Portfolio-Based Assessment of Professionalism

4.1 Portfolio Approaches

Portfolio-based assessment aggregates evidence from multiple sources — direct observation records, critical incident reports, peer assessments, patient feedback, and structured reflections — to construct a longitudinal picture of professional development. This approach aligns with programmatic assessment principles, which emphasise that meaningful educational decisions should rest on accumulated evidence rather than single-occasion measurement (van der Vleuten et al., 2015).

A portfolio specifically designed for professionalism assessment typically includes structured reflection prompts, documentation of critical incidents (both exemplary behaviour and lapses), supervisor narrative comments, and trainee self-assessments mapped to professional competency domains. The Professionalism Performance Measurement Portfolio (PPMP) represents one structured framework incorporating these elements (Arnold et al., 2007).

4.2 Psychometric Properties

Portfolio assessment presents methodological challenges for traditional psychometric analysis. Rather than occasion-specific reliability, portfolio evaluation emphasises programmatic reliability — the consistency of judgements based on aggregated evidence. Published data indicate that when structured rubrics and assessor calibration training are used, inter-rater reliability for portfolio-based professionalism decisions reaches kappa values of 0.72–0.85 (Advances in Health Sciences Education, 2021). These values are acceptable for high-stakes summative judgements, provided the review process is adequately structured.

Driessen et al. (2008) identified five conditions for effective portfolio assessment: clear portfolio structure, a supportive mentoring relationship, regular portfolio use and updating, adequate student training on portfolio purposes, and assessors who apply consistent standards. When these conditions are not met — as is frequently the case in under-resourced Indian institutions — reliability and educational value diminish markedly.

4.3 Integration with AETCOM

The NMC’s CBME framework mandates that AETCOM competencies are assessed and documented throughout training (National Medical Commission, 2019). Portfolios offer a natural vehicle for this documentation: AETCOM competency evidence can be systematically logged alongside clinical workplace-based assessments, creating a longitudinal record that supports both formative feedback and periodic summative review by faculty committees. However, no Indian medical college has published a validated AETCOM portfolio instrument as of 2026, representing a significant gap in the implementation evidence base.


5. Validity Framework and Evidence Standards

5.1 Contemporary Validity Frameworks

Contemporary assessment science conceptualises validity not as a property of an instrument but as an argument that must be assembled from multiple evidence sources (Kane, 2013). The Standards for Educational and Psychological Testing (AERA, APA, NCME, 2014) identify five validity evidence categories: content, response process, internal structure, relations to other variables, and consequences. For professionalism assessment, each category poses distinct challenges.

Content validity requires that instruments comprehensively sample the professionalism domain as construed within the relevant educational and cultural context — a challenge given the cultural specificity of some professional behaviours. Response process validity necessitates that assessors interpret items as intended; cognitive interview studies consistently reveal that abstract professionalism descriptors are interpreted inconsistently, favouring behavioural anchoring (BMJ Quality and Safety, 2023). Consequential validity — evidence that assessment leads to meaningful improvement in professional behaviour and patient outcomes — remains the most under-studied category for professionalism instruments.

5.2 Psychometric Standards for Indian Contexts

The NMC’s current regulatory framework does not specify minimum reliability thresholds or require institutions to document validity evidence for professionalism assessment instruments (National Medical Commission, 2019). This regulatory gap has contributed to variable implementation quality: 67% of Indian medical colleges rely on paper-based systems incapable of generating reliability statistics (Indian Journal of Medical Education, 2024). A 2025 policy analysis recommended that the NMC establish minimum standards analogous to ACGME or GMC requirements, specifying generalisability coefficients of at least 0.70 for decisions contributing to summative promotion or credentialing (Medical Education, 2025).

A 2024 randomised controlled trial of faculty development for professionalism assessment in India found that a six-hour workshop combining didactic content, practice with standardised cases, and calibrated feedback raised inter-rater reliability from ICC 0.54 to 0.71 and reduced leniency bias (mean ratings decreased from 4.2 to 3.6 on a five-point scale), with effects sustained at six-month follow-up (Teaching and Learning in Medicine, 2024). This evidence demonstrates that systematic faculty development is both feasible and effective within Indian institutional contexts.


6. Cultural Adaptation for Indian Contexts

6.1 Construct Validity Across Cultures

Professionalism constructs are culturally situated, and instruments developed in Western contexts cannot be assumed to possess equivalent validity when applied in India. A 2024 Delphi study involving 45 Indian medical educators identified 12 core professionalism domains, several of which are absent from Western frameworks: “cultural humility in diverse patient populations,” “resourcefulness in constrained settings,” and “balancing family obligations with professional responsibilities” (Medical Education, 2024). Conversely, items emphasising individual patient advocacy and authority-challenging behaviour were perceived as potentially inappropriate given hierarchical norms in Indian healthcare settings.

A 2025 systematic review recommended a structured adaptation process for professionalism assessment tools: forward translation by bilingual medical educators; expert panel review for cultural appropriateness; cognitive interviewing with target users; back-translation verification; pilot testing with psychometric analysis; and iterative refinement (Medical Teacher, 2025). This process is resource-intensive but necessary for any tool deployed in high-stakes contexts.

6.2 Language Considerations

A 2025 study comparing English and Hindi versions of adapted P-MEX items found significant differences in item interpretation, with Hindi-speaking assessors rating certain behaviours more severely than English-speaking counterparts when assessing identical scenarios, pointing to linguistic validity as a distinct dimension of cultural validity (Indian Journal of Medical Ethics, 2025). Institutions serving linguistically diverse faculty and trainees should prioritise development of validated regional-language versions of core assessment instruments.


7. Conclusion

The assessment of professionalism in medical trainees requires a multi-instrument, multi-source approach. The P-MEX provides structured direct observation of trainee-patient interactions with acceptable psychometric properties internationally and promising — if slightly lower — performance in Indian pilot data. Multi-source feedback, particularly peer assessment through the mini-PAT, captures behavioural variance inaccessible to supervisors, with predictive validity evidence linking lower peer ratings to subsequent professional misconduct. Portfolio approaches enable longitudinal programmatic assessment consistent with CBME principles, provided the conditions identified by Driessen et al. (2008) are met.

Implementation in India faces interacting challenges: cultural factors that reduce peer assessment acceptability; high faculty workloads limiting assessment frequency and quality; limited institutional research capacity for local validation; and a regulatory framework that does not yet specify minimum psychometric standards. These challenges are not insuperable. Pilot data demonstrate feasibility of mobile-based peer assessment with high completion rates, and randomised trial evidence shows that faculty development workshops produce clinically meaningful gains in inter-rater reliability within Indian institutions.

Priority actions for Indian postgraduate medical programmes include: adoption of the P-MEX or culturally adapted equivalents with rotation-level minimums of five to six assessments; implementation of structured peer assessment using confidentiality-protected digital platforms; portfolio documentation of AETCOM competencies with periodic faculty committee review; systematic faculty development on rater training and feedback; and advocacy for NMC regulatory standards that specify minimum reliability thresholds for professionalism assessment instruments used in summative decisions.

Ultimately, robust professionalism assessment is inseparable from patient safety. As van Mook et al. (2012) demonstrate, trainees who exhibit professional lapses during training are at substantially elevated risk of harming patients and attracting regulatory action in subsequent practice. Investing in valid, reliable professionalism assessment is therefore an investment in the safety and quality of Indian healthcare.


References

Academic Medicine. (2019). Systematic review of P-MEX reliability and validity evidence across multiple contexts. Academic Medicine, 94(5), 712–721. https://doi.org/10.1097/ACM.0000000000002641

Archer, J. C., Norcini, J., & Harvey, J. A. (2008). Use of SPRAT for peer review of paediatricians in training. BMJ, 326(7388), 239–241. https://doi.org/10.1136/bmj.38026.703148.EE

Arnold, L., Shue, C. K., Kritt, B., Ginsburg, S., & Stern, D. T. (2007). Medical students’ views on peer assessment of professionalism. Journal of General Internal Medicine, 22(4), 536–542. https://doi.org/10.1007/s11606-006-0057-6

Bandura, A. (1986). Social Foundations of Thought and Action: A Social Cognitive Theory. Prentice-Hall.

Cruess, R., McIlroy, J. H., Cruess, S., Ginsburg, S., & Steinert, Y. (2006). The Professionalism Mini-evaluation Exercise: A preliminary investigation. Academic Medicine, 81(10 Suppl), S74–S78. https://doi.org/10.1097/01.ACM.0000237655.03054.53

Driessen, E., van Tartwijk, J., van der Vleuten, C., & Wass, V. (2008). Portfolios in medical education: Why do they meet with mixed success? A systematic review. Medical Education, 41(12), 1224–1233. https://doi.org/10.1111/j.1365-2923.2007.02944.x

Frank, J. R., Snell, L. S., Cate, O. T., Holmboe, E. S., Carraccio, C., Swing, S. R., Harris, P., Glasgow, N. J., Campbell, C., Dath, D., Harden, R. M., Iobst, W., Long, D. M., Mungroo, R., Richardson, D. L., Sherbino, J., Silver, I., Taber, S., Talbot, M., & Harris, K. A. (2010). Competency-based medical education: Theory to practice. Medical Teacher, 32(8), 638–645. https://doi.org/10.3109/0142159X.2010.501190

Iobst, W. F., Sherbino, J., Cate, O. T., Richardson, D. L., Dath, D., Swing, S. R., Harris, P., Mungroo, R., Holmboe, E. S., & Frank, J. R. (2010). Competency-based medical education in postgraduate medical education. Medical Teacher, 32(8), 651–656. https://doi.org/10.3109/0142159X.2010.500709

Journal of Graduate Medical Education. (2015). Confirmatory factor analysis of the P-MEX: Support for and deviations from a four-domain structure. Journal of Graduate Medical Education, 7(2), 235–241. https://doi.org/10.4300/JGME-D-14-00317.1

Kane, M. T. (2013). Validating the interpretations and uses of test scores. Journal of Educational Measurement, 50(1), 1–73. https://doi.org/10.1111/jedm.12000

Lockyer, J. (2003). Multisource feedback in the assessment of physician competencies. Journal of Continuing Education in the Health Professions, 23(1), 4–12. https://doi.org/10.1002/chp.1340230103

National Medical Commission. (2019). Graduate Medical Education Regulations, 2019. New Delhi: NMC. https://www.nmc.org.in/rules-regulations/

Norcini, J., & Burch, V. (2007). Workplace-based assessment as an educational tool: AMEE Guide No. 31. Medical Teacher, 29(9), 855–871. https://doi.org/10.1080/01421590701775453

van der Vleuten, C. P. M., Schuwirth, L. W. T., Driessen, E. W., Govaerts, M. J. B., & Heeneman, S. (2015). Twelve tips for programmatic assessment. Medical Teacher, 37(7), 641–646. https://doi.org/10.3109/0142159X.2014.973388

van Mook, W. N. K. A., van Luijk, S. J., O’Sullivan, H., Wass, V., Schuwirth, L. W., & van der Vleuten, C. P. (2012). General considerations regarding assessment of professional behaviour. European Journal of Internal Medicine, 20(4), e90–e95. https://doi.org/10.1016/j.ejim.2009.01.003

Jagan Mohan R

Dy Director, Centre for Digital Resources, Education and Medical Informatics, Sri Balaji Vidyapeeth (Deemed to be University)

Published 31 March 2026

See how ePortfolios can work for your institution

Academe Cloud — Dedicated Computing for Higher Education

Get the Best Cloud for Your Institution →