Article Access Statistics | | Viewed | 2216 | | Printed | 138 | | Emailed | 0 | | PDF Downloaded | 19 | | Comments | [Add] | |
|

 Click on image for details.
|
|
|
LETTER |
|
|
|
Year : 2021 | Volume
: 67
| Issue : 1 | Page : 57-58 |
The power of subjectivity in competency-based assessment
M Gopalakrishnan, MK Garg
Department of Internal Medicine, All India Institute of Medical Sciences, Jodhpur, Rajasthan, India
Date of Submission | 19-Nov-2020 |
Date of Decision | 13-Dec-2020 |
Date of Acceptance | 15-Dec-2020 |
Date of Web Publication | 21-Jan-2021 |
Correspondence Address: M Gopalakrishnan Department of Internal Medicine, All India Institute of Medical Sciences, Jodhpur, Rajasthan India
 Source of Support: None, Conflict of Interest: None  | Check |
DOI: 10.4103/jpgm.JPGM_1251_20
How to cite this article: Gopalakrishnan M, Garg M K. The power of subjectivity in competency-based assessment. J Postgrad Med 2021;67:57-8 |
We congratulate Virk A et al. on their recent review- “The power of subjectivity in competency-based assessment” as it has reopened the timely question of subjectivity in assessments in the context of the recent competency-based medical education (CBME) curriculum implementation in India.[1] We agree with the authors that assessing clinical competence is more than mere stacking of numbers, and the assessment scenario must be holistic, more than just the sum of parts.
We also agree with the authors about the misplaced assumption that reliability in the checklist-mediated approach, especially in ?Objective Structured Clinical Examinations (OSCEs), was attributed to “objectivity,” while it has turned out to be a “red-herring” while the “black box” that works there is actually wider sampling, i.e., multiple examiners and multiple assessment stations.[2] There are several points in the review which need further examination.
The authors have argued that objective assessments are “norm-referenced,” while subjective assessments are criterion referenced. Both these claims appear to be misleading. The norm and criterion referencing are useful and applicable in different situations.[3] For example, we routinely use a classical “objective” assessment like a multiple choice question test (MCQ) with a “pass mark” (say 40%) for the test which makes it criterion referenced. When the same MCQ is interpreted as a student doing better than 60% of all the test takers, it becomes norm referenced. A norm-referenced interpretation may be useful in situations like entrance exams, while a criterion is referenced for assessing the mastery of subjects by the students. They have nothing to do with the assessment being subjective or objective.
The authors have also stated that objective assessment is comparable to a cross-sectional study and subjective to a longitudinal study. This is also misleading in the sense that both types of assessments can have data points over time or at one point making them longitudinal or cross-sectional. For example, the results of three MCQ tests or OSCEs over three months is a longitudinal objective assessment. Also, the claim that all standard setting is purely expert opinion seems far-fetched as several well-established frameworks exist for the same.[4]
We believe that the argument here must not be objective versus subjective assessment but how both can complement each other in creating valuable richer data about the Indian Medical Graduates for fulfilling the dual objectives of feedback for learning and making defensible high stake pass–fail judgments. Hence, we wish to reframe the dichotomy with the lens of programmatic assessment.
Programmatic assessment is a set of broad principles which incorporate various aspects which the authors have touched upon in their article. These include collecting a continuum of rich low-stake data points (each representing a “pixel” in the whole picture) involving qualitative feedbacks (including subjective assessments) and mentorship for the learner, resulting in trustworthy decision making.[5] The principles of programmatic assessment align well with CBME. The real challenge is to figure out the implementation strategies, especially at the undergraduate level and see how concepts like workplace-based assessments and portfolios can be modified and adapted to our needs and settings.
Financial support and sponsorship
Nil.
Conflicts of interest
There are no conflicts of interest.
:: References | |  |
1. | Virk A, Joshi A, Mahajan R, Singh T. The power of subjectivity in competency-based assessment. J Postgrad Med 2020;66:200-5.  [ PUBMED] [Full text] |
2. | Wilkinson TJ, Tweed MJ. Deconstructing programmatic assessment. Adv Med Educ Pract 2018;9:191-7. |
3. | Ricketts C. A plea for the proper use of criterion-referenced tests in medical assess?ment. Med Educ 2009;43:1141-6. |
4. | Downing SM. Item response theory: Applications of modern test theory in medical education. Med Educ 2003;37:739-45. |
5. | Van Der Vleuten CPM, Schuwirth LWT, Driessen EW, Govaerts MJB, Heeneman S. Twelve tips for programmatic assessment. Med Teach 2015;37:641-6. |
|