A pilot survey of student perceptions on the benefit of the OSCE and MCQ modalities
Stefan Müller 1Utz Settmacher 1
Ines Koch 2
Uta Dahmen 3
1 Jena University Hospital, Department of General, Visceral and Vascular Surgery, Jena, Germany
2 Jena University Hospital, Department of Gynaecology and Reproductive Medicine, Jena, Germany
3 Jena University Hospital, Department of General, Visceral and Vascular Surgery, Experimental Transplantation Surgery, Jena, Germany
Abstract
Objective: The objective structured clinical examination (OSCE) has become widely accepted as a form of assessment in medical education. At the same time, the more traditional multiple choice question (MCQ) examinations remain a central modality of student assessment. This pilot survey aimed to investigate students’ perceptions about the benefits of the OSCE and MCQs to yield data supporting the implementation of this assessment strategy into the national medical licensing examination in Germany.
Methods: A questionnaire was delivered electronically to 34 German medical schools. Students in years 3-6 were invited to rate 11 items about objectives of good medical assessment. All items were presented for both the OSCE and MCQs using a 5-point Likert Scale (1=strongly disagree to 5=strongly agree). Factor analysis was used to identify underlying components in the ratings. Average scores of items that belonged to a component were computed.
Results: Data analysis included 1,082 students from 32 medical schools. For the OSCE, factor analysis revealed two components, which were labelled “educational impact” and “development of clinical competence”. The average scores of items were 3.37 and 3.55, respectively. For the MCQ modality, also two components emerged. These were labelled “perceived weaknesses of MCQs” and “perceived strengths of MCQs” (consisting of items such as “promotes my theoretical knowledge”). The average scores for these components were 1.85 and 3.62.
Conclusion: The results of this pilot survey indicate that students consider both OSCE and MCQs as useful assessments for the purposes for which they were designed. The assessment strategy thus appears appropriate and it should be used in the national licensing examination.
Keywords
medical assessment, OSCE and MCQ modalities, perceptions, students
1. Introduction
Assessment is a core aspect of the medical education process. Besides making judgements about a candidate’s competence or performance, assessment influences the curriculum and, more importantly, drives students’ learning. For producing good physicians, medical assessment should adapt to clinical practice and be able to cover the necessary spectrum of competencies [1], [2], [3], [4]. Traditional (written) assessment practices, however, focus on testing students’ knowledge instead of appraising performance related skills [5]. This may explain why medical school graduates are often not adequately prepared for clinical work [6], [7].
The objective structured clinical examination (OSCE) is a performance-based assessment that was developed to appraise a student’s clinical performance. An OSCE typically includes a series of stations where examinees are required to apply their knowledge and skills in simulated settings. At each station, examinees’ performance is rated according to pre-established criteria. The modality has become widely accepted as a form of assessment in many countries [8]. Although OSCEs are now widespread in use, the more traditional multiple choice question (MCQ) examinations continue to play a central role in undergraduate medical education [9], [10].
1.1. Medical education in Germany
All 36 medical schools in Germany (i.e. schools founded before 2012) have a six-year undergraduate programme that is based on the statutory provisions of the German licensing regulations for physicians [
In conjunction with the amendment of the licensing regulations for physicians [
The national medical licensing examination in Germany is a three-part examination. Part I of the examination after year 2 assesses basic medical science subjects. Part II after year 5 is a written assessment to test a student’s clinical knowledge. This part of the examination lasting three days consists of standardised MCQ examinations, which are organised centrally by the Institute for Medical and Pharmaceutical Examination Questions [
1.2. Purpose of the survey
In this pilot survey, we aimed to investigate students’ perceptions about the benefits of the OSCE and MCQ modalities to yield data that support the incorporation of the OSCE into the German medical licensing examination. We focussed on the OSCE and MCQs, as these tools are commonly used methods of assessment at medical schools in Germany, and are, moreover, the key components of other national licensing examinations, for example the Medical Council of Canada Qualifying Examination (MCCQE) and the United States Medical Licensing Examination (USMLE) [15], [16]. We focussed only on the students’ views. The attitudes of the teaching staff were out of scope and thus were not considered in this survey.
2. Methods
2.1. Survey population
According to data from the Federal Statistical Office [Statistisches Bundesamt], more than 85,000 students (87,863) were enrolled at German medical schools in 2014/15. The majority of them were female students (53,352; 60.7%) [
2.2. Material
An 11-item set was developed. The process of development first involved informal interviews with students from the local medical school. During the interviews, the students were asked to indicate what they expect from good medical assessment in terms of their vocational preparation. The statements obtained were then used to construct the set of items. The set was pretested to ensure that the items were clear and understandable. The complete item set was presented on separate pages for both the OSCE and MCQs, each with the heading “What does … [the modality] do for you?” within a larger questionnaire.
- A “gives me an understanding of medical care”
- B “demonstrates the practices and principles of medical treatment”
- C “gives me feedback on my performance level”
- D “reveals my strengths in medical practice”
- E “reveals my weaknesses in medical practice”
- F “shows me gaps in my education”
- G “enhances my problem-solving and decision-making abilities”
- H “promotes my theoretical knowledge”
- J “reflects the requirements of the medical profession”
- K “allows me to assess my own ability to work as a medical professional”
- L “helps me with my speciality choice”
Moreover, demographic data on gender, age, academic year, and medical school affiliation were collected at the end of the questionnaire.
2.3. Procedure
The research was performed between February and April 2015. All 34 medical schools using the OSCE modality were solicited to forward a cover letter with the link for accessing the questionnaire to their students in years 3 to 6. Participation in the survey was voluntary and anonymous. No incentives were offered for survey completion. Formal approval from the local research ethics committee was not required.
For ensuring that respondents had personal experience with the OSCE modality, we crosschecked each data set received with the assessment schedule of the relevant medical school. Data were then analysed descriptively. To identify underlying components in the students’ ratings, we entered the data into an exploratory factor analysis with oblique rotation for each of the two assessment methods. We summed up the responses to all items loading highly onto a factor and divided the sums by the number of items to calculate average scores.
3. Results
3.1. Participants
In all, 1,189 participants completed the questionnaire. One hundred and seven participants were identified who did not have experience with the OSCE modality, gave the same rating for all the items in the set or did not specify their academic year or medical school affiliation (necessary for crosschecking), so their data were excluded from further analysis. Data of 1,082 participants, which corresponds to 3.1% of the target population, from 32 of the 34 medical schools (from two schools there were no respondents at all) were analysed. Of this total, 747 (69.0%) were female students. The age of participants ranged between 19 and 45 years, with a mean of 25.3 and a median of 25 years, covering the full age range of medical students in Germany. The percentages of students in years 4, 5, and 6 were equal (28.0%, 29.8% or 27.0%). The percentage of students in year 3, however, was smaller (15.2%), because of less experience with the OSCE modality.
3.2. Ratings in the 11-item set
Table 1 [Tab. 1] displays the ratings in the 11-item set for the OSCE and MCQs. The OSCE was rated positive (means above 3.50) or weakly positive (means between 3.25 and 3.50) for items B (3.81), A (3.75), F (3.53), E (3.36), D (3.31), and C (3.28). The MCQ modality, by contrast, received positive or weakly positive ratings for items H (4.21), C (3.34), and F (3.32).
Table 2 [Tab. 2] records the item ratings related to the students’ demographic characteristics. The ratings for the items did not vary substantially with respect to gender, age or stage of training (academic year) of participants.
3.3. Factor analysis
Table 3 [Tab. 3] shows the results of the factor analysis including the rotated factor coefficients. The Kaiser-Meyer-Olkin measure of sampling adequacy was 0.90 for the OSCE and 0.87 for the MCQs, respectively, suggesting that data were appropriate for factor analysis. Two factors or components with eigenvalues >1 were found for both modalities. For the OSCE, items E, D, F, and C loaded highly (coefficients >0.60) on factor 1 with loadings of 0.95, 0.84, 0.83 or 0.68. The average score of the four items was 3.37. After inspecting the content of the items, we labelled the component “educational impact”. Items B and A, as well as item H (mean 3.10) loaded highly on factor 2 (factor loadings of 0.80, 0.74, and 0.77, respectively). The average item score was 3.55. Based on the content of the items, we labelled this component “development of clinical competence”. For the MCQ modality, items D, E, K, J, and G (each with a mean below 2.00) had high loadings on factor 1 (factor coefficients of 0.92, 0.89, 0.78, 0.73, and 0.62, respectively), while items H, C, and F loaded highly on factor 2 (coefficients of 0.85, 0.79 or 0.70). The average scores for the two components were 1.85 and 3.62. The components were labelled “perceived weaknesses of MCQs” and “perceived strengths of MCQs”.
4. Discussion
This pilot study investigated the benefits of the OSCE and MCQ modalities perceived by students. In the following paragraphs, we discuss the survey findings classified by themes.
Educational impact is considered an important feature of assessment [17], [18]. An assessment, even when it is used for summative purposes, should have a positive effect on students’ future learning by providing feedback to the learners about their strengths and weaknesses. The assessment should also identify areas of weakness in teaching practices or the curriculum so educators can make adjustments. In our study, students rated the items C to F (items that contributed to the component “educational impact”) positive or weakly positive for the OSCE. This indicates that the OSCE modality can serve as a tool to improve both students’ learning and the curriculum, which is in line with previous research [19], [20]. However, it should be pointed out that the provision of detailed feedback to students is difficult to accomplish in a high-stakes licensing OSCE. Nonetheless, the experience gained from taking the assessment as well as the grades achieved can be used by students to enhance their future learning [21].
A practicing physician should be competent in the domains medical knowledge, clinical skills (e.g. taking history from a patient or performing a physical examination), practical procedures (e.g. establishing intravenous access), patient management, communication, and professional behaviour [22], [23]. Helping students develop competencies in the required domains is a major objective of medical education. The item ratings suggest that students perceive the OSCE modality as useful for developing competencies in the domains other than medical knowledge. This conclusion can be drawn from positive ratings for items A and B, vs. a merely mediocre rating for item H (all items that were assigned to the component “development of clinical competence”). Bearing in mind that the OSCE is a performance measure appraising skills and behaviours that are needed in the clinical workplace [8], the present findings are consistent with the OSCE modality’s intended focus.
The greatest perceived strength of the MCQs appears to be in the fostering of the acquisition of knowledge. A strong positive rating for the item “promotes my theoretical knowledge” underpins that. However, our findings indicate that students perceive the MCQ modality to be only suitable to measure lower order cognitive skills (e.g. factual knowledge). This raises the issue of whether medical schools in Germany too often rely on context-free MCQs, which typically consist of discrete questions that aim to test factual knowledge. If so, and in order to include higher order cognitive skills (e.g. processes of problem solving or decision making), it would be advisable to expand the use of context-rich MCQs, in which the questions are directly related to a clinical case presentation [24], [25].
Furthermore, similar to the OSCE, the MCQ modality seems to enable students to evaluate both their own achievements and the content of the curriculum, which is most likely driven by the assessment results. Weakly positive ratings for the items “gives me feedback on my performance level” and “shows me gaps in my education”; vs. low ratings under “reveals my strengths in medical practice” or “reveals my weaknesses in medical practice” suggest this inference.
There are several limitations to this study. First, the study is limited to a relatively small sample representing less than 5% of the target population. This is probably because most medical schools did not inform their students by email, but merely put the cover letter with the link for accessing the questionnaire on their websites. For this reason, a large number of students were likely uninformed about the survey. Second, the study was a pilot survey. The item set presented is not exhaustive and gives an overview only. Despite its limitations, we believe that this survey, the first to investigate the perceptions of students across Germany, provides insight into how and to what extent the OSCE and MCQs are useful assessments. Moreover, our sample was quite representative in terms of gender and age, and we had great diversity regarding the stages of training and medical school affiliations of participants, which make it probable that our results can be generalised to the whole medical student population in Germany.
5. Conclusions
The findings of this pilot survey suggest that students consider both assessment modalities, the OSCE and MCQs, to be valuable tools. In summary, the OSCE may have an impact on the educational process and support the development of skills and behaviours required for clinical practice, while the MCQ modality fosters the acquisition of knowledge. Whilst the employment of assessment programmes including a battery of tests will be the most robust strategy to create a global appraisal of a candidate’s knowledge and skills [2], [3], [26], the use of the OSCE and MCQs appears to be an appropriate assessment strategy. This is further evidence of the need to incorporate the OSCE into the German medical licensing examination in addition to the existing MCQs.
Notes
Acknowledgements
The authors wish to acknowledge the deans’ offices of the medical schools, all the participating students, and Prof. Dr. Thomas Kessler and Prof. Dr. Rolf Steyer, both Institute of Psychology at Jena University, for providing support in carrying out this research.
Competing interests
The authors declare that they have no competing interests.
References
[1] Norcini J, Anderson B, Bollela V, Burch V, Costa MJ, Duvivier R, Galbraith R, Hays R, Kent A, Perrott V, Roberts T. Criteria for good assessment: Consensus statement and recommendations from the Ottawa 2010 Conference. Med Teach. 2011;33(3):206-214. DOI: 10.3109/0142159X.2011.551559[2] Schuwirth LWT, Ash J. Principles of assessment. In: Walsh K (Hrsg). Oxford Textbook of Medical Education. Oxford: Oxford University Press; 2013. S.409-420. DOI: 10.1093/med/9780199652679.003.0035
[3] Shumway JM, Harden RM. AMEE Guide No. 25: The assessment of learning outcomes for the competent and reflective physician. Med Teach. 2003;25(6):569-584. DOI: 10.1080/0142159032000151907
[4] Sturmberg JP, Farmer L. Educating capable doctors–A portfolio approach. Linking learning and assessment. Med Teach. 2009;31(3):e85-89. DOI: 10.1080/01421590802512912
[5] Crossley J, Humphris G, Jolly B. Assessing health professionals. Med Educ. 2002;36(9):800-804. DOI: 10.1046/j.1365-2923.2002.01294.x
[6] Burch VC, Nash RC, Zabow T, Gibbs T, Aubin L, Jacobs B, Hift RJ. A structured assessment of newly qualified medical graduates. Med Educ. 2005;39(7):723-731. DOI: 10.1111/j.1365-2929.2005.02192.x
[7] Lypson ML, Frohna JG, Gruppen LD, Woolliscroft JO. Assessing residents' competencies at baseline: identifying the gaps. Acad Med. 2004;79(6):564-570. DOI: 10.1097/00001888-200406000-00013
[8] Humphrey-Murto S, Touchie C, Smee S. Objective structured clinical examinations. In: Walsh K (Hrsg). Oxford Textbook of Medical Education. Oxford: Oxford University Press; 2013. S.524-536. DOI: 10.1093/med/9780199652679.003.0045
[9] Fowell SL, Maudsley G, Maguire P, Leinster SJ, Bligh J. Student assessment in undergraduate medical education in the United Kingdom, 1998. Med Educ. 2000;34 Suppl 1:S1-49. DOI: 10.1046/j.1365-2923.2000.0340s1001.x
[10] Mavis BE, Cole BL, Hoppe RB. A survey of student assessment in U.S. medical schools: The balance of breadth versus fidelity. Teach Learn Med. 2001;13(2):74-79. DOI: 10.1207/S15328015TLM1302_1
[11] Chenot JF. Undergraduate medical education in Germany. GMS Ger Med Sci. 2009;7:Doc02. DOI: 10.3205/000061
[12] Müller S, Dahmen U, Settmacher U. Application of the Objective Structured Clinical Examination (OSCE) in German medical schools: An inventory. Gesundheitswesen. 2016. DOI: 10.1055/s-0042-116435
[13] Huber-Lang M, Palmer A, Grab C, Boeckers A, Boeckers TM, Oechsner W. Visions and reality: the idea of competence-oriented assessment for German medical students is not yet realised in licensing examinations. GMS J Med Educ. 2017;34(2):Doc25. DOI: 10.3205/zma001102
[14] Berendonk C, Schirlo C, Balestra G, Bonvin R, Feller S, Huber P, Jünger E, Monti M, Schnabel K, Beyeler C, Guttormsen S, Huwendiek S. The new final Clinical Skills examination in human medicine in Switzerland: Essential steps of exam development, implementation and evaluation, and central insights from the perspective of the national Working Group. GMS Z Med Ausbild. 2015;32(4):Doc40. DOI: 10.3205/zma000982
[15] Bordage G, Meguerditchian AN, Tamblyn R. Practice indicators of suboptimal care and avoidable adverse events: a content analysis of a national qualifying examination. Acad Med. 2013;88(10):1493-1498. DOI: 10.1097/ACM.0b013e3182a356af
[16] Swanson DB, Roberts TE. Trends in national licensing examinations in medicine. Med Educ. 2016;50(1):101-114. DOI: 10.1111/medu.12810
[17] Shepard LA. The role of assessment in a learning culture. Educ Res. 2000;29(7):4-14. DOI: 10.3102/0013189X029007004
[18] Van der Vleuten CP. The assessment of professional competence: Developments, research and practical implications. Adv Health Sci Educ Theory Pract. 1996;1(1):41-67. DOI: 10.1007/BF00596229
[19] König S, Wagner P, Markus PM, Becker H. Anders prüfen – anders studieren: Motivation durch OSCE. Med Ausbild. 2002;19(2):73-76.
[20] Tervo RC, Dimitrievich E, Trujillo AL, Whittle K, Redinius P, Wellman L. The Objective Structured Clinical Examination (OSCE) in the clinical clerkship: an overview. S D J Med. 1997;50(5):153-156.
[21] Wood DF. Formative assessment. In: Walsh K (Hrsg). Oxford Textbook of Medical Education. Oxford: Oxford University Press; 2013. S.478-488. DOI: 10.1093/med/9780199652679.003.0041
[22] Simpson JG, Furnace J, Crosby J, Cumming AD, Evans PA, Friedman Ben David M, Harden RM, Lloyd D, McKenzie H, McLachlan JC, McPhate GF, Percy-Robb IW, MacPherson SG. The Scottish doctor – learning outcomes for the medical undergraduate in Scotland: a foundation for competent and reflective practitioners. Med Teach. 2002;24(2):136-143. DOI: 10.1080/01421590220120713
[23] Swing SR. The ACGME outcome project: retrospective and prospective. Med Teach. 2007;29(7):648-654. DOI: 10.1080/01421590701392903
[24] McConnell MM, St-Onge C, Young ME. The benefits of testing for learning on later performance. Adv Health Sci Educ Theory Pract. 2015;20(2):305-320. DOI: 10.1007/s10459-014-9529-1
[25] Schuwirth LW, van der Vleuten CM. Different written assessment methods: what can be said about their strengths and weaknesses? Med Educ. 2004;38(9):974-979. DOI: 10.1111/j.1365-2929.2004.01916.x
[26] Van der Vleuten CP, Schuwirth LW. Assessing professional competence: from methods to programmes. Med Educ. 2005;39(3):309-317. DOI: 10.1111/j.1365-2929.2005.02094.x