Learning physical examination skills outside timetabled training sessions: what happens and why?
Lack of published studies on students’ practice behaviour of physical examination skills outside timetabled training sessions inspired this study into what activities medical students undertake to improve their skills and factors influencing this. Six focus groups of a total of 52 students from Years 1–3 using a pre-established interview guide. Interviews were recorded, transcribed and analyzed using qualitative methods. The interview guide was based on questionnaire results; overall response rate for Years 1–3 was 90% (n = 875). Students report a variety of activities to improve their physical examination skills. On average, students devote 20% of self-study time to skill training with Year 1 students practising significantly more than Year 3 students. Practice patterns shift from just-in-time learning to a longitudinal selfdirected approach. Factors influencing this change are assessment methods and simulated/real patients. Learning resources used include textbooks, examination guidelines, scientific articles, the Internet, videos/DVDs and scoring forms from previous OSCEs. Practising skills on fellow students happens at university rooms or at home. Also family and friends were mentioned to help. Simulated/real patients stimulated students to practise of physical examination skills, initially causing confusion and anxiety about skill performance but leading to increased feelings of competence. Difficult or enjoyable skills stimulate students to practise. The strategies students adopt to master physical examination skills outside timetabled training sessions are self-directed. OSCE assessment does have influence, but learning takes place also when there is no upcoming assessment. Simulated and real patients provide strong incentives to work on skills. Early patient contacts make students feel more prepared for clinical practice.
KeywordsUndergraduate Medical students Practical Physical examination skills Clinical skills
Recent years have seen increasing attention for physical examination skills in the medical curriculum. Pleas from clinicians to restore the art of physical diagnosis as a core competency for students have not gone unheard (Anderson et al. 2001). In recommendations for changes in the content and delivery of the undergraduate medical curriculum, various bodies have emphasised the importance of physical examination skills and recommended essential skills in which students have to demonstrate competence and proficiency before graduation. (General Medical 1993) (Colleges 1998; Metz et al. 2001).
All this attention has encouraged many medical schools to revise their curricula to include clinical skills and to establish clinical skill centres to provide appropriate training. (Du Boulay and Medway 1999; Dent 2001; Stark and Fortune 2003).
There is some disagreement about the scheduling of physical examination skills in the curriculum: in the preclinical (Ledingham 1998; Bradley and Bligh 1999) or in the clinical (Corbett et al. 2008) phase of the undergraduate curriculum. Remmen showed that students in medical schools offering longitudinal skills programmes were better prepared for clinical rotations (Remmen et al. 2001). Others have argued that skills training should be integrated with real life practice (Silverman and Wood 2004), and clinical clerkship directors have indicated that the majority of skills should be learned during clerkships. (O’Brien et al. 2007).
Students appreciate skills training as preparation for future medical practice (Lam et al. 2002). It also appears to ease the transition from the theoretical phase to the clinical phase of undergraduate medical education (Prince et al. 2000a, b).
Students learn skills not only during formal training sessions but also by engaging in a variety of activities either in preparation for or complementary to regular training sessions. The existing evidence about students’ preparation for skills training is ambiguous. Most studies found that students prepare for training sessions and practise outside timetabled teaching activities. Rudland et al. recently suggested that students mostly practise skills on each other and rehearse routines (Rudland et al. 2008). However, a study by Mavis revealed that students who practised for more than 3 h spent only 20% of that time actually performing skills, while the remaining time was used to review textbooks and class notes. (Mavis 2000) There is anecdotal support for the hypothesis that students’ learning and practising outside timetabled sessions is limited. As much as one-third of students never practise outside training sessions. (Bhoopatkar and Weam 2008).
which activities do medical students undertake outside regular training sessions to improve their physical examination skills?
how much time do they spend on skill practice, what factors influence their practice behaviour and why?.
Examples of physical examination skills taught in years 1–3
Examination of foot/ankle (distorsion)
Auscultation of the lungs
Central venous pressure, Peripheral arterial circulation and blood
Percussion and palpation of the abdomen
Pediatric abdominal examination
Taking cervical smear test
Examination in pregnancy, normal labour and delivery
Motor development infant
Examination of the (Painful) shoulder
Andrological and urological examination
Years 4–5 consist of clinical rotations and year 6 consists of an 18-week research internship and an 18-week clinical internship in one clinical department. Skills training in these years is mainly focused on refreshing and rehearsing skills, because the essential clinical skills have been addressed during the previous years. The skills training programme is developed, organised and delivered by the Skillslab, a specialised educational facility (Van Dalen 1997).
During the first 3 years students take part in a communication skills programme, consisting of simulated patient encounters and group sessions, in which students provide feedback on the recorded encounters of their peers. Simulated patients are primarily used for teaching communications skills; when students are required to take a physical examination during simulated patient encounters they will have received formal training in the Skillslab beforehand. (Duvivier 2010b)In Year 3, students have real patient encounters on 1 day per week in the community (general practice) or the university hospital. The patient contacts serve as the starting point for self-study and group discussions, as described in detail elsewhere. (Diemers et al. 2007) Students sit in on consultations, and gradually take on more responsibilities such as history taking and physical examination.
At the end of years 1, 3 and 5 there is a compulsory Objective Structured Clinical Examination (OSCE) Performance is scored on checklists by trained observers. Passing the OSCE is prerequisite for advancing to the next stage of the curriculum.
Sequential study design
We used a sequential study approach with focus groups interviews guided by questionnaire results. This method allows for a broad and possibly superficial collection of themes (statistical analysis of quantitative data) as well as in-depth exploration of experiences (using qualitative focus groups).
The questionnaire asked about the amount of time devoted to skill practice and about learning resources used. The results were used to identify important factors to be pursued during the focus group interviews.
We developed a questionnaire based on the literature and on expert opinions. After piloting among four students (2nd and 3rd Year) and review by four medical education experts (all professors and holding PhDs), some modifications were made. The final questionnaire consisted of thirteen questions and took 10 min to complete. It consisted of two parts: questions about age, sex and year of study, and questions about the amount of time devoted to studying and skill practice (current year only) See appendix 1 for the full questionnaire. Before completing the questionnaire, the participating students received an explanation of the purpose of the questionnaire. Students participated voluntarily and received no compensation.
We approached all students in Years 1–3, because after the 3 year students no longer attend scheduled skill training sessions. Not all students could be reached: 875 out of 972 completed the questionnaire giving an overall response rate of 90%. The sample comprised 298 first-year students (92% of the Year 1 cohort), 292 s-year students (87% of the Year 2 cohort) and 285 third-year students (90% of that cohort). The data of 45 students who failed to state their year of training were not used in the analysis that considered year of study as a variable.
SPSS Software (version 15.0 for MS Windows) was used to analyse the questionnaire data. We calculated mean scores, standard deviations, and performed t-tests to determine significant differences. A p value < 0.05 was considered significant.
We used a qualitative approach for in-depth exploration of students’ skill training activities, because it can elicit a wide variety of opinions and includes multiple perspectives. We organised focus group interviews in accordance with guidelines described by Morgan. (Morgan 1997) We used an interview scheme based on the results of the questionnaire. It contained general open questions and prompting questions to be asked when elaboration of a topic did not occur spontaneously. See appendix. Recruitment was done via an e-mail sent to all students of Years 1–3 explaining the purpose and procedure of the study and inviting the students to participate, followed by a reminder e-mail 2 weeks later. Our aim was to collect data until saturation had been reached. We randomly interviewed 52 students (12 male, 40 female) from the total that had responded. We randomly assigned participants to two groups per study year: six groups in total. The participants received a small financial compensation for participating in the study. The participants did not significantly differ in terms of study results from the general student population. Gender distribution was significantly different: 64% of questionnaire respondents were female while 77% of focus group participants were female.
Twelve sessions in total (two per group, two groups per study year 1–3) were held and facilitated by one researcher (AJJAS or JvD). A second researcher (KvG) observed and took notes, interrupting to ask additional questions, if necessary. First sessions lasted 60–70 min. During the second session (20–30 min) the findings of the first session were verified and clarified. The sessions were recorded and transcribed literally. Participants were allowed to review the transcribed sessions and make additions (member check). The transcripts were independently analysed by three researchers (RJD, KvG, JvD). Comments were coded and analyzed for emerging themes.
Two of the researchers (RJD and KvG) were medical students. In order to prevent bias (due to preconceptions and experiences that might affect data collection and analysis) we adopted an iterative analytical process.The central approach of the analysis was constant comparison: issues of interest in the data were compared for similarities and differences. The researchers discussed the process afterwards and reached consensus on any inconsistencies (Strauss and Corbin 1990).
At the time of study, educational research studies reporting students’ opinions did not require approval from the ethics committee at Maastricht University. However, relevant ethical issues were carefully considered by the Department of Educational Research and Development.
We present the results of the quantitative analysis of the questionnaire data, followed by the themes that emerged from the qualitative analysis of the focus group transcripts, with illustrative quotes.
The results are grouped under two categories and several subheadings, providing answers to the questions: how do students practise physical examination skills outside training sessions and what influences these activities?
How do students practise physical examination skills?
The mean age of the respondents was 20.3 years (SD = 1.50 years) and 64% were female. This reflects the overall age and gender distribution in the medical student population of Maastricht University.
Time spent practising skills outside timetabled training sessions
Time spent on study-related activities
Mean number of hours of self study per week
15.6 h (SD 6.3) 68% of total study time
16.4 h (SD 6.4) 71% of total study time
15.8 h(SD 6.2) 65% of total study time
17.0 h (SD 6.3) 72% of total study time
Mean number of hours per week for skills
3.5 h (SD 2.3) 22% of self study time
3.7 h (SD 2.4) 23% of self study time
2.9 h (SD 1.8) 18% of self study time
3.4 h (SD 2.2) 20% of self study time
The focus group sessions revealed six categories of resources used by the participants: textbooks, examination guidelines, scientific articles, the Internet, videos/DVDs and scoring forms from previous OSCEs.
I watched the DVDs quite a lot. You read the guidelines but you are not quite sure what exactly it is you are supposed to do and it is very good to actually see what you have to do. I think there is real added value to that (year 2, group 2, student 3). Those DVDs are quite helpful: it is easier than reading about it and you see someone demonstrating the skills. Year2, group 2, student 5).
General textbooks were not frequently used by Year 1 students. Students in Year 3 reported using them to link complaint, disease and physical examination. Some students said they only studied the guidelines and never used any other learning resources when practising clinical skills.
There is so much theory and the question is where to start and then scoring forms are useful, for you do know you have to know all the examinations. But in this way you can see what they are going to ask (year 2, group 2, student 6).
Students can use university rooms, equipped with all the necessary materials to practise physical examination skills. Some students said they had never used these rooms. As inhibiting factors they mentioned suboptimal materials or outdated learning resources and having to reserve the room in advance, which they considered cumbersome and inflexible.
It is nicer to be in your underwear with just two girlfriends and to just chat around a bit. (year 2, group 1, student 4 [female student]).
My friend is a 2 year medical student and it is really easy to practise with him for he can tell me like ‘you’re doing this the wrong way, you’re doing that the wrong way’, when I practise with my mother, she is thinking ‘oh, you are doing really well, you are just like a real doctor’, but she cannot tell me when I get it wrong (year 1, group 2, student 3).
If you get it wrong there is nobody there to tell you, so you never know if you are getting it right and that is quite frustrating (year 2, group 2, student 4).
You may practise abdominal examination but when you do percussion the wrong way there is nobody there to tell you that that’s not how it is done (year 3, group 1, student 6).
Students also valued the opportunity to be examined and to feel the effect of certain procedures: too hard, too soft etc.
I sometimes go to my grandmother. She is very old and that makes such a difference. She thinks it’s great. She is 90 years old, so when you listen to her lungs you hear almost nothing because it is all very muted. Or my friend’s grandmother. They all have lungs. Everybody has to know what I am studying (year 1, group 2, student 5).
What influences practising?
Influence of OSCEs
Responses from 1 year students showed that the uncertainty surrounding the OSCE is a strong incentive to practise physical examination skills. In later years students seemed to have become more familiar with the purpose of the exam, but there was still considerable anxiety about the expected level of competence.
The immediate prospect of the OSCE. That you write more down during training. Really much more, yes (year 1, group 1, student 8).
Influence of simulated and real patients
In our group no feedback is given about content aspects. So I say to someone with an ankle, well let me feel your ankle, come over here. Sit down please and then I do a fake examination which is wrong in every possible way (year 1, group 2, student 4).
This year I haven’t given any thought to real patients not at any moment [when practising skills] But maybe next year I will think more about real patients when I am going to a general practice. (year 2, group 1, student 2).
Real patient contacts stimulate students to practise clinical skills.
Real patient contacts cause confusion about “what is the correct way”.
Real patient contacts lessen the “shock of practice”/make students feel more prepared for the reality of clinical practice.
In the first 2 years it was only for the test, like I have to start cramming to pass the test and now, in third year, with a patient … you do not want to be there not knowing how to do a lung examination. (year 3, group 1, student 6).
It is not so much for the supervisor, it is more towards the patient (year 3, group 1, students 4).
It is simply not acceptable to mess around (year 3, group 1, student 7).
It is quite important when you are with a patient that you know what to do and that you are not standing there aimlessly. For me that is the main motivation to prepare really well. (year 3, group 2, student 7).
Influence of clinical environment
Students described several confusing experiences when they first entered the clinical environment. Overall, students found that the reality of clinical practice in Year 3 was rather shocking because it confronted them with deficiencies in their knowledge and skills. Being confronted with a different culture, an unfamiliar role and novel interaction with healthcare staff also made them feel uncertain.
And with physical examination you are taught to listen for so long and at so many places and then you are with a GP and he puts his stethoscope at 4 places within 30 s … it is a matter of more experience as well. (year 3, group 1, students 5).
With the abdomen you have a skillslab and you are taught to determine the outline of the liver and we are taught to do that in such a way that you think … about here, for I know it should be approximately there … but he showed us a way that works really well. You put your stethoscope here and then you brush with your finger and then you hear the difference where you are brushing, you hear it extremely clearly. It seems strange that we should not learn this in the skillslab … (year 3, group 1, students 3).
Differences between years
You know more and more and at a certain point things start to fall into place. Right now I notice this very strongly with skills: you are thinking more and more like I hear this and I hear that and you can make connections […] and in the first year I just outright memorised lists. (year 3, group 1, student 4).
You integrate different steps of the examination. When you have seen something on inspection, you think during auscultation ‘hey I just have to hear that (year 3, group 1, student 5).
In the clinic I never really feel that I might be deficient in examination skills, I think that I am more likely to have deficiencies in knowledge to interpret… I am quite capable of doing a physical examination (year 3, group 2, student 8).
I think it is good that you learn the basics and then move on but I think it is a pity that you have to put on an act for the OSCE….“It should be done precisely so” and even if the doctor at the exam does things very differently himself but it is not on the scoring form, it has o be done exactly as it is written. You feel as if you are in a play. (year 3, group 1, student 4).
When you do an examination you do it because you have an idea at the back of your head and you want to find something in the patient and then you think of a diagnosis and the observer says that you did not look at the pattern of the hair on the legs in relation to circulation deficiencies. You could not see it because the legs were shaven but when you would have said it would have meant an extra mark. It’s just so stupid, things like that I just don’t need them (year 3, group 2, students 3).
Influence of the nature of the examination
The students reported a wide variety of study habits with regard to clinical skills.
Skills that are difficult to perform.
Skills that are enjoyable to perform.
I think that unconsciously it plays a role. It is not so much embarrassment but because you feel you have to be able to do it. You do not want to seem backward or something. (year 2, group 1, student 6).
Because you practise on someone and it is quite stressful the first time (year 2, group 2, student 6).
What I also do is practising at home on my girlfriend, my father, my mother. Anyone who feels like it comes under the stethoscope (year 3, group 1, student 2).
This study aimed to explore how students work outside scheduled training sessions to improve their physical examination skills, and what factors play a role in this. Using a combination of qualitative and quantitative methods, we discovered that students report a variety of activities to improve their physical examination skills.
On average, students devote 20% of self-study time to skill training with Year 1 students practising significantly more than Year 3 students. Learning resources used include textbooks, examination guidelines, scientific articles, the Internet, videos/DVDs and scoring forms from previous OSCEs. Student practise skills on fellow students, family members or friends at university rooms or at home.
The motivation to practise is influenced by both external and internal factors. Assessment provides a strong incentive for students to practise, caused by anxiety about the format of the exam and the expected level of competence. Real and simulated patient encounters also motivate students to practise their skills. Students mention patient pressure and supervisors’ expectations as drivers for learning. In later years the impact of those external factors gradually fades and is replaced by intrinsic motivation. Students report that practice increases their self-confidence and perceived level of competence.
In conclusion, students’ practice patterns shift from just-in-time learning to a longitudinal self-directed approach during the first three years of the curriculum. Factors influencing this change are A) assessment methods and B) authenticity of the learning environment by introducing simulated/real patients.
These outcomes are in line with other research on the transition from preclinical to clinical training (Prince et al. 2000a, b; Prince et al. 2004; Diemers et al. 2008). However, the current study is the first to address the acquisition of physical examination skills and the strategies students adopt to master these. We will consider theoretical explanations for the reported findings, and discuss possible implications for medical education.
Influence of assessment methods
The influence of OSCEs on students’ practising behaviour provides partial support for the conventional assertion that assessment drives learning. (Muijtjens et al. 1998; McLachlan 2006) However, our results also reveal significant incentives for learning that are not related to assessment. The OSCE does influence learning strategies, but learning takes place also when there is no upcoming assessment, as is demonstrated by the influence of real patients. This shows that students will work on their physical examination skills, even when they do not know if and when they will be formally assessed. Intrinsic motivation to practise for (simulated) patient encounters is high. We therefore propose to modify the maxim “assessment drives learning” to include “and patients drive practising” within the context of physical examination skills. The influence of early patient contacts on learning physical examination skills extends the growing body of evidence on patient-based education into the skills domain. (Dornan et al. 2006a, b).
Contrary to earlier studies, which investigated time spent preparing for an OSCE (Mavis 2000; Rudland et al. 2008), we found that the larger part of practice time is dedicated to skill performance rather than reviewing textbooks and notes. Students in Year 3 practise less outside timetabled training sessions as they (1) feel increasingly competent and (2) have opportunities to practise on real patients in general practice and hospital settings. A recent study also found that students regularly practised skills on each other outside time-tabled training sessions. (Rudland et al. 2008) These spontaneous, self-directed group learning activities seemed to be inspired by a PBL-environment. This cooperative/collaborative learning may be seen as a manifestation of the ‘hidden curriculum’ (Snyder 1971) and as such warrants more research.
Authenticity of the learning environment
The reported shift towards clinically oriented thinking was addressed in earlier studies on knowledge acquisition in the theoretical curriculum. (Diemers et al. 2008; Vanhell et al. 2008) Our findings complement the theory that students progressively integrate their knowledge and skills while problem-solving. Cognitive psychological theories of clinical reasoning have emphasised that the structure of knowledge differs in experts (clinicians) and novices (medical students) (Schmidt and Boshuizen 1993). As expertise develops and medical students acquire extensive clinical experience, they develop the ability to recognize and manage symptoms efficiently and effectively. Our findings suggest that early introduction of skills training contributes to the development of competence and confidence in physical examination skill performance. Remmen et al. demonstrated earlier that longitudinal skills training is more effective than clerkships to learn physical examination skills (Remmen et al. 2001).
Early exposure to physical examination skills would therefore assist students in developing an extensive cognitive database of possible clinical findings as preparation for clerkship rotations. During their clerkships this database will develop further. Seeing patients in clinical care will spur the development of ‘illness scripts’, cognitive structures that clinicians use for storing and organizing information about patient symptoms and conditions in memory (Feltovich and Barrows 1984). Illness scripts contain not only clinically relevant information about patient cases, but also the context in which the case was seen. In everyday practice, the resulting ‘pattern recognition’ allows relevant information to be retrieved efficiently thereby contributing to a doctor’s ability at differential diagnosis (Eva 2005).
In this model, based on cognitive psychological research, the more opportunities students have to see patients or to be exposed to clinical material, the richer their database of illness scripts will be. This will provide more opportunity to organise schemes of relevant information in memory readily available for quick retrieval when needed (Schmidt et al. 1990).
Consequently, including clinical scenarios or (simulated) patients in physical examination skills teaching will assist the development of clinical competence.
Physical examination skills are demonstrably an important part of the diagnostic process (Sackett 1992). There is evidence that diagnostic accuracy improves with training and time spent on a task (Patel and Groen 1991). Exposing medical students to a large number of clinical cases improves their performance (Vosti et al. 1997; Papa et al. 1999), especially when combined with quality feedback (Papa et al. 1999).
Although feedback during performance is important, research in skills acquisition has shown that humans learn best if feedback is incorporated in training designs (Fischman and Oxendine 1998). Students in our study emphasised the motivational effect that (simulated) patients have on their practising behaviour, but they felt that feedback on their physical examination skills was sometimes lacking. Evidence shows that individual feedback and remedial instruction in general improve learning outcomes (Bloom et al. 1987). Provision of feedback immediately after summative tests has been associated with improved competence (Newble and Jaeger 1983).
In addition to the large body of literature on feedback in general, several studies have described feedback during physical examination skills training (Wigton et al. 1986; Dunnington et al. 1992; Tape et al. 1992; Schwind et al. 2001). The learning effect is boosted when feedback is specific, for example when teachers highlight which aspects of physical examination skills are redundant in reaching a correct diagnosis, or when teachers point out areas where more work is needed. Those examples were deduced from research on preclinical teaching, but evidence suggests that the underlying principles would also be valid in the clinical learning environment (Johnson and Carpenter 1986).
One of the central themes emerging in our study was the motivational effect (simulated) patients have on students. Next to the resulting drive-to-practise this has on students, additional benefits can be bolstered with literature.
The transition to the clinical phase of undergraduate medical education has often been described as a stressful period (Prince et al. 2000a, b). Students experience stress due to increased working hours, uncertainty as to what is expected of them and self-perceived lack of knowledge (Godefrooij et al. 2010).
Boshuizen described how the resulting ‘shock of practice’ is associated with a temporary decrease in students’ ability to properly use biomedical knowledge in clinical reasoning (Boshuizen 1996). Students in our study described similar confusing experiences during their first patient contacts. However, in contrast to the literature on the “shock of practice” these first patient contacts took place before the clerkship phase. Our findings suggest that early patient contacts in physical examination skills teaching may improve students’ preparedness for clinical clerkships.
Recent research adds more evidence to the value of early patient contacts in overcoming the ‘shock of practice’ (Littlewood et al. 2005; Dornan et al. 2006a, b; Diemers et al. 2008; Yardley et al. 2010).
Learning physical examination skills in a longitudinal integrated skills programme could facilitate a student’s ability to perform those skills in the ‘real world’: the clinical workplace (Remmen et al. 2001). Our findings suggest that students feel well-prepared to do so and show that students use the skills learned when examining (simulated) patients. When learning in one situation (the preclinical skills programme) enhances performance in another situation (the clinical workplace), positive transfer has occurred. In contrast, when something learned in one situation (the preclinical skills programme) hinders learning or performance in a second situation (the clinical workplace) then negative transfer has occurred (Ormrod 2004).
Examples of both positive and negative transfer are discussed in our study. Factors contributing to positive transfer are the increasing complexity of skills and using (simulated) patients as authentic stimuli for learning. Conditions adding to negative transfer include socialisation processes, uncertainty about what is expected and conflicting techniques being used in practice.
The implications of our findings for medical education are fourfold.
Firstly, considerations for skills teachers. Our findings show that students in later years are capable of identifying areas that need work, especially when they are confronted with real patients. Students can create learning opportunities to address their shortcomings. Skills teachers could accommodate this process by using a student-centred approach, empowering students to take an active role in their learning (Martens 2009).
Secondly, considerations with regard to the learning environment. Educators should note that authentic learning environments stimulate students to learn physical examination skills. Bringing the real world into the classroom (using simulated or real patients) or taking the classroom to the real world (using outpatient departments to practise) can contribute greatly to the learning of skills. Experiential learning should be considered as an educational strategy for physical examination skills. (Kolb 1984).
Thirdly, considerations for course constructors and the curriculum at large. Building on the considerations for the learning environment, the statements from our students emphasised the importance of alignment of training sessions with (simulated) patient encounters. Curriculum alignment includes both the place of skills training in a course and the arrangement of skills training with communication sessions. Course constructors should safeguard the longitudinal integration of physical examination skills training in the curriculum.
Fourthly, considerations for assessment procedures. Examiners should be aware that students tailor their learning to what they are tested on, and adjust assessment accordingly. This strategy, also called ‘measurement-driven instruction’ can have powerful educational consequences. (Popham 1987; Vleuten 1996; Rust 2002). One such adjustment might be to revise current OSCE models, and abandon the checklist approach in favour of a more clinically oriented test which allows for variation in thinking. (Hodges et al. 1999) Such a test puts different requirements on examiners, and this has implications for staff training.
The limitations of this study are that we relied on students’ self-reports and thereby on their perceptions. Another limitation is that the confidence or lack of confidence of the students in their own abilities is not known and this would influence their perception of their learning of clinical skills. It should also be taken into account that the students who accepted the invitation to participate in our study might be the most enthusiastic ones. Even though saturation was reached in all focus groups and all groups discussed similar themes, it cannot be ruled out that NOT all themes were appropriately covered. Nevertheless, we think self-reporting is a strength too, since the impact of various influences on practising behaviour will be mediated through self-perceptions. Another strength is our sequential approach, combining self-reporting by questionnaire and focus groups. The high response rate on the questionnaire and the iterative procedure adopted for the qualitative analysis assure representativeness of our findings. The translation of our findings to other curriculum structures needs further clarification.
This exploratory study has indicated areas for further study: the introduction of early patient contacts, clinical reasoning in undergraduate years, innovative assessment of clinical skills. Further research should also explore if and how self-reported study behaviours are related to performance in clinical situations and achievement on formal assessments.
This article is distributed under the terms of the Creative Commons Attribution Noncommercial License which permits any noncommercial use, distribution, and reproduction in any medium, provided the original author(s) and source are credited.
- Bloom, B., Madaus, G., et al. (1987). Evaluation to improve learning. McGraw-Hill: New York.Google Scholar
- Boshuizen, H. (1996). The shock of practice: Effects on clinical reasoning.Google Scholar
- Colleges’, A. O. A. M. (1998). Report 1:Learning objectives for medical student education. Guidelines for medical schools. Washington: Medical School Objectives Project.Google Scholar
- Dent, J. A. (2001). Current trends and future implications in the developing role of clinical skills centres. Medical Teacher, 23(5), 483–489.Google Scholar
- Duvivier, R. J., Dalen J. van, Bartholomeus, P., Scherpbier, A. J. J. A., & Verwijnen, G. M. (2010a). Skills training. In H. Van Berkel, A. J. J. A. Scherpbier, H. H. Hillen, & C. P. M. Van der Vleuten (Eds.), Lessons from problem based learning. Oxford University Press.Google Scholar
- Duvivier, R. J., Rethans, J. J., & Dalen J. van. (2010b). Communication skills training. In H. Van Berkel, A. J. J. A. Scherpbier, H. H. Hillen, & C. P. M. Van der Vleuten (Eds.), Lessons from problem based learning. Oxford University Press.Google Scholar
- Feltovich, P., & Barrows, H. (1984). Issues of generality in medical problem solving. Tutorials in problem-based learning: a new direction in teaching the health professions. In H. Schmidt & M. De Volder (Eds.), Assen, Van Gorcum (pp. 128–142).Google Scholar
- Fischman, M., & Oxendine J. (1998). Motor skill learning for effective coaching and performance. Applied sport.Google Scholar
- General Medical, C. (1993). Tomorrow’s doctors: Recommendations on undergraduate medical education.Google Scholar
- Kolb, D. A. (1984). Experiential learning: Experience as the source of learning and development. NJ: Prentice-Hall Englewood Cliffs.Google Scholar
- Metz, J. C. M., Verbeek-Weel, A. M. M., et al. (2001). Blueprint 2001: Training of doctors in The Netherlands. Utrecht: NFU.Google Scholar
- Morgan, D. L. (1997). Focus groups as qualitative research, Sage pubns.Google Scholar
- Ormrod, J. (2004). Human learning. Upper Sadle River: Pearson Education.Google Scholar
- Popham, W. J. (1987). The merits of measurement-driven instruction. Phi Delta Kappan, 68(9), 679–682.Google Scholar
- Snyder, B. R. (1971). The hidden curriculum, Knopf New York.Google Scholar
- Strauss, A. L., & Corbin, J. (1990). Basics of qualitative research: Grounded theory procedures and techniques. CA: Sage Newbury Park.Google Scholar
- Van Dalen, J. (1997) Skillslab: Centre for training of skills. Maastricht University, Maastricht; (3rd Ed.).Google Scholar
- Wigton, R. S., Patil, K. D., et al. (1986). The effect of feedback in learning clinical diagnosis. Journal of Medical Education, 61(10), 816–822.Google Scholar