gms | German Medical Science

GMS Journal for Medical Education

Gesellschaft für Medizinische Ausbildung (GMA)

ISSN 2366-5017

Workplace-based assessment of communication skills: A pilot project addressing feasibility, acceptance and reliability

article Assessment

  • corresponding author Simone Weyers - Heinrich-Heine-Universität Düsseldorf, Universitätsklinikum Düsseldorf - Centre for Health and Society (CHS), Institut für Medizinische Soziologie, Düsseldorf, Germany
  • Iman Jemi - Johannes Gutenberg-Universität Mainz, Mainz, Germany
  • author André Karger - Heinrich-Heine-Universität Düsseldorf, Universitätsklinikum Düsseldorf, Klinisches Institut für Psychosomatische Medizin und Psychotherapie, Düsseldorf, Germany
  • author Bianca Raski - Heinrich-Heine-Universität Düsseldorf, Universitätsklinikum Düsseldorf, Klinisches Institut für Psychosomatische Medizin und Psychotherapie, Düsseldorf, Germany
  • author Thomas Rotthoff - Heinrich-Heine-Universität Düsseldorf, Universitätsklinikum Düsseldorf, Studiendekanat, Düsseldorf, Germany; Heinrich-Heine-Universität Düsseldorf, Universitätsklinikum Düsseldorf, Klinik für Endokrinologie und Diabetologie, Düsseldorf, Germany
  • author Michael Pentzek - Heinrich-Heine-Universität Düsseldorf, Universitätsklinikum Düsseldorf, Institut für Allgemeinmedizin, Düsseldorf, Germany
  • author Achim Mortsiefer - Heinrich-Heine-Universität Düsseldorf, Universitätsklinikum Düsseldorf, Institut für Allgemeinmedizin, Düsseldorf, Germany

GMS J Med Educ 2016;33(5):Doc70

doi: 10.3205/zma001069, urn:nbn:de:0183-zma0010699

This is the English version of the article.
The German version can be found at: http://www.egms.de/de/journals/zma/2016-33/zma001069.shtml

Received: January 22, 2016
Revised: June 24, 2016
Accepted: July 22, 2016
Published: November 15, 2016

© 2016 Weyers et al.
This is an Open Access article distributed under the terms of the Creative Commons Attribution 4.0 License. See license information at http://creativecommons.org/licenses/by/4.0/.


Abstract

Background: Imparting communication skills has been given great importance in medical curricula. In addition to standardized assessments, students should communicate with real patients in actual clinical situations during workplace-based assessments and receive structured feedback on their performance. The aim of this project was to pilot a formative testing method for workplace-based assessment. Our investigation centered in particular on whether or not physicians view the method as feasible and how high acceptance is among students. In addition, we assessed the reliability of the method.

Method: As part of the project, 16 students held two consultations each with chronically ill patients at the medical practice where they were completing GP training. These consultations were video-recorded. The trained mentoring physician rated the student’s performance and provided feedback immediately following the consultations using the Berlin Global Rating scale (BGR). Two impartial, trained raters also evaluated the videos using BGR. For qualitative and quantitative analysis, information on how physicians and students viewed feasibility and their levels of acceptance was collected in written form in a partially standardized manner. To test for reliability, the test-retest reliability was calculated for both of the overall evaluations given by each rater. The inter-rater reliability was determined for the three evaluations of each individual consultation.

Results: The formative assessment method was rated positively by both physicians and students. It is relatively easy to integrate into daily routines. Its significant value lies in the personal, structured and recurring feedback. The two overall scores for each patient consultation given by the two impartial raters correlate moderately. The degree of uniformity among the three raters in respect to the individual consultations is low.

Discussion: Within the scope of this pilot project, only a small sample of physicians and students could be surveyed to a limited extent. There are indications that the assessment can be improved by integrating more information on medical context and student self-assessments. Despite the current limitations regarding test criteria, it is clear that workplace-based assessment of communication skills in the clinical setting is a valuable addition to the communication curricula of medical schools.

Keywords: Communication, competency, workplace based assessment, formative assessment


1. Introduction

1.1 Background

In the international and German recommendations for designing curricula for medical degree programs, teaching communication skills is identified as one of the core objectives for successful medical education [1], [2], [3]. Programs have already been implemented at many German medical schools to teach and assess social and communicative competencies [4], [5], [6], [7]. At the University of Düsseldorf the interdisciplinary and longitudinal curriculum CoMeD has been put into place for teaching and testing communication skills. Assessment procedures have previously focused on basic and practical knowledge, as well as the application of what has been learned during an OSCE (Objective Structured Clinical Examination) using actors [8].

In addition, workplace-based learning at stations and GP training practices is seen as essential to the acquisition of communication skills and competencies [9], [10]. During clinical clerkships and practicums, physicians-to-be need to practice their skills communicating with patients in supervised surroundings. Students frequently feel they are left on their own and desire more active instruction and feedback from experienced physicians [11], [12]. In contrast to the often well-structured and supervised training sessions and seminars in the Skills Lab and the theoretical subjects [13], the learning process that takes place during a practicum is not adequately guided [14], [15]. Consequently, the question arises as to which formative assessment methods can be employed during the practicum to foster and structure the student learning process as it relates to communication.

The aim of a workplace-based assessment is to observe specific student performance during routine clinical work and to provide feedback to guide learning [16], [17], [18]. In medical education, structured observation and feedback in the workplace often concentrate on practical skills such as anamnesis or physical examination techniques. At present there are very few instruments available for differentiated workplace-based assessment of communicative competencies in complex doctor-patient interactions. Global ratings are more likely to be recommended for this purpose, as are methods for giving direct feedback to students [19], [20]. In this project the Berlin Global Rating (BGR) scale, which has been successfully used in summative simulation assessments [8], has been transferred to the practicum and drawn upon for the workplace-based, formative assessment of communication skills. To our knowledge, no prior experience with this has been gathered.

1.2 Project aims

The goal of this pilot project was to test the possibilities and limits of a workplace-based assessment of doctor-patient communication using BGR that evaluates student performance in the clinical setting and provides structured feedback. Answers to the following questions were sought:

1.
Is a formative workplace-based assessment using BGR feasible within the setting of a general medical practice?
2.
What is the level of acceptance among students for this assessment method?
3.
How reliable is BGR as a test instrument for the workplace-based assessment?

2. Methods

2.1 Project planning and recruitment

Representatives from the subjects covered by the communication curriculum participated in regular meetings concerning the design and development of the workplace-based assessment. Organization and realization took place over the course of a year through the efforts of a graduate assistant.

The resulting assessment was administered as part of the 14-day practicum in general practice during the fourth year of study for which students are prepared in case-based seminars on general medicine and training sessions on communication [7]. One of the main learning objectives is independently holding patient consultations. Students are supervised and guided by mentoring physicians who serve as medical student trainers in one-on-one teaching sessions. Students also have access to an online logbook [21].

First, only mentoring physicians with GBR training were recruited for the pilot project. All students who were doing their GP practicum at one of these practices were then invited to voluntarily participate in the project. Efforts were made to increase the response rate by holding informational meetings and approaching individuals personally. As a result, 16 students agreed to participate and, using a set of guidelines, were informed about the project’s aims and procedure. Student participation was rewarded with a 20 Euro gift certificate for books. The students were each assigned to one of nine medical practices. The participating nine mentoring physicians were then informed in detail a second time and given instructions based on the guidelines.

The mentoring physician selected and briefed the patients on-site at the practice. Patients were given written information explaining the study procedure before they gave their consent to participate and be recorded on video; participating patients also acknowledged the provisions for data privacy.

2.2 Assessment of communicative competencies using the Berlin Global Rating Scale

The BGR scale developed by Hodges and McIlroy [22] and validated in German by Scheffer [23] was used as a test instrument for the workplace-based assessment. The BGR scale measures the quality of communication with four items: empathy, structuring doctor-patient conversations, verbal and nonverbal expression. Using a five-point Likert scale, student performance regarding these aspects is rated, with clear definition of the extremes (e.g. for empathy: “The student does not respond or responds inappropriately to obvious verbal and nonverbal cues or the needs of the patient”=1 point versus “The student consistently responds with understanding to verbal and nonverbal cues and the needs of the patient or responds in an appropriate manner”=5 points). Students can earn a maximum of 20 points.

2.3 Data collection and analysis

Students were assigned the task of holding and recording two consultations at the practice with chronically ill patients (e.g. within the context of a disease management program or check-up). Recording devices were borrowed from the university; students were explicitly asked not to use cell phones to record. The mentoring physician observed and rated the student’s performance according to the four BGR items. Detailed feedback was then given at the end. As an aid, there was an information sheet with concrete recommendations for giving feedback on each of the items (see attachment 1 [Attach. 1]).

After completion of the practicum, the video recordings were given to the project team so that consultations could be blindly evaluated by two independent raters. This task was also assumed by other persons who were familiar with BGR. As a result each consultation was evaluated by three different raters.

Finally, the nine mentoring physicians were surveyed using a partially standardized questionnaire regarding the feasibility of the assessment method in the context of routine clinical practice. Information was collected on sociodemographic profile, amount of time (four-point scale from “very time intensive” to needing “very little time”), and the overall evaluation of the assessment method (expressed as a standardized academic grade). In open-ended responses physicians were also able to comment on the reactions of the patients, practicability in terms of daily workloads, added value, transferability, and suggestions for improvement.

Students were asked in writing about their experiences and overall evaluation of the assessment method (also expressed as a standardized academic grade). In open-ended responses they were able to share what they considered positive about the assessment method, and what they thought was problematic.

2.4 Analyses

The quantitative information contained in the physician surveys was scrutinized to ascertain the ease with which the assessment method could be incorporated into daily routines. The open-ended responses were paraphrased, parsed and condensed using summarizing content analysis. To determine the level of acceptance among students, the content of their open-ended responses was also evaluated by content analysis.

To verify the test-retest reliability, both overall evaluations of each rater were correlated using Pearson’s correlation coefficient (p<=0.05). The inter-rater reliability, meaning the correlation between the mentoring physician and the two impartial raters, who only viewed the video material, was calculated as an intra-class correlation (ICC) (type C, model with mixed two-way effects, 95% confidence interval; see attachment 2 [Attach. 2]). Values greater than 0.75 indicate good reliability [24]. The quantitative analyses were performed with IBM SPSS Statistics 22.


3. Results

3.1 Project participants and data material

Of the mentoring physicians, three were female and six male. The mean age was 49.6 years with the mean length of professional experience in general practice amounting to 17.7 years.

In addition, 12 female and four male students participated in the pilot and were assigned among nine GP training practices. At the time of the project the students were in their eighth semester of medical study. With only one exception, each student recorded two consultations resulting in a total of 31 consultations for analysis. The results are presented in terms of the three main issues.

3.2 Feasibility

Practicability on a daily basis: seven of nine physicians reported having no problems implementing the assessment method. Two physicians stated that it can be difficult to implement depending on the patients coming into the practice, for instance, if there is insufficient space available for the assessment, or if the mentoring physician does not have enough time due to emergencies. The amount of time needed was viewed as being neither great, nor particularly small (see Table 1 [Tab. 1]).

Added value: In response to the question about the added value of the assessment method, two respondents named the feedback, two respondents the self-observation and gain in self-knowledge, and one person identified the reward of empathetic communication techniques. One person stated that the OSCE to assess communicative competencies with actors standing in for patients was sufficient.

Transferability: Eight of nine physicians were of the opinion that the method can be transferred to other communicative contexts, for example, for anamnesis or setting therapy goals. One person viewed the transferability as challenging.

Improvements: In response to the question about what could be added to the evaluation criteria, one person suggested a comparison between the self-assessment and those by observers. Another person recommended giving more detailed information on the context for the consultation; a third person proposed that the medical expertise of the students also be evaluated.

On average, the mentoring physicians rated the assessment method with “good.”

3.3 Acceptance

On average the students also rated the assessment method positively (see Table 1 [Tab. 1]). The feedback provided by the students on the positive and negative aspects of the workplace-based assessment yielded the following points:

Positive aspects

Authenticity: Students frequently praised the fact that during the workplace-based assessment they had spoken with actual patients in an authentic setting under real conditions. The doctor-patient exchange was viewed as being more authentic than the interchange with actors posing as patients, as takes place in the OSCE.

Time: Several students wrote that they found it good that they were able to take time for the doctor-patient consultation.

Feedback: Most students stated that they had received direct, personal and differentiated feedback from their mentoring physician. This feedback was given in part directly after the consultation and later as part of the joint analysis of the recorded material. One participant wrote that this was the first time he had experienced this during the clinical phase of his medical education. Some students had the impression that the physicians addressed their strengths and weaknesses in a differentiated manner. It was frequently reported that BGR was helpful in structuring the feedback according to different aspects. Students also reported that repeating the assessment was helpful and presented an opportunity to apply the criticism from the first consultation. Progress was subjectively noticeable based on the second round of feedback. One student stated that he also found the feedback given by the patient useful. This was not intended to be part of this assessment, but does suggest a possible variation of the assessment method.

One student was of the opinion that all students, regardless of any participation in the research project, should have the opportunity to experience such a workplace-based assessment.

Two particularly indicative statements by students are presented as quotations in the following Figure 1 [Fig. 1].

Negative aspects

Setting: One student found that it was impossible to plan for the consultation’s context, and that time was needed to adjust. Some students stated that the situation felt artificial. They felt inhibited by the running camera or the presence of the mentoring physician. One student reported that this also affected the patients.

Assigned task: Two students stated that the assignment was unclear. One felt a need for a more clearly formulated objective for the consultation; another found the physician was not clear on how to handle the videos.

Role of the mentoring physician: Some students reported that the assessment depended heavily on the teaching skills and commitment of the physician. They also expressed the concern that the evaluation was affected by the personal acquaintance between physician and student making it less objective.

BGR: On the topic of BGR one student responded that the scale was so finely graduated that it led to discrepancies between evaluations.

3.4 Reliability

For the 31 patient consultations, the students earned a mean score of 15.55 out of a possible total of 20 points on the BGR scale. The two evaluations by the raters correlated with r=0.669, r=0.520, and r=0.653. The correlation between the mentoring physician and the two impartial raters was ICC=0.445. The correlation between pairs was ICC=0.619, ICC=0.301, and ICC=0.443 (see Table 2 [Tab. 2]).


4. Discussion

The aim of this project was to pilot a workplace-based assessment of communication skills at the University of Düsseldorf. The results are discussed here according to the three main issues.

4.1 Feasibility

The responses of the mentoring physicians indicate general concurrence. One very positive aspect is the near consensus among the physicians regarding the ease of integrating the evaluation and BGR feedback into the daily workloads shouldered in medical practices. No one indicated that the method demands “great amounts of time”, however, it is also true that no one stated that the method could be implemented with just a little effort here and there. A suitable point in time must be found, and it depends primarily on the presence of patients in the practice. Half of the nine physicians perceived added value in the new assessment method identifying, above all, the feedback. This corresponds not only with the philosophy underlying workplace-based assessments but also with statements made by the participating students. In this context, an interesting suggestion was made to include self-assessment on the evaluation sheet. Learners can give themselves feedback and compare their observations with the evaluations of other observers. Furthermore, the desire for more information on the reason for the consultation is understandable since the evaluation of the communication skills depends upon the context [25]. Suitable instruments could augment the formative assessment in valuable ways [26].

4.2 Acceptance

Students consistently rated the assessment positively. Some criticism was expressed, but no one rejected the assessment method. As described in the literature [18], there is the impression that students rarely receive feedback on their clinical competencies. They particularly valued the chance to interact with “real” patients in “real” situations. A significant aspect of this workplace-based assessment is that students are given feedback immediately following their performance. In terms of the evaluation, some students felt that the BGR scale defined the grading criteria well and provided a structured and performance-centered evaluation, allowing the student to know their level of proficiency. Others, in contrast, felt that the mentoring physicians’ evaluations were too subjective and depended on how well they were trained or how long mentor and student had known each other. This is not an issue if the BGR is used for formative assessment. The opportunity of the second assessment to apply the criticism from the first round and potentially make progress was viewed as particularly positive.

According to Crooks [27], workplace-based assessment should be integrated into the learning process, feedback should be given immediately, students should have multiple chances to take assessments, and grading criteria should be clearly defined and formulated. The BGR scale can certainly serve as a framework here to organize some of the basic aspects of doctor-patient communication: structuring the medical consultation, responding to the patient, verbal and nonverbal communication. It clearly defines the expectations for each grade category and provides guidance for the feedback.

4.3 Reliability

The results regarding the reliability of the assessment method are sobering. There is a moderate correlation among the overall evaluations of raters. This is to be expected with the short interval between the two consultations. In addition, there is a low level of correlation between the evaluations of the mentoring physicians and the impartial raters. From this it can be derived that it is urgently necessary to provide for physician training, joint standard setting, and clear definition of the context for consultations if this instrument is used for summative workplace-based assessment.

4.4 Limitations

Despite multiple recruiting measures, only a small number of students could be acquired for participation in the pilot project. Consequently, the sample size is too small to draw any firm conclusions from the statistical results. The participants were not chosen randomly, but rather because of their interest and are not representative of the full student body. In terms of the participating medical practices, a selection bias must be assumed in favor of physicians who are generally open to the formative assessment method presented here. The students here involve a group that is interested in the topic of doctor-patient communication in a way that exceeds the average level of interest exhibited by medical students. However, a comparison with the OSCE scores shows that the student participants do not distinguish themselves significantly from the cohort average (results not shown here).

No validated survey instrument was available for the physician or student evaluations. The evaluation based on open-ended written responses is limited to a descriptive analysis of content; to evaluate more deeply held attitudes or motives, interviews or focus groups using qualitative research methods would be necessary.

4.5 Conclusion

Despite various methodological limitations, there are indications for continued development of a workplace-based assessment of communicative competencies in the clinical setting. It has become clear that this formative assessment is a valuable addition to the communication curricula of medical schools. As part of the model study program at the University of Düsseldorf, the communication training program using actors to stand in as patients is now supplemented by a training program (with online logbook) using real patients that takes place during the clerkships of the fourth and fifth years of medical study and the 14-day practicum at a GP training practice [21]. The assessment presented here should continue to be tried out as a formative assessment method and developed further in response to evaluation results.

The results also demonstrate that use as a summative assessment requires a considerably greater amount of training and coordination than could be easily mastered by many medical practices. Together with other assessments, the concept discussed here involving observation and feedback yields a longitudinal portfolio of communicative competencies. It also poses a chance for and a challenge to medical schools to implement an improved culture of feedback in routine clinical practice.


Acknowledgements

We wish to extend our gratitude to the QVM Kommission der Medizinischen Fakultät Düsseldorf for supporting this educational project. A special thanks goes to the students and general practitioners who participated in this project.


Competing interests

The authors declare that they have no competing interests.


Informed Consent

Approval was given for this project by the Ethics Commission of the University of Düsseldorf’s Medical School (no. 7194).


References

1.
Bachmann C, Barzel A, Dunkelberg S, Schrom K, Erhardt M, van den Bussche, H. Fachübergreifendes Kommunikationstraining mit Simulationspatienten: ein Pilotprojekt ins Curriculum. GMS Z Med Ausbild. 2008;25(1):Doc58. Zugänglich unter/available from: http://www.egms.de/static/en/journals/zma/2008-25/zma000542.shtml External link
2.
Frank JR. The CanMEDS 2005 physician competency framework. Better standards - better physicians - better care. Ottawa: Royal College of Physicians and Surgeons of Canada; 2005. Zugänglich unter/available from: http://www.royalcollege.ca/portal/page/portal/rc/common/documents/canmeds/resources/publications/framework_full_e.pdf External link
3.
Kiessling C, Dieterich A, Fabry G, Hölzer H, Langewitz W, Mühlinghaus I, Pruskil S, Scheffer S, Schubert S. Basler Consensus Statement "Kommunikative und soziale Kompetenzen im Medizinstudium": Ein Positionspapier des GMA-Ausschusses Kommunikative und soziale Kompetenzen. GMS Z Med Ausbild. 2008;25(2):Doc83. Zugänglich unter/available from: http://www.egms.de/static/de/journals/zma/2008-25/zma000567.shtml External link
4.
Jünger J, Lauber H, Schultz JH, Herzog W. 5 Jahre Kommunikations- und Interaktionstraining von Medizinstudenten (Medi-KIT) von der Insellösung zum longitudinalen Curriculum. GMS Z Med Ausbild. 2007;24(1):Doc68. Zugänglich unter/available from: http://www.egms.de/static/de/journals/zma/2007-24/zma000362.shtml External link
5.
Roch K, Trubrich A, Haidinger G, Mitterauer L, Frischenschlager O. Unterricht in ärztlicher Gesprächsführung - eine vergleichende Erhebung in Deutschland, Österreich und der Schweiz. GMS Z Med Ausbild. 2010;27(3):Doc48. DOI: 10.3205/zma000685 External link
6.
Langewitz W. Evaluation eines zweijährigen Curriculums in Psychosozialer und Psychosomatischer Medizin–Umgang mit Emotionen und patientenzentrierter Gesprächsführung. Psychother Psychos Med Psychol. 2010;60(11):451-456. DOI: 10.1055/s-0030-1251980 External link
7.
Mortsiefer A, Rotthoff T, Schmelzer R, Immecke J, Ortmanns B, in der Schmitten, J, Altiner A, Karger A. Implementation of the interdisciplinary curriculum Teaching and Assessing Communicative Competence in the fourth academic year of medical studies (CoMeD). GMS Z Med Ausbild. 2012;29(1):Doc06. DOI: 10.3205/zma000776 External link
8.
Mortsiefer A, Immecke J, Rotthoff T, Karger A, Schmelzer R, Raski B, Schmitten JL, Altiner A, Pentzek M. Summative assessment of undergraduates' communication competence in challenging doctor-patient encounters. Evaluation of the Düsseldorf CoMeD-OSCE. Pat Educ Couns. 2014;95(3):348–355. DOI: 10.1016/j.pec.2014.02.009 External link
9.
Jolly BC, Macdonald MM. Education for practice: the role of practical experience in undergraduate and general clinical training. Med Educ. 1989;23(2):189–195. DOI: 10.1111/j.1365-2923.1989.tb00885.x External link
10.
Evans DE, Wood DF, Roberts CM. The effect of an extended hospital induction on perceived confidence and assessed clinical skills of newly qualified pre-registration house officers. Med Educ. 2004;38(9):998–1001. DOI: 10.1111/j.1365-2929.2004.01908.x External link
11.
Daelmans R, Hoogenboom A, Donker A, Scherpbier A, Stehouwer C, Van Der Vleuten, C. Effectiveness of clinical rotations as a learning environment for achieving competences. Med Teach. 2004;26(4):305–312. DOI: 10.1080/01421590410001683195 External link
12.
Köhl-Hackert N, Krautter M, Andreesen S, Hoffmann K, Herzog W, Jünger J, Nikendei C. Workplace learning: an analysis of students' expectations of learning on the ward in the Department of Internal Medicine. GMS Z Med Ausbild. 2014;31(4):Doc43. DOI: 10.3205/zma000935 External link
13.
Wimmers PF, Schmidt HG, Splinter TA. Influence of clerkship experiences on clinical competence. Med Educ. 2006;40(5):450–458. DOI: 10.1111/j.1365-2929.2006.02447.x External link
14.
Remmen R, Denekens J, Scherpbier A, Hermann I, van der Vleuten C, van Royen P, Bossaert L. An evaluation study of the didactic quality of clerkships. Med Educ. 2000;34(6):460–464. DOI: 10.1046/j.1365-2923.2000.00570.x External link
15.
Schrauth M, Weyrich P, Kraus B, Jünger J, Zipfel S, Nikendei C. Lernen am späteren Arbeitsplatz: Eine Analyse studentischer Erwartungen und Erfahrungen im Praktischen Jahr. Z Evid Fortbild Qual Gesundheitswes. 2009;103(3):169–174. DOI: 10.1016/j.zefq.2008.05.005 External link
16.
Van Der Vleuten AJ, Scherpbier DH, Dolmans LW, Schuwirth GM, Verwijnen HA, Wolfhagen CP. Clerkship assessment assessed. Med Teach. 2000;22(6):592–600. DOI: 10.1080/01421590050175587 External link
17.
Norcini JJ. The Mini-CEX: A Method for Assessing Clinical Skills. Ann Int Med. 2003;138(6):476. DOI: 10.7326/0003-4819-138-6-200303180-00012 External link
18.
Norcini J, Burch V. AMEE Guide Workplace-based assessment as an educational tool: AMEE Guide No. 31. Med Tech. 2007;29(9):855–871. DOI: 10.1080/01421590701775453 External link
19.
Langewitz W. Zur Erlernbarkeit der Arzt-Patienten-Kommunikation in der Medizinischen Ausbildung. Bundesgesundheitsbl. 2012;55(9):1176–1182. DOI: 10.1007/s00103-012-1533-0 External link
20.
Kurtz SM, Silverman J, Draper J. Teaching and learning communication skills in medicine. 2nd ed. Oxford, San Francisco: Radcliffe Pub.; 2005.
21.
Gummersbach E, Abholz HH, Altiner A, Stock K, Schürer-Maly C. Online-Logbuch im Hausarztpraktikum – Bericht über die erste Nutzung in Deutschland. Z Allgemeinmed. 2010;2:73–79.
22.
Hodges B, McIlroy JH. Analytic global OSCE ratings are sensitive to level of training. Med Educ. 2003;37(11):1012–1016. DOI: 10.1046/j.1365-2923.2003.01674.x External link
23.
Scheffer S. Validierung des Berlin Global Rating" (BGR) - ein Instrument zur Prüfung kommunikativer Kompetenzen Medizinstudierender im Rahmen klinisch-praktischer Prüfungen (OSCE): Dissertation. Berlin: Charité; 2009.
24.
Portney LG, Watkins MP. Foundations of clinical research. Applications to practice. New Jersey: Prentice Hall Inc.; 2000.
25.
Baig LA, Violato C, Crutcher RA. Assessing clinical communication skills in physicians: are the skills context specific or generalizable. BMC Med Educ. 2009;9:22. DOI: 10.1186/1472-6920-9-22 External link
26.
Essers G, Kramer A, Andriesse B, van Weel C, Van Der Vleuten C, van Dulmen S. Context factors in general practitioner - patient encounters and their impact on assessing communication skills - an exploratory study. BMC Fam Pract. 2013;14:65. DOI: 10.1186/1471-2296-14-65 External link
27.
Crooks J. The impact of classroom evaluation practices on students. Rev Educ Res. 1988;58(4):438–481. DOI: 10.3102/00346543058004438 External link