General principles for the assessment of communication skills final by T5vwZJ


									           General principles for the assessment of communication skills
Claudia Kiessling (Germany), Geurt Essers (The Netherlands), Tor Anvik (Norway), Katarzyna Jankowska
(Poland), Rute Meneses (Portugal), Zoi Tsimtsiou (Greece), Marcy Rosenbaum (Iowa, U.S.A.), Jonathan
Silverman (UK), for the tEACH Assessment Subgroup

Aim of the paper
This paper provides a summary of the current literature on assessment of communication skills within
health care education. The aim is to help teachers who are thinking about a way to assess their students
and the results of their communication skills courses to grasp the most important aspects of a fair and
state-of-the-art assessment system.

Assessment drives learning
Assessment helps you as a teacher to decide whether your students are fit for later professional life and
have acquired sufficient skills to be able meet the demands of clinical reality (Hulsman, 2011). For learners,
assessment helps to identify their learning needs. Content and format of your assessment tools will influ-
ence students’ learning behaviour, implicitly and explicitly. Students tend to study what they expect to be
tested. If we do not assess communication skills, students could assume that communication skills are not
important: assessment legitimises the subject. On the other hand, the way we assess communication skills
will send out a strong signal to the students about what we as professionals think good communication
skills are in clinical practice. Therefore, it is of utmost importance to be aware of these considerations.

Establish a coherent entity of teaching, training, and assessment
The content and form of assessment need to be aligned with the purpose and desired outcomes of the
teaching programme (Norcini et al., 2011). The core message is: assess what you teach and train. A mature
communication skills curriculum is based on theoretical principles and scientific evidence of effective com-
munication. Educational objectives1 should be derived from these principles. Teaching methods should be
selected to help students to meet the educational objectives. The curriculum should encourage cumulative
learning and self-reflection and prepare students for their later professional life. The system of assessment
should be based on the same theoretical principles and educational objectives (Duffy et al., 2002). The
methods of assessment should mirror the methods of instruction. Students should be assessed according to
the educational objectives and their level of competence. This can be achieved by a careful blueprinting2
where the content of the assessment is planned to be congruent with the conceptual frameworks and edu-
cational objectives of the curriculum. Blueprinting enables sampling from across the curriculum and guar-
antees that assessment is seen as an integral part of the communication skills curriculum as a whole.

Assess communication skills within the clinical context
Like other clinical competencies, communication skills seem to be content and context bound (Wass et al.,
2001; Schuwirth et al., 2010). Learning should take place as close as possible to the later professional reality
(Schuwirth & Van der Vleuten, 2004). This means that for health care professionals, communication skills
should be taught and learned within the clinical context, either in clinical practice or in clinically relevant
simulations (Silverman, 2009; Kurtz et al., 2003). Even in the early years of medical training, when students
do not have a lot of patient contact, it is helpful to use clinical examples and simulations to teach basic
communication skills. For students, it does not make intuitive sense to isolate communication skills from
later professional practice. Similarly, the assessment of communication skills should also be made within a
meaningful context e.g. combined with and integrated into the assessment of other clinical skills, clinical
knowledge and clinical reasoning. For instance, assessing the communication process skills of gathering
information could be undertaken in an integrated fashion with the content of history taking and the per-

 An educational objective is a statement that provides a clear description of the outcome or performance you expect from learners
as a result of a lesson or course. An objective should be SMART: specific, measurable, attainable, relevant, and timely.
  A test blueprint can be described as a template of specifications that shows the structure of the test. It can include the test con-
tent, knowledge domains, or levels of expertise. In medicine, it can also include age and gender of patients, setting of care etc.
ceptual processes of clinical reasoning so that learners are assessed approaching a clinically realistic task in
which the communication skills are necessary to achieve a clinical end.

Establish a system of assessment with formative and summative assessment
Assessment can have a formative purpose which means detecting strengths and weaknesses of students
and guiding future learning, providing reassurance and providing motivation (“forming” students’ learning).
Formative assessment is often informal, integrated into a teaching course, timely, and low stakes. Summa-
tive assessment usually is intended to achieve another purpose. The focus is on making a decision about
competence and fitness for clinical practice (“summing up” students’ learning). Summative assessment acts
as a barrier to protect patients and public by identifying incompetent health care providers (Epstein, 2007).
Both formative and summative aspects of assessment are important for learners becoming a professional
and should be taken into consideration when planning a comprehensive assessment system.

Provide feedback that enhances learning
There is a growing awareness of the relation between assessment, feedback and continuous learning (Nor-
cini, 2011). Assessment, especially formative assessment, provides helpful feedback to learners and teach-
ers. Assessment needs to provide as much constructive feedback as possible. Moreover, although summa-
tive assessment concentrates on pass – fail decisions, it is still useful to provide feedback to students. Adult
learners seek feedback from external sources and provide helpful feedback for other learners. If you man-
age to strengthen the self-regulation of your learners, feedback and self-assessment will accompany and
complement the formal assessment methods set by the institution. Constructive feedback from teachers
and peers will help learners understand and accept the criteria that teachers apply in the assessment of
their communication skills. A broader understanding of assessment would therefore include feedback, self-
assessment and peer-assessment into an assessment system (Dochy et al., 1999; Gielen, 2007).

Establish a multi-source and longitudinal assessment programme
All assessment methods have strengths and weaknesses (Schuwirth & Van der Vleuten, 2003). “Each single
assessment is a biopsy, and a series of biopsies will provide a more complete, more accurate picture” (Van
der Vleuten et al. 2010). Therefore, it is important to try to establish the use of different assessment for-
mats, multiple observations, and independent times of measurement in different settings. Epstein provides
a helpful overview about different commonly used methods of assessment (Epstein, 2007). Depending on
available resources, it is helpful to try to assess students with different methods over a longer period of
time with a broad sampling. Among others, Rider and colleagues (Rider et al., 2006) have published an in-
teresting example of a model for communication skills assessment across the undergraduate curriculum in
medical education.

Use methods of assessment according to your goals and levels of competence
There are some general criteria to consider when you implement an assessment strategy (Schuwirth & van
der Vleuten, 2004):
     Reliability (the degree to which the measurement is reproducible and consistent)
     Validity (whether the assessment measures what it claims to measure)
     Educational impact (impact on future learning and practice)
     Credibility (to students and faculty)
     Feasibility and costs (to individual trainee, the institution, the society)
Assessment is always a compromise between what is desirable and what is achievable, involving a trade-off
between the weightings attached to each of the five components listed above. If the assessment is summa-
tive, the tools need to meet higher requirements, especially of reliability. Reproducibility of decisions based
upon the test results is essential when we “sum up” students’ learning and draw decisions about their pro-
fessional advancement. The following two statements highlight the balance between assessment purpose
and assessment criteria in a concise way: “What is important is the relationship between the accuracy of
the scoring and the level at which decisions are made” (Schuwirth & Van der Vleuten, 2004 p.804). “The
stakes of the evaluation determine the requirements for the psychometric rigor of the measurement”
(Duffy et al., 2004 p.500).
 Use an established standard setting procedure3
In summative assessments, the cut-off point for pass and fail needs to be defined (Friedman Ben David
2000). A transparent and robust standard setting guarantees a fair judgment of an examinee’s perform-
ance. There are different methods on how to set a cut-off score, with a large body of relevant literature.
The choice of method is dependent on the assessment method and tool you are going to use, the resources
you have and on the consequences of misclassifying examinees as having passed or failed (Wass et al.,
2001; Friedman Ben David, 2000).

Categorising assessment tools according to the stages of clinical competence
A selection of different assessment tools has been provided by the tEACH website. All of them are applica-
ble to the assessment of communication skills within an educational programme. Most of the tools in the
tEACH website are directed at assessing communication in an entire patient–provider consultation. Some
tools have been developed for specific professions; some are aimed at assessing specific topics (e.g. pa-
tient’s perspective and health beliefs, reasoning and decision-making). (Hyperlink)
There are different ways of categorising assessment strategies and instruments. A simple way is to divide
them according to the level of competence you want to assess: basic and applied knowledge, performance
in standardised quasi-real situations, and performance or action in the work-place (Miller, 1990).

 Knowledge and written assessment: to assess basic and applied knowledge
Multiple methods of written assessments have been developed over the last decades. Questions can be
open-ended or multiple choice (which means, selecting the best possible answer out of the choices from a
list); context-rich (e.g. with a clinical vignette) and context-poor; media-enhanced (e.g. video, audio); pa-
per-and-pencil-based or computer-based. Written assessment can be administered in a relatively short
time and with high standardisation. With clear constructing and grading guidelines, different types of writ-
ten assessments tend to reach a high reliability per hour of testing. Written assessment allows you to test
different types of cognitive abilities: knowledge and understanding of facts, processes, and concepts. It
does not allow you to test skills although it may be able to predict the performance of skills (Van Dalen et
al., 2002). Nevertheless, there seems to be a place for written assessment in the field of communication,
especially in the beginning of training and especially following observation of prepared videos (Humphris &
Kaney, 2000; Hulsman et al., 2004). Newer written assessment methods like the script concordance test
(Charlin & Van der Vleuten, 2004), key feature assessment (Page et al., 1995), situational judgement test
(Strahan et al., 2005), and reflective portfolio assessment (Rees & Sheard, 2004) may well bring new stimuli
to the field of assessing communication skills.

 Performance in clinical simulations: performance in standardised quasi-real situations
The best known assessment method using clinical simulations is the Objective Structured Clinical Examina-
tion (OSCE) (Hodges et al., 2002; Schuwirth & Van der Vleuten, 2003; Newble, 2004). Its development and
implementation into medical education has become a world-wide success in the last decades. Students
complete a number of stations with standardized patients who are trained to consistently repeat typical
clinical situations. Students have to perform medical interviews, physical examinations, clinical procedures
etc. Raters - either standardized patients or clinicians - use either a checklist or global ratings to evaluate
the students’ performance. In order to achieve acceptable reliability, it is important to consider the number
of cases, stations, examiners and patients. For example, with an overall testing time of 3 to 4 hours and a
minimum of 10 stations, OSCEs have shown to be reliable enough for high stakes examinations, e.g. a fed-
eral licensing examination (Van der Vleuten & Schuwirth 2005; Wass et al., 2001). Although OSCEs have
been proven to be a worthwhile method for the assessment of clinical skills, there is still a need for further
evidence. Additional research is needed around scoring and standard setting (Norcini et al., 2011). There is
an on-going discussion about the use of checklists versus global ratings for the assessment of communica-
  Standard setting is a method used in assessment to define levels of achievements or proficiency and the cut-off scores corre-
sponding to these levels. A cut-off score classifies students whose score are below into one level and students whose scores are
above into the next higher level (e.g. level “fail” and level “pass”). A variety of standard setting procedures are available to set a
reasonable cut-off score for assessment purposes. A comprehensive introduction about different methods is given by Friedman Ben
David (2001) and PMETB (2007).
tion skills. Important aspects for consideration include among others reliability (e.g. between different rat-
ers, between different OSCE stations), validity (e.g. reducing complex competencies to easy measurable but
trivial patterns of behaviours), and feasibility (e.g. amount of time for rater trainings). At the moment,
there seems to be a slightly positive tendency for global ratings in concerns of validity (Regehr et al., 1998;
Hodges et al., 1999; Van der Vleuten et al., 2010). However, the scoring method should match your educa-
tional objectives and teaching models for maximal educational impact (Duffy et al., 2004).
Simulation, the imitation of a real-world task or process, has become increasingly important for the as-
sessment of such things as clinical reasoning and teamwork (Epstein, 2007). Assessment situations based
on simulation can include simulated patients, computer simulations, mannequins, and high fidelity simula-
tors. One of the most important tasks of simulation-based education is providing feedback to learners (Is-
senberg et al., 2005; Bradley, 2006). Therefore, these new developments also offer innovative research and
assessment opportunities for training and assessing communication skills in health care settings.

    Direct observation of performance or action in every-day clinical situations: performance
     or action in the work-place
Unstructured and structured observations from supervisors, peers, co-workers, or other health professions
are commonly used to evaluate learners’ performance with patients (Kogan et al., 2009). Many of these
strategies are called “workplace-based assessments” (Govaerts, 2011). For example, the mini-clinical-
evaluation exercise (mini-CEX) has become a helpful instrument to support structured feedback for post-
graduate training in medicine (Norcini & Burch, 2007; Kogan et al., 2009). It turns the unstructured observa-
tion into a structured formative assessment, which provides feedback to learners and teachers to enhance
learning and teaching. These kinds of workplace-based observations can be combined with written exer-
cises and oral presentations, and documentation of these may well be collected in a learning portfolio
(Friedman Ben David et al., 2001). The formative character of providing multisource feedback to guide fu-
ture learning is certainly one of the most valuable advantages of all kinds of direct observations. Neverthe-
less, the need for effective supervision and follow-up of direct observation and feedback should not be
Patients’ ratings may also be valuable for formative assessment. At the moment, more evidence is needed
on how to establish a reliable and valid assessment system that includes patients as observers and raters of
clinical performance (Epstein, 2007; Duffy et al., 2004). Patient-related outcomes can serve as an indicator
of the quality of doctor/provider-patient communication, but to include patient ratings into an assessment
programme is not an easy task. For example, patients and doctors do not necessarily agree about the qual-
ity of the interaction, and they may have contradictory goals. Patient satisfaction may not be an adequate
indicator for high quality medical care. A satisfied patient can still have an incompetent doctor. “Defining
functions and endpoints and connecting those to theory will improve our understanding of the effective-
ness of communicative behaviour. Eventually, this will be of uttermost importance to teaching and clinical
practice” (De Haes & Bensing, 2009). Nevertheless, some validated patient rating instruments are promis-
ing for medical education (Reinders, 2009). Pre-planned but unannounced standardised patients who pre-
sent incognito in real clinical settings (Siminoff et al. 2011) may also be particularly valuable for postgradu-
ate or higher level training, and can be used for the assessment of diagnostic reasoning, treatment deci-
sion, and communication skills.

 Ratings based on videotaped encounters
Another approach to assess the communication and interaction with patients and standardised patients is
the rating of videotaped real-life or simulated encounters. It allows students to review and analyse their
own behaviour, third persons (including peers) can be invited to assess the performance, and raters can
explain or justify their scoring in a documented way. Similar to simulation-based assessments, the quality of
real-life encounters can be measured by rating scales and checklists. In addition, measurement is also pos-
sible with interaction analysis coding systems (Boon & Stewart, 1998). This kind of performance assessment
is particularly suited to postgraduate training. If an adequate case mix is made, a sample of eight consulta-
tions can suffice for a valid and reliable assessment (Ram, 1999).
 Attitudes, values, and “multidimensional constructs”
The most difficult challenge in education is assessing the students’ attitudes, values, and “multidimensional
constructs”. Multidimensional constructs are, for example, empathy or patient-centeredness when knowl-
edge, skills, and attitudes all work together to form a whole. Stepien and Baernstein (Stepien & Baernstein,
2006) describe four dimensions of clinical empathy: the emotive dimension (the ability to imagine patients’
emotions and perspectives), the moral dimension (the physician’s internal motivation to empathize), the
cognitive dimension (the intellectual ability to identify and understand patients’ emotions and perspec-
tives), and the behavioural dimension (the ability to convey understanding of those emotions and perspec-
tives back to the patient) (see also Mercer & Reynolds, 2002). In comparison to technical skills, like suturing
or venepuncture, multidimensional constructs are much more difficult to measure. Hemmerdinger and
colleagues (Hemmerding et al., 2007) recommend classifying instruments measuring such things as empa-
thy from three different dimensions: self-ratings (first person assessment), patient-ratings (second person
assessment), and observer ratings (third person assessment). First and second person instruments are of-
ten used in the field of education research. They may well have a potential to enhance self-reflection in
communication skills trainings and further formative assessment. Third person instruments typically focus
on the behavioural dimension for example demonstration of verbal or non-verbal clinical empathy. They
are commonly used in OSCEs to assess behaviourally measurable skills rather than intention (e.g. Hodges et
al., 2002).

Search for evidence and look for advice
The last decades have brought a large array of assessment tools that can be used for assessing communica-
tion. Sometimes when you implement a new assessment system, it is difficult to anticipate the effects on
students, other stakeholders, and the institution. There are always unintended effects of testing. Increas-
ingly, “there is little agreement on ideal assessment tools” (Schirmer et al., 2005 p.185). And perhaps the
ideal assessment tool does not exist. They all have their advantages and disadvantages and they need to
match the educational context. Therefore, if you are considering communication skills assessment, apart
from reading the literature, it is worthwhile talking to people who are experienced and experts in the field.
Implementing assessment in the curriculum will need careful change management to guarantee sustain-
ability and continuous improvement. One of the main goals will be to gain support from students, col-
leagues, superiors, and teachers for your plans.

We sincerely thank Dr. Götz Fabry (Freiburg, Germany), Anja Görlitz (Munich, Germany), Dr. Robert L.
Hulsman (Amsterdam, The Netherlands), and Tanja Pander (Munich, Germany) for critically reviewing the

   Association of American Medical Colleges (AAMC). Report III Contemporary Issues in Medicine: Com-
    munication in Medicine. Medical School Objectives Project. October 1999.
    20Medicine%20Report%20III%20.pdf (accessed 12.06.2012)
   Bradley P. The history of simulation in medical education and possible future directions. Med Educ
    2006; 40: 254–262.
   Boon H, Stewart M. Patient-physician communication assessment instruments: 1986 to 1996 in review.
    Pat Educ Couns 1998; 35: 161-76.
   Charlin B, Van der Vleuten CPM. Standardized assessment of reasoning in contexts of uncertainty. The
    script concordance test. Evaluation and the Health Professions 2004; 27:304-319.
   De Haes H, Bensing J. Endpoints in medical communication research, proposing a framework of func-
    tions and outcomes. Patient Education and Counseling 2009; 74:287–294.
   Dochy F, Segers M, Sluijsmans D. The use of self-, peer and co-assessment in higher education: A re-
    view. Studies in Higher Education 1999; 24:3, 331-350.
   Duffy FD, Gordon GH, Whelan G, Cole-Kelly K, Frankel R. Assessing Competence in Communication and
    Interpersonal Skills: The Kalamazoo II Report. Acad Med 2004; 79:495–507.
   Epstein RM. Assessment in medical education. N Engl J Med 2007; 356:387-396.
   Friedman Ben David MH. AMEE Guide No. 18: Standard setting in student assessment. Med Teach
    2000; 22(2): 120-130.
   Friedman Ben David MH, Davis MH, Harden RM, Howie PW, Ker J, Pippard MJ. AMEE Medical Education
    Guide No. 24: Portfolios as a method of student assessment. Med Teach 2001; 23 (6): 535-551.
   Gielen S. Peer assessment as a tool for learning (dutch). Dissertation. Leuven 2007. (accessed
   Govaerts M. Climbing the pyramid. Towards Understanding Performance Assessment. Mheer 2001. (accessed 06.08.2012).
   Hemmerdinger JM, Stoddart SD, Lilford RJ. A systematic review of tests of empathy in medicine, BMC
    Med Educ 2007; 7:24.
   Hodges B, Regehr G, McNaughton N, Tiberius R, Hanson M. OSCE checklists do not capture increasing
    levels of expertise, Acad Med 1999; 74:1129-34.
   Hodges B, Hanson M, McNaughton N, Regehr G. Creating, monitoring, and improving a psychiatry
    OSCE: a guide for faculty, Acad Psychiatry 2002; 26:134-61.
   Hulsman RL, Mollema ED, Hoos AM, De Haes JCJM, Donnison-Speijer JD. Assessment of medical com-
    munication skills by computer: assessment method and student experiences. Med Educ 2004; 38:813–
   Hulsman RL. The art of assessment of medical communication skills. Pat Educ Couns 2011;83:143-1444.
   Humphris GM, Kaney S. The Objective Structured Video Exam for assessment of communication skills.
    Med Educ 2000;34:939-945.
   Issenberg SB, McGaghie WC, Petrusa ER, Gordon DL, Scalesi RJ. Features and uses of high-fidelity medi-
    cal simulations that lead to effective learning. Med Teach 2005; 27(1):10–28.
   Kogan, JR, Holmboe ES, Hauer KE. Tools for direct observation and assessment of clinical skills of medi-
    cal trainees: a systematic review, JAMA 2009; 302:1316-26.
   Kurtz S, Silverman J, Benson J, Draper J. Marrying Content and Process in Clinical Method Teaching:
    Enhancing the Calgary–Cambridge Guides. Acad Med 2003;78:802–809.
   Mercer SW, Reynolds WJ. Empathy and quality of care. Br J Gen Pract. 2002;52(Suppl):S9-13.
   Miller GE. The assessment of clinical skills/competence/performance. Acad Med 1990; 65: S63-S67.
   Newble D. Techniques for measuring clinical competence: objective structured clinical examinations,
    Med Educ 2004; 38:199-203.
   Norcini J, Burch V. Workplace-based assessment as an educational tool: AMEE Guide No. 31, Med
    Teach 2007; 29: 855-71.
   Norcini JJ, Anderson B, Bollela V, Burch V, Costa MJ, Duvivier R, et al. Criteria for good assessment: con-
    sensus statement and recommendations from the Ottawa 2010 Conference. Med Teach 2011;33:206-
   Page G, Bordage G, Allen T. Developing key-feature problems and examinations to assess clinical deci-
    sion-making skills. Acad Med 1995;70:194-201.
   Postgraduate Medical Education and Training Board (PMETB). Guidance. Developing and maintaining
    an assessment system - a PMETB guide to good practice. 2007. http://www.gmc- (accessed 06.08.2012)
   Ram P, Grol R, Rethans JJ, Schouten B, van der Vleuten C, Kester A. Assessment of general practitioner
    by video observation of communicative and medical performance in daily practice: issues of validity, re-
    liability and feasibility. Med Educ 1999; 33: 447-454.
   Rees CE, Sheard CE. The reliability of assessment criteria for undergraduate medical students’ commu-
    nication skills portfolios: the Nottingham experience. Med Educ 2004; 38: 138–144.
   Regehr G, MacRae H, Reznick RK, Szalay D. Comparing the psychometric properties of checklists and
    global rating scales for assessing performance on an OSCE-format examination, Acad Med 1998;
   Reinders ME, Blankenstein AH, Knol DL, de Vet, HCW, van Marwijk HWJ. Validity aspects of the Patient
    Feedback questionnaire on Consultation skills (PFC), a promising learning instrument in medical educa-
    tion. Pat Educ Couns 2009;76(2):2002-6.
   Rider EA, Hinrichs MM, Lown BA. 'A model for communication skills assessment across the undergradu-
    ate curriculum', Med Teach 2006, 28: 127-134.
   Schirmer JM, Mauksch L, Lang F, Marvel MK, Zoppi K, Epstein RM, Brock D, Pryzbylski M. Assessing
    Communication Competence: A Review of Current Tools. Family Med 2005; 184-192.
   Schuwirth LWT, Van der Vleuten CPM. ABC of learning and teaching in medicine. Written assessment.
    BMJ 2003; 326:643-645.
   Schuwirth LWT, Van der Vleuten CPM. Changing education, changing assessment, changing research?
    Med Educ 2004; 38:805-812.
   Silverman J. Teaching clinical communication: A mainstream activity or just a minority sport? Pat Educ
    Couns 2009; 76: 361–367.
   Siminoff LA, Rogers HL, Wallera AC, Haywood SH, Epstein RM, Borrell Carrio F, Gliva-McConveye G,
    Longof DR. The Advantages and Challenges of Unannounced Standardized Patients Methodology to As-
    sess Healthcare Communication. Patient Educ Couns. 2011; 82: 318–324.
   Strahan J, Fogarty GJ, Machin MA. Predicting Performance on a Situational Judgement Test: The Role of
    Communication Skills, Listening Skills, and Expertise. Proceedings of the 40 Annual Conference of the
    Australian Psychological Society 2005. 323-327.
   Stepien KA, Baernstein A. Educating for Empathy. A Review. J Gen Intern Med 2006; 21:524–530.
   van Dalen J, Kerkhofs E, Verwijnen GM, van Knippenberg-van den Berg BW, van den Hout HA,
    Scherpbier AJ and van der Vleuten CP. Predicting communication skills with a paper-and-pencil test,
    Med Educ, 36 (2002) 148-53.
   Van der Vleuten CPM, Schuwirth LWT. Assessing professional competence: from methods to pro-
    grammes. Med Educ 2005; 39: 309–317.
   Van der Vleuten CPM, Schuwirth LWT, Scheele F, Driessen EW, Hodges B. The assessment of profes-
    sional competence: buildings blocks for theory development. Clin Obstet Gyn 2010;24:703-719.
   Wass V, Van der Vleuten CPM, Shatzer J, Jones R. Assessment of clinical competence. Lancet 2001;
Further reading
Numerous studies have been made on how and why to assess competencies in communication. We give
only a few references, mainly reviews or consensus statements, from the vast literature on this subject.

Boon H, Stewart M. Patient-physician communication assessment instruments: 1986 to 1996 in review. Pat
Educ Couns 1998; 35: 161-76.

Duffy FD, et al. Assessing Competence in Communication and Interpersonal Skills: The Kalamazoo II Report.
Acad Med. 2004;79:495–507.

Elwyn G, Edwards A, Mowle S, Wensing M, Wilkinson C, Kinnersley P, Grol R. Measuring the involvement of
patients in shared decision making: a systematic review of instruments. Pat Educ Couns 2001; 43:5-22.

Kogan, J.R., Holmboe, E.S. and Hauer, K.E., Tools for direct observation and assessment of clinical skills of
medical trainees: a systematic review, JAMA 2009, 302:1316-26.

Laidlaw A, Hart J. Communication skills: An essential component of medical curricula. Part I: Assessment of
clinical communication: AMEE Guide No. 511. (accessed

Norcini J, Anderson B, Bollela V,Burch V, Costa MJ, Duvivier R, et al. Criteria for good assessment: consensus
statement and recommendations from the Ottawa 2010 Conference. Med Teach 2011; 33: 206-214.

Norcini, J. and Burch, V., Workplace-based assessment as an educational tool: AMEE Guide No. 31, Med
Teach, 29 (2007) 855-71.

Postgraduate Medical Education and Training Board (PMETB). Guidance. Developing and maintaining an
assessment system - a PMETB guide to good practice. 2007. http://www.gmc- (accessed 06.08.2012)

Rider EA. Chapter 1 - Competency 1: Interpersonal and Communication Skills. In: Rider EA, Nawotniak RH,
Smith G. A practical guide. Teaching and Assessing the ACGME Core Competencies. Marblehead: HCPro, Inc.
2007; pp1-84.

Schirmer JM, Mauksch L, Lang F, Marvel MK, Zoppi K, Epstein RM, Brock D, Pryzbylski M. Assessing Commu-
nication Competence: A Review of Current Tools. Familiy Medicine 2005.

Teamwork and Communication Working Group. Improving patient safety with effective teamwork and
communication: Literature review needs assessment, evaluation of training tools and expert consultations.
Edmonton (AB): Canadian Patient Safety Institute; 2011. (accessed June
26th, 2012)

Wass V, Van der Vleuten CPM, Shatzer J, Jones R. Assessment of clinical competence. The Lancet 2001;

To top