Interprofessional Learning (IPL) has evolved into an important component of health professional education. It aims to cultivate collaborative competencies across diverse professions and enhance patient-centred healthcare delivery. [1-4] The value of student teamwork is recognised by IPL competency frameworks, which specifically require opportunities for students to work in interprofessional groups, with an emphasis on collaborative practice and collective decision-making. [5] [6] [7] However, amidst the drive for collaborative excellence, there is a significant and persisting challenge in how to effectively assess the performance of interprofessional teams in authentic large-scale IPL activities.[8] This situation presents challenges, as it hinders universities from delivering high-stakes assessments that adequately align with the importance of achieving collaborative learning outcomes critical for healthcare. [9]
There is a critical gap in the literature; a scarcity of robust assessment methodologies that can comprehensively assess the teamwork, communication, and decision-making skills of students engaged in large scale IPL settings ( N>300). [10] [2] [7] [11] [12] Many current IPL assessments have been designed to test if students can function expertly as competent individuals, or independent practitioners, [13, 14] [15] and focus on short-term knowledge acquisition, impact on attitudes to other professions, and student satisfaction. [16] Several studies report self-assessment, which are not optimal given an individual’s inability to assess such gains accurately. [17] [18] There is a small but growing literature providing validity evidence for the generation and interpretation of team scores intended for a large scale assessment of teamwork performances [18, 19] [20, 21] [15]
In this paper, we extend our research on large scale interprofessional assessment [10] [22] by investigating the psychometrics of a novel IPL assessment that has the potential to be optimised to provide a high stakes assessment of collaborative teamwork. An opportunity to study this arose in a secondary analysis of data from a purposively designed interprofessional activity based around a video-based assessment for learning task. Within our context, multiple student teams embarked on two linked missions: first, to creatively document their collective problem-solving and decision-making processes for a complex patient case, and secondly, to rate the quality of their peers' work as depicted in their videos.
Assessment in IPL
We now summarise the literature on the key issues that informed our choice of a peer rated video-based teamwork assessment task. Following on, we provide an overview of our chosen validity framework and in the methods, describe the measurement model through which we address our research questions. Assessment profoundly shapes student learning and experience in higher education. Assessing team performance encompassing aspects like leadership, communication, collaboration, and decision-making can be complex. Factors that impact any team-based assessment include the specific task (e.g. case management), team composition, and cultural and organizational contexts. [4] Several tools have been developed to observe and measure student team performance either in real-time (in clinical settings or simulation) or video recordings. [7] [19] [15] In the work-based assessment literature, there are similar distinctions to be made between an assessment designed to measure the collective competence of a team or the individual competency of team members. [23] [24] [25] Challenges also arise in assessing student interprofessional team performance due to resource demands and limited curricular opportunities for team-based clinical activities. [2] [8] [26].
Interprofessional collaborative practice operates under the assumption that a student's professional background should not hinder their ability to evaluate peers from other professions. However, established professional perspectives can hinder collaborative learning, [27] [28] [29] particularly in the area of assessment. [2] [7] [14] Currently, there is little evidence as to what extent a student’s profession impacts their assessment of teamwork performance. There is some evidence from the work-based assessment literature that professional designation does influence individual judgement in terms of both stringency and perspective. [30] However, it is not known whether those differences arise during training or employment, or at the student level.
Peer Assessment
Although the affordances of peer assessment in IPL has been recognised for some time, [7] [31] [32] [33] there are several challenges to overcome in implementation. Peer assessment is a recognised teaching and learning strategy to develop student skills in giving and receiving feedback. It can facilitate students to become better self-regulated learners because when they review the work of their peers, they tend to reflect back on their own work and consider ways of improving it. [34] Whilst tutors have only limited time to observe each student, students have many opportunities to observe each other. [35, 36] Students are more likely to see authentic aspects of interprofessional collaboration than tutors when engaged in, for example, practice-based learning in the clinical setting. [37] Several studies have measured uniprofessional small group learning effectiveness based on student self-assessment of particular behaviours. [38] [39] [40] [41] Despite concerns that the quality of evidence on peer assessment assessments is impacted by the varying methodological approaches, [42] it is thought that peer ratings are reliable and perhaps more valid than self- or tutor-based assessments. [33] [43] [44] Both the giving and receiving of peer feedback have good educational impact for the student assessor. [41]
Student-produced video in assessment for learning
There is a long tradition of using video to assess student clinical skills in the laboratory setting [45] [46] and of using video to develop students’ knowledge or skills acquisition. [47] There is a developing literature on student-produced video as a format in the assessment for learning and of learning that calls for educators to embrace student videography. It is a form of assessment that provides students with important graduate capabilities, to actively apply knowledge in authentic contexts, and promote self-reflection and the giving and receiving feedback from others. [48] Examples include science education, [49, 50] health promotion education, [51] [52] [53] and communication skills training in both dentistry. [54] and in pharmacy. [55], and interprofessional settings [56] [57]
In the broader educational literature, studies of student-produced video that include details of the assessment, focus on the adequacy of topic content, communication of the key message, and the technical quality of the video product. [10] A recent literature review of the utility of student-generated video in the health professions suggested projects suffered from having unclear purposes and weak pedagogical considerations which risks undermining the intended learning.[58]
Kane’s Validity Framework
Several interprofessional competency statements have been published internationally that can inform academic content of an IPL assessment rubric. [59-62] To evaluate the validity of the assessment in this study, we used Kane's argument-based validity framework, which emphasizes key inferences as the assessment progresses from a single observation to a final decision. We were interested in four aspects of the validity of the collaborative teamwork assessment tool. [63] [64] These were;
- Scoring - the observation of the video was translated into one or more scores,
- Generalisation - the observed scores were used to generate an overall test score representing collaborative teamwork in the IPL setting,
- Extrapolation - an inference was drawn regarding what the test score might imply for the collaborative teamwork of students, and
- Implications - if the scores were credible and reasonably free from error, they could be used as an assessment of collaborative teamwork.
As part of the generalisation argument, reliability refers to the reproducibility or consistency of scores from one assessment to another, and is a necessary, but not sufficient, component of validity. It can be measured by a generalizability co-efficient which estimates multiple sources of error and provides a method which is generalizable to virtually any setting in which the reliability of the instrument could be assessed.[65] As part of the generalisation argument, valid “for decision making” entails generating a defensible standard.
In this paper, we posed two research questions. First, what factors impact the generalisability of a peer assessment of a student-produced video depicting interprofessional collaborative teamwork in a large-scale IPL activity? Second, to what extent does the student assessor’s profession and the patient case on which the team worked impact student ratings?