In Shakespeare’s Macbeth (1710), King Duncan tries, and later fails, to make sense of the defection of a trusted companion. He finally contends that it is impossible to truly know or predict the behaviours of another simply by looking at their face. The scientific literature is seemingly at odds with these musings. Psychological research consistently informs us that humans extrapolate a wide range of information from facial movements (Barrett et al., 2019; Calvo & Nummenmaa, 2016). The contraction of facial muscles creates folds, wrinkles, and areas of tightness which are assumed to fulfil communicative and adaptive functions (Rinn, 1984). While thousands of potential configurations are possible, most research focuses on deriving communicative value from either individual muscle movements (e.g., Tian et al., 2000) or configurations of movements based on theoretical assumptions relating to categories of emotion (e.g., Ekman, 1992; Morais, 2022). Evidence suggests that individuals navigate their everyday social world by making inferences about an interaction partner’s state of mind from these clusters of facial muscle movements (de Melo et al., 2012).
However, there is little consensus about the precise nature of the information that facial muscle movements provide (Barrett & Satpute, 2019; Scarantino, 2017; Siegel et al., 2018). In addition, expressions in everyday interactions often do not resemble the patterns of muscle movements supposedly associated with basic emotions (e.g., Ekman, 1992), and perceivers’ interpretation of these non-prototypical facial configurations is correspondingly less consistent (Hess & Hareli, 2017; Hoegen et al., 2019). These findings have led to a shift in research focus towards assessing the facial movements that are actually produced during realistic interpersonal interactions without making restrictive assumptions about the patterns that faces should adopt during emotional experience (e.g., Hyniewska et al., 2019; Stratou et al., 2017; Zhou et al., 2020). Such a shift facilitates exploration of how facial configurations vary based on intrapersonal, interpersonal, and contextual cues as opposed to predefined emotion categories. Recent research highlights that context is important for recognition of facial expressions in interpersonal interactions (Hoegen et al., 2019). The study reported here therefore aimed to examine the configurations of facial muscle movements occurring during a specific interpersonal context (iterated prisoner’s dilemma game) and to evaluate what information these expressions may convey within that context.
As noted above, researchers disagree about the kinds of information that can be extracted from facial activity. Some categorical conceptualisations suggest that faces display information about a person’s emotional state. In particular, Basic Emotion Theory (e.g., Ekman, 1992; Ekman & Keltner, 1997) holds that a limited number of canonical facial expressions clearly and reliably signal specific emotion categories, providing a means through which interaction partners can express and interpret otherwise unknowable emotional qualia (for discussion, see Barrett & Satpute, 2019). According to some recent exponents of this view, upwards of 20 specific emotions can be reliably diagnosed from dynamic expressions (Keltner et al., 2019). Each expression allows interaction partners to draw systematic inferences about the expresser’s inclinations, guided by the specific emotion expressed. A wealth of literature in this area focuses on the recognition and perception of prototypical, static basic emotion expressions (Scherer et al., 2011), and suggests that humans consistently associate specific emotions with posed prototypical facial expressions when relevant categories are provided as response options (Lopes et al., 2017).
Despite the high levels of consistency in categorising static images of prototypical facial expressions, correlations between basic emotions and predicted facial expressions are generally low in both naturalistic(Fernández-Dols & Crivelli, 2013) and laboratory settings (Durán & Fernández-Dols, 2021; Reisenzein et al., 2013). The fact that prototypical basic emotion expressions occur relatively infrequently during interpersonal interactions(Gaspar et al., 2014) may suggest that facial muscle movements serve a range of purposes in addition to any role they might play in emotion expression. Indeed, research suggests that patterns of muscle activity vary across eliciting contexts, such that diverse patterns of facial movements convey specific information relevant to the current social interaction (Aviezer et al., 2008; Crivelli & Fridlund, 2019). According to this functional perspective, facial movements are flexible, context dependent, and contingent upon the dynamics of the interpersonal interactions in which they occur (Crivelli & Fridlund, 2018).
To determine different dimensions of naturalistic facial activity during ongoing social interactions a previous study conducted an exploratory factor analysis on a large dataset of naturalistic expressions produced during dyadic experimental tasks (Stratou et al., 2017). Frames from these videos were processed, using commercial software based on the computer expression recognition toolbox (CERT; Littlewort et al., 2011). The resulting output provided likelihood scores for 16 specific facial muscle movements, with higher scores indicating a higher likelihood that the specific muscle had been activated in that frame. These Action Units (AUs; (Ekman & Friesen, 1978) derived from the Facial Action unit Coding System (FACS) were then analysed to elucidate meaningful groupings of facial muscles. These likelihood scores were subsequently subjected to exploratory factor analysis.
From this analysis, six consistent configurations of AU activation were identified, described as Enjoyment Smile, Eyebrows Up, Open Mouth, Mouth Tightening, Eye Tightening, and Mouth Frown (Stratou et al., 2017). These configurations mainly failed to map directly onto basic emotion categories. For data derived from Prisoner’s Dilemma gameplay, the authors later correlated each participant’s factor scores with specific outcomes (e.g., number of rounds resulting in mutual cooperation), and concluded that the documented 6 configurations were psychologically meaningful, as they were likely related to contextual cues in social interactions rather than to theoretical emotion categories.
The present study builds upon this prior research by focusing on a circumscribed interpersonal situation during an experimental game (Prisoner’s Dilemma, PD) in order to clarify the specific interpersonal contingencies that give rise to particular patterns of changes to facial musculature. While Stratou and colleagues (2017) assessed facial muscle configurations across entire interactions (i.e., IPD gameplay, negotiations and diagnostic interviews), the current research focuses on a specific period in IPD gameplay (when the round outcome was revealed to both players simultaneously). By limiting the analysis of facial activity to this period, we were able to assess the impact of different gameplay outcomes such as mutual cooperation or mutual defection.
As operationalized in our study, PD is a two-player task, where outcomes are predicated on the simultaneous choices of both players (Poundstone, 1993). Cooperation yields the largest mutual reward, but successful defection yields a greater individual reward, thus creating a dilemma of trust. We used an Iterated prisoner’s dilemma (IPD) task in which this dilemma is repeated across multiple rounds, weaving a dynamic narrative of trust and deceit distinct to each dyad. To aid decision-making, players typically use a range of cues to draw inferences about their opponent to predict their most likely action. There is evidence that inferences made from facial expressions significantly impact decisions in IPD (Hoegen et al., 2019; Lei et al., 2020).
As facial muscle movements are highly context-dependent (e.g., Parkinson, 2013, it is important to delineate how they are influenced by variations in gameplay outcome. By identifying clusters of facial muscle movements which convey important information about intentions and orientations in situations where trust can be violated, research may shed light on how those facial configurations can predict or promote mutually beneficial outcomes.
Using a similar approach to Stratou and colleagues (2017) we processed our videos of participants engaged in dyadic IPD gameplay using the automated Facial Action Coding System AFFDEX (McDuff et al., 2016). AFFDEX automatically provides intensity scores for specific facial actions across 34 AUS which were then analysed to elucidate meaningful groupings of facial activity associated with specific outcome contexts. Following the same procedure as Stratou et al (2017), we then conducted EFA on the data; we extended this approach by undertaking multiple EFAs and a CFA to assess the extent to which our findings were a robust reflection of patterns in facial muscle movements, rather than random noise. When assessing latent structures, multiple EFAs on separate datasets are recommended as a form of cross-validation (Thompson, 2004). Subsequent CFA of unrelated data provides a further test of model consistency. We therefore divided our video data into three sets and explored the first two using EFA before applying CFA to the final set. To establish the meaning of any discovered facial action factors, the relationship between the factors, game states and decisions were explored subsequent to factor analysis. Through assessing the relationships between patterns of facial muscle movements to specific game states, the current study aims to assess whether there are reliable patterns of facial activity that respond to these contexts and that provide meaningful context-related information to interaction partners.