Hostname: page-component-78c5997874-j824f Total loading time: 0 Render date: 2024-11-10T05:09:04.385Z Has data issue: false hasContentIssue false

Representations of facial expressions since Darwin

Published online by Cambridge University Press:  28 April 2022

David Perrett*
Affiliation:
School of Psychology and Neuroscience, University of St Andrews, St Mary's Quad, St Andrews, Fife KY169JP, UK
*
*Corresponding author. E-mail: dp@st-andrews.ac.uk

Abstract

Darwin's book on expressions of emotion was one of the first publications to include photographs (Darwin, The expression of the emotions in Man and animals, 1872). The inclusion of expression photographs meant that readers could form their own opinions and could, like Darwin, survey others for their interpretations. As such, the images provided an evidence base and an ‘open source’. Since Darwin, increases in the representativeness and realism of emotional expressions have come from the use of composite images, colour, multiple views and dynamic displays. Research on understanding emotional expressions has been aided by the use of computer graphics to interpolate parametrically between different expressions and to extrapolate exaggerations. This review tracks the developments in how emotions are illustrated and studied and considers where to go next.

Type
Review
Creative Commons
Creative Common License - CCCreative Common License - BY
This is an Open Access article, distributed under the terms of the Creative Commons Attribution licence (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted re-use, distribution and reproduction, provided the original article is properly cited.
Copyright
Copyright © The Author(s), 2022. Published by Cambridge University Press

Social media summary: This review tracks the development in illustration and study of facial expressions of emotions.

Introduction

Darwin wrote his book, The expression of the emotions in Man and animals in just four months after completing the page proofs for the Descent of Man (Ekman, Reference Ekman2009). One feature that was unusual for that time was the inclusion of photographs of facial expressions in the book. This was pioneering in the publishing world, partly because printing photographs was costly. The purpose of this review is to discuss the impact of the inclusion of photographs in Darwin's initial work and to chart the development of emotion illustrations (particularly those of facial expressions) since that time. The review will cover the utility of depictions within the discipline of psychology.

Darwin as a psychologist

Darwin collected reactions to photographs in a way that was a prototype for many psychological experiments. He presented an image, for example, a man's face stimulated by Duchenne's electrical probes, to 20 educated persons of various ages and both sexes, and asked them what emotion or feeling the man was displaying. He recorded the participants’ answers in the words that they had used (Darwin, Reference Darwin1872; see Wilson, Reference Wilson2006: 1267). He used these psychological experiments in the narrative of The expression of the emotions, referring to photographic illustrations both posed and spontaneous and discussed his survey of responses to them. He concluded that the public were consistent in their interpretations: for example, 14 out of 15 people recognised a furrowed forehead as ‘despairing sorrow’, ‘suffering endurance’ or ‘melancholy’. The inclusion of photographs in his book thus invited readers to make judgments for themselves and, in this way, Darwin drew his readers into a conversation. The photographs thus represented a move towards ‘open science’ (Foster & Deardorff, Reference Foster and Deardorff2017) because others could use the depicted expressions to verify or falsify Darwin's conclusions about the public's reactions.

In Darwin's time, lithographs were the standard alternative to photographs (see Figures 1–21, Darwin, Reference Darwin1872), but photographic images offer greater verisimilitude. A lithographic engraving, however skilled, may exaggerate or fail to capture particular details that are present in real life or in the photograph on which they are based. Of course, photographs are not infallible; they too can miss particular expressive details because of perspective view or lighting shadows. Camera perspective can also modify the apparent expression of the face, e.g. when the head is bowed slightly, the chin appears smaller and less threatening (Zhang et al., Reference Zhang, Lin and Perrett2020). Photographic capture may also be timed poorly relative to the apex of a dynamic expression. Yet despite each of these shortcomings, photography increases the accuracy of representation over lithographs.

Figure 1. Images of health and sickness from the nineteenth and twenty-first centuries. (a) Galton's composite photographs of ‘health’ – a combination of 23 Royal Engineers, and ‘sickness’ – combinations of six and nine cases of tubercular disease (Galton, Reference Galton1883). (b) Composite images of 22 individuals 2 hours after an injection of a placebo (left) or a bacterial endotoxin (right). Note the subtle change in expression after the toxin. Reproduced from Axelsson et al. (Reference Axelsson, Sundelin, Olsson, Sorjonen, Axelsson, Lasselin and Lekander2018), Proceedings of the Royal Society B: Biological Sciences published under creative commons. Permissions for reproduction were obtained from https://www.copyright.com/. The author's permission was provided by email.

Multiple emotions

Darwin grouped emotional expressions into six categories. Ekman, a century later, concurred that a basic set of six emotions (happiness, surprise, fear, sadness, anger and disgust combined with contempt; Ekman et al., Reference Ekman, Friesen, Ellsworth, Goldstein and Krasner2013) or seven (these six emotions plus interest; Ekman, Reference Ekman1972) are recognised universally. Others have argued that emotional expressions are not recognised with cultural universality (Russell, Reference Russell1994). Some authors have argued that the number of expressions that are recognisable is likely to be more than the basic six (see Keltner et al., Reference Keltner, Sauter, Tracy and Cowen2019). Others have argued that fewer than six are recognised across cultures (Jack et al., Reference Jack, Sun, Delis, Garrod and Schyns2016). Perhaps part of the problem is that there is not 100% agreement in the labelling of facial expressions of emotions in any society. For example, using perhaps the most standard set of expression photographs from the Facial Expressions of Emotion Stimuli and Test, normal adult recognition accuracy of all 10 examples of each of six emotional expressions is about 80% (Young et al., Reference Young, Perrett, Calder, Sprengelmeyer and Ekman2002). Data for 227 individuals (aged 20–70 years with IQs of 90 and above) were used to define the average and spread of recognition performance. These data provide a cut-off score defining the border between normal range and impaired performance for each emotional expression (i.e. significantly different from the mean, p = 0.05; Young et al., Reference Young, Perrett, Calder, Sprengelmeyer and Ekman2002). For individual emotional expressions such as fear, a recognition score of 4/10 would lie at the boundary of normal and abnormal recognition. While 40% is higher than chance (1/6 or 16.7%), it is not particularly impressive. Indeed, the cut-off drops further to 3/10 (30% correct) when considering individuals older than 60 years. So, photographic examples of the most standard facial expressions are not that well recognised even within a single culture. Indeed, Barrett (Reference Barrett2011) argues that the photographs of these prototypical expressions the actors have adopted (often muscle by muscle) are culturally agreed symbols of emotions (somewhat similar to emoticons ) rather than being actual expressions that we come across in everyday use. The symbolic nature of posed expressions becomes more apparent from the work of Dawel et al. (Reference Dawel, Wright, Irons, Dumbleton, Palermo, O'Kearney and McKone2017). These investigators asked observers to judge posed expressions on a 15-point scale (–7 = completely fake; 0 = don't know; +7 = completely genuine). The majority of the posed expressions from standard expression sets such as those of Young et al. (Reference Young, Perrett, Calder, Sprengelmeyer and Ekman2002) were not seen as genuine despite the fact that the same expressions could be correctly categorised as representing the emotional states angry, disgusted, fearful, sad and happy. The issue remains as to why expression recognition is not better but, as we will see, recognition of emotion depends on a variety of factors, including context.

Composite portraits

Following on from Darwin's work on facial expressions, the technique of composite photography was developed in the nineteenth century. Composite construction is detailed here because it has had a pervasive influence on facial expression research. Darwin's cousin, Francis Galton (Galton, Reference Galton1878, Reference Galton1879), was a pioneer in making composite photographs. Galton sought to extract the generic features of different groups or types. His notorious views on eugenics may have been the drive to differentiate socially inferior types of man from healthy types (Levy & Peart, Reference Levy and Peart2004). With the visual specification of healthy types one could then ‘encourage as far as practicable the breed of those who conform most nearly to the central type, and to restrain as far as may be the breed of those who deviate widely from it’ (Galton, Reference Galton1907). Taking multiple exposures of different faces with each face aligned by the pupils of the eye would progressively average out idiosyncratic features and emphasise features common to faces in the group. For example, Galton combined photographs of convicted felons in an effort to generate the average criminal face. His drive here was consistent with his views on eugenics to breed out unwanted traits like criminality once he had identified what constituted a criminal type (Green, Reference Green1984), although had he succeeded in his endeavour, it would also be tempting to supply the resultant image to the police force who could then pre-emptively round up any would-be criminal on the basis of their facial appearance. By his own admission, the resulting composite did not look criminal, instead ‘the features of the composites are much better looking than those of the components. The special villainous irregularities in the latter have disappeared and the common humanity that underlies them has prevailed’ (Galton,1878). In short, the image appeared to be a man who looked quite handsome. Galton speculated that the composite might represent ‘the man who is liable to fall into crime’, which is hardly useful as this might apply to anyone given particular circumstances.

The finding that the process of averaging multiple faces increases the attractiveness of the resulting image has been upheld by modern Psychology (Langlois and Roggman, Reference Langlois and Roggman1990). The evidence for this contention has improved with the development of computer graphic techniques (Little & Hancock, Reference Little and Hancock2002). Practical tips on using such software to modify facial appearance are given by Sutherland et al. (Sutherland & Young, Reference Sutherland and Young2015; Sutherland et al., Reference Sutherland, Rhodes and Young2017). Computer composites are now made in several stages. First the structure of each component face image is defined with landmarks placed at corresponding feature points (e.g. the tip of the nose). The average face shape across the group is then calculated as the mean positions of the landmarks. Next, each component face image is reshaped to the average face shape. Finally, the colour and texture information from the corresponding positions in the reshaped images are averaged together (Tiddeman et al., Reference Tiddeman, Burt and Perrett2001). The resulting composites are neatly in focus and appear lifelike, with no particular identity showing through (see Figure 1b).

The notion that beauty is, to a large extent, averageness might appear an enigma because it implies that beauty is nothing special. In fact, although the average-shaped face is indeed attractive, it is not the most attractive face shape (Perrett et al., Reference Perrett, May and Yoshikawa1994; DeBruine et al., Reference DeBruine, Jones, Unger, Little and Feinberg2007). Indeed, several factors that move faces away from averageness contribute to facial attractiveness; these include femininity, apparent health, youthfulness and positive expression (Perrett et al., Reference Perrett, Lee, Penton-Voak, Rowland, Yoshikawa, Burt and Akamatsu1998; Jones et al., Reference Jones, Little, Burt and Perrett2004; Fink et al., Reference Fink, Grammer and Matts2006; Tatarunaite et al., Reference Tatarunaite, Playle, Hood, Shaw and Richmond2005).

Galton also made composite images of healthy individuals (army personnel) and those suffering from disease (e.g. tuberculosis) (Galton, Reference Galton1883; Figure 1a). His composites of sickness do suggest a pallor associated with consumption. With computer averaging, composite images can be produced with greater clarity, making it possible to read subtle expression changes accompanying sickness, as well as changes in pallor. Figure 1b compares the composite images of 22 volunteers 2 hours after the injection of either a placebo and/or a bacterial toxin. The reaction induced by the toxin includes a slightly drawn expression with downturned mouth and drooping eyelids, in addition to the skin losing redness and the lips becoming bluer (Axelsson et al., Reference Axelsson, Sundelin, Olsson, Sorjonen, Axelsson, Lasselin and Lekander2018).

Computer graphic manipulations of expressions

A given expression has characteristic cues but these different cues may not be equally evident in all people making the expression. One way of depicting a typical emotional expression is, again, to generate a composite, averaging together many examples of that expression. Figure 2 illustrates the combination of 35 male faces from the Karolinska Directed Emotional Faces (KDEF) collection (Lundqvist et al., Reference Lundqvist, Flykt and Öhman1998). In Figure 2 top row, far left column, all of the men have a neutral expression whereas in the fifth column of the top row all of the men have a happy expression. While composite expressions are more representative than a single actor's pose, composites suffer from the same criticism raised by Barrett (Reference Barrett2011). Each male actor displayed what they took to be the agreed symbol of the emotion whereas what is expressed in natural circumstances can differ.

Figure 2. Composite faces posing happy, afraid and disgust expressions. Row 1: a composite of images of 35 males posing with neutral expression (0%) and a happy expression (100%). The differences in shape, colour and texture between the neutral and happy face images are used to transform the neutral image in 25% steps. This creates images where the happy expression gradually emerges in intensity. The 125 and 150% images represent an extrapolation of the series with caricature exaggeration of the happy expression by 25 and 50%. Row 2: as row 1 for the expression of fear. Row 3: as row 1 for the expression of disgust. Composite images and emotion transforms produced by the author.

Computer graphics can be used to enhance expressions in calculated steps. First the structure of a neutral or resting face is defined by a collection of (200) landmarks allocated manually to the corners of the mouth, the eyes and so on. The procedure is then repeated for a given expression, for example happiness (Figure 2 row 1, column 5). The difference between the position of the two corresponding sets of landmarks can then be measured. For example, with a smile present in a happy expression, one would expect the corners of the lips to be raised by contraction of the zygomatic muscles. Additionally, during a genuine smile the eyelids would close slightly owing to constriction of the orbicularis occuli muscles. Indeed, there would be a whole configuration of changes in the position of each of the facial landmarks. With computer graphics the landmarks are used to deform the photographic image, much as a sheet of rubber can be stretched by anchors at given points. The movement of the landmarks can be exaggerated (or diminished) computationally. The result is an expression of hyper happiness for the caricaturing exaggeration (125 or 150% Happy in Figure 2). Alternatively, diminution of the movement anchors results in a more subtle (25 or 50%) smile.

This technique has been used extensively in neuroscience in attempts to map out the brain systems involved in analysis and recognition of particular emotions. For example, the amygdala becomes more activated as expressions move between neutral to fear and exaggerated fear (Figure 2 row 2; Morris et al., Reference Morris, Frith, Perrett, Rowland, Young, Calder and Dolan1996). Likewise, the insular cortex shows increasing blood flow as expressions viewed move from neutral through disgust to exaggerated disgust (see Figure 2 row 3; Phillips et al., Reference Phillips, Young, Senior, Brammer, Andrew, Calder and David1997). Thus, computer exaggeration of facial expressions has enabled various brain systems to be implicated in the processing of emotions.

The use of standardised expressions of emotion has also revealed problems in emotion recognition. Such testing can involve the presentation of full-blown expressions, yet some emotions such as happiness are relatively easy to recognise and performance is therefore often at the ceiling (i.e. 100% correct). A more sensitive way of testing recognition has been to employ subtle expressions produced by computer graphic transformation of neutral faces in small steps towards the full expression (e.g. 25, 50 and 75% expressions in Figure 2, columns 2–5).

The misperception or failure to recognise individual or multiple expressions of emotions, particularly at low intensity levels, has been associated with a variety of conditions following brain injury or neurodegenerative diseases, including amygdala damage (Adolphs et al., Reference Adolphs, Tranel, Hamann, Young, Calder, Phelps and Damasio1999), Huntington's disease (Sprengelmeyer et al., Reference Sprengelmeyer, Young, Calder, Karnat, Lange, Hömberg and Rowland1996), Wilson's disease (Wang et al., Reference Wang, Hoosain, Yang, Meng and Wang2003) and Parkinson's disease (Coundouris et al., Reference Coundouris, Adams, Grainger and Henry2019). Recognition of emotion in facial expressions is compromised by a great range of issues affecting mental health, including autism (Law-Smith et al., Reference Law-Smith, Montagne, Perrett, Gill and Gallagher2010; Eack et al., Reference Eack, Mazefsky and Minshew2015, although see Cook et al., Reference Cook, Brewer, Shah and Bird2013), alcoholism (Frigerio et al., Reference Frigerio, Burt, Montagne, Murray and Perrett2002; Kornreich et al., Reference Kornreich, Philippot, Foisy, Blairy, Raynaud, Dan and Verbanck2002), alexithymia (Cook et al., Reference Cook, Brewer, Shah and Bird2013), bipolar disorder (Venn et al., Reference Venn, Gray, Montagne, Murray, Burt, Frigerio and Young2004), borderline personality disorder (Daros et al., Reference Daros, Zakzanis and Ruocco2013), depression (Krause et al., Reference Krause, Linardatos, Fresco and Moore2021), psychopathy (Montagne et al., Reference Montagne, van Honk, Kessels, Frigerio, Burt, van Zandvoort and de Haan2005; Hastings et al., Reference Hastings, Tangney and Stuewig2008; Dawel et al., Reference Dawel, O'Kearney, McKone and Palermo2012), schizophrenia (Kohler et al., Reference Kohler, Walker, Martin, Healey and Moberg2010) and social phobia (Bell et al., Reference Bell, Bourke, Colhoun, Carter, Frampton and Porter2011). This list is extensive but is by no means exhaustive. It indicates the pervasiveness of problems in understanding expressions of emotion. Indeed, emotion recognition failures may exacerbate mental health problems by frustrating normal social interaction.

Bias in interpreting ambiguous expressions

Another way that computer graphics have been harnessed to investigate the perception of emotions is to ‘morph’ or dissolve slowly from one expression to a different expression (e.g. in Figure 3 a happy expression is changed in stages to an angry expression). Despite the continuous nature of the morphing process, the way we interpret the images is discontinuous or categorical. We are likely to label all the images on the left of the sequence as belonging to the category ‘happy’ and the steps on the right of the sequence as belonging to the category ‘angry’. At some point along the continuum the category switches. Images at this point are the most ambiguous and have the longest reaction times for observers to assign them a name or category (Young et al., Reference Young, Rowland, Calder, Etcoff, Seth and Perrett1997).

Figure 3. Happy to angry facial expression continuum. Five steps are illustrated progressing from 100% happy to 100% angry. The central image is ambiguous showing both characteristics of happiness and of anger. The upper part of the figure illustrates the categorical boundary between the images being categorised as angry or happy before training (see text). The lower section illustrates that, post training, the boundary is shifted such that more ambiguous expressions are classified as happy. Reproduced from figure 1 in Penton-Voak et al. (Reference Penton-Voak, Thomas, Gage, McMurran, McDonald and Munafò2013) Psychological Science, 24, 688–697 with permission from the author.

The ambiguity of the facial configuration at the midpoint or categorical boundary of two expressions is itself valuable as it can be used to demonstrate biases in the identification of emotions. In depression there is a bias to interpret ambiguous emotions negatively. Depressed individuals are more likely to report ambiguous or mixed expressions as sad and less likely to report them as happy (Bourke et al., Reference Bourke, Douglas and Porter2010).

For juvenile offenders with conduct disorder there is often a bias to see others as threatening (Dodge et al., Reference Dodge, Price, Bachorowski and Newman1990). This hostile attribution bias is rational in an adverse environment but may be self-reinforcing. In a vicious cycle, a negative interpretation of another's expression can lead to an aggressive reaction which is in turn reciprocated. Penton-Voak et al. (Reference Penton-Voak, Thomas, Gage, McMurran, McDonald and Munafò2013) reasoned that it may be possible to reverse the process by establishing a virtuous cycle where a shift in the interpretation may allow positive reactions to be reinforced.

Penton-Voak et al. (Reference Penton-Voak, Thomas, Gage, McMurran, McDonald and Munafò2013) used intermediate or ambiguous expressions in an attempt to rehabilitate conduct disorder. Adolescent participants were given a task to decide whether an image was happy or angry. The stimuli presented were a series of images gradually morphing between anger and happiness (see Figure 3). The participants received biased feedback. When they identified ambiguous stimuli as happy they received ‘correct’ as feedback, reciprocally when they labelled the ambiguous emotion as angry they received ‘incorrect’ as feedback. After training, the category boundary between angry and happy expressions shifted so that more intermediate expressions were classified as happy. Of greater significance, participants reported less anger and staff reported less aggressive behaviour in the two weeks after training (Penton-Voak et al., Reference Penton-Voak, Thomas, Gage, McMurran, McDonald and Munafò2013). The malleability of emotion categories used to label the same facial expression reinforces the role of learning in the interpretation of emotions (Barrett et al., Reference Barrett, Adolphs, Marsella, Martinez and Pollak2019).

The colour of happiness

Benitez-Quiroz et al. (Reference Benitez-Quiroz, Srinivasan and Martinez2018) claim that colour can be used to support recognition of facial expressions of emotion. This capacity to use colour for emotion recognition may depend on guessing the valence of the underlying emotion with happiness representing positive valence and anger, fear, disgust and sadness all representing negative valence emotions. The strongest colour cue comes from happy expressions because of the blood flow and reflection changes in the cheek area. Hence, red cheek colour is a cue to a positively valanced emotion. The diagnostic value of colour shows the importance of colour photography in the reproduction of facial expressions. Of course colour use adds to the expense of printing.

Adding depth to emotions

A development in representation of faces with emotional expressions comes from the shift from two to three dimensions. A three-dimensional model also allows measures of responses to facial expressions aimed towards or away from the observer. One can determine the cone of view in which people judge expressions to be directed at them. Most observers have an egotistical bias and interpret someone's happy expression as directed at themselves while more negative emotions are interpreted as directed away from them and towards someone else (Lobmaier & Perrett, Reference Lobmaier and Perrett2011). The tendency to see negative expressions as self-directed is heightened in people who are socially anxious and could exacerbate their anxiety (Schulze et al., Reference Schulze, Lobmaier, Arnold and Renneberg2013).

To view items in 3D, it is common to think of stereoscopic displays where the left and the right eyes receive separate views of two different photographs. Publications with such requirements are clumsy (e.g. issuing red–green viewers and printing two photographs in red and green ink; Julesz, Reference Julesz1971). Alternative strategies for displaying 3D images include the use of motion parallax, for example, by presenting a video of the head rotating (Holzleitner & Perrett, Reference Holzleitner and Perrett2016).

The advantage of a 3D head model is that it allows viewing from more than one perspective. This advantage can be achieved by rendering two or more 2D views of the 3D model and presenting them together. The aim here is not to achieve a stereoscopic view but to provide readers with the increased visual information that comes from the 3D model. Viewers can be presented with the front view and also a view turned towards the profile. The combined views give more information about the chin, nose, brow and forehead than is available from a frontal 2D image Figure 4 (see Figure 5 for example).

Figure 4. A face with and without additional diagnostic colour information for the emotion of happiness. With the augmented colour information, the images were easier to classify as happy. Reproduced under Creative Commons License cropping the original image to show only the face pair from figure S6 from Benitez-Quiroz et al. (Reference Benitez-Quiroz, Srinivasan and Martinez2018) Proceedings of the National Academy of Sciences, 115, 3581–3586. With permission from the author.

Figure 5. 3D faces varying in apparent trustworthiness. Frontal and half profile views of male and female 3D head models varying in apparent personality. The head models were constructed by averaging together the 3D surface shape and texture of male and female faces separately (middle row). A collection of 118 faces (male = 50, female = 68) were rated for how trustworthy they looked while being rotated to reveal their 3D structure. For each gender, an average 3D head shape was formed from those faces that appeared high in trustworthiness. Separately an average was formed from those that appeared low in trustworthiness. These two averages defined a trustworthiness trajectory in 3D shape space for men and for women. Male and female composite faces were then transformed in shape along this trajectory to decrease apparent trustworthiness (top row) or to increase apparent trustworthiness (bottom row). Methods for averaging and transforming have been presented elsewhere (Holzleitner et al., Reference Holzleitner, Hunter, Tiddeman, Seck, Re and Perrett2014). 3D head models and apparent trait transforms models produced by the author.

First impressions

We readily make judgments of a person's character even when we have not observed any behaviour. While there are many adjectives that we might use to describe a person's character (intelligent, dominant, warm, charismatic and so on), these judgments can be boiled down to just two or three dimensions (Todorov et al., Reference Todorov, Said, Engell and Oosterhof2008; Sutherland et al., Reference Sutherland, Oldmeadow, Santos, Towler, Burt and Young2013). The most important of these dimensions reflects emotional expression. The first dimension of character judgment is referred to as trustworthiness, warmth and sometimes valence. A person with a slightly positively valanced emotional expression (with upturned lip corners and slightly raised eyebrows, see Figure 5 bottom row) is seen as trustworthy (Todorov, Reference Todorov2008). Reciprocally, a slightly negatively valanced emotional expression (i.e. with downturned mouth and lowered eyebrows) is seen as untrustworthy (see Figure 5 top row). This trustworthy dimension refers to a person's intentions: warmth and approachable vs. hostile and unapproachable.

A second dimension is often referred to as power or competence and may reflect the impression of capacity to carry out intensions. This dimension is most strongly linked to maturity and masculinity. A large male with masculine physique has the power to act antagonistically and with impunity, whereas a baby-faced adult looks weaker in physique and seems to have little power to act dangerously even if angry.

Face images are often collected with individuals posing with a neutral expression (e.g. Todorov et al., Reference Todorov, Said, Engell and Oosterhof2008), nonetheless the apparent expression of neutral or resting faces varies considerably. A person's mouth may appear to curve down or upwards depending on underlying anatomy or posture. Likewise, eyebrows may appear lowered or raised, depending on anatomy and pose. These subtle differences in apparent expression play a major role in the attribution of personality (Todorov, Reference Todorov2008). Zebrowitz et al. (Reference Zebrowitz, Kikuchi and Fellous2010) found that individuals who displayed a resting expression resembling anger were seen as less trustworthy and less competent. Resemblance to expressions influenced trait impressions even when statistically controlling possible confounding influences of attractiveness and baby-facedness.

It is important to note that trait attributions need not show any relation to actual personality. Furthermore, we have systematic biases when interpreting expressions and other facial cues. As noted, we assume that angry-looking individuals are less trustworthy even though there is no causal link between trustworthiness and the tendency for anger. Indeed, there is a danger that artificial intelligence software for classifying facial expressions of emotion (Pantic & Rothkrantz, Reference Pantic and Rothkrantz2000) will be misused to differentiate presumed trustworthy and untrustworthy individuals. Attempts to classify people into different types from face photographs is a misguided endeavour and will unwittingly reinforce the biases that we hold (see y Arcas et al., Reference y Arcas, Mitchell and Todorov2017; Bowyer et al., Reference Bowyer, King, Scheirer and Vangara2020). For example, individuals convicted of crime may be more likely to come from lower socio-economic sections of society. Any cue related to low socio-economic status, e.g. low health or mood at the time of photograph, could contribute to criminal face image classification even when the cue is not causally linked to crime. The sources of images of criminals are likely to be systematically different to sources of images for non-criminals. Comparing mugshots of supposed criminals taken at a police station with web-based images of supposed non-criminals would be an obvious error but even when this difference in photography is controlled for there are further confounds (Bowyer et al., Reference Bowyer, King, Scheirer and Vangara2020).

Emotion in context

Aviezer et al. (Reference Aviezer, Hassin, Ryan, Grady, Susskind, Anderson and Bentin2008) placed isolated facial expressions of disgust on the body of an actor/actress in different contexts (displaying disgust, anger, fear and sadness). Participants were greatly affected by the context and frequently miscategorised the emotion. For example, in the angry context participants labelled the disgust facial expression as anger on 80% of trials. In a second experiment, disgust face expressions were placed in a negatively valanced disgust context (e.g. a man holding soiled underwear) or in a positively valanced pride context (e.g. a body builder showing off his muscular torso, see Figure 6). Eighty per cent of disgust expressions were categorised as negatively valanced in the disgust context but 0% of the disgust expressions were categorised as having a negative valence in the ‘pride’ context (Aviezer et al., Reference Aviezer, Hassin, Ryan, Grady, Susskind, Anderson and Bentin2008). This experiment makes it clear that facial expressions of emotion are not interpreted in isolation from other environmental cues, including body posture. Meeren et al. (Reference Meeren, van Heijnsbergen and de Gelder2005) note that ‘When face and body convey conflicting emotional information, judgment of facial expression is hampered and becomes biased towards the emotion expressed by the body’. Indeed, the accompanying environmental context presented alongside a facial expression is remembered better when observers are able to categorise the emotion displayed (Barrett & Kensinger, Reference Barrett and Kensinger2010). That is, context is key to understanding and remembering the displayed emotion (Barrett et al., Reference Barrett, Mesquita and Gendron2011).

Figure 6. Disgust expression modified by context. An isolated facial expression of disgust was placed in a ‘disgust’ context (left) or in a ‘pride’ context (right). While the disgust expression was accurately categorised as negatively valanced in the disgust context, it was never categorised as having a negative valence in the pride context. Reproduced from figure 4a, Aviezer et al., (Reference Aviezer, Hassin, Ryan, Grady, Susskind, Anderson and Bentin2008) Psychological Science, 19, 724–732 with permission from the author. Permissions for reproduction were obtained from https://www.copyright.com/. The author's permission was provided by email.

Thus, one aspect of ecological validity in presenting expressions of emotions is to provide an environmental context, and to include information from body posture. Perhaps that is why Darwin chose to display both the face and the body in many his photographic plate illustrations of emotions (Darwin, Reference Darwin1872).

Dynamic expressions across cultures

Darwin (Reference Darwin1872) sent a questionnaire to variety of colleagues abroad to examine the extent of commonality of expressions across different societies. He argued for consistency in expressions and their recognition across cultures and, indeed, reviews of cross-cultural perception have concluded that there is considerable commonality across peoples (Elfenbein & Ambady, Reference Elfenbein and Ambady2002; Sauter & Eisner, Reference Sauter and Eisner2013). More recent assessment across cultures suggests that there is also diversity in interpretation (Gendron et al., Reference Gendron, Crivelli and Barrett2018). Evidence for some diversity in facial expressions has come in part from research using dynamic expressions (Jack et al., Reference Jack, Garrod, Yu, Caldara and Schyns2012). Jack et al. (Reference Jack, Garrod, Yu, Caldara and Schyns2012) showed 15 European and 15 Chinese participants’ dynamic facial images with combinations of different muscles contracting to build up a visual representation of the expression of each of the six basic emotions. In general, there was more information about the nature of the expression from the mouth and lower parts of the face for the Western participants and conversely more information about the expression from the eyes and upper part of the face for the Eastern participants (see Figure 7). For the European observers, the six expressions appeared to be based on unique sets of muscle contractions. For Chinese observers, the muscles used for the different expressions were less consistent and showed some overlap across the six emotions. For happy, angry and disgust expressions Chinese participants represented the intensity of emotion with movements of the eye region while this was less true for the European participants. Hence Jack et al. (Reference Jack, Garrod, Yu, Caldara and Schyns2012) concluded that expressions vary with culture.

Figure 7. Comparing the representation of three expressions for one European (left) and one Chinese participant (right). The mouth region is more informative for the European and the eye region is more informative for the Chinese participant. Reproduced from Movie S2, Jack et al. (Reference Jack, Garrod, Yu, Caldara and Schyns2012) Proceedings of the National Academy of Sciences, 109, 7241–7244 with permission from the author. For the dynamic movie see http://www.pnas.org/lookup/suppl/doi:10.1073/pnas.1200155109/-/DCSupplemental/sm02.avi.

Regardless of the commonality or diversity in expression interpretation, Jack's research shows how the methods of expression illustration have developed. The expression stimuli used by Jack et al. (Reference Jack, Garrod, Yu, Caldara and Schyns2012) included internal facial movement. The skin and features were made to deform in a manner corresponding to those that take place when a given face muscle contracts, holds and relaxes. Chen et al. (Reference Chen, Crivelli, Garrod, Schyns, Fernández-Dols and Jack2018) illustrate similar methods for combining different facial muscles contracting with different time courses (http://movie-usa.glencoesoftware.com/video/10.1073/pnas.1807862115/video-1). Chen et al. (Reference Chen, Crivelli, Garrod, Schyns, Fernández-Dols and Jack2018) also show the process of averaging together different instances of the same class of dynamic expression (https://movie-usa.glencoesoftware.com/video/10.1073/pnas.1807862115/video-2). With these methods Chen et al. (Reference Chen, Crivelli, Garrod, Schyns, Fernández-Dols and Jack2018) show cultural generality in the mental representations of pain and culturally different ‘accents’ in the representations of pleasure (orgasm).

Future illustrations

Darwin's book, The expression of the emotions in Man and animals (1872), was one of the first mass-market books to contain photographs. The publisher John Murray warned Darwin that including the photographs would ‘would poke a terrible hole in the profits’ of the book. Costs have always been a factor to consider when printing illustrations. Over the last few decades, authors have often had to bear the cost of colour printing and reproduction. Two to four colour illustrations could have easily run into thousands of pounds, which has been prohibitive for many. With the rise of online publishing, dynamic and colour images should be more accessible, as there should be no extra cost for reproducing these online.

More specifically, there is no reason why publications cannot include 3D images that have depth through motion (e.g. rigid rotation about one or two axes) or through dynamic internal changes to shape (e.g. Holzleitner & Perrett, Reference Holzleitner and Perrett2016, extras, https://ars.els-cdn.com/content/image/1-s2.0-S1090513815001208-mmc5.mp4). These films (mp4 clips) are currently included in appendices or links but when publishing electronic versions there is no reason not to include the movies within the text – something like the Daily Prophet in the Harry Potter films. While technological development might be needed to allow holographic representation, portrayal of depth can easily be achieved with current technology. Some journals have begun to publish video presentations of methods, for example the Journal of Visual Experiments (https://www.jove.com/), but no journals are integrating videos into text in a manner that can be found in newsfeeds.

All of the expression images described here have been based on images captured from real faces. Computer graphic techniques have made huge progress in building models of faces and synthesising expressions on the models. In many cases an actor may drive the 3D representation as an avatar to produce the correct expression dynamics. To achieve this, landmarks are placed on an actor's face around the eyebrows and mouth, etc. The landmarks are filmed at high speed and resolution in two or three dimensions. The movement trajectory of each landmark can then be used to drive the corresponding landmark on a 2D or 3D model of a face (Tiddeman & Perrett, Reference Tiddeman and Perrett2002; Cao et al., Reference Cao, Weng, Lin and Zhou2013). Automated facial feature finding allows the same to be achieved for actors without them wearing landmarks, with only a slight loss of fidelity. The avatar model can be a different identity, sex, age or species as long as the corresponding landmarks can be found. The avatar can also be a transformed version of the actor, for example the actor depicted with increased age. In effect, this process systematically transfers facial movements and expressions onto the avatar. The avatar retains some of the actor's identity because idiosyncrasies of expression are transposed from the actor to the avatar.

The models of faces driven are increasingly realistic and include skin properties such as albedo, elasticity, dynamic texture and sub-surface light scattering (Weyrich et al., Reference Weyrich, Matusik, Pfister, Bickel, Donner, Tu and Gross2006; Chandran et al., Reference Chandran, Bradley, Gross and Beeler2020; Chen et al., Reference Chen, Garrod, Schyns and Jack2020). Facial representations have passed through the uncanny valley where being approximately correct looks weird (Mori et al., Reference Mori, MacDorman and Kageki2012); we are now in a state in which we can no longer differentiate reality and (re)construction (e.g. https://www.youtube.com/watch?v=HjHiC0mt4Ts). Typing hyper-realistic facial animations into a search engine will show the latest developments (e.g. https://www.youtube.com/watch?v=W_rphISMMzs&list=TLPQMTQxMTIwMjEqrW9Bb9oJAQ&index=3). With realism comes believability and a new problem of ‘deepfakes’. It is relatively straightforward to create an avatar that has a likeness to anyone, be they a celebrity or national leader. Convincing deepfakes are built from a corpus of utterances and expressions made by the target (see https://www.youtube.com/watch?v=p1b5aiTrGzY). The impersonating avatar or deepfake can then be controlled to say or do just about anything, complete with idiosyncratic facial expressions and mannerisms (e.g. https://www.youtube.com/watch?v=ttGUiwfTYvg). Techniques for detecting and mitigating the effects of deepfakes are under development but deepfake videos are already of sufficient quality to attract millions of viewers, many of whom accept their authenticity (Kietzmann et al., Reference Kietzmann, Lee, McCarthy and Kietzmann2020). Such realism has yet to make it through to research on facial expression. When it does, we can expect to see ever greater contextual effects and individual differences in the nuances of interpretating emotions.

One of the characteristics of individuals with schizophrenia (e.g. Jeannerod, Reference Jeannerod2003) is that they are less able to differentiate movements or voices created through their own actions from those created by others (Jeannerod, Reference Jeannerod2003; Johns et al., Reference Johns, Rossell, Frith, Ahmad, Hemsley, Kuipers and McGuire2001). In effect, individuals with schizophrenia are less able to discriminate self from others. While these are clinical symptoms there are parallel problems that could potentially be introduced by technology.

It will not be long before avatars become used as our representatives in work and in social interactions. As explained, technology allows our dynamic facial expressions to be mimicked on an avatar. It is not a big step for the avatar's emotional expression to be caricatured so that a look of mild displeasure may become amplified to outright anger. Conversely, the same expression could be muted so that the avatar's repertoire allows full anger management in the face. Further, one can imagine tweaking the baseline emotional disposition from normal to a cheerier or a more sullen avatar. Such control could allow individuals to be more assertive, or more friendly, than they are in real life.

These advances seem potentially beneficial, yet there are all too obvious dangers. With a new wave of avatars acting and expressing on our behalf, there are likely to be inaccuracies and biases in the sense of self and sense of ideal. Indeed, there is plenty of evidence that, in the domain of body shape, the divergence of people's sense of self and their sense of ideal has been to the detriment of body image, and has encouraged eating disorders, excessive exercise and steroid use (Derenne & Beresin, Reference Derenne and Beresin2006; Harvey & Robinson, Reference Harvey and Robinson2003). The move to expressive avatars representing us in social interactions is likely to bring with it a new set of problems.

Future directions

There has been a plethora of research on the perception of static expression photographs. Work with dynamic expressions is developing (Dawel et al., Reference Dawel, Miller, Horsburgh and Ford2021), but there are many more avenues to explore with dynamic expression stimuli. For example, how do the dynamics of expression affect attribution of personality?

It is important that assessments of expression recognition in clinical groups be made with dynamic facial expressions that are perceived as genuine. Spontaneous genuine (dynamic) expressions are easier to discriminate compared with artificially posed static expressions (Namba et al., Reference Namba, Kabir, Miyatani and Nakao2018). Most of the facial expression stimuli that have been used to date are not perceived as genuine. Thus, our current understanding of how mood, personality and clinical conditions affect the interpretation of emotional expressions in social interactions is sure to need revision.

Dawel et al. (Reference Dawel, Miller, Horsburgh and Ford2021) make a plea for studing natural human expressions made ‘in the wild’ rather than studying synthetic expressions made by avatars. There is a need for study of both since, like it or not, we are going to be experiencing many more synthetic faces in the near future. While it is likely that the use of synthetic expressions could be detrimental, there are also avenues of potential benefit.

Questions such as how many expressions are recognised and what factors (e.g, culture) account for variance in recognition have already been addressed. With new technologies, it is important to address new questions, particularly those concerning the experience of seeing oneself and others making emotional expressions on avatars. Do we feel a different intensity of positive expression when portrayed by a predominantly happy or grumpy avatar? Can the prevailing mood of an avatar we control influence our own mood? If so, this would suggest avenues to explore for therapy of mood disorders. Can avatar expressions be exaggerated or made more typical in real time to benefit individuals with interpretation difficulties (e.g. alexithymia or macular degeneration, Lane et al., Reference Lane, Robbins, Rohan, Crookes, Essex, Maddess and McKone2019)?

A welcome recent trend has been for journals to require that data and stimuli be made freely available from repositories. Expression databases and facial control programs (e.g. KDEF, Lundqvist et al., Reference Lundqvist, Flykt and Öhman1998; The Emotion Recognition Task, Montagne et al., Reference Montagne, Kessels, De Haan and Perrett2007; FEEST, Young et al., Reference Young, Perrett, Calder, Sprengelmeyer and Ekman2002; FaceGen, Todorov et al., Reference Todorov2008; Psychomorph, Tiddeman et al., Reference Tiddeman, Burt and Perrett2001; Sutherland et al., Reference Sutherland, Rhodes and Young2017) have helped research by providing both standardisation and flexibility in use by non-computer scientists. The community needs equivalent openness in access to expressive avatars that can be driven easily and equivalently by different research groups. Such facilities are likely to emerge to service appetites for their use in social media.

Acknowledgements

Thanks to Louise Barrett for guidance on the first draft of the manuscript and Anne Perrett for proof reading.

Author contributions

The manuscript was written by the author

Financial support

There was no financial support for the author.

Conflicts of interest declarations in manuscripts

The author declares that there are no conflicts of interest.

Research transparency and reproducibility

A data availability statement: there are no data accompanying the manuscript. Illustrations have been prepared with Morphanalyser 2.4 https://cherry.dcs.aber.ac.uk/morphanalyser/version2.4/launch2.4.html and Java Psychomorph https://users.aber.ac.uk/bpt/jpsychomorph/

References

Adolphs, R., Tranel, D., Hamann, S., Young, A. W., Calder, A. J., Phelps, E. A., … Damasio, A. R. (1999). Recognition of facial emotion in nine individuals with bilateral amygdala damage. Neuropsychologia, 37(10), 11111117. https://doi.org/10.1016/S0028-3932(99)00039-1CrossRefGoogle ScholarPubMed
Aviezer, H., Hassin, R. R., Ryan, J., Grady, C., Susskind, J., Anderson, A., … Bentin, S. (2008). Angry, disgusted, or afraid? Studies on the malleability of emotion perception. Psychological Science, 19(7), 724732. https://doi.org/10.1111/j.1467-9280.2008.02148.xCrossRefGoogle ScholarPubMed
Axelsson, J., Sundelin, T., Olsson, M. J., Sorjonen, K., Axelsson, C., Lasselin, J., & Lekander, M. (2018). Identification of acutely sick people and facial cues of sickness. Proceedings of the Royal Society B: Biological Sciences, 285(1870), 20172430. https://doi.org/10.1098/rspb.2017.2430Google ScholarPubMed
Barrett, L. F. (2011). Was Darwin wrong about emotional expressions? Current Directions in Psychological Science, 20(6), 400406. https://doi.org/10.1177/0963721411429125CrossRefGoogle Scholar
Barrett, L. F., & Kensinger, E. A. (2010). Context is routinely encoded during emotion perception. Psychological Science, 21(4), 595599. https://doi.org/10.1177/0956797610363547CrossRefGoogle ScholarPubMed
Barrett, L. F., Mesquita, B., & Gendron, M. (2011). Context in emotion perception. Current Directions in Psychological Science, 20(5), 286290. https://doi.org/10.1177/0963721411422522CrossRefGoogle Scholar
Barrett, L. F., Adolphs, R., Marsella, S., Martinez, A. M., & Pollak, S. D. (2019). Emotional expressions reconsidered: Challenges to inferring emotion from human facial movements. Psychological Science in the Public Interest, 20(1), 168. https://doi.org/10.1177/1529100619832930CrossRefGoogle ScholarPubMed
Bell, C., Bourke, C., Colhoun, H., Carter, F., Frampton, C., & Porter, R. (2011). The misclassification of facial expressions in generalised social phobia. Journal of Anxiety Disorders, 25(2), 278283. https://doi.org/10.1016/j.janxdis.2010.10.001CrossRefGoogle ScholarPubMed
Benitez-Quiroz, C. F., Srinivasan, R., & Martinez, A. M. (2018). Facial color is an efficient mechanism to visually transmit emotion. Proceedings of the National Academy of Sciences, 115(14), 35813586. https://doi.org/10.1073/pnas.1716084115CrossRefGoogle ScholarPubMed
Bourke, C., Douglas, K., & Porter, R. (2010). Processing of facial emotion expression in major depression: A review. Australian & New Zealand Journal of Psychiatry, 44(8), 681696. https://doi.org/10.3109/00048674.2010.496359CrossRefGoogle ScholarPubMed
Bowyer, K. W., King, M. C., Scheirer, W. J., & Vangara, K. (2020). The ‘Criminality from face’ illusion. IEEE Transactions on Technology and Society, 1(4), 175183. https://doi.org/10.1109/TTS.2020.3032321CrossRefGoogle Scholar
Cao, C., Weng, Y., Lin, S., & Zhou, K. (2013). 3D shape regression for real-time facial animation. ACM Transactions on Graphics (TOG), 32(4), 110. https://doi.org/10.1145/2461912.2462012CrossRefGoogle Scholar
Chandran, P., Bradley, D., Gross, M., & Beeler, T. (2020). Semantic deep face models. In 2020 international conference on 3D vision (3DV), Fukuoka, Japan (pp. 345354). IEEE. https://doi.org/10.1109/3DV50981.2020.00044CrossRefGoogle Scholar
Chen, C., Crivelli, C., Garrod, O. G., Schyns, P. G., Fernández-Dols, J. M., & Jack, R. E. (2018). Distinct facial expressions represent pain and pleasure across cultures. Proceedings of the National Academy of Sciences, 115(43), E10013E10021. https://doi.org/10.1073/pnas.1807862115CrossRefGoogle ScholarPubMed
Chen, C., Garrod, O. G., Schyns, P. G., & Jack, R. E. (2020, October). Dynamic face movement texture enhances the perceived realism of facial expressions of emotion. In Proceedings of the 20th ACM international conference on intelligent virtual agents (pp. 13). https://doi.org/10.1145/3383652.3423912CrossRefGoogle Scholar
Cook, R., Brewer, R., Shah, P., & Bird, G. (2013). Alexithymia, not autism, predicts poor recognition of emotional facial expressions. Psychological Science, 24(5), 723732. https://doi.org/10.1177/0956797612463582CrossRefGoogle Scholar
Coundouris, S. P., Adams, A. G., Grainger, S. A., & Henry, J. D. (2019). Social perceptual function in Parkinson's disease: A meta-analysis. Neuroscience & Biobehavioral Reviews, 104, 255267. https://doi.org/10.1016/j.neubiorev.2019.07.011CrossRefGoogle ScholarPubMed
Daros, A. R., Zakzanis, K. K., & Ruocco, A. C. (2013). Facial emotion recognition in borderline personality disorder. Psychological Medicine, 43(9), 19531963. http://dx.doi.org/10.1017/S0033291712002607CrossRefGoogle ScholarPubMed
Darwin, C. (1872). The expression of the emotions in Man and animals. John Murray.CrossRefGoogle Scholar
Dawel, A., O'Kearney, R., McKone, E., & Palermo, R. (2012). Not just fear and sadness: Meta-analytic evidence of pervasive emotion recognition deficits for facial and vocal expressions in psychopathy. Neuroscience & Biobehavioral Reviews, 36(10), 22882304. https://doi.org/10.1016/j.neubiorev.2012.08.006CrossRefGoogle ScholarPubMed
Dawel, A., Wright, L., Irons, J., Dumbleton, R., Palermo, R., O'Kearney, R., & McKone, E. (2017). Perceived emotion genuineness: Normative ratings for popular facial expression stimuli and the development of perceived-as-genuine and perceived-as-fake sets. Behavior Research Methods, 49(4), 15391562. https://doi.org/10.3758/s13428-016-0813-2CrossRefGoogle ScholarPubMed
Dawel, A., Miller, E. J., Horsburgh, A., & Ford, P. (2021). A systematic survey of face stimuli used in psychological research 2000–2020. Behavior Research Methods. https://doi.org/10.3758/s13428-021-01705-3CrossRefGoogle Scholar
DeBruine, L. M., Jones, B. C., Unger, L., Little, A. C., & Feinberg, D. R. (2007). Dissociating averageness and attractiveness: Attractive faces are not always average. Journal of Experimental Psychology: Human Perception and Performance, 33(6), 1420. https://doi.org/10.1037/0096-1523.33.6.1420Google Scholar
Derenne, J. L., & Beresin, E. V. (2006). Body image, media, and eating disorders. Academic Psychiatry, 30(3), 257261. https://doi.org/DOI:10.1176/appi.ap.30.3.257CrossRefGoogle ScholarPubMed
Dodge, K. A., Price, J. M., Bachorowski, J. A., & Newman, J. P. (1990). Hostile attributional biases in severely aggressive adolescents. Journal of Abnormal Psychology, 99(4), 385. https://doi.org/10.1037/0021-843X.99.4.385CrossRefGoogle ScholarPubMed
Eack, S. M., Mazefsky, C. A., & Minshew, N. J. (2015). Misinterpretation of facial expressions of emotion in verbal adults with autism spectrum disorder. Autism, 19(3), 308315. https://doi.org/10.1177/1362361314520755CrossRefGoogle ScholarPubMed
Ekman, P. (1972). Universals and cultural differences in facial expressions of emotions. In W. J. Cole (Ed.), Nebraska symposium on motivation (vol. 19, pp. 207283). Lincoln: University of Nebraska Press.Google Scholar
Ekman, P. (2009). Darwin's contributions to our understanding of emotional expressions. Philosophical Transactions of the Royal Society B: Biological Sciences, 364(1535), 34493451. http://doi.org/10.1098/rstb.2009.0189CrossRefGoogle ScholarPubMed
Ekman, P., Friesen, W. V., & Ellsworth, P. (2013). Emotion in the human face: Guidelines for research and an integration of findings (vol. 11), Goldstein, A. P. & Krasner, L. (Eds). Elsevier.Google Scholar
Elfenbein, H. A., & Ambady, N. (2002). On the universality and cultural specificity of emotion recognition: A meta-analysis. Psychological Bulletin, 128(2), 203. https://doi.org/10.1037//0033-2909.128.2.203CrossRefGoogle ScholarPubMed
Fink, B., Grammer, K., & Matts, P. J. (2006). Visible skin color distribution plays a role in the perception of age, attractiveness, and health in female faces. Evolution and Human Behavior, 27(6), 433442. https://doi.org/10.1016/j.evolhumbehav.2006.08.007CrossRefGoogle Scholar
Foster, E. D., & Deardorff, A. (2017). Open science framework (OSF). Journal of the Medical Library Association: JMLA, 105(2), 203. https://doi.org/10.5195/jmla.2017.88CrossRefGoogle Scholar
Frigerio, E., Burt, D. M., Montagne, B., Murray, L. K., & Perrett, D. I. (2002). Facial affect perception in alcoholics. Psychiatry Research, 113(1–2), 161171. https://doi.org/10.1016/S0165-1781(02)00244-5CrossRefGoogle ScholarPubMed
Galton, F. (1878). Composite portraits. Nature 18, 97100.Google Scholar
Galton, F. (1879). Composite portraits, made by combining those of many different persons into a single resultant figure. The Journal of the Anthropological Institute of Great Britain and Ireland, 8, 132144.CrossRefGoogle Scholar
Galton, F. (1883). Inquiries into human faculty and its development. Macmillan.CrossRefGoogle Scholar
Galton, F. (1907). Inquiries into human faculty and its development (2nd ed.), J.M. Dent.Google Scholar
Gendron, M., Crivelli, C., & Barrett, L. F. (2018). Universality reconsidered: Diversity in making meaning of facial expressions. Current Directions in Psychological Science, 27(4), 211219. https://doi.org/10.1177/0963721417746794CrossRefGoogle ScholarPubMed
Green, D. (1984) Veins of resemblance: Photography and eugenics. Oxford Art Journal, 7(2), 316 https://doi.org/10.1093/oxartj/7.2.3CrossRefGoogle Scholar
Harvey, J. A., & Robinson, J. D. (2003). Eating disorders in men: Current considerations. Journal of Clinical Psychology in Medical Settings, 10(4), 297306. https://doi.org/10.1023/A:1026357505747CrossRefGoogle Scholar
Hastings, M. E., Tangney, J. P., & Stuewig, J. (2008). Psychopathy and identification of facial expressions of emotion. Personality and Individual Differences, 44(7), 14741483. https://doi.org/10.1016/j.paid.2008.01.004CrossRefGoogle ScholarPubMed
Holzleitner, I. J., & Perrett, D. I. (2016). Perception of strength from 3D faces is linked to facial cues of physique. Evolution and Human Behavior, 37(3), 217229. https://doi.org/10.1016/j.evolhumbehav.2015.11.004 https://ars.els-cdn.com/content/image/1-s2.0-S1090513815001208-mmc5.mp4CrossRefGoogle Scholar
Holzleitner, I. J., Hunter, D. W., Tiddeman, B. P., Seck, A., Re, D. E., & Perrett, D. I. (2014). Men's facial masculinity: When (body) size matters. Perception, 43(11), 11911202. https://doi.org/10.1068/p7673CrossRefGoogle ScholarPubMed
Jack, R. E., Garrod, O. G., Yu, H., Caldara, R., & Schyns, P. G. (2012). Facial expressions of emotion are not culturally universal. Proceedings of the National Academy of Sciences, 109(19), 72417244. https://doi.org/10.1073/pnas.1200155109CrossRefGoogle Scholar
Jack, R. E., Sun, W., Delis, I., Garrod, O. G., & Schyns, P. G. (2016). Four not six: Revealing culturally common facial expressions of emotion. Journal of Experimental Psychology: General, 145(6), 708. http://dx.doi.org/10.1037/xge0000162Google Scholar
Jeannerod, M. (2003). The mechanism of self-recognition in humans. Behavioural Brain Research, 142(1–2), 115. https://doi.org/10.1016/S0166-4328(02)00384-4CrossRefGoogle ScholarPubMed
Johns, L. C., Rossell, S., Frith, C., Ahmad, F., Hemsley, D., Kuipers, E., & McGuire, P. K. (2001). Verbal self-monitoring and auditory verbal hallucinations in patients with schizophrenia. Psychological Medicine, 31(4), 705715. https://doi.org/10.1017/S0033291701003774CrossRefGoogle ScholarPubMed
Jones, B. C., Little, A. C., Burt, D. M., & Perrett, D. I. (2004). When facial attractiveness is only skin deep. Perception, 33(5), 569576. https://doi.org/10.1068/p3463CrossRefGoogle ScholarPubMed
Julesz, B. (1971). Foundations of cyclopean perception. University of Chicago Press.Google Scholar
Keltner, D., Sauter, D., Tracy, J., & Cowen, A. (2019). Emotional expression: Advances in basic emotion theory. Journal of Nonverbal Behavior, 43(2) 133160. https://doi.org/10.1007/s10919-019-00293-3CrossRefGoogle ScholarPubMed
Kietzmann, J., Lee, L. W., McCarthy, I. P., & Kietzmann, T. C. (2020). Deepfakes: Trick or treat? Business Horizons, 63(2), 135146. https://doi.org/10.1016/j.bushor.2019.11.006CrossRefGoogle Scholar
Kohler, C. G., Walker, J. B., Martin, E. A., Healey, K. M., & Moberg, P. J. (2010). Facial emotion perception in schizophrenia: A meta-analytic review. Schizophrenia Bulletin, 36(5), 10091019. https://doi.org/10.1093/schbul/sbn192CrossRefGoogle ScholarPubMed
Kornreich, C., Philippot, P., Foisy, M. L., Blairy, S., Raynaud, E., Dan, B., … Verbanck, P. (2002). Impaired emotional facial expression recognition is associated with interpersonal problems in alcoholism. Alcohol and Alcoholism, 37(4), 394400. https://doi.org/10.1093/alcalc/37.4.394CrossRefGoogle ScholarPubMed
Krause, F. C., Linardatos, E., Fresco, D. M., & Moore, M. T. (2021). Facial emotion recognition in major depressive disorder: A meta-analytic review. Journal of Affective Disorders, 293, 320328. https://doi.org/10.1016/j.jad.2021.06.053CrossRefGoogle ScholarPubMed
Lane, J., Robbins, R. A., Rohan, E. M., Crookes, K., Essex, R. W., Maddess, T., … McKone, E. (2019). Caricaturing can improve facial expression recognition in low-resolution images and age-related macular degeneration. Journal of Vision, 19(6), 1818. https://doi.org/10.1167/19.6.18CrossRefGoogle ScholarPubMed
Langlois, J. H., & Roggman, L. A. (1990). Attractive faces are only average. Psychological Science, 1(2), 115121. https://doi.org/10.1111/j.1467-9280.1990.tb00079.xCrossRefGoogle Scholar
Law-Smith, M. J. L., Montagne, B., Perrett, D. I., Gill, M., & Gallagher, L. (2010). Detecting subtle facial emotion recognition deficits in high-functioning autism using dynamic stimuli of varying intensities. Neuropsychologia, 48(9), 27772781. https://doi.org/10.1016/j.neuropsychologia.2010.03.008CrossRefGoogle ScholarPubMed
Levy, D. M., & Peart, S. J. (2004). Statistical prejudice: From eugenics to immigrants. European Journal of Political Economy, 20(1), 522. https://doi.org/10.1016/j.ejpoleco.2003.01.003CrossRefGoogle Scholar
Little, A. C., & Hancock, P. J. (2002). The role of masculinity and distinctiveness in judgments of human male facial attractiveness. British Journal of Psychology, 93(4), 451464. https://doi.org/10.1348/000712602761381349CrossRefGoogle ScholarPubMed
Lobmaier, J. S., & Perrett, D. I. (2011). The world smiles at me: Self-referential positivity bias when interpreting direction of attention. Cognition and Emotion, 25(2), 334341. https://doi.org/10.1080/02699931003794557CrossRefGoogle ScholarPubMed
Lundqvist, D., Flykt, A., & Öhman, A. (1998). The Karolinska directed emotional faces (KDEF). CD ROM from Department of Clinical Neuroscience, Psychology section, Karolinska Institutet, 91(630), 2.Google Scholar
Meeren, H. K., van Heijnsbergen, C. C., & de Gelder, B. (2005). Rapid perceptual integration of facial expression and emotional body language. Proceedings of the National Academy of Sciences, 102(45), 1651816523. https://doi.org/10.1073/pnas.0507650102CrossRefGoogle ScholarPubMed
Montagne, B., van Honk, J., Kessels, R. P., Frigerio, E., Burt, M., van Zandvoort, M. J., … de Haan, E. H. (2005). Reduced efficiency in recognising fear in subjects scoring high on psychopathic personality characteristics. Personality and Individual Differences, 38(1), 511. https://doi.org/10.1016/j.paid.2004.02.008CrossRefGoogle Scholar
Montagne, B., Kessels, R. P., De Haan, E. H., & Perrett, D. I. (2007). A paradigm to measure the perception of facial emotional expressions at different intensities. Perceptual and Motor Skills, 104(2), 589598. https://doi.org/10.2466/pms.104.2.589-598CrossRefGoogle ScholarPubMed
Mori, M., MacDorman, K. F., & Kageki, N. (2012). The uncanny valley [from the field]. IEEE Robotics & Automation Magazine, 19(2), 98100. https://doi.org/10.1109/MRA.2012.2192811CrossRefGoogle Scholar
Morris, J. S., Frith, C. D., Perrett, D. I., Rowland, D., Young, A. W., Calder, A. J., & Dolan, R. J. (1996). A differential neural response in the human amygdala to fearful and happy facial expressions. Nature, 383(6603), 812815. https://doi.org/10.1038/383812a0CrossRefGoogle ScholarPubMed
Namba, S., Kabir, R. S., Miyatani, M., & Nakao, T. (2018). Dynamic displays enhance the ability to discriminate genuine and posed facial expressions of emotion. Frontiers in Psychology, 9, 672. https://doi.org/10.3389/fpsyg.2018.00672CrossRefGoogle ScholarPubMed
Pantic, M., & Rothkrantz, L. J. M. (2000). Automatic analysis of facial expressions: The state of the art. IEEE Transactions on Pattern Analysis and Machine Intelligence, 22(12), 14241445. https://doi.org/10.1109/34.895976CrossRefGoogle Scholar
Penton-Voak, I. S., Thomas, J., Gage, S. H., McMurran, M., McDonald, S., & Munafò, M. R. (2013). Increasing recognition of happiness in ambiguous facial expressions reduces anger and aggressive behavior. Psychological Science, 24(5), 688697. https://doi.org/10.1177/0956797612459657CrossRefGoogle ScholarPubMed
Perrett, D. I., May, K. A., & Yoshikawa, S. (1994). Facial shape and judgements of female attractiveness. Nature, 368(6468), 239242. https://doi.org/10.1038/368239a0CrossRefGoogle ScholarPubMed
Perrett, D. I., Lee, K. J., Penton-Voak, I., Rowland, D., Yoshikawa, S., Burt, D. M., … Akamatsu, S. (1998). Effects of sexual dimorphism on facial attractiveness. Nature, 394(6696), 884887. https://doi.org/10.1038/29772CrossRefGoogle ScholarPubMed
Phillips, M. L., Young, A. W., Senior, C., Brammer, M., Andrew, C., Calder, A. J., … David, A. S. (1997). A specific neural substrate for perceiving facial expressions of disgust. Nature, 389(6650), 495498. https://doi.org/10.1038/39051CrossRefGoogle ScholarPubMed
Russell, J. A. (1994). Is there universal recognition of emotion from facial expression? A review of the cross-cultural studies. Psychological Bulletin, 115(1), 102141. https://doi.org/10.1037/0033-2909.115.1.102CrossRefGoogle ScholarPubMed
Sauter, D. A., & Eisner, F. (2013). Commonalities outweigh differences in the communication of emotions across human cultures. Proceedings of the National Academy of Sciences, 110(3), E180E180. https://doi.org/10.1073/pnas.1209522110CrossRefGoogle ScholarPubMed
Schulze, L., Lobmaier, J. S., Arnold, M., & Renneberg, B. (2013). All eyes on me?! Social anxiety and self-directed perception of eye gaze. Cognition & Emotion, 27(7), 13051313. https://doi.org/10.1080/02699931.2013.773881CrossRefGoogle ScholarPubMed
Sprengelmeyer, R., Young, A. W., Calder, A. J., Karnat, A., Lange, H., Hömberg, V., … Rowland, D. (1996). Loss of disgust: Perception of faces and emotions in Huntington's disease. Brain, 119(5), 16471665. https://doi.org/10.1093/brain/119.5.1647CrossRefGoogle ScholarPubMed
Sutherland, C. A., & Young, A. W. (2015). A basic guide to Psychomorph. https://www.researchgate.net/publication/297124784_A_basic_guide_to_PsychomorphGoogle Scholar
Sutherland, C. A., Oldmeadow, J. A., Santos, I. M., Towler, J., Burt, D. M., & Young, A. W. (2013). Social inferences from faces: Ambient images generate a three-dimensional model. Cognition, 127(1), 105118. https://doi.org/10.1016/j.cognition.2012.12.001CrossRefGoogle ScholarPubMed
Sutherland, C. A., Rhodes, G., & Young, A. W. (2017). Facial image manipulation: A tool for investigating social perception. Social Psychological and Personality Science, 8(5), 538551. https://doi.org/10.1177/1948550617697176CrossRefGoogle Scholar
Tatarunaite, E., Playle, R., Hood, K., Shaw, W., & Richmond, S. (2005). Facial attractiveness: A longitudinal study. American Journal of Orthodontics and Dentofacial Orthopedics, 127(6), 676682. https://doi.org/10.1016/j.ajodo.2004.01.029CrossRefGoogle ScholarPubMed
Tiddeman, B., Burt, M., & Perrett, D. (2001). Prototyping and transforming facial textures for perception research. IEEE Computer Graphics and Applications, 21(5), 4250. https://doi.org/10.1109/38.946630CrossRefGoogle Scholar
Tiddeman, B., & Perrett, D. (2002). Transformation of dynamic facial image sequences using static 2D prototypes. The Visual Computer, 18(4), 218225. https://doi.org/10.1007/s003710100142CrossRefGoogle Scholar
Todorov, A. (2008). Evaluating faces on trustworthiness: An extension of systems for recognition of emotions signaling approach/avoidance behaviors. Annals of the New York Academy of Sciences, 1124(1), 208224. https://doi.org/10.1196/annals.1440.012CrossRefGoogle ScholarPubMed
Todorov, A., Said, C. P., Engell, A. D., & Oosterhof, N. N. (2008). Understanding evaluation of faces on social dimensions. Trends in Cognitive Sciences, 2(12), 455460. https://doi.org/10.1016/j.tics.2008.10.001CrossRefGoogle Scholar
Venn, H. R., Gray, J. M., Montagne, B., Murray, L. K., Burt, M. D., Frigerio, E., … Young, A. H. (2004). Perception of facial expressions of emotion in bipolar disorder. Bipolar Disorders, 6(4), 286293. https://doi.org/10.1111/j.1399-5618.2004.00121.xCrossRefGoogle ScholarPubMed
Wang, K., Hoosain, R., Yang, R. M., Meng, Y., & Wang, C. Q. (2003). Impairment of recognition of disgust in Chinese with Huntington's or Wilson's disease. Neuropsychologia, 41(5), 527537. https://doi.org/10.1016/S0028-3932(02)00171-9CrossRefGoogle ScholarPubMed
Weyrich, T., Matusik, W., Pfister, H., Bickel, B., Donner, C., Tu, C., … Gross, M. (2006). Analysis of human faces using a measurement-based skin reflectance model. ACM Transactions on Graphics (ToG), 25(3), 10131024. https://doi.org/10.1145/1141911.1141987CrossRefGoogle Scholar
Wilson, E. O. (Ed) (2006). From so simple a beginning: The four great books of Charles Darwin. WW Norton.Google Scholar
y Arcas, B. A., Mitchell, M., & Todorov, A. (2017). Physiognomy's new clothes. Medium (6 May). https://medium.com/@blaisea/physiognomys-new-clothes-f2d4b59fdd6aGoogle Scholar
Young, A. W., Rowland, D., Calder, A. J., Etcoff, N. L., Seth, A., & Perrett, D. I. (1997). Facial expression megamix: Tests of dimensional and category accounts of emotion recognition. Cognition, 63(3), 271313. https://doi.org/10.1016/S0010-0277(97)00003-6CrossRefGoogle ScholarPubMed
Young, A. W., Perrett, D. I., Calder, A. J., Sprengelmeyer, R., Ekman, P. (2002) Facial expressions of emotion: Stimuli and tests (FEEST). Thames Valley Test Company.Google Scholar
Zebrowitz, L. A., Kikuchi, M., & Fellous, J.-M. (2010). Facial resemblance to emotions: Group differences, impression effects, and race stereotypes. Journal of Personality and Social Psychology, 98(2), 175189. https://doi.org/10.1037/a0017990CrossRefGoogle ScholarPubMed
Zhang, D., Lin, H., & Perrett, D. I. (2020). Apparent emotional expression explains the effects of head posture on perceived trustworthiness and dominance, but a measure of facial width does not. Perception, 49(4), 422438. https://doi.org/10.1177/0301006620909286CrossRefGoogle ScholarPubMed
Figure 0

Figure 1. Images of health and sickness from the nineteenth and twenty-first centuries. (a) Galton's composite photographs of ‘health’ – a combination of 23 Royal Engineers, and ‘sickness’ – combinations of six and nine cases of tubercular disease (Galton, 1883). (b) Composite images of 22 individuals 2 hours after an injection of a placebo (left) or a bacterial endotoxin (right). Note the subtle change in expression after the toxin. Reproduced from Axelsson et al. (2018), Proceedings of the Royal Society B: Biological Sciences published under creative commons. Permissions for reproduction were obtained from https://www.copyright.com/. The author's permission was provided by email.

Figure 1

Figure 2. Composite faces posing happy, afraid and disgust expressions. Row 1: a composite of images of 35 males posing with neutral expression (0%) and a happy expression (100%). The differences in shape, colour and texture between the neutral and happy face images are used to transform the neutral image in 25% steps. This creates images where the happy expression gradually emerges in intensity. The 125 and 150% images represent an extrapolation of the series with caricature exaggeration of the happy expression by 25 and 50%. Row 2: as row 1 for the expression of fear. Row 3: as row 1 for the expression of disgust. Composite images and emotion transforms produced by the author.

Figure 2

Figure 3. Happy to angry facial expression continuum. Five steps are illustrated progressing from 100% happy to 100% angry. The central image is ambiguous showing both characteristics of happiness and of anger. The upper part of the figure illustrates the categorical boundary between the images being categorised as angry or happy before training (see text). The lower section illustrates that, post training, the boundary is shifted such that more ambiguous expressions are classified as happy. Reproduced from figure 1 in Penton-Voak et al. (2013) Psychological Science, 24, 688–697 with permission from the author.

Figure 3

Figure 4. A face with and without additional diagnostic colour information for the emotion of happiness. With the augmented colour information, the images were easier to classify as happy. Reproduced under Creative Commons License cropping the original image to show only the face pair from figure S6 from Benitez-Quiroz et al. (2018) Proceedings of the National Academy of Sciences, 115, 3581–3586. With permission from the author.

Figure 4

Figure 5. 3D faces varying in apparent trustworthiness. Frontal and half profile views of male and female 3D head models varying in apparent personality. The head models were constructed by averaging together the 3D surface shape and texture of male and female faces separately (middle row). A collection of 118 faces (male = 50, female = 68) were rated for how trustworthy they looked while being rotated to reveal their 3D structure. For each gender, an average 3D head shape was formed from those faces that appeared high in trustworthiness. Separately an average was formed from those that appeared low in trustworthiness. These two averages defined a trustworthiness trajectory in 3D shape space for men and for women. Male and female composite faces were then transformed in shape along this trajectory to decrease apparent trustworthiness (top row) or to increase apparent trustworthiness (bottom row). Methods for averaging and transforming have been presented elsewhere (Holzleitner et al., 2014). 3D head models and apparent trait transforms models produced by the author.

Figure 5

Figure 6. Disgust expression modified by context. An isolated facial expression of disgust was placed in a ‘disgust’ context (left) or in a ‘pride’ context (right). While the disgust expression was accurately categorised as negatively valanced in the disgust context, it was never categorised as having a negative valence in the pride context. Reproduced from figure 4a, Aviezer et al., (2008) Psychological Science, 19, 724–732 with permission from the author. Permissions for reproduction were obtained from https://www.copyright.com/. The author's permission was provided by email.

Figure 6

Figure 7. Comparing the representation of three expressions for one European (left) and one Chinese participant (right). The mouth region is more informative for the European and the eye region is more informative for the Chinese participant. Reproduced from Movie S2, Jack et al. (2012) Proceedings of the National Academy of Sciences, 109, 7241–7244 with permission from the author. For the dynamic movie see http://www.pnas.org/lookup/suppl/doi:10.1073/pnas.1200155109/-/DCSupplemental/sm02.avi.