Abstract
There is growing concern that artificial intelligence conversational agents (e.g., Siri, Alexa) reinforce voice-based social stereotypes. Because little is known about social perceptions of conversational agents’ voices, we investigated (1) the dimensions that underpin perceptions of these synthetic voices and (2) the role that acoustic parameters play in these perceptions. Study 1 (N = 504) found that perceptions of synthetic voices are underpinned by Valence and Dominance components similar to those previously reported for natural human stimuli and that the Dominance component was strongly and negatively related to voice pitch. Study 2 (N = 160) found that experimentally manipulating pitch in synthetic voices directly influenced dominance-related, but not valence-related, perceptions. Collectively, these results suggest that greater consideration of the role that voice pitch plays in dominance-related perceptions when designing conversational agents may be an effective method for controlling stereotypic perceptions of their voices and the downstream consequences of those perceptions.
Similar content being viewed by others
Introduction
Perceptions of human faces and voices influence important social outcomes1,2. For example, people prefer to date, mate with, hire, and vote for individuals perceived as being particularly attractive1,2. These social perceptions (e.g., our impressions of other people’s attractiveness or trustworthiness) can also influence life-or-death outcomes. For example, untrustworthy-looking defendants are more likely to receive death sentences3.
Several recent studies suggest that social perceptions are underpinned by two key dimensions. For example, Principal Component Analysis (PCA) of ratings of natural human faces on a variety of different traits on which people spontaneously assess faces (trustworthiness, emotional stability, responsibility, sociability, caringness, attractiveness, intelligence, confidence, weirdness, unhappiness, meanness, aggressiveness, dominance) reveals two key underlying dimensions4,5. The first of these dimensions, often labelled valence, is highly correlated with ratings of pro-social traits, such as trustworthiness, and is thought to reflect perceptions of an individual’s willingness to inflict harm on others4,5. The second of these dimensions, often labelled dominance, is highly correlated with traits such as dominance and aggressiveness and is thought to reflect perceptions of an individual’s capacity to inflict harm on others4,5. Comparable dimensions have also been observed in work that used similar data-reduction methods to reveal the dimensions that underpin social perceptions of natural human voices6,7.
Conversational agents (e.g., Siri and Alexa) have become ubiquitous and communicate with users via voices generated by artificial intelligence algorithms (i.e., synthetic voices). A recent report by UNESCO (United Nations Educational, Scientific and Cultural Organization) proposed that these synthetic voices have the potential to reinforce voice-based gender stereotypes and argued that research is urgently needed to ameliorate this issue8. Crucially, understanding how people perceive synthetic voices is necessary to both understand and address this issue. While there is a large literature examining the factors that influence perceptions of natural social stimuli (natural human faces and voices), it is unclear whether our understanding of the factors that shape perceptions of natural social stimuli is applicable to perceptions of the synthetic voices employed by conversational agents. Indeed, previous work has shown qualitative differences in how people process and perceive natural human faces and synthetic faces9,10. Such results suggest that the factors that underpin perceptions of natural human stimuli may not necessarily underpin perceptions of synthetic stimuli. Indeed, some previous research suggests that natural human voices are perceived to be more expressive, understandable, and likeable than synthetic voices, suggesting that people can distinguish between natural human voices and synthetic voices and may process them differently11. However, other work suggests that the expectation that synthetic voices possess robotic qualities can also cause listeners to have difficulty when they are asked to classify voices as natural or artificial12. In light of these points, Study 1 first investigated whether trait-ratings of the synthetic voices used by conversational agents are underpinned by valence and dominance dimensions similar to those observed for natural human stimuli in previous work4,5.
Work on social perceptions of natural human voices has suggested that acoustic properties of voices can be key predictors and determinants of voice perceptions13,14. The majority of this work has focused on fundamental frequency (the acoustic correlate of perceived pitch) and formant frequencies (a correlate of vocal tract length and body size,13). For example, studies have reported strong negative relationships between dominance perceptions and voice pitch and/or that voices in which pitch was lowered are perceived to be more dominant than voices in which pitch was raised13,14. In light of findings such as these, Study 1 also investigated possible relationships between the dimensions that underpin social perceptions of synthetic voices and both pitch and formant frequencies.
Study 2 built directly on our results of Study 1 by testing whether Study 1’s results for correlations between social perceptions and measured acoustic properties of synthetic voices also occur when we experimentally manipulated acoustic parameters of voices. Whereas stimuli in Study 1 were individual words, stimuli in Study 2 were full sentences.
Collectively, these studies may be an important first step in identifying how appropriate the large literature on perceptions of natural social stimuli is to understand how we perceive and interact with the type of synthetic voices employed by artificial conversational agents. They may also identify mechanisms through which designers of conversational agents can better control stereotypic perceptions of conversational agents’ voices and the downstream consequences of these perceptions.
Study 1
Study 1 aimed to (1) identify the perceptual dimensions that underpin social judgments of synthetic voices and (2) investigate how these perceptual dimensions are related to voice pitch and formant frequencies.
Methods
Ethics
All procedures were approved by the School of Psychological Sciences and Health (University of Strathclyde) Ethics Commitee, all work was undertaken in accordance with the Declaration of Helsinki, and all participants provided informed consent.
Stimuli
Forty-six synthetic (i.e., computer-generated) voices (17 male, 29 female) were downloaded from online voice synthesis platforms (28 voices) or were provided to us by companies working on synthetic voice production (18 voices). Because studies that measured pitch and formant frequencies from recordings of natural human voices have generally used /α/, /ε/, /i/, /o/, and /u/ vowel sounds13, the voices used in the current study spoke the words ‘Father, See, Bet, Note, Boot’. All stimuli were amplitude normalized to 70 dB prior to both acoustic analyses and collection of trait ratings. Voices spoke in English and reflected a range of accents. Stimuli are publicly available at https://osf.io/4zgrf/.
Trait-rating procedure
Five hundred and six participants (213 men, 291 women, two participants did not report their gender; mean age = 34.0 years, standard deviation = 11.0 years) were recruited through the Prolific participant recruitment platform. All participants reported having English as their first language.
Participants were randomly allocated to rate all 46 voice stimuli for one of 17 traits (trustworthy, emotionally stable, responsible, sociable, caring, attractive, intelligent, confident, weird, unhappy, mean, aggressive, dominant, competent, old, masculine, feminine) on a 1 (not at all) to 7 (very) scale. The order in which stimuli were presented for rating was fully randomized and participants could play each voice as many times as they wanted to before rating it. Participants had to click a play button to play each voice and could not rate the voice and proceed to the next trial until the voice recording had been played in full. Ratings were made by clicking on labelled buttons, preventing participants from entering invalid responses. Mean stimulus duration was 3.42 s (SD = 0.42 s). Traits were chosen to reflect those used in previous studies of the dimensions underpinning perceptions of social stimuli (trustworthy, emotionally stable, responsible, sociable, caring, attractive, intelligent, confident, weird, unhappy, mean, aggressive, dominant, old,4,5,7). Masculinity and femininity were also included as separate traits, rather than being treated as opposite ends of a single continuum, because of recent work suggesting that they can have independent effects on perceptions of social stimuli and that including these characteristics as separate traits increased the predictive accuracy of models of social perception15. Competence was included because of research suggesting it plays an important role in a range of social outcomes (see, e.g.,16 for a recent review of this literature). Age, masculinity, and femininity were included because they had been included in previous research that used similar data-reduction methods to identify the perceptual dimensions underpinning social judgments7,17. Words spoken by each voice were presented in a single file for rating.
Table 1 shows the number of raters who rated stimuli on each trait, inter-rater agreement for those ratings (indicated by Cronbach’s alpha), the mean rating for each trait, and standard deviation. Because inter-rater agreement in ratings was relatively high for all traits, we calculated the mean rating for each voice. These mean ratings were calculated separately for each trait and were used in subsequent analyses. All data and analysis code are publicly available at https://osf.io/4zgrf/.
Acoustic analyses
Vowel sounds were extracted from each voice recording and analyzed using PRAAT18. The methods used to measure pitch (f0) and formant frequencies were identical to those used to measure acoustic properties of natural speech in recent work on social judgments of human voices13. f0, f1, f2, f3, and f4 were measured separately from each vowel sound. Next, all measures were converted to z-scores. Standardized f0, averaged across vowel sounds, was then used as our measure of pitch in our statistical analyses. The remaining standardized measures (f1, f2, f3, and f4) were first averaged for each vowel sound and then averaged across vowel sounds. This latter score was used as our measure of formants in our statistical analyses.
Results
All analyses were carried out using R19 and the packages tidyverse 1.3.020, readxl 1.3.121, psych 2.0.1222, paran 1.5.223, kableExtra 1.3.424, knitr 1.3125, and jtools 2.1.326. All data and analysis code are publicly available at https://osf.io/4zgrf/.
First, mean trait ratings for each voice were subject to Principal Component Analysis (PCA) with oblimin rotation. This analysis revealed two Principal Components (PCs), explaining 45 and 29% of the variance in ratings, respectively. Factor loadings of the individual traits on both PCs are shown in Table 2.
Consistent with previous research using natural human voices and faces as stimuli4,5,7, the first PC was highly correlated with pro-social traits, such as trustworthiness, competence, responsibility, emotional stability, and sociableness, but weakly correlated with dominance and aggressiveness, and the second PC was highly correlated with dominance and aggressiveness, but weakly correlated with trustworthiness, competence, responsibility, emotional stability, and sociableness. Following previous research showing this pattern of results, we labelled these PCs Valence and Dominance, respectively. There was a non-significant, very weak, positive correlation between scores on both these components (r(46) = 0.07, p = 0.655).
Next, we used regression analyses to investigate possible relationships between PC scores and the acoustic characteristics pitch and formants. Separate models were run for Valence and Dominance PC scores. PC scores were our outcome variable. Predictors were voice gender (effect coded so that − 0.5 = male and 0.5 = female), pitch, formants, all two-way interactions, and the three-way interaction.
Full results from these regression analyses are summarized in Table 3 (Valence PC scores) and Table 4 (Dominance PC scores). For Valence, the regression model was not significant (F(7,38) = 1.04, adjusted R squared = 0.01, p = 0.42) and the model showed no significant effects for any of the predictors (Table 3). For Dominance, the regression model was significant (F(7, 38) = 21.8, adjusted R squared = 0.76, p < 0.001) and a strong significant negative effect of voice pitch was observed (Table 4). Figure 1 shows the negative relationship between pitch and Dominance PC scores. Pitch explained 74% of the variance in Dominance PC scores.
Robustness checks in which ratings of male and female voices were analyzed separately showed the same patterns of results as the analyses described above. That is, separate PCAs of ratings of male and female voices revealed Valence and Dominance PCs that were similar to those we obtained from PCA of all voices. Further analyses also showed that, for both male and female voices, pitch was strongly and negatively related to scores on the Dominance PC. Full results for these analyses are given in the supplemental analyses for Study 1 (https://osf.io/4zgrf/).
Study 2
Study 1 demonstrated that trait-ratings of synthetic voices are underpinned by Valence and Dominance components similar to those previously reported for perceptions of natural human stimuli. Study 1 also found a strong negative relationship between pitch and scores on the Dominance PC. By contrast, there was no significant relationship between pitch and scores on the Valence PC. Study 2 attempted to validate the results of Study 1 by investigating the effects of manipulating voice pitch on perceptions of dominance and aggressiveness (i.e., the two individual traits most strongly correlated with Dominance PC scores in Study 1) and perceptions of trustworthiness and competence (i.e., the two individual traits most strongly correlated with Valence PC scores in Study 1). Whereas voices in Study 1 spoke individual words, stimuli in Study 2 spoke a full sentence commonly used in studies of social judgments of natural voices that have used full sentences as stimuli (“When the sunlight strikes raindrops in the air, they act as a prism and form a rainbow.”).
Methods
Stimuli
Stimuli were manufactured from recordings of the same 46 voices used in Study 1, this time speaking the sentence “When the sunlight strikes raindrops in the air, they act as a prism and form a rainbow”. This sentence was chosen because it has been used in previous studies of social judgments of natural voices27,28.
Two versions of each recording were created; one version (raised-pitch version) in which the pitch of the recording was raised by 0.5 equivalent rectangular bandwidths (ERBs) and another version (lowered-pitch version) in which the pitch of the recording had been lowered by 0.5 ERBs. The ERB scale corrects for the difference between perceived pitch and actual fundamental frequency. Mean pitch for the lowered-pitch versions was 142.06 Hz (SD = 8.71 Hz) and mean pitch for the raised-pitch versions was 184.83 Hz (SD = 9.09 Hz). Pitch was manipulated using the same methods used in previous studies of the effect of manipulating pitch on social judgments of natural voices29,30,31. Only the raised- and lowered-pitch versions were used in the study. All stimuli were amplitude normalized to 70 dB prior to collection of trait ratings. Stimuli are publicly available at https://osf.io/4zgrf/. Voices spoke in English and reflected a range of accents.
Trait-rating procedure
One hundred and sixty participants (59 men, 93 women, eight participants did not report their gender; mean age = 31.5 years, standard deviation = 11.6 years) were recruited through the Prolific participant recruitment platform. All participants reported having English as their first language.
Participants were randomly allocated to rate all 92 voice stimuli (the 46 raised-pitch and 46 lowered-pitch versions) for either competence (36 raters), trustworthiness (40 raters), dominance (44 raters), or aggressiveness (40 raters) on a 1 (not at all) to 7 (very) scale. The order in which stimuli were presented for rating was fully randomized and participants could play each voice as many times as they wanted to before rating it. Participants had to click a play button to play each voice and could not rate the voice and proceed to the next trial until the voice recording had been played in full. Ratings were made by clicking on labelled buttons, preventing participants from entering invalid responses. Mean stimulus duration was 5.40 s (SD = 0.42 s). Inter-rater agreement was high for all traits (all Cronbach’s alpha > 0.93).
Results
All analyses were carried out using R19 and the packages tidyverse 1.3.020, broom.mixed 0.2.632, lmerTest 3.1-333, psych 2.0.1222, and jtools 2.1.326. All data and analysis code are publicly available at https://osf.io/4zgrf/.
We used linear mixed models to investigate possible effects of the pitch manipulation on competence, trustworthiness, dominance, and aggressiveness ratings. Separate models were run for each trait. Ratings were the outcome variables and predictors were voice gender (effect coded so that − 0.5 = male and 0.5 = female), pitch manipulation (effect coded so that lowered pitch = − 0.5 and raised pitch = 0.5), and the interaction between voice gender and pitch manipulation. The models included random intercepts for participant and stimulus. Random slopes were specified maximally.
Results of these analyses are summarized in Table 5 (competence ratings), Table 6 (trustworthiness ratings), Table 7 (dominance ratings), and Table 8 (aggressiveness ratings). While our analyses of competence and trustworthiness ratings showed no significant effect of pitch manipulation, our analyses of dominance and aggressiveness ratings showed that the lowered-pitch versions of voices were judged significantly more dominant and aggressive than the raised-pitch versions. Figure 2 shows the negative effects of the pitch manipulation on dominance and aggressiveness perceptions.
Discussion
PCA of ratings of synthetic voices used by conversational agents on a range of traits (trustworthiness, emotional stability, responsibility, sociability, caringness, attractiveness, intelligence, confidence, weirdness, unhappiness, meanness, aggressiveness, dominance, competence, age, masculinity, femininity) produced two components. The first component, which explained the bulk of the variance in ratings, was highly correlated with ratings of pro-social traits, such as trustworthiness, competence, responsibility, emotional stability, and sociable, and weakly correlated with dominance and aggressiveness ratings. The second component, which explained substantially less of the variance in ratings, was highly correlated with dominance and aggressiveness ratings and weakly correlated with trustworthiness, competence, responsibility, emotional stability, and sociable ratings. This pattern of results is extremely similar to those obtained when ratings of natural human faces and voices were subject to PCA in previous studies4,5,6,7, suggesting that social perceptions of synthetic voices are underpinned by valence and dominance dimensions similar to those previously found to underpin social perceptions of natural human stimuli.
Analyses of acoustic properties of synthetic voices and their relationship to the PCs in Study 1 revealed a strong negative correlation between pitch and scores on the Dominance component. Consistent with results of this correlational analysis, experimentally manipulating the pitch of synthetic voices in Study 2 had a strong effect on dominance and aggressiveness perceptions (the two traits most strongly correlated with the Dominance component). Synthetic voices with lowered pitch were judged significantly more dominant and aggressive than those with raised pitch (Fig. 2). These results are consistent with previously reported results for both measured and manipulated pitch and dominance perceptions of natural human voices13,14.
Previous studies of formant frequencies and social perceptions of natural human voices have typically reported that voices with lower formant frequencies are perceived to be more dominant34. Previous studies of perceptions of natural human voices have also often reported associations between attractiveness and both voice pitch and formant frequencies31,34. In our study, we did not observe significant associations between scores on the Dominance PC and measured formants (Study 1) or between scores on the Valence PC (very highly correlated with attractiveness) and measured pitch or formants (Study 1). It is currently unclear whether these potential differences in results for synthetic voices in the current studies and natural human voices in previous studies reflect differences in how these two classes of stimuli are perceived or methodological differences among studies (e.g., larger number of voices tested in studies of natural human voices). While further work is needed to address this issue, our results clearly indicate that pitch is particularly strongly related to (and, as our results from Study 2 indicate, directly influences) dominance-related perceptions of synthetic voices. Indeed, in Study 1, pitch explained ~ 74% of the variance in Dominance PC scores. In a recent study of natural human voices, Schild et al.35 found that trustworthiness ratings were not significantly correlated with pitch, but that pitch was a very good predictor of dominance ratings. Our findings for conversational agents’ voices very closely align with Schild et al.’s results.
Three key aspects of the research reported here suggest that our results are likely to have good generalisability. First, we show that dominance-related perceptions of voices are both highly correlated with voice pitch (Study 1) and directly influenced by experimentally manipulating voice pitch (Study 2). This pattern of results suggests that our results generalise well across two different types of study design. Second, in Study 2 we analysed responses using linear mixed models that take into account variability in responses across both raters and stimuli. This analytical strategy is known to produce results that generalise better to new sets of stimuli than those of analytical approaches in which responses are aggregated across stimuli36. Third, the similarity in our results for dominance-related perceptions of synthetic voices to those reported elsewhere for natural human voices suggest the large literature on dominance-related perceptions of natural human voices is likely to be useful for understanding perceptions of synthetic voices.
To summarize, our results for social perceptions of conversational agents’ synthetic voices highlight two clear similarities in the characteristics of judgments of synthetic voices and those previously reported for natural human stimuli. Like natural human voices, our results show that perceptions of synthetic voices appear to be underpinned by Valence and Dominance dimensions and that dominance-related perceptions are both strongly related to voice pitch and directly affected by experimental manipulations of voice pitch. Collectively, these results suggest that greater consideration of the role that voice pitch plays in dominance-related social perceptions when designing conversational agents will be effective in controlling stereotypic perceptions of their voices and the downstream consequences of those perceptions.
Data availability
Data, analysis code, and stimuli for both studies are publicly available on the Open Science Framework (https://osf.io/4zgrf/).
References
Olivola, C. Y., Funk, F. & Todorov, A. Social attributions from faces bias human choices. Trends Cogn. Sci. 18(11), 566–570 (2014).
Rhodes, G. The evolutionary psychology of facial beauty. Annu. Rev. Psychol. 57, 199–226 (2006).
Wilson, J. P. & Rule, N. O. Facial trustworthiness predicts extreme criminal-sentencing outcomes. Psychol. Sci. 26(8), 1325–1331 (2015).
Jones, B. C. et al. To which world regions does the valence–dominance model of social perception apply?. Nat. Hum. Behav. 5(1), 159–169 (2021).
Oosterhof, N. N. & Todorov, A. The functional basis of face evaluation. Proc. Natl. Acad. Sci. 105(32), 11087–11092 (2008).
Baus, C., McAleer, P., Marcoux, K., Belin, P. & Costa, A. Forming social impressions from voices in native and foreign languages. Sci. Rep. 9(1), 1–14 (2019).
McAleer, P., Todorov, A. & Belin, P. How do you say ‘Hello’? Personality impressions from brief novel voices. PLoS ONE 9(3), e90779 (2014).
West, M., Kraut, R. & Chew, H. E. I’d blush if I could: Closing gender divides in digital skills through education. UNESCO Technical Report (2019).
Balas, B. & Pacella, J. Trustworthiness perception is disrupted in artificial faces. Comput. Hum. Behav. 77, 240–248 (2017).
Balas, B., Tupa, L. & Pacella, J. Measuring social variables in real and artificial faces. Comput. Hum. Behav. 88, 236–243 (2018).
Cabral, J. P., Cowan, B. R., Zibrek, K. & McDonnell, R. The influence of synthetic voice on the evaluation of a virtual character. In INTERSPEECH 229–233 (2017).
Wester, M., Aylett, M. P. & Braude, D. A. Bot or not: Exploring the fine line between cyber and human identity. In Proceedings of the 19th ACM International Conference on Multimodal Interaction 506–507 (2017).
Armstrong, M. M., Lee, A. J. & Feinberg, D. R. A house of cards: Bias in perception of body size mediates the relationship between voice pitch and perceptions of dominance. Anim. Behav. 147, 43–51 (2019).
Aung, T. & Puts, D. Voice pitch: A window into the communication of social power. Curr. Opin. Psychol. 33, 154–161 (2020).
Hester, N., Jones, B. C. & Hehman, E. Perceived femininity and masculinity contribute independently to facial impressions. J. Exp. Psychol. Gen. (2020).
Oh, D., Buck, E. A. & Todorov, A. Revealing hidden gender biases in competence impressions of faces. Psychol. Sci. 30(1), 65–79 (2019).
Sutherland, C. A. et al. Social inferences from faces: Ambient images generate a three-dimensional model. Cognition 127(1), 105–118 (2013).
Boersma, P. & Weenink, D. Praat: Doing phonetics by computer (2018).
R Core Team. R: A language and environment for statistical computing. R Foundation for Statistical Computing, Vienna, Austria. Retrieved from http://www.R-project.org/ (2021).
Wickham, H. & Bryan, J. tidyverse: Easily Install and Load the 'Tidyverse' (Version 1.3.0) [Computer software]. Retrieved from https://cran.r-project.org/web/packages/tidyverse/index.html (2021).
Wickham, H. & Bryan, J. readxl: Read excel files. R Package Version (2019).
Revelle, W. Psych: Procedures for personality and psychological research (Version 1.6.12) [Computer software]. Retrieved from https://cran.r-project.org/packagepsych (2016).
Dinno, A. paran: Horn’s Test of Principal Components/Factors [R package version 1.5.2] (2018).
Zhu, H. KableExtra: Construct complex table with ’kable’ and pipe syntax. R Package Version (2019).
Xie, Y. knitr: A comprehensive tool for reproducible research in R. In Implementing Reproducible Research (eds Stodden, V. et al.) 3–29 (CRC Press, 2014).
Long, J. A. jtools: Analysis and Presentation of Social Scientific Data. (Version 2.1.3). Retrieved from https://cran.r-project.org/web/packages/jtools (2021).
Puts, D. A., Gaulin, S. J. & Verdolini, K. Dominance and the evolution of sexual dimorphism in human voice pitch. Evol. Hum. Behav. 27(4), 283–296 (2006).
Hodges-Simeon, C. R., Gaulin, S. J. & Puts, D. A. Different vocal parameters predict perceptions of dominance and attractiveness. Hum. Nat. 21(4), 406–427 (2010).
Apicella, C. L. & Feinberg, D. R. Voice pitch alters mate-choice-relevant perception in hunter–gatherers. Proc. R. Soc. B Biol. Sci. 276(1659), 1077–1082 (2009).
Feinberg, D. R., Jones, B. C., Little, A. C., Burt, D. M. & Perrett, D. I. Manipulations of fundamental and formant frequencies influence the attractiveness of human male voices. Anim. Behav. 69(3), 561–568 (2005).
Jones, B. C., Feinberg, D. R., DeBruine, L. M., Little, A. C. & Vukovic, J. A domain-specific opposite-sex bias in human preferences for manipulated voice pitch. Anim. Behav. 79(1), 57–62 (2010).
Bolker, B. & Robinson, D. broom.mixed: Tidying Methods for Mixed Models (Version 0.2.6) [Computer software]. Retrieved from https://cran.r-project.org/web/packages/broom.mixed/index.html (2021).
Kuznetsova, A., Brockhoff, P. B. & Christensen, R. H. B. lmerTest package: Tests in linear mixed effects models. J. Stat. Softw. 82(13), 1–26. https://doi.org/10.18637/jss.v082.i13 (2017).
Schild, C. et al. Linking human male vocal parameters to perceptions, body morphology, strength and hormonal profiles in contexts of sexual selection. Sci. Rep. 10(1), 1–16 (2020).
Schild, C., Braunsdorf, E., Steffens, K., Pott, F. & Stern, J. Gender and context-specific effects of vocal dominance and trustworthiness on leadership decisions. Adapt. Hum. Behav. Physiol. 1–19 (2022).
Barr, D. J., Levy, R., Scheepers, C. & Tily, H. J. Random effects structure for confirmatory hypothesis testing: Keep it maximal. J. Mem. Lang. 68(3), 255–278 (2013).
Acknowledgements
This research was supported by the EPSRC grant ‘Designing Conversational Assistants to Reduce Gender Bias’ (EP/T023783/1), awarded to Benedict Jones. Daria Altenburg was supported by Grant BOF.24Y.2019.0006.01 of Ghent University, awarded to Adriaan Spruyt. For the purpose of Open Access, the authors have applied a Creative Commons Attribution (CC BY) to any Author Accepted Manuscript (AAM) version arising from this submission.
Author information
Authors and Affiliations
Contributions
All authors designed the studies. V.S., B.J., and A.L. carried out analyses. B.J. and V.S. wrote the first draft of the manuscript, revising in light of feedback from D.A., A.L., and D.F. All authors reviewed the mansurciot.
Corresponding author
Ethics declarations
Competing interests
The authors declare no competing interests.
Additional information
Publisher's note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Shiramizu, V.K.M., Lee, A.J., Altenburg, D. et al. The role of valence, dominance, and pitch in perceptions of artificial intelligence (AI) conversational agents’ voices. Sci Rep 12, 22479 (2022). https://doi.org/10.1038/s41598-022-27124-8
Received:
Accepted:
Published:
DOI: https://doi.org/10.1038/s41598-022-27124-8
Comments
By submitting a comment you agree to abide by our Terms and Community Guidelines. If you find something abusive or that does not comply with our terms or guidelines please flag it as inappropriate.