Literature DB >> 21713180

Gender affects body language reading.

Arseny A Sokolov1, Samuel Krüger, Paul Enck, Ingeborg Krägeloh-Mann, Marina A Pavlova.   

Abstract

Body motion is a rich source of information for social cognition. However, gender effects in body language reading are largely unknown. Here we investigated whether, and, if so, how recognition of emotional expressions revealed by body motion is gender dependent. To this end, females and males were presented with point-light displays portraying knocking at a door performed with different emotional expressions. The findings show that gender affects accuracy rather than speed of body language reading. This effect, however, is modulated by emotional content of actions: males surpass in recognition accuracy of happy actions, whereas females tend to excel in recognition of hostile angry knocking. Advantage of women in recognition accuracy of neutral actions suggests that females are better tuned to the lack of emotional content in body actions. The study provides novel insights into understanding of gender effects in body language reading, and helps to shed light on gender vulnerability to neuropsychiatric and neurodevelopmental impairments in visual social cognition.

Entities:  

Keywords:  biological motion; gender; social cognition; visual perception

Year:  2011        PMID: 21713180      PMCID: PMC3111255          DOI: 10.3389/fpsyg.2011.00016

Source DB:  PubMed          Journal:  Front Psychol        ISSN: 1664-1078


Introduction

Body language reading is of immense importance for adaptive social behavior and non-verbal communication. This ability constitutes a central component of social competence. Healthy perceivers are able to infer emotions and dispositions of others represented by point-light body movements that minimize availability of other cues (Pollick et al., 2001; Atkinson et al., 2004; Heberlein et al., 2004; Clarke et al., 2005; Ikeda and Watanabe, 2009; Rose and Clarke, 2009). Perceivers can reliably judge emotional content of dance represented by a few moving dots placed on the dancer's body (Dittrich et al., 1996). Visual sensitivity to camouflaged point-light human locomotion is modulated by the emotional content of gait with the highest sensitivity to angry walking (Chouchourelou et al., 2006). Observers can discriminate between deceptive and true intentions conveyed by body motion, and true information is precisely detected despite misleading endeavors (Runeson and Frykholm, 1983; Grèzes et al., 2004a,b). But how do we know whom to trust or who is attracted to us? Such judgments are vital to social interaction, and men and women appear to show profound differences in cues attended to. Yet research on sex differences in visual social cognition has been mainly limited to static face images, in particular, still photographs. In accordance with widespread beliefs, females exhibit higher sensitivity to non-verbal cues: they better discriminate friendliness from sexual interest (Farris et al., 2008) and are more proficient in recognition of facial emotions (Montagne et al., 2005). Females without and with Asperger syndrome are better at recognizing emotions from dynamic faces than males (Golan et al., 2006). Moreover, females tend to better recognize emotions from faces than from voices, whereas males exhibit the opposite tendency. As a rule, however, facial expressions and static body postures can only signal emotional states and affect, but do not provide information about how to deal with it. Dynamic body expressions, gestures, and actions of others are a richer and more ecologically valid source of information for social interaction (De Gelder, 2006, 2009; Pavlova, 2009). The other important advantage of bodily expressions is that whereas face expressions (similarly to a verbal information flow) are believed to be easily kept under control, body movements reveal our true feelings. When emotions expressed by faces and bodies are incongruent, recognition of facial expressions is affected by emotions revealed by body (Meeren et al., 2005). Brain imaging indicates that emotions expressed by dynamic bodies as compared to faces elicit greater activation in a number of brain areas including the superior temporal sulcus (STS), a cornerstone of the social brain (Kret et al., 2010). Experimental evidence obtained primarily in patients with lesions and cortical blindness favors the assumption that emotional body language can be processed automatically, without visual awareness and attention (for review, see Tamietto and de Gelder, 2010). To a great surprise, however, gender impact on body language reading is largely unknown. A few studies conducted at the beginning of the 80s based on the profile of non-verbal sensitivity (PONS) test, which includes body motion (neck to knees) video clips, point to the superiority of females in body language reading (e.g., Blank et al., 1981). However, this test has some serious methodological limitations; for example, it is based on body motion video clips of only one female actor. Although sex differences represent a rather delicate topic, underestimation, or exaggeration of possible effects can retard progress in the field. The present work intends to make an initial step in filling the gap, and to clarify whether, and, if so, how perceiver's gender affects recognition of emotional expressions conveyed by actions of others. More specifically, we ask (i) whether gender affects recognition of emotions represented by body motion, or, in other words, whether females excel in recognition of emotional actions; and (ii) whether gender effects depend on emotional content of actions. To this end, healthy young females and males were presented with point-light displays portraying knocking at a door with different emotional expressions (happy, neutral, and angry). We took advantage of a point-light technique that helps to isolate information revealed by motion from other cues (shape, color, etc.). Perceivers saw only a few bright dots placed on the main joints of an otherwise invisible arm (Figure 1) so that all other clues except for motion characteristics were abandoned.
Figure 1

Three static frames taken from the dynamic sequence representing knocking motion by a set of dots placed on the arm joints, shoulder, and head of an otherwise invisible actor. Actors were seen facing right, in a sagittal view, and struck the surface directly in front of them.

Three static frames taken from the dynamic sequence representing knocking motion by a set of dots placed on the arm joints, shoulder, and head of an otherwise invisible actor. Actors were seen facing right, in a sagittal view, and struck the surface directly in front of them.

Materials and Methods

Participants

Thirty four healthy adults, students of the University of Tübingen Medical School (aged 20–36), were enrolled in the study. Mean age of females (20 participants) was 23.8 ± 3.7 years, and mean age of males (14 participants) was 22.9 ± 2.0 years. There was no age difference between female and male participants (t32 = 0.95, p = 0.35, ns). The groups were also comparable in terms of educational and socio-economic status. All participants had normal or corrected-to-normal vision and heterosexual orientation. None had a history of neurological or psychiatric disorders including autistic spectrum disorders (ASD), schizophrenia, head injures, or medication for anxiety or depression. None had previous experience with such displays or tasks. Participants were run individually. Informed written consent was obtained in accordance with the requirements of the local Ethical Committee at the University of Tübingen Medical School.

Stimuli and procedure

We used point-light displays portraying knocking arm motion (Pollick et al., 2001, 2002). Point-light displays were recorded during performance of knocking with different emotional content (happy, neutral, and angry). We chose to use animations with happy and angry motions, because happiness and anger are reported to be quite similar on the activation dimension, and these animations tended to have fast and jerky movements (Pollick et al., 2001). Display creation is described in detail elsewhere (Pollick et al., 2001). In brief, recording was performed using a 3D position measurement system at a rate of 60 Hz (Optotrak, Northern Digital Inc., Waterloo, ON, Canada). Each display consisted of six point-light dots placed on the head, shoulder, elbow, wrist, and the first and forth metacarpal joints of an otherwise invisible right hand (Figure 1). Point-light actors were seen facing right, in a sagittal view, and struck the surface directly in front of them. The size of all point-light knocking stimuli was standardized in such a way that in the first frame, the distance from the head to the first metacarpal joint was identical for all actors. For each emotion, six different displays with equal number of knocking performed by female and male actors were created. By using the Presentation software (Neurobehavioral Systems Inc., Albany, CA, USA), each video was displayed five times per experimental session resulting in 30 trials per emotion. The whole experimental session consisted of a set of 90 displays representing three emotions in a random order, and took about 15–20 min per participant. Each display was shown for 1 s. We used a three alternative-forced choice paradigm. On each trial, participants indicated (by pressing with their dominant hand one of three respective keys on a computer keyboard) whether a display portrayed happy, neutral, or angry knocking. Positions of the keys were counterbalanced between participants. Participants were told that they have to perform the task as accurately as possible. No immediate feedback was given regarding performance.

Results

Percentage correct in recognition of emotions conveyed by knocking is represented in Figure 2A. In both females and males, recognition of all emotional expressions was above chance level (p < 0.001). However, recognition of happy knocking was less accurate than of neutral and angry actions. This is consistent with the outcome of previous studies on emotion recognition through point-light human locomotion (Chouchourelou et al., 2006; Ikeda and Watanabe, 2009) and dance (Dittrich et al., 1996) that show better recognition of angry over happy motion.
Figure 2

Recognition of happy, neutral, and angry point-light knocking by females and males. (A) Percentage correct: Males outperformed in recognition of happy knocking (p < 0.015), whereas females excelled in recognition of neutral knocking (p < 0.016) and tended to over-perform in recognition of angry knocking (p < 0.07). Bold horizontal line indicates chance level. Significant differences are indicated by an asterisk; (B) Error rate: The lack of gender differences in error rate demonstrates that gender differences in recognition accuracy of emotional content of knocking were not caused by gender-related bias for mistaking one emotion for another. Each bar represents an average ratio of the number of errors of particular type to the overall number of errors made for a display type (e.g., leftmost bar represents an average ratio of number of trials when happy knocking was mistaken for neutral knocking to the number of trials when happy knocking was mistaken for both neutral and angry knocking); (C) Response time to happy, neutral, and angry point-light knocking by females and males. Females and males do not differ in response time. Vertical bars represent ± SE.

Recognition of happy, neutral, and angry point-light knocking by females and males. (A) Percentage correct: Males outperformed in recognition of happy knocking (p < 0.015), whereas females excelled in recognition of neutral knocking (p < 0.016) and tended to over-perform in recognition of angry knocking (p < 0.07). Bold horizontal line indicates chance level. Significant differences are indicated by an asterisk; (B) Error rate: The lack of gender differences in error rate demonstrates that gender differences in recognition accuracy of emotional content of knocking were not caused by gender-related bias for mistaking one emotion for another. Each bar represents an average ratio of the number of errors of particular type to the overall number of errors made for a display type (e.g., leftmost bar represents an average ratio of number of trials when happy knocking was mistaken for neutral knocking to the number of trials when happy knocking was mistaken for both neutral and angry knocking); (C) Response time to happy, neutral, and angry point-light knocking by females and males. Females and males do not differ in response time. Vertical bars represent ± SE. Individual number of correct responses was submitted to a 2 × 3 repeated-measures ANOVA (as assessed by the Shapiro–Wilk test, the data were normally distributed) with factors Gender (female/male) and Emotional expression of knocking (happy/neutral/angry). This analysis revealed the lack of a main effect of gender (F(1,32) = 0.21, p = 0.648, ns). However, a main effect of emotional expression (F(2,32) = 82.94, p < 0.0001) and interaction between the factors Gender × Emotional expression (F(2,32) = 6.23, p < 0.003) were highly significant. Planned pair-wise comparisons indicated that males outperformed in recognition of happy knocking (t32 = 2.58, p < 0.015, one-tailed, here and below Bonferroni corrected for multiple comparisons; d = 0.84), whereas females tended to over-perform in recognition of angry knocking (t32 = 1.87, p < 0.07, one-tailed) and excelled in recognition of neutral knocking (t32 = 2.54, p < 0.016, one-tailed, d = 0.88). The data, therefore, reveals the lack of advantage of females in recognition accuracy. Instead, the findings indicate that sex effects in recognition accuracy are modulated by emotional contents of actions. Error analysis (Figure 2B) indicated that by both females and males, happy knocking was mistaken for neutral knocking in more than 80% of wrong responses (error rate 0.84 and 0.86 for females and males, respectively; gender difference: t32 = 0.42, p = 0.68, two-tailed, ns. Error rate was calculated as an average ratio of the number of errors of particular type to the overall number of errors made for a display type). In turn, with a lack of gender differences, neutral knocking was misperceived as happy actions in about 70% of error responses (error rate 0.68 and 0.67 for females and males, respectively; gender differences: t32 = 0.02, p = 0.99, two-tailed, ns). In about 80% of error trials in response to angry knocking, both females and males mistook angry knocking for neutral knocking (error rate 0.79 and 0.8 for females and males, respectively; gender difference: t32 = 0.14, p = 0.88, two-tailed, ns). The lack of gender differences in error rate suggests that gender effects in recognition accuracy of emotional content of knocking observed in the present study are not caused by gender-related bias for mistaking one emotion for another. For response time analyses, a 2 × 3 repeated-measures ANOVA was performed on individual values (as assessed by the Shapiro–Wilk test, the data were normally distributed) with factors Gender (female/male) and Emotional expression (happy/neutral/angry). This analysis did not reveal any effect of gender (F(1,32) = 1.56, p = 0.22, ns) as well as any interaction of factors Gender × Emotional expression on response time (F(2,32) = 1.42, p = 0.25, ns; Figure 2C). However, a main effect of emotional expression was significant (F(2,32) = 35.16, p < 0.0001), with the fastest response to angry knocking, and the slowest response to neutral knocking (Figure 2B). This shows that recognition of neutral knocking was more difficult than that of angry and happy knocking. Post hoc pair-wise comparisons showed no gender difference in response time to happy (t32 = 0.09, p = 0.93, two-tailed, ns, average 2.00 ± 0.39 and 1.99 ± 0.28 s from the stimulus onset, for females and males, respectively), neutral (t32 = 1.21, p = 0.24, two-tailed, ns; average 2.15 ± 0.33 and 2.28 ± 0.3 s, for females and males, respectively), and angry knocking (t32 = 0.14, p = 0.89, two-tailed, ns; average 1.84 ± 0.32 and 1.85 ± 0.28, for females and males, respectively). Taken together, the findings suggest that gender does not affect speed of body language reading. For both females and males, however, the swiftness of response to body language depends on the emotional content of actions. Since it is difficult to interpret negative findings within a relatively small sample size that might be considered a limitation of the study, the lack of sex differences in error rate and response time has to be further explored.

Discussion

The outcome of the study indicates that gender affects accuracy rather than speed of body language reading. To the best of our knowledge, the present work delivers the first evidence for sex effects in body language reading. The gender effect, however, is modulated by the emotional content of actions. Females tend to excel in recognition accuracy of angry knocking, whereas males over-perform in recognition of happy actions. Furthermore, females clearly surpass males in recognition of emotionally neutral knocking. The lack of gender differences in error rate suggests that gender effects in recognition accuracy are not caused by gender-related bias. Based on popular wisdom, one can expect that while women possess soft skills in social perception including high sensitivity to positive emotional signals and subtle details, men might outperform in recognition of negative menacing expressions. This assumption is based on the different evolutionary and socio-cultural roles of both genders (e.g., Biele and Grabowska, 2006; Proverbio et al., 2008). High sensitivity of women to positive emotions has been related to their role as primary offspring care providers. Social cognition in men is presumably connected with active interactions and immediate reactions, and, therefore, emotion perception is likely associated with motor programs. Anger detection is usually associated with a need to act, for example, escape from a person or prepare to confront a person. However, the data available are controversial. In the present study, males over-perform in recognition of emotionally positive happy actions. These data agree with findings showing that men appear to exhibit stronger brain activation in response to positive pictures (depicting landscapes, sport activities, families, and erotic scenes) than women (Wrase et al., 2003; Sabattineli et al., 2004; Gasbarri et al., 2007). Moreover, males are equally responsive to happiness conveyed through static and dynamic happy faces (males rate the intensity of dynamic and static expressions of happiness equally high), whereas females are less responsive to happiness in static faces (Biele and Grabowska, 2006). Presumably, this indicates that males are better tuned to subtle expressions of happiness in faces and actions. This might hold true, at least, for a population of young men with a high social status and educational level as those participated in the present study. The prominent outcome of the study is that females had a clear advantage in recognition of neutral knocking. This suggests that women are better tuned to the lack of emotional content in body actions. Future research should clarify whether gender effects in body language reading occur with other repertoires of actions, and with other arrays of emotions. What is the nature of gender effects in body language reading? One possibility is that gender differences have neurobiological sources (Cahill, 2006; Jazin and Cahill, 2010), and brain mechanisms underpinning body language reading are sex-specific. The social cognition network, commonly referred to as the social brain, primarily involves the parieto-temporal junction, temporal cortices including the fusiform face area and the STS, orbitofrontal cortices, the amygdala (Adolphs, 2003), and the left lateral cerebellum (Sokolov et al., 2010). The right STS is a cornerstone for processing of meaningful body motion (Grossman and Blake, 2002; Pavlova et al., 2004; Pelphrey et al., 2004). Is the social brain sex-specific? This is an open question. To date, studies of sex effects on the social brain have been limited to investigation of face expressions or body actions represented in still photographs. Brain activation in females is reported to be more bilaterally distributed, presumably providing greater contribution of both hemispheres to identification of facial affect (Bourne, 2005; Proverbio et al., 2010). Females show stronger event-related potential (ERP) response to emotional faces (Orozco and Ehlers, 1998). However, the findings are controversial. Sex effects are found in the blood oxygen level dependent (BOLD) response of the amygdala to happy, but not to fearful faces (Killgore and Yurgelun-Todd, 2001). On the other hand, a significant correlation between functional magnetic resonance imaging (fMRI) activity of the amygdala and behavioral response to fearful faces is observed in males only (Derntl et al., 2009a). Both behavioral and amygdala responses to threat-related face expressions are correlated with testosterone level (Derntl et al., 2009b). In accordance with widespread belief, it is reported that the female brain is more responsive to social stimuli represented in still images (Proverbio et al., 2009). Recent ERP findings indicate that in females, processing of actions’ goals occurs earlier (Proverbio et al., 2010). Neuroimaging reveals that gender effects are not evident in the neural circuitry underpinning visual processing of social interaction, but rather in the regions engaged in perceptual decision making: the neuromagnetic gamma response over the left prefrontal cortex peaks earlier in females (Pavlova et al., 2010a). Gender effects at behavioral level do not necessarily imply that there is sex-related difference in brain activation subserving body language reading. Moreover, gender differences in performance on social cognition tasks can be impacted by socio-cultural stereotypes (Pavlova et al., 2010b). Several types of interrelations between behavioral measures and brain mechanisms engaged in social perception should be taken into account: (i) sex differences both in behavioral and brain responses; (ii) sex differences detectable either at behavioral level or only in brain activation; and (iii) absence of sex differences both at behavioral and brain levels (Pavlova, 2009). Noteworthy, gender-related dimorphism in the brain may not only elicit but also prevent behavioral differences if they are maladaptive (De Vries, 2004). Future research should be directed at uncovering sex differences in brain activity during body language reading. Such investigation would also shed light on sex differences in neuropsychiatric conditions characterized by impairments in social cognition such as ASD, depression, and schizophrenia. It is known that males are more commonly affected by ASD than females, with a ratio of about 4:1 (Newschaffer et al., 2007). Females, however, are affected much more severely, and, therefore, in high functioning autistic individuals this ratio is even much higher. Although there is some behavioral evidence that individuals with ASD have difficulties in revealing information about emotions from point-light body movements (Moore et al., 1997; Hubert et al., 2007; Parron et al., 2008), it is unclear whether females and males with ASD differ in body language reading. The lack of studies in females with ASD calls for a thorough investigation of their profile. The other important issue for future research is sex differences in visual social cognition in survivors of premature birth. Males are at a 14–20% higher risk of premature birth (Melamed et al., 2010) and of its complications in the brain and cognition. Adolescents who were born prematurely are likely to exhibit difficulties in visual social cognition (Pavlova et al., 2008), but gender effects are largely unknown. Clarification of gender impact on body language reading and underlying brain networks would provide novel insights into understanding of gender vulnerability to neuropsychiatric and neurodevelopmental impairments in visual social cognition.

Conflict of Interest Statement

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
  48 in total

1.  Brain Areas Active during Visual Perception of Biological Motion.

Authors:  Emily D Grossman; Randolph Blake
Journal:  Neuron       Date:  2002-09-12       Impact factor: 17.173

2.  Inferring false beliefs from the actions of oneself and others: an fMRI study.

Authors:  J Grèzes; C D Frith; R E Passingham
Journal:  Neuroimage       Date:  2004-02       Impact factor: 6.556

3.  Look who's talking: visual detection of speech from whole-body biological motion cues during emotive interpersonal conversation.

Authors:  David Rose; Tanya J Clarke
Journal:  Perception       Date:  2009       Impact factor: 1.490

4.  Grasping the intentions of others: the perceived intentionality of an action influences activity in the superior temporal sulcus during social perception.

Authors:  Kevin A Pelphrey; James P Morris; Gregory McCarthy
Journal:  J Cogn Neurosci       Date:  2004-12       Impact factor: 3.225

5.  Face coding is bilateral in the female brain.

Authors:  Alice Mado Proverbio; Federica Riva; Eleonora Martin; Alberto Zani
Journal:  PLoS One       Date:  2010-06-21       Impact factor: 3.240

6.  Recognition of biological motion in children with autistic spectrum disorders.

Authors:  Carole Parron; David Da Fonseca; Andreia Santos; David G Moore; Elisa Monfardini; Christine Deruelle
Journal:  Autism       Date:  2008-05

7.  Estimating the efficiency of recognizing gender and affect from biological motion.

Authors:  Frank E Pollick; Vaia Lestou; Jungwon Ryu; Sung-Bae Cho
Journal:  Vision Res       Date:  2002-09       Impact factor: 1.886

8.  Affective picture perception: gender differences in visual cortex?

Authors:  Dean Sabatinelli; Tobias Flaisch; Margaret M Bradley; Jeffrey R Fitzsimmons; Peter J Lang
Journal:  Neuroreport       Date:  2004-05-19       Impact factor: 1.837

9.  Dissociable cortical processing of recognizable and non-recognizable biological movement: analysing gamma MEG activity.

Authors:  Marina Pavlova; Werner Lutzenberger; Alexander Sokolov; Niels Birbaumer
Journal:  Cereb Cortex       Date:  2004-02       Impact factor: 5.357

10.  Cerebellar engagement in an action observation network.

Authors:  Arseny A Sokolov; Alireza Gharabaghi; Marcos S Tatagiba; Marina Pavlova
Journal:  Cereb Cortex       Date:  2009-06-22       Impact factor: 5.357

View more
  28 in total

1.  Structural and effective brain connectivity underlying biological motion detection.

Authors:  Arseny A Sokolov; Peter Zeidman; Michael Erb; Philippe Ryvlin; Karl J Friston; Marina A Pavlova
Journal:  Proc Natl Acad Sci U S A       Date:  2018-12-04       Impact factor: 11.205

Review 2.  Empathy: gender effects in brain and behavior.

Authors:  Leonardo Christov-Moore; Elizabeth A Simpson; Gino Coudé; Kristina Grigaityte; Marco Iacoboni; Pier Francesco Ferrari
Journal:  Neurosci Biobehav Rev       Date:  2014-09-16       Impact factor: 8.989

3.  Peripersonal space in social context is modulated by action reward, but differently in males and females.

Authors:  Maria Francesca Gigliotti; Patrícia Soares Coelho; Joana Coutinho; Yann Coello
Journal:  Psychol Res       Date:  2019-09-06

Review 4.  Empathy as a Concept from Bench to Bedside: A Translational Challenge.

Authors:  Nazan Uysal; Ulaş M Çamsari; Mehmet ATEş; Sevim Kandİş; Aslı Karakiliç; Gamze B Çamsari
Journal:  Noro Psikiyatr Ars       Date:  2019-11-28       Impact factor: 1.339

5.  Sex differences in the development of brain mechanisms for processing biological motion.

Authors:  L C Anderson; D Z Bolling; S Schelinski; M C Coffman; K A Pelphrey; M D Kaiser
Journal:  Neuroimage       Date:  2013-07-20       Impact factor: 6.556

6.  Brain circuits signaling the absence of emotion in body language.

Authors:  Arseny A Sokolov; Peter Zeidman; Michael Erb; Frank E Pollick; Andreas J Fallgatter; Philippe Ryvlin; Karl J Friston; Marina A Pavlova
Journal:  Proc Natl Acad Sci U S A       Date:  2020-08-06       Impact factor: 11.205

7.  Anterior Temporal Lobectomy Impairs Neural Classification of Body Emotions in Right Superior Temporal Sulcus and Reduces Emotional Enhancement in Distributed Brain Areas without Affecting Behavioral Classification.

Authors:  Laura Van de Vliet; Jan Jastorff; Yun-An Huang; Wim Van Paesschen; Mathieu Vandenbulcke; Jan Van den Stock
Journal:  J Neurosci       Date:  2018-09-18       Impact factor: 6.167

Review 8.  Emotional expressions beyond facial muscle actions. A call for studying autonomic signals and their impact on social perception.

Authors:  Mariska E Kret
Journal:  Front Psychol       Date:  2015-05-27

9.  Face-n-Food: Gender Differences in Tuning to Faces.

Authors:  Marina A Pavlova; Klaus Scheffler; Alexander N Sokolov
Journal:  PLoS One       Date:  2015-07-08       Impact factor: 3.240

10.  Individual differences in reading social intentions from motor deviants.

Authors:  Daniel Lewkowicz; Francois Quesque; Yann Coello; Yvonne N Delevoye-Turrell
Journal:  Front Psychol       Date:  2015-08-17
View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.