| Literature DB >> 32695154 |
Chong Wu1, Zhenan Feng2, Jiangbin Zheng3, Houwang Zhang2.
Abstract
In this paper, we want to find out whether gender bias will affect the success and whether there are some common laws driving the success in show business. We design an experiment, set the gender and productivity of an actor or actress in a certain period as the independent variables, and introduce deep learning techniques to do the prediction of success, extract the latent features, and understand the data we use. Three models have been trained: the first one is trained by the data of an actor, the second one is trained by the data of an actress, and the third one is trained by the mixed data. Three benchmark models are constructed with the same conditions. The experiment results show that our models are more general and accurate than benchmarks. An interesting finding is that the models trained by the data of an actor/actress only achieve similar performance on the data of another gender without performance loss. It shows that the gender bias is weakly related to success. Through the visualization of the feature maps in the embedding space, we see that prediction models have learned some common laws although they are trained by different data. Using the above findings, a more general and accurate model to predict the success in show business can be built.Entities:
Mesh:
Year: 2020 PMID: 32695154 PMCID: PMC7368965 DOI: 10.1155/2020/8842221
Source DB: PubMed Journal: Comput Intell Neurosci
Figure 1The process of subsequence generation.
The details of training data and test data for each model.
| Training data including validation data | Test data | ||
|---|---|---|---|
| Model 1: MAO_ours | 70% data of an actor (AM ≥ 5, | 30% data of an actor (AM ≥ 5, | |
|
| |||
| Model 2: MAE_ours | 70% data of an actress (AM ≥ 5, | 30% data of an actress (AM ≥ 5, | |
|
| |||
| Model 3: MM_ours | 70% mixed data of an actor and actress (AM ≥ 5, | 30% mixed data of an actor and actress (AM ≥ 5, | |
The validation data are included in the training data. Note. MM_ours denotes the prediction model trained by the mixed data of an actor and actress, MAO_ours denotes the prediction model trained by the data of an actor only, and MAE_ours denotes the prediction model trained by the data of an actress only.
Figure 2The structure of RNN and details of the LSTM unit.
Figure 3The workflow of our model. It has an end-to-end structure and can be divided into two parts: (1) encoder; (2) binary classifier. The encoder of our model is a single LSTM layer which is used to embed different sequences to an n-dim embedding space. The binary classifier is a fully connected neural network.
Figure 4Sequences sorted by the sequence length.
The details of each layer's configuration in our model.
| Layer's name | Input size | Output size | No. of hidden units |
|---|---|---|---|
| Sequence input | 1 | 1 | 30 |
| LSTM | 1 | — | |
| Fully connected layer | 30 | 2 | |
| Softmax layer | 2 | 2 | |
| Classification layer | 2 | 2 |
Performance comparison of our methods and a recent study NatComm19 [1] in the prediction of the AM on the subset (AM ≥ 5) of the test data.
| Actor | Actress | Actor | Actress | |
|---|---|---|---|---|
| L ≥ 20, AM ≥ 5 |
| 5 = < | 5 = < | |
|
| 0.8074 | 0.6136 | 0.7888 | 0.5655 |
| Baseline accuracy | 0.6702 | 0.7221 | 0.7034 | 0.7487 |
|
| ||||
| MM_ours | ||||
|
| 0.9102 | 0.9262 | 0.8891 | 0.9173 |
| Precision | 0.8866 | 0.9079 | 0.8570 | 0.9037 |
| Recall | 0.9350 | 0.9452 | 0.9237 | 0.9313 |
| Accuracy | 0.8978 | 0.9067 | 0.8702 | 0.8917 |
|
| ||||
| MAO_ours | ||||
|
| 0.9082 | 0.9254 | 0.9045 | 0.9272 |
| Precision | 0.9010 | 0.9267 | 0.8845 | 0.9203 |
| Recall | 0.9156 | 0.9241 | 0.9254 | 0.9341 |
| Accuracy | 0.8992 | 0.9077 | 0.8897 | 0.9048 |
|
| ||||
| MAE_ours | ||||
|
| 0.9104 | 0.9268 | 0.9021 | 0.9265 |
| Precision | 0.8958 | 0.9203 | 0.8537 | 0.8966 |
| Recall | 0.9255 | 0.9334 | 0.9564 | 0.9584 |
| Accuracy | 0.8992 | 0.9087 | 0.8828 | 0.9020 |
|
| ||||
| NatComm19MM | ||||
|
| 0.7956 | 0.7878 | 0.7442 | 0.7436 |
| Precision | 0.8930 | 0.8346 | 0.6092 | 0.6074 |
| Recall | 0.7174 | 0.7459 | 0.9562 | 0.9585 |
| Accuracy | 0.8338 | 0.8453 | 0.7100 | 0.7099 |
|
| ||||
| NatComm19MAO | ||||
|
| 0.7942 | 0.7872 | 0.7457 | 0.7438 |
| Precision | 0.8902 | 0.8347 | 0.6103 | 0.6075 |
| Recall | 0.7169 | 0.7448 | 0.9582 | 0.9588 |
| Accuracy | 0.8332 | 0.8464 | 0.7116 | 0.7111 |
|
| ||||
| NatComm19MAE | ||||
|
| 0.7707 | 0.7770 | 0.7766 | 0.7409 |
| Precision | 0.9176 | 0.8803 | 0.6630 | 0.6057 |
| Recall | 0.6643 | 0.6954 | 0.9371 | 0.9540 |
| Accuracy | 0.8238 | 0.8474 | 0.7622 | 0.7591 |
MM_ours denotes the prediction model trained by the mixed data of an actor and actress; MAO_ours denotes the prediction model trained by the data of an actor only; MAE_ours denotes the prediction model trained by the data of an actress only; MM_NatComm19 denotes the model of NatComm19 [1] trained by the mixed data of an actor and actress, and the learned threshold d = 6.1523; MAO_NatComm19 denotes the model of NatComm19 [1] trained by the data of an actor, and the learned threshold d = 6.9580; MAE_NatComm19 denotes the model of NatComm19 [1] trained by the data of an actress, and the learned threshold d = 5.6640.
Performance comparison of our methods and a recent study NatComm19 [1] in the prediction of the AM on the subset (AM ≥ 10) of the test data.
| Actor | Actress | Actor | Actress | |
|---|---|---|---|---|
|
|
| 5 = < | 5 = < | |
|
| 0.6481 | 0.4169 | 0.6053 | 0.3348 |
| Baseline accuracy | 0.7275 | 0.7968 | 0.7668 | 0.8173 |
|
| ||||
| MM_ours | ||||
|
| 0.9409 | 0.9591 | 0.9202 | 0.9530 |
| Precision | 0.9355 | 0.9612 | 0.9418 | 0.9780 |
| Recall | 0.9463 | 0.9571 | 0.8995 | 0.9293 |
| Accuracy | 0.9279 | 0.9422 | 0.9024 | 0.9313 |
|
| ||||
| MAO_ours | ||||
|
| 0.9389 | 0.9557 | 0.9276 | 0.9563 |
| Precision | 0.9551 | 0.9729 | 0.9538 | 0.9836 |
| Recall | 0.9232 | 0.9391 | 0.9029 | 0.9306 |
| Accuracy | 0.9270 | 0.9387 | 0.9118 | 0.9359 |
|
| ||||
| MAE_ours | ||||
|
| 0.9396 | 0.9559 | 0.9377 | 0.9643 |
| Precision | 0.9414 | 0.9664 | 0.9338 | 0.9761 |
| Recall | 0.9378 | 0.9457 | 0.9415 | 0.9528 |
| Accuracy | 0.9264 | 0.9386 | 0.9217 | 0.9467 |
|
| ||||
| NatComm19MM | ||||
|
| 0.7688 | 0.8008 | 0.8114 | 0.7607 |
| Precision | 0.9299 | 0.8879 | 0.7321 | 0.6371 |
| Recall | 0.6552 | 0.7292 | 0.9101 | 0.9439 |
| Accuracy | 0.8460 | 0.8916 | 0.8367 | 0.8478 |
|
| ||||
| NatComm19MAO | ||||
|
| 0.7681 | 0.7989 | 0.8085 | 0.7559 |
| Precision | 0.9280 | 0.8841 | 0.7250 | 0.6297 |
| Recall | 0.6552 | 0.7287 | 0.9137 | 0.9453 |
| Accuracy | 0.8453 | 0.8909 | 0.8371 | 0.8449 |
|
| ||||
| NatComm19MAE | ||||
|
| 0.7377 | 0.7790 | 0.8127 | 0.7616 |
| Precision | 0.9373 | 0.8956 | 0.7518 | 0.6431 |
| Recall | 0.6082 | 0.6892 | 0.8843 | 0.9337 |
| Accuracy | 0.8330 | 0.8823 | 0.8429 | 0.8502 |
MM_ours denotes the prediction model trained by the mixed data of an actor and actress; MAO_ours denotes the prediction model trained by the data of an actor only; MAE_ours denotes the prediction model trained by the data of an actress only; MM_NatComm19 denotes the model of NatComm19 [1] trained by the mixed data of an actor and actress, and the learned threshold d = 6.1523; MAO_NatComm19 denotes the model of NatComm19 [1] trained by the data of an actor, and the learned threshold d = 6.9580; MAE_NatComm19 denotes the model of NatComm19 [1] trained by the data of an actress, and the learned threshold d = 5.6640.
Performance comparison of our methods and a recent study NatComm19 [1] in the prediction of the AM on the subset (AM ≥ 15) of the test data.
| Actor | Actress | Actor | Actress | |
|---|---|---|---|---|
|
|
| 5 = < | 5 = < | |
|
| 0.5271 | 0.3253 | 0.6292 | 0.3429 |
| Baseline accuracy | 0.7683 | 0.8439 | 0.7940 | 0.8467 |
|
| ||||
| MM_ours | ||||
|
| 0.9563 | 0.9725 | 0.9236 | 0.9583 |
| Precision | 0.9600 | 0.9756 | 0.9548 | 0.9883 |
| Recall | 0.9527 | 0.9694 | 0.8945 | 0.9301 |
| Accuracy | 0.9434 | 0.9584 | 0.9021 | 0.9358 |
|
| ||||
| MAO_ours | ||||
|
| 0.9533 | 0.9697 | 0.9336 | 0.9618 |
| Precision | 0.9750 | 0.9832 | 0.9692 | 0.9934 |
| Recall | 0.9326 | 0.9566 | 0.9005 | 0.9322 |
| Accuracy | 0.9401 | 0.9548 | 0.9159 | 0.9414 |
|
| ||||
| MAE_ours | ||||
|
| 0.9560 | 0.9710 | 0.9340 | 0.9638 |
| Precision | 0.9632 | 0.9794 | 0.9306 | 0.9758 |
| Recall | 0.9489 | 0.9627 | 0.9375 | 0.9520 |
| Accuracy | 0.9425 | 0.9562 | 0.9179 | 0.9458 |
|
| ||||
| NatComm19MM | ||||
|
| 0.7647 | 0.7990 | 0.8161 | 0.7425 |
| Precision | 0.9303 | 0.8555 | 0.7581 | 0.6161 |
| Recall | 0.6492 | 0.7495 | 0.8837 | 0.9340 |
| Accuracy | 0.8600 | 0.9099 | 0.8593 | 0.8614 |
|
| ||||
| NatComm19MAO | ||||
|
| 0.7608 | 0.8043 | 0.7978 | 0.7588 |
| Precision | 0.9230 | 0.8660 | 0.7282 | 0.6345 |
| Recall | 0.6470 | 0.7508 | 0.8822 | 0.9437 |
| Accuracy | 0.8610 | 0.9096 | 0.8491 | 0.8639 |
| NatComm19MAE | ||||
|
| 0.7361 | 0.7883 | 0.7954 | 0.7552 |
| Precision | 0.9268 | 0.8586 | 0.7455 | 0.6391 |
| Recall | 0.6105 | 0.7286 | 0.8525 | 0.9230 |
| Accuracy | 0.8530 | 0.9059 | 0.8489 | 0.8671 |
MM_ours denotes the prediction model trained by the mixed data of an actor and actress; MAO_ours denotes the prediction model trained by the data of an actor only; MAE_ours denotes the prediction model trained by the data of an actress only; MM_NatComm19 denotes the model of NatComm19 [1] trained by the mixed data of an actor and actress, and the learned threshold d = 6.1523; MAO_NatComm19 denotes the model of NatComm19 [1] trained by the data of an actor, and the learned threshold d = 6.9580; MAE_NatComm19 denotes the model of NatComm19 [1] trained by the data of an actress, and the learned threshold d = 5.6640.
Figure 5The workflow of the model in NatComm19 [1]. d is a scalar threshold which is learnable. The target of this model is to get an optimal d to separate two classes in the original feature space.
Figure 6Feature maps of the original feature space. Note. There are a few outliers (sequences with a length over 100). It is caused by a few films that in some sense exist but have not been released. Since they are so rare and are the correct data, they are also considered as in [1]: (a) actor, AM ≥ 5, L ≥ 20; (b) actress, AM ≥ 5, L ≥ 20; (c) actor, AM ≥ 5.5 ≤ L < 20; (d) actress, AM ≥ 5.5 ≤ L < 20.
Figure 7Feature maps of the testing data of an actor and actress in the embedding space obtained by different models. Blue line denotes class 1, and red line denotes class 2. It can be seen that the curves of different datasets show the same distribution and shape in the same embedding space. And, the boundary between two classes is clearer than the original feature space. Although it seems that the embedding spaces of different models are different, they are actually equivalent because they are different approximations of the global optimum obtained by the neural network. And, the curves of each feature's weight show that there is one feature dominating the classification. Note that it is like the eigen decomposition. Hence, the order of these weights has no meaning. And, the dominative feature of each model shows a similar floating range, and there is a clear boundary between two classes in this feature. It further proves that three models have learned a similar feature.