| Literature DB >> 36065448 |
Shaina Raza1, Deepak John Reji2, Chen Ding3.
Abstract
Because of the increasing use of data-centric systems and algorithms in machine learning, the topic of fairness is receiving a lot of attention in the academic and broader literature. This paper introduces Dbias (https://pypi.org/project/Dbias/), an open-source Python package for ensuring fairness in news articles. Dbias can take any text to determine if it is biased. Then, it detects biased words in the text, masks them, and suggests a set of sentences with new words that are bias-free or at least less biased. We conduct extensive experiments to assess the performance of Dbias. To see how well our approach works, we compare it to the existing fairness models. We also test the individual components of Dbias to see how effective they are. The experimental results show that Dbias outperforms all the baselines in terms of accuracy and fairness. We make this package (Dbias) as publicly available for the developers and practitioners to mitigate biases in textual data (such as news articles), as well as to encourage extension of this work. © Crown 2022.Entities:
Keywords: Bias; Classification; Deep learning; Entity recognition; Fairness; Masking; Transformer-based models
Year: 2022 PMID: 36065448 PMCID: PMC9434100 DOI: 10.1007/s41060-022-00359-4
Source DB: PubMed Journal: Int J Data Sci Anal
Fig. 1De-bias—a fair ML pipeline for news articles
Fig. 2Dbias and its de-biasing workflow
Some biases in news domain
| Bias | Example |
|---|---|
| Gender | All |
| Age | Apply if you are a |
| Racial/ Ethnicity | Police are looking for any |
| Disability | Genuine concern for the |
| Mental health | Any experience working with |
| Other biases addressed: Religion, education, political ideology (liberal, conservative) | |
Fig. 3Bias detection module
Fig. 4Bias recognition pipeline
Fig. 5De-biasing example with MLM
Distribution of identities based on biased words
| Identity | Count |
|---|---|
| Female | 5548 |
| Male | 2494 |
| Black | 4536 |
| White | 1761 |
| English Speaker | 810 |
| Non- English Speaker | 3744 |
| Young | 1301 |
| Older | 3636 |
| College degree | 945 |
| High school | 3573 |
Fig. 6Biased words in the news articles
Fig. 7(a) Bigrams of biased words in the news (b) Trigrams of biased words in news
Baseline methods
| Model | Description |
|---|---|
| Fairness Pre-processing models | |
| Disparate impact remover [ | Disparate Impact Remover is a pre-processing approach for increasing fairness between groups (privileged and unprivileged). This technique edits the feature values (e.g., the features that are privileged, unprivileged) so that the data can be made unbiased while preserving relevant information in the data. Once this algorithm has been implemented, any machine learning or deep learning model can be built using the repaired data. The Disparate Impact metric is then used to validate if the model is unbiased (or within an acceptable threshold). In this baseline method, we use a couple of methods using AutoML and reporting the results with the best performing model. For this baseline, the Logistic Regression gave us the best results |
| Reweighing [ | Reweighing is a pre-processing technique that weighs the examples in each group (such as privileged, unprivileged groups) to ensure fairness before classification. This algorithm transforms the dataset to have more equity in positive outcomes on the protected attribute(s) for both privileged and unprivileged groups. We run a couple of algorithms on the transformed data and report the result with the best performing model, which is Support Vector Machine (SVM) in this experiment |
| Fairness In-processing Models | |
| Adversarial De-biasing [ | Adversarial de-biasing is based on the generative adversarial network (GAN) model. Through training, this model de-biases the word and general feature embeddings. This is an in-processing technique that learns the definitions of fairness, such as demographic parity, equality of odds, and quality of opportunity, so that a discriminator (part of GAN) has been tasked with predicting the protected attribute encoded in the bias of the original feature vector, while a competing generator (part of GAN) has been tasked with producing more de-biased embeddings to compete with the discriminator |
| Exponentiated Gradient Reduction [ | Exponentiated gradient reduction is an in-processing technique that reduces the fair classification down into a series of cost-sensitive classification problems. It also returns a randomized classifier with the lowest empirical error (approximation of the expected error), if the fair classification rules are |
| met Fairness post-processing models | |
| Calibrated Equalized Odds Post-processing [ | Calibrated equalized odds post-processing is a post-processing technique that optimizes over calibrated classifier score outputs to find probabilities for changing output labels with an equalized odds objective |
| Equalized Odds post-processing [ | Equalized odds post-processing is a post-processing technique that uses a linear program to find probabilities for changing output labels in order to optimize equalized odds |
| Our approach | |
| Dbias | Our approach mitigates biases during the pre-processing stage and ensures that fairness is carried on throughout the ML pipeline to give a fair representation of data |
Confusion Matrix
| Actual Fake | Actual Real | |
|---|---|---|
| Predicted Fake | TP | FP |
| Predicted Real | FN | TN |
Comparison of our framework with the baseline methods
| Model | ||||||||||||
|---|---|---|---|---|---|---|---|---|---|---|---|---|
| PREC | REC | F1 | ACC | DI | PREC | REC | F1 | ACC | DI | G-AUC | ||
| Disparate impact remover | 0.593 | 0.549 | 0.570 | 0.587 | 0.702 | 0.532 | 0.414 | 0.466 | 0.541 | 0.804 | 0.634 | |
| Reweighing | 0.613 | 0.535 | 0.572 | 0.619 | 0.702 | 0.591 | 0.524 | 0.555 | 0.604 | 0.832 | 0.653 | |
| Adversarial De-biasing | 0.624 | 0.600 | 0.612 | 0.641 | 0.702 | 0.592 | 0.587 | 0.590 | 0.610 | 0.923 | 0.679 | |
| Exponentiated Gradient Reduc | 0.612 | 0.587 | 0.599 | 0.626 | 0.702 | 0.589 | 0.557 | 0.573 | 0.606 | 0.896 | 0.645 | |
| Calibrated Equalized Odds | 0.568 | 0.479 | 0.520 | 0.560 | 0.702 | 0.563 | 0.479 | 0.518 | 0.523 | 0.829 | 0.610 | |
| Equalized Odds | 0.498 | 0.487 | 0.492 | 0.577 | 0.702 | 0.487 | 0.488 | 0.487 | 0.505 | 0.818 | 0.598 | |
| Dbias | 0.735 | 0.784 | 0.759 | 0.776 | 0.702 | 0.690 | 0.704 | 0.697 | 0.743 | 1.012 | 0.780 | |
Effectiveness of different classification models
| Model | PREC | REC | F1 |
|---|---|---|---|
| LG-TFIDF | 0.62 | 0.61 | 0.61 |
| RF-TFIDF | 0.65 | 0.64 | 0.64 |
| GBM—TFIDF | 0.65 | 0.66 | 0.65 |
| LG- ELMO | 0.66 | 0.68 | 0.67 |
| MLP- ELMO | 0.69 | 0.67 | 0.68 |
| BERT-base | 0.72 | 0.69 | 0.70 |
| RoBERTa-base | 0.75 | 0.70 | 0.72 |
| DistilBERT | 0.76 | 0.74 | 0.75 |
Fig. 8probability p vs exact bias mask
Evaluation of different NER pipelines
| Model | PREC | REC | F1 | ACC |
|---|---|---|---|---|
| Core-sm | 0.59 | 0.27 | 0.37 | 0.37 |
| Core-md | 0.61 | 0.45 | 0.52 | 0.53 |
| Core-lg | 0.60 | 0.62 | 0.60 | 0.67 |
| Core-trf | 0.66 | 0.65 | 0.63 | 0.72 |
Dbias tasks
| Feature | Output |
|---|---|
| Text de-biasing | Returns de-biased news recommendations with bias probability |
| Bias classification | Classifies whether a news article is biased or not with probability |
| Bias words recognition | Extract Biased words or phrases from the news fragment |
| Bias masking | Returns the news fragment with biased words masked out |
Fig. 9Working of Dbias
Fig. 10Bias probability in a news article