Survey on Machine Learning Biases and Mitigation Techniques
<p>Steps of systematic review method.</p> "> Figure 2
<p>Patent papers in scholar year (2020–2023).</p> "> Figure 3
<p>Citations papers in scholar year (2015–2023).</p> "> Figure 4
<p>Co-related keywords.</p> "> Figure 5
<p>Quantify bias in ML.</p> "> Figure 6
<p>Most common sources of bias.</p> "> Figure 7
<p>Strategies of selection bias.</p> "> Figure 8
<p>Sample bias.</p> "> Figure 9
<p>Volunteer bias.</p> "> Figure 10
<p>Strategies for avoiding confirmation bias.</p> "> Figure 11
<p>Strategies to overcome confirmation bias.</p> "> Figure 12
<p>Types of algorithm bias.</p> "> Figure 13
<p>Adversarial training on NN.</p> "> Figure 14
<p>Causes of data bias.</p> ">
Abstract
:1. Introduction
2. Method
2.1. Database Selection
- IEEE Xplore: This database is a great resource for articles on machine learning and artificial intelligence. It includes articles from over 4000 journals, conference proceedings, and technical standards. Use keywords such as “machine learning bias” and “algorithmic fairness” to retrieve relevant articles.
- ACM Digital Library: This database is a comprehensive resource for computer science and information technology research. It includes articles from over 50 ACM journals and conference proceedings. Use keywords such as “machine learning” and “bias mitigation” to retrieve relevant articles.
- ArXiv: This database is a repository for articles in physics, mathematics, computer science and other related fields. It includes articles on machine learning bias and fairness. Use keywords such as “algorithmic bias” and “fairness in machine learning” to retrieve relevant articles.
- Google Scholar: This database is a free resource that includes articles, theses, books, and other academic literature. It is particularly useful for retrieving articles that may not be available in other databases. Use a combination of keywords and Boolean operators to retrieve the most relevant articles.
- ScienceDirect: This database is a comprehensive resource for scientific research. It includes articles from over 3800 journals and book series. Use keywords such as “machine learning” and “bias correction” to retrieve relevant articles.
- Springer Link: This database is a comprehensive resource for scientific research. It includes articles from over 2500 journals and book series. Use keywords such as “machine learning” and “algorithmic fairness” to retrieve relevant articles.
2.2. Keyword Selection
2.3. Collection of Documents and Filtering (Inclusion/Exclusion Criteria)
2.4. Bibliometric and Document Analysis
Publication and Citation Frequency
2.5. Source Analysis
2.6. Keywords Statistics
2.7. Document Analysis
3. Machine Learning Bias
- Dataset Bias
- Measure the distribution of the data: Analyzing the frequency of different attributes across the dataset helps identify potential biases. By calculating the proportions or counts of attribute categories, you can understand their representation. Over-representation or under-representation of certain attributes may indicate bias in the data. For example, if a dataset used for college admissions contains a significantly higher proportion of students from affluent backgrounds, it could indicate socioeconomic bias. For example, if a dataset used for college admissions contains a significantly higher proportion of students from affluent backgrounds compared to the general population, it could indicate socioeconomic bias. This bias may stem from inequitable access to resources or opportunities in the admissions process.
- Check for imbalances in the target variable: Target variables can lead to biased predictions, particularly for under-represented groups. It is crucial to examine the distribution of the target variable to ensure fairness. Identify whether there are significant disparities in the number of samples belonging to different target categories. For instance, in a medical diagnosis model, if the dataset has a disproportionate number of healthy patients compared to patients with a particular disease, the model might struggle to accurately predict the disease cases.
- Use statistical tests for assessing attribute distribution: Statistical tests like chi-squared tests or t-tests can provide quantitative insights into the differences in attribute distribution across different groups. These tests help determine whether there is a significant association between two categorical variables. They can be used to assess whether observed differences in attribute distribution across groups are statistically significant or due to chance. By applying these tests, you can quantify the extent of bias and ascertain if the observed differences are statistically significant or if they can be attributed to random variations.
- Scraped Data Bias
- Evaluate the sources and methods used to scrape the data to identify potential biases or inaccuracies in the data. Assess the reliability and credibility of the data sources. Consider the reputation, authority, and transparency of the sources to ensure the data are trustworthy. Evaluate the methodology employed for data scraping. Determine whether it adhered to ethical guidelines, respected user privacy, and obtained consent if required. Consider potential biases in the data sources. If the sources are known to have inherent biases or limitations, then these can impact the quality and representations of the scraped data.
- Check for missing data or errors in the scraped data that could affect the model’s predictions. Examine the scraped data for missing values or errors that can affect the model’s predictions. Missing or erroneous data can introduce bias or distort the analysis. Identify the types and patterns of missing data. Determine whether they are missing at random or if certain attributes or groups are more affected. Systematic missing can lead to biased results. Investigate the potential causes of missing data, such as technical issues during scraping or limitations in the data sources. Addressing missing data appropriately is crucial in avoiding biased or inaccurate predictions.
- Analyze the distribution of the scraped data to identify any under-represented groups or biases. Assess the distribution of attributes within the scraped data to identify under-represented groups or biases. Understanding the representation of different groups is vital for fair modeling. Calculate the frequencies or proportions of attribute categories and compare them to known distributions or benchmarks. Look for significant disparities or imbalances in attribute representation. Under-represented groups may be susceptible to biased predictions or exclusion from the modeling process. Analyzing attribute distribution helps identify potential biases, such as gender, race, ethnicity, or socioeconomic disparities, which may exist in the data.
- Abstract Data Bias
- Evaluate the methods used to generate or extract abstract data: When assessing potential biases or inaccuracies in abstract data, it is essential to scrutinize the methods used for data generation or extraction. This involves understanding the data collection process, including the sources, instruments, and techniques employed. For example, if the data were collected through surveys, evaluate whether the survey design could introduce response or sampling biases. If the data were obtained from online sources, consider the limitations of web scraping techniques and potential biases associated with the sampled websites or platforms.
- Check for missing data or errors in the abstract data: Missing data or errors can significantly impact the accuracy and validity of a model’s predictions. Carefully examine the abstract data for any missing values, outliers, or inconsistencies. Missing data can occur due to various reasons, such as non-response, data entry errors, or unintentional omissions. Investigate whether the missing data are random or if there is a systematic pattern to its absence, as this pattern could introduce biases. Depending on the extent of missing data, imputation techniques such as mean imputation, regression imputation, or multiple imputations can be employed to address the gaps and minimize bias.
- Analyze the distribution of the abstract data: Analyzing the distribution of abstract data is an essential step in understanding potential biases and under-represented groups within the dataset. It is the distribution of the abstract data that helps identify any under-represented groups or biases within the dataset. Start by examining the demographic or categorical variables in the data and determine whether they adequately represent the diversity of the target population. Look for disparities or imbalances across different groups, such as gender, race, age, or socioeconomic status. Unequal representation or significant variations in the distribution can indicate potential biases or under-representation of certain groups, which can lead to unfair predictions or outcomes. Addressing such biases may require collecting more data from under-represented groups or applying bias mitigation techniques during model training.
- Machine Learning Model Bias
- Evaluate the performance of the machine learning model across different groups to identify any disparities in the predictions. Assess the model’s performance separately for each group to understand any disparities. Calculate standard evaluation metrics such as accuracy, precision, recall, F1-score, or area under the ROC curve (AUC) for each group. By comparing these metrics across groups, you can identify variations in performance.
- Check for any bias or inaccuracies introduced during the training or evaluation of the model. Examine the pre-processing steps applied to the data during training and evaluation. Pre-processing techniques such as normalization, feature scaling, or imputation can unintentionally introduce biases if not carefully implemented. Evaluate whether the pre-processing steps are appropriate for the data and ensure they are applied consistently across different groups. Data Augmentation: Assess the use of data augmentation techniques during training. Data augmentation can help increase the diversity and robustness of the training data. However, it is important to ensure that the augmentation techniques do not introduce biases or distort the underlying distribution of the data. Regularly review and validate the augmented data to verify its quality and fairness. Model Architecture: Examine the architecture of the machine learning model itself. Biases can be introduced if the model is designed in a way that disproportionately favors certain groups or if it relies on discriminatory features. Validation and Cross-Validation: Use appropriate validation strategies during model training and evaluation. Employ techniques such as k-fold cross-validation or stratified sampling to ensure that the performance metrics are consistent across different groups. Sensitivity Analysis: Conduct sensitivity analysis to evaluate the model’s performance across different thresholds or decision boundaries. External Validation and Auditing: Seek external validation and auditing of the model’s performance. Engage independent experts or domain specialists to assess the model’s predictions and evaluate potential biases. By following these steps and considering these factors, you can thoroughly evaluate the model for bias or inaccuracies introduced during training or evaluation.
- Use fairness metrics such as demographic parity, equalized odds, and equal opportunity to measure bias in the predictions. Demographic parity measures whether the predictions of a model are independent of sensitive attributes such as gender, race, or age. It ensures that individuals from different demographic groups have equal chances of receiving positive outcomes. To evaluate demographic parity, you can compare the proportion of positive predictions across different groups. Equalized odds assess whether the model’s predictions are consistent across different groups, considering both false positives and false negatives. It focuses on maintaining equal false positive rates and equal true positive rates across different subgroups. Equal opportunity evaluates whether the model provides an equal opportunity for positive outcomes across different groups, specifically focusing on the true positive rates. These fairness metrics help quantify and measure bias in machine learning models by focusing on the disparate impact on different groups. It is important to note that the choice of fairness metrics depends on the specific context and the sensitive attributes relevant to the problem at hand. In addition to these metrics, other fairness measures such as predictive parity, treatment equality, or counterfactual fairness may also be considered, depending on the requirements and constraints of the application.
- Prediction Bias
- Analyzing predictions across different groups: To identify disparities or inaccuracies in the model’s predictions, it is crucial to conduct a thorough analysis across different groups. Divide the dataset into subgroups based on relevant attributes such as race, gender, age, or socioeconomic status. Evaluate the model’s performance metrics, such as accuracy, precision, recall, or F1 score, for each subgroup. Compare these metrics across groups to identify any significant variations or disparities in the model’s predictions. Visualizations, such as confusion matrices or ROC curves, can help in understanding the prediction behavior across different groups.
- a.
- Chi-square test: This test can determine whether the differences in prediction outcomes across groups are statistically significant.
- b.
- t-test or ANOVA: These tests can be applied to compare prediction scores or probabilities between different groups and evaluate if the differences are statistically significant.
- c.
- Fairness metrics: Demographic parity, equalized odds, and equal opportunity are fairness metrics that quantify disparities in prediction outcomes across different groups. Calculating these metrics and comparing them between groups can help identify bias in the model’s predictions.
- Checking for biases or inaccuracies in the data: Biases or inaccuracies in the data used for making predictions can lead to biased model outcomes. It is crucial to check for potential biases or inaccuracies in the data and address them appropriately. Consider the following aspects:
- a.
- Data Collection Bias: Assess whether the data used for training the model are representative of the target population. Biases can arise if certain groups are under-represented or over-represented in the training data.
- b.
- Labeling Bias: Examine the quality and accuracy of the labels or annotations in the training data. If stereotypes, cultural biases, or subjective judgments influence the labeling process, then biases may occur.
- c.
- Feature Selection Bias: Evaluate whether the features used for prediction are fair and unbiased. Biases can be unintentionally encoded in the features if they correlate with protected attributes or capture societal prejudices.
- Data used for training: Because machine learning (ML) models are data-driven, they may be biased if the training data are not diverse or representative of the community being studied. A facial recognition algorithm, for instance, may have trouble correctly identifying people with darker skin tones if it has been trained mainly on images of white people. In the case of facial recognition algorithms, which are widely used in various applications such as identity verification and surveillance systems, biased training data can result in significant disparities in performance across different demographic groups. For example, if the training data predominantly consists of images of white individuals, then the algorithm may struggle to accurately identify people with darker skin tones.
- Data selection: A machine learning (ML) model’s training data may not be a representative sample of the entire community. This may occur if the data are gathered in an unfair manner, such as by excluding some categories or oversampling some groups. One common scenario where data selection bias can occur is when data collection processes systematically exclude or under-represent certain categories or groups. For example, in a healthcare dataset, if data are primarily collected from a specific demographic or geographic region, then they may not accurately capture the experiences and health conditions of other populations. This can result in biased predictions or limited generalizability of the model to broader populations.
- Architecture of the algorithm: The ML algorithm’s architecture can introduce bias. For instance, an algorithm may be more prone to bias if it heavily depends on a single trait that is associated with a specific group. Bias can arise when the algorithm heavily relies on a single trait or feature that is associated with a specific group, leading to discriminatory or unfair predictions.
- Feedback loops: Feedback loops can happen when a machine learning (ML) model’s predictions are used to inform choices that are then fed back into the model. It can perpetuate and amplify biases over time if the input reinforces pre-existing biases in the model. Feedback loops in machine learning models can contribute to the perpetuation and amplification of biases. When a model’s predictions are used to inform decisions or actions, and those decisions are subsequently fed back into the model as new data, it can create a cycle that reinforces pre-existing biases.
- Human biases: Last but not least, human biases can be incorporated into ML algorithms. This might occur if the people in charge of creating or training the model have prejudices of their own that affect the choices they make.
How Does It Work?
4. Bias Reduction Strategy
- Diverse and representative training data: Using diverse and representative training data are one of the most efficient methods to reduce bias. This can make sure that the data used to train the ML model represents the complete range of experiences and viewpoints of the population being studied. Utilizing diverse and representative training data is crucial in minimizing bias. This can be achieved by ensuring that the training dataset, denoted as D, contains a wide range of examples from different subgroups or classes. Mathematically, we can represent this as Equation (2):
- Data pre-processing: Techniques for data pre-processing can be used to find and eliminate prejudice in the training data. To balance the representation of various subgroups in the training data, methods like oversampling or undersampling may be used. Mathematically, this can be represented as Equation (3):
- Algorithmic transparency: By making it simpler to spot and correct any possible biases in the ML model, ensuring algorithmic transparency can help to mitigate bias. This might entail employing strategies like interpretability methods, which can make the ML model’s decision-making process more visible. Mathematically, this can be represented as Equation (4):
- Regular assessment and monitoring: Monitoring and evaluating the ML model on a regular basis can help to spot any biases that may exist and help to correct them as needed. This might entail methods like fairness measures, which are useful for assessing how well the ML model performs across various subgroups. Mathematically, we can represent this Equation (5):This equation signifies that the fairness of the ML model is determined by the assessment conducted on it.
- Adversarial training: It entails purposefully introducing bias into the training data to increase the ML model’s resistance to bias. This can ensure that the model can handle biased data more effectively when they are encountered in the real world. See mathematical Equation (6)In Equation, the symbol represents the total loss. L denotes the original loss function, represents the predicted output, y represents the true label, is the hyperparameter controlling the weight given to the adversarial loss term, and represents the adversarial loss term. The adversarial loss term measures the difference between the model’s predictions on the perturbed input and the true label y.
4.1. Types of Bias and Their Reduction Strategies
4.2. Selection Bias
4.2.1. Categories of Selection Bias
- Sampling bias: Sampling bias occurs when the data are not chosen at random, resulting in a non-representative sample. This occurs when the data are in a non-representative sample. This can lead to inaccurate and misleading conclusions. One can mitigate sampling bias through employing various techniques (Figure 8).
- Volunteer bias: Volunteer bias is a type of bias that occurs when individuals who choose to participate in a study are not representative of the population being studied. Specifically, volunteer bias occurs when individuals who volunteer for a study are systematically different from those who do not volunteer. This can result in a biased sample that does not accurately reflect the population of interest [37]. Volunteer bias can occur for a variety of reasons. Volunteer bias is represented in Figure 9.
- Survivorship bias: This occurs when the sample is biased towards people who have survived a particular event or process. For example, if a study on the long-term effects of a particular treatment only includes people who have survived for a certain amount of time, then it may not represent the entire population of people who received the treatment [40].
- Time interval bias: Time interval bias arises when the time intervals or durations of observation or follow-up are systematically different. Time interval bias is a type of selection bias that can occur in studies where the exposure and outcome occur over different time intervals. This bias arises when the time intervals used to measure exposure and outcome are not aligned or are different for different study subjects [45]. Time interval bias can lead to incorrect conclusions about the relationship between the exposure and outcome, and it is important to consider this potential bias when designing and interpreting study results. To avoid time interval bias, researchers should consider aligning the time intervals for measuring exposure and outcome or adjusting for any differences in the time intervals when analyzing the data [46]. Time interval bias can affect the validity and generalizability of research findings by leading to inaccurate or biased results. It is important to minimize time interval bias to ensure accurate and reliable research findings.
- Berkson’s bias: This occurs when the sample is biased because of the way participants were selected. For example, if a study on the relationship between two medical conditions only includes people who have been admitted to a hospital, then it may not represent the general population because hospital patients are likely to have multiple medical conditions. This is a type of selection bias that can occur in statistical studies [48]. It occurs when the selection criteria for a study create a non-random sample that is different from the general population in a way that affects the relationship between two variables. Specifically, it occurs when the sample includes only individuals who have a particular condition or disease and also have a particular unrelated attribute or risk factor that is not present in the general population. This can create a spurious or inflated relationship between the condition or disease and the unrelated attribute or risk factor [49].
- Healthy user bias: This occurs when the sample is biased because of the characteristics of the participants. For example, if a study on the health effects of a particular supplement only includes people who take the supplement regularly, then it may not represent the general population because people who take supplements regularly may also have other healthy habits. There are various strategies for mitigating healthy user bias. One approach is to use randomization to assign participants to different groups, including a control group that does not take the supplement. By randomly assigning participants, researchers can help ensure that the characteristics of the participants are balanced across the groups, reducing the impact of healthy user bias.
- Prevalence–incidence bias: Prevalence–incidence bias is a type of bias that can occur in cross-sectional studies when the prevalence of a disease or condition influences the measurement of its incidence. Prevalence refers to the proportion of individuals in a population who have a particular disease or condition at a specific point in time, while incidence refers to the number of new cases of a disease or condition that occur over a specific period of time [53]. In cross-sectional studies, both prevalence and incidence may be measured simultaneously, which can create a bias if the prevalence of the disease or condition is related to the duration of the disease or condition.
4.2.2. Examples of Selection Bias
- Bias in facial recognition technology: The training process for facial recognition algorithms typically involves feeding the system a large dataset of facial images to learn patterns and features for accurate identification and matching. Facial recognition technology has been found to have a bias against people with darker skin tones, due to the way the algorithms were trained [56]. This is because the training data used to develop the algorithms did not include a diverse enough sample of individuals with different skin tones. As a result, the technology may not accurately identify or match individuals with darker skin tones [57].
- Bias in hiring algorithms: AI-powered hiring algorithms may introduce selection bias if the data used to train the algorithms contain biases against certain groups of people [58]. For example, if the data used to train the algorithm contain a disproportionate number of resumes from men candidates, then the algorithm may favor men candidates over female candidates. Organizations must be aware of the potential biases. It is possible to mitigate selection bias and create a more inclusive and equitable hiring process.
- Bias in predictive policing: Predictive policing algorithms use historical crime data to predict future crime patterns and allocate police resources accordingly [59]. However, if the historical data contain biases against certain groups of people, the algorithm may perpetuate or amplify these biases by targeting certain neighborhoods or individuals more heavily than others [60].
- Bias in chatbot: Chatbot is a program that uses AI to imitate human-like discussions with users. It understands user inputs, generates relevant responses, and provides information. Chatbots may exhibit selection bias if they are trained on a biased sample of conversations [61]. For example, if a chatbot is trained on conversations between customers and customer service representatives, it may not accurately respond to non-native English speakers or people with different communication styles [62].
- Bias in healthcare: Healthcare algorithms are used to treat patients may introduce selection bias if the training data contain biases against groups of patients [63]. For example, if the data used to train an algorithm only include data from white patients, the algorithm may not accurately diagnose or treat patients from other racial or ethnic backgrounds [64]. This can result in disparities in healthcare outcomes and access for marginalized communities.
- Inherent Sampling Challenges: Real-world datasets are multidimensional, it is difficult to eliminate bias in careful sample selection.
- Complicated Source Identification: Selection bias is difficult task that makes the mitigation process more complicated, particularly when dealing with datasets that have a large number of variables.
4.2.3. Minimize Selection Bias
4.2.4. Limitation
4.3. Confirmation Bias
4.3.1. Measurement of Confirmation Bias
- Self-Report Measures: Self-report measures are indeed commonly used to assess confirmation bias. These measures involve asking individuals directly about their attitudes, beliefs, and decision-making processes related to information processing. Questionnaires and interviews are two common methods employed to gather these self-reported data.Questionnaires typically consist of a series of statements or items that individuals respond to by indicating their level of agreement or disagreement. These statements are designed to assess various aspects of confirmation bias, such as the tendency to seek information that supports one’s existing beliefs while dismissing or ignoring conflicting evidence. Interviews, on the other hand, involve direct conversations between researchers and participants. Researchers can ask open-ended questions to explore individuals’ thought processes, information-seeking behaviors, and their inclination toward confirming their pre-existing beliefs. Self-report measures provide valuable insights into individuals’ subjective experiences and perceptions of confirmation bias [66]. However, it is important to note that self-reported data can be influenced by social desirability bias, where individuals may provide responses that align with societal norms or what they believe is expected of them. To address potential biases in self-report measures, researchers may employ additional techniques to complement the self-report data. This can include behavioral observations, cognitive tasks, or physiological measures to provide a more comprehensive assessment of confirmation bias.
- Cognitive Tasks: Researchers have developed a variety of cognitive tasks to measure confirmation bias. One common task is the “Wason selection task”, which asks participants to evaluate a rule by selecting cards that could confirm or dis-confirm it. The task measures the degree to which people are biased toward selecting information that confirms their preexisting beliefs [67]. In conclusion, cognitive tasks, including the Wason selection task, are useful tools for researchers to measure confirmation bias. They allow for controlled assessment of individuals’ bias in selecting and interpreting information, providing valuable insights into the cognitive mechanisms underlying confirmation bias.
- Implicit Association Test: The Implicit Association Test (IAT) is a popular method for measuring unconscious biases, including confirmation bias [67]. The IAT measures the speed with which people categorize words or images as either confirming or disconfirming their beliefs. In the context of confirmation bias, the IAT can be used to assess individuals’ automatic associations between belief confirmation and disconfirmation. It typically involves presenting participants with a series of words or images related to confirming or disconfirming beliefs. Participants are then required to categorize these stimuli as quickly as possible into relevant categories.
- Behavioral Measures: Researchers can also measure confirmation bias through participants’ behavior [68]. For example, they may observe whether participants selectively seek out and attend to information that confirms their beliefs while ignoring information that does not support them. By analyzing participants’ behavioral responses, researchers can gain insights into the manifestation of confirmation bias in real-world decision making and information processing. This approach provides direct evidence of individuals’ biased behaviors and preferences, highlighting the impact of confirmation bias on their interactions with information.
4.3.2. Strategies to Overcome Confirmation Bias
- a.
- Self-awareness: Recognize that you have biases and that they can influence your perception of information. Acknowledge that your beliefs and assumptions may be based on incomplete or inaccurate information.For example, we have a strong belief that a particular political party is always right, and any information that contradicts this belief is automatically dismissed. By being aware of your own bias, you can recognize when you are automatically dismissing information that challenges your beliefs.
- b.
- Seek out diverse perspectives: Make an effort to seek out information and views that challenge your preconceptions. Expose yourself to a variety of opinions and perspectives.For example, we are researching a topic and find that most of the information we are finding supports a particular view. By seeking out diverse perspectives, you can find information that challenges this view and gain a more complete understanding of the topic.
- c.
- Information sources: Evaluate the source of information and consider whether it is credible, reliable, and unbiased. Check the author’s credentials, reputation, and potential conflicts of interest.For example, we come across a blog post that makes a strong argument for a particular viewpoint. However, upon further investigation, you discover that the author of the post has a strong bias and a financial stake in promoting that viewpoint. By considering the source of information, you can recognize when information is biased and evaluate it accordingly.
- d.
- Evidence evaluation: Evaluate the evidence objectively and consider whether it supports your preconceptions or not. Look for counter-evidence and consider alternative explanations. For example, you are presented with evidence that seems to support your preconceived beliefs about a particular topic. However, upon closer examination, you realize that the evidence is based on flawed assumptions. By evaluating the evidence objectively, you can recognize when evidence is flawed and avoid being misled.
- e.
- Critical thinking: Ask questions and challenge assumptions to avoid jumping to conclusions. Use logic and reason to evaluate information.For example, you come across a news story that seems to support a particular viewpoint. However, upon further examination, you realize that the story is based on incomplete information and makes unwarranted assumptions. By engaging in critical thinking, you can recognize when information is incomplete or inaccurate and avoid being misled.
- f.
- Practice empathy: Try to understand the perspectives of others, even if they are different from your own. Put yourself in their shoes and consider how they arrived at their beliefs.For example, you hold strong beliefs about a particular topic and are frustrated by people who hold different beliefs. By practicing empathy, you can recognize that people come from different backgrounds and have different experiences that shape their beliefs.
- g.
- Time management: Avoid making quick judgments and take the time to carefully consider information. Be patient and gather all the relevant information before making a decision.For example, you are presented with information that seems to support a particular viewpoint, but you are not sure if it is accurate. By taking your time, you can gather more information, evaluate the evidence, and avoid jumping to conclusions.
- h.
- Seek feedback: Seek feedback from others to gain a different perspective and challenge your own thinking. Consider the opinions of people who disagree with you and be open to constructive criticism.For example, you are working on a project and have developed a particular hypothesis. By seeking feedback, you can obtain input from others who hold different viewpoints, challenge your assumptions, and improve the quality of your work.
4.3.3. Limitation
4.4. Algorithmic Bias
4.4.1. Types of Algorithm Bias
- Sample Bias: This occurs when the training dataset is not representative of the population being modeled. If the training dataset favors one group of people over another, the model that is made might not do well with other groups [71].
- 2.
- Labeling bias: Labeling bias is a type of algorithm bias that occurs when the labels or categories assigned to the training data are not accurate or representative of the population being predicted. This can cause the machine learning model to make incorrect predictions and fail to capture important patterns in the data.
- 3.
- Model bias: Model bias is a type of algorithm bias that occurs when the assumptions and constraints of a machine learning model do not accurately represent the real-world problem being solved. This can cause the model to make incorrect predictions and fail to capture important patterns in the data.
- 4.
- Measurement Bias: This occurs when the way that data are collected or measured is biased. Measurement bias is a type of algorithm bias that occurs when the measurements used to train a machine learning model are inaccurate or biased. This can happen if the instruments used to collect the data are faulty or if the data collectors have a bias that affects their measurements [72].
- 5.
- Feedback Loop Bias: This occurs when the output of a machine learning model is used to inform future decisions, which can create a feedback loop that reinforces any biases in the data or the model itself [73]. Feedback loop bias is a type of algorithm bias that occurs when the output of a machine learning model is used to make decisions that affect the input data used to train the model. This can create a self-reinforcing loop where the model becomes more biased over time.
- 6.
- Confirmation Bias: Confirmation bias is a cognitive bias that refers to the tendency of people to search for, interpret, and remember information in a way that confirms their preexisting beliefs, while ignoring or downplaying information that contradicts those beliefs. In other words, people tend to seek out information that confirms their beliefs and ignore information that challenges them.
- 7.
- Overfitting Bias: Overfitting bias is a type of algorithm bias that occurs when a machine learning model becomes too complex and captures noise or random fluctuations in the training data, rather than the underlying pattern. This can cause the model to perform well on the training data but poorly on new data.
- 8.
- Underfitting Bias: Underfitting bias is a type of algorithm bias that occurs when a machine learning model is too simple and does not capture the underlying pattern in the training data. This can cause the model to perform poorly on both the training data and new data.
4.4.2. Causes of Algorithm Bias
- a.
- Biased training data: Machine learning algorithms rely on large datasets to learn patterns and make predictions [36]. If the training data are biased or unrepresentative of the real-world population, the algorithm may learn and reproduce those biases in its outputs.
- b.
- Biased algorithm design: The way algorithms are designed can also introduce bias, such as using features that are highly correlated with protected characteristics (e.g., race, gender, age) or weighting certain features more heavily than others. Biases can also arise from the choice of performance metrics used to evaluate the algorithm [70].
- c.
- Biased human decision making: Human decision making can also contribute to algorithmic bias, such as in the selection and labeling of training data, the choice of features and performance metrics, and the decision to deploy the algorithm in certain contexts. Biases can also arise from human cognitive biases, such as confirmation bias, anchoring bias, and group [36].
- d.
- Lack of diversity and inclusion: The lack of diversity and inclusion in the technology industry can also contribute to algorithmic bias, such as in the composition of development teams, the selection of training data, and the deployment of the algorithm in different contexts.
4.4.3. Mitigate of Algorithm Bias
4.4.4. Limitation
4.5. Adversarial Training
- Generate adversarial: Use an algorithm, such as the Fast Gradient Sign Method (FGSM) or Projected Gradient Descent (PGD) to generate adversarial examples from the original dataset. Adversarial examples are created by making small perturbations to the original data points in a way that maximizes the model’s prediction error.
- Fast Gradient Sign Method: The Fast Gradient Sign Method (FGSM) is a simple gradient-based algorithm for generating adversarial examples. Given an input data point x and a neural network classifier with loss function where are the model’s parameters, y is the ground truth label, and J measures the difference between the predicted and ground truth labels; FGSM generates an adversarial example by adding a small perturbation to the input data point that maximizes the loss function. The perturbation is computed as Equation (21):The perturbed data point is then given by (22):This new data point is then passed through the neural network, resulting in the misclassification of the input point. FGSM is fast and easy to implement, but may not always generate robust adversarial examples that are resistant to other perturbation techniques.
- Projected Gradient Descent (PGD): This is an iterative algorithm for generating adversarial examples that build on the FGSM approach. Given an input data point x, a neural network classifier with loss function , and a maximum perturbation size ; PGD generates an adversarial example by iteratively applying the FGSM method with a small step size until convergence or until the maximum number of iterations T is reached. At each iteration t, the perturbed data point is projected onto the -ball centered at x, ensuring that the perturbation size does not exceed the specified limit.The iterative update rule for PGD is given by Equation (23):The perturbed data point is then given by Equation (25):PGD can be seen as a stronger attack than FGSM because it iteratively refines the perturbation to find an adversarial example that is more effective in fooling the classifier. However, PGD requires more computational resources and may take longer to converge than FGSM.
- CW attack: The Carlini–Wagner (CW) attack is an optimization-based approach to generating adversarial examples. Here are the mathematical details:Let x be the original input, be the output of the target classifier with parameters , and y be the true label of x. The goal of the CW attack is to find a perturbed input that maximizes the loss function . The optimization problem can be formulated as Equation (26):To solve the optimization problem, Carlini and Wagner proposed a differentiable surrogate loss function that upper bounds the true loss function. They also introduced a change of variables that maps the perturbation to the unconstrained space of a real vector w such that . The optimization problem becomes Equation (28):The optimization problem is solved using an iterative algorithm that alternates between updating w using gradient descent and projecting w onto the feasible set. The projection step enforces the constraints on the perturbation size and the valid input range. CW attack is a very powerful and effective attack method that can often evade state-of-the-art defense mechanisms. However, it is also more computationally expensive and difficult to implement compared to other attack methods.
- DeepFool algorithm: The DeepFool algorithm is an iterative algorithm for generating adversarial examples. It works by finding the closest decision boundary to the input data point and then iteratively moving the data point toward that decision boundary until it is misclassified. Here are the mathematical details of the DeepFool algorithm:Given an input data point x and a classifier f, the goal of DeepFool is to find a small perturbation such that is misclassified by the classifier f. The algorithm starts by initializing the perturbation to zero and then iteratively updates it using the following Equation (29):
- JSMA: Jacobian-based Saliency Map Attack (JSMA) is a gradient-based adversarial attack method that perturbs the input data by identifying and modifying the most salient features of the data. The key idea behind the JSMA attack is to find the minimal set of features in the input that needs to be modified in order to change the classification output of the model. This is achieved by computing the Jacobian matrix of the model’s output with respect to the input and selecting the input features that have the highest influence on the output. The JSMA attack is performed in two steps:
- a.
- Compute the saliency map: The saliency map is a matrix that represents the sensitivity of the model’s output to changes in each input feature. It is computed by taking the absolute value of the Jacobian matrix and multiplying it element-wise with the sign of the difference between the model’s predicted class and the target class.
- b.
- Perturb the input: The input is then perturbed by modifying the most salient features identified in the saliency map until the model’s prediction changes to the target class.
The optimization problem for the JSMA attack can be written as Equation (31):- (a)
- Compute the saliency map:, where t is the true label of the input data.
- (b)
- Find the most salient features: Sort the features in the saliency map in descending order of importance.
- (c)
- Perturb the input: Modify the input data by adding a perturbation vector that maximally changes the model’s prediction to the target class label while keeping the perturbation vector sparse. perturbation vector sparse.
- (d)
- Repeat steps 1–3 until the model’s prediction changes to the target class label.
The JSMA attack is effective in generating small perturbations that are hard to detect, but it can be computationally expensive as it requires multiple iterations to find the most salient features and perturb the input data.
- Add adversarial examples to the training set: Add the generated adversarial examples to the original training set. This creates a new, larger dataset that includes both the original examples and the adversarial examples.
- Train the model on the combined dataset: Train the model on the new, combined dataset that includes both the original examples and the adversarial examples. This helps the model learn to be more robust to bias and better generalize to unseen examples.
- Evaluate the model on the original test set: After training the model on the combined dataset, evaluate its performance on the original test set. This will give you an idea of how well the model generalizes to new, unseen examples.
- Repeat the process: If necessary, repeat the process of generating new adversarial examples and adding them to the training set and retrain the model until you are satisfied with its performance.
4.5.1. Benefits and Drawbacks of Adversarial Training
4.5.2. Common Techniques for Defending against Adversarial Attacks
4.5.3. Purposes of Use Adversarial Attacks
4.5.4. Limitation
4.6. Fairness Constraints
- Demographic parity: This calls for the model to produce the same results for all groups, independent of their delicate characteristics. This can be mathematically stated as Equation (32):
- Equalized odds: This calls for the model to offer comparable rates of true positives and false positives for each category. Mathematically, this can be expressed as Equation (33):
- Conditional independence: This demands that, given the other inputs, the model’s output be independent of the sensitive attribute. This can be stated mathematically as Equation (34):
4.6.1. Importance of Fairness Constraints
4.6.2. Fairness Constraints Incorporated into Machine Learning Algorithms
4.6.3. Future Evolution and Research Directions for Fairness Constraints
- a.
- Intersectional Fairness: This involves considering multiple dimensions of identity, such as race and gender when evaluating fairness. This approach recognizes that individuals may experience discrimination or bias due to the intersection of multiple factors, rather than a single factor alone.
- b.
- Fairness in Contextual Decision Making: Contextual decision making refers to decisions that are made in specific contexts or situations. Fairness constraints can be difficult to apply in these contexts, as they may require a nuanced understanding of the factors that influence decision making.
- c.
- Fairness in Deep Learning: Deep learning algorithms are becoming increasingly popular, but they can be challenging to ensure fairness. Researchers are exploring ways to develop fairness constraints that can be applied to deep learning algorithms.
- d.
- Fairness in Reinforcement Learning: Reinforcement learning involves learning through trial and error, and fairness constraints can be difficult to apply in this context. Researchers are exploring ways to develop fairness constraints that can be applied to reinforcement learning algorithms.
- e.
- Fairness in Privacy-Preserving Machine Learning: Privacy-preserving machine learning involves using cryptographic techniques to ensure that data remains private while being used to train machine learning algorithms. Researchers are exploring ways to develop fairness constraints that can be applied in the context of privacy-preserving machine learning.
- f.
- Fairness in Human–AI Collaboration: As AI systems become more integrated into our daily lives, it is important to ensure that they are fair and equitable. Researchers are exploring ways to develop fairness constraints that can be applied in the context of human–AI collaboration.
4.6.4. Common Type of Fairness Constraints
- 1.
- Demographic parity: This type of fairness constraint requires that the algorithm produces similar outcomes for different demographic groups. For example, if the algorithm is being used to make lending decisions, demographic parity would require that people from different races or genders are approved for loans at the same rate. Demographic parity does not take into account any differences in the underlying characteristics or risk factors of the different groups.
- 2.
- Equal opportunity: This type of fairness constraint ensures that the algorithm provides equal opportunity to people from different demographic groups. This means that all qualified individuals should have an equal chance of being selected or chosen, regardless of their demographic group. For example, if the algorithm is being used to screen job applicants, the equal opportunity would require that people from different races or genders are selected for interviews at the same rate, provided that they meet the qualifications for the job.
- 3.
- lEqualized odds: This type of fairness constraint ensures that the algorithm achieves similar levels of accuracy across different demographic groups. This means that the algorithm should produce similar rates of true positives, false positives, true negatives, and false negatives for different groups. For example, if the algorithm is being used to diagnose a medical condition, equalized odds would require that the algorithm achieves similar rates of correct diagnoses for people from different races or genders.
- 4.
- Individual fairness: This type of fairness constraint ensures that the algorithm treats similar individuals in a similar way, regardless of their demographic group. This means that the algorithm should produce similar outcomes for people who have similar characteristics or risk factors, regardless of their race, gender, or other characteristics. For example, if the algorithm is being used to determine credit scores, individual fairness would require that people with similar credit histories receive similar scores, regardless of their race or gender.
4.6.5. Evaluate and Measure Fairness
- a.
- Statistical Parity: Statistical parity refers to the proportion of individuals from different groups who receive a positive outcome (e.g., job offer, loan approval) from the algorithm. If the proportion of individuals receiving positive outcomes is the same across all groups, then algorithm is considered fair in terms of statistical parity.
- b.
- Equal Opportunity: Equal opportunity refers to the proportion of individuals from different groups who are qualified for a positive outcome (e.g., meet job qualifications, have good credit score) and receive it from the algorithm. If the proportion of qualified individuals receiving positive outcomes is the same across all groups, the algorithm is considered fair in terms of equal opportunity.
- c.
- Predictive Parity: Predictive parity refers to the accuracy of the algorithm in predicting outcomes for different groups of individuals. If the accuracy is the same across all groups, the algorithm is considered fair in terms of predictive parity.
- d.
- Group Fairness: Group fairness refers to the fairness of the algorithm for each group of individuals. If the algorithm is fair for each group of individuals, it is considered fair in terms of group fairness.
4.6.6. Real-World Applications and Impact of Fairness Constraints
- Hiring: Fairness constraints are being used to ensure that hiring algorithms do not discriminate against individuals on the basis of race, gender, or other protected characteristics.
- Credit scoring: Fairness constraints are being used to ensure that credit scoring algorithms do not discriminate against individuals on the basis of race, gender, or other protected characteristics.
- Criminal justice: Fairness constraints are being used to ensure that algorithms used in the criminal justice system do not discriminate against individuals.
4.6.7. Limitation
4.7. Data Label Bias
4.7.1. Common Causes of Data Label Bias
- a.
- Human error: Human error can occur due to various reasons, such as fatigue, lack of attention, or distractions. It is important to have a quality assurance process in place to review the labeled data for errors and inconsistencies and correct them before using the data to train machine learning models. Additionally, providing clear instructions and guidelines to annotators and training them on the labeling process can help minimize human errors in data labeling. Human bias refers to the influence of personal beliefs, values, and experiences on decision making. It can be conscious or unconscious and can affect the collection, analysis, and interpretation of data.
- b.
- Lack of diversity: Data may not be representative of the entire population, leading to biases in labeling. For example, if a dataset is biased toward a particular demographic or geographic region, it may lead to biased labeling. Suppose you have a dataset of medical images used to train a machine learning model to diagnose skin cancer. If the dataset contains mostly images of light-skinned patients, it may lead to biased labeling and inaccurate model predictions for patients with darker skin tones. This can be a serious issue since skin cancer can manifest differently in people with different skin colors, and inaccurate predictions can lead to delayed or incorrect diagnosis and treatment. Similarly, if a dataset used to train a speech recognition model only contains audio recordings of men’s voices, the model may have difficulty recognizing and transcribing female voices accurately, leading to biased model performance.
- c.
- Subjectivity: In some cases, the labeling process may be subjective and open to interpretation. This can lead to different annotators labeling data differently, resulting in inconsistencies and bias. Suppose you have a dataset of product reviews and are asked to label them as positive or negative. In some cases, a review may have both positive and negative aspects, and different annotators may label it differently based on their interpretation of the review’s overall sentiment. For example, one annotator may label a review that says, “The product works well, but the packaging was damaged” as positive, while another may label it as negative. Similarly, if you have a dataset of customer support chat transcripts and are asked to label them based on the customer’s satisfaction level, different annotators may interpret the customer’s tone and language differently, leading to inconsistent and biased labeling.
- d.
- Insufficient training data: If there is not enough training data, annotators may rely on assumptions or biases when labeling data, leading to incorrect or biased labels. Suppose you have a dataset of medical images used to train a machine learning model to diagnose a rare medical condition. If the dataset contains only a few examples of rare conditions, annotators may rely on their assumptions or biases when labeling new data, leading to incorrect or biased labels. Similarly, if you have a dataset of audio recordings used to train a speech recognition model, but the dataset is not diverse enough in terms of accents or languages, annotators may rely on assumptions or biases when labeling new data, leading to biased model performance.
- e.
- Prejudice: In some cases, annotators may have conscious or unconscious biases towards certain groups or characteristics, leading to biased labeling. Prejudice in data label bias occurs when the labels assigned to the data used to train a machine learning algorithm are influenced by pre-existing biases and stereotypes, which can lead to biased predictions by the algorithm. Here are some examples of prejudice in data label bias:
- Gender Bias: If a machine learning algorithm is trained on data that are biased towards a particular gender, it may lead to biased predictions. For instance, if a dataset used to train a hiring algorithm contains more men candidates than women candidates, the algorithm may be biased toward hiring men.
- Racial Bias: Prejudice in data label bias can also result in racial bias. For example, if a facial recognition algorithm is trained on a dataset that has predominantly white faces, it may struggle to accurately recognize the faces of people of different skin colors, leading to biased predictions.
- Socioeconomic Bias: Prejudice in data label bias can also lead to socioeconomic bias. For instance, if a credit scoring algorithm is trained on data that are biased toward individuals with high incomes, it may unfairly deny loans to individuals with low incomes.
- Cultural Bias: Cultural bias is another form of prejudice in data label bias. For example, if a natural language processing algorithm is trained on text written in one language, it may struggle to accurately understand text written in another language, leading to biased predictions.
4.7.2. Mitigate Prejudice in Data Label Bias
- Incomplete labeling: If some data are not labeled or missing, it can lead to bias in the final dataset. For example, if a dataset contains only positive samples, it may lead to biased predictions and inaccurate model performance. Incomplete labeling occurs when the labels assigned to the data used to train a machine learning algorithm are incomplete or missing, which can result in a model that is less accurate and less reliable.
- Partially Labeled Data: Partially labeled data are a common form of incomplete labeling. For instance, if a dataset used to train a machine learning algorithm has missing labels for some data points, the model may not be able to learn patterns from these data points, leading to a less accurate model.
- Noisy Data: Noisy data are another form of incomplete labeling. If a dataset contains data points that have labels that are inaccurate or incorrect, the model may learn patterns from these data points, leading to a less reliable model.
- Limited Labeling: Limited labeling occurs when the labels assigned to the data are not sufficient to capture all of the information present in the data. For example, if a dataset used to train a sentiment analysis algorithm only has binary labels for positive or negative sentiment, it may not be able to capture more nuanced sentiments, leading to a less accurate model.
- Outdated Data: Outdated data are another form of incomplete labeling. If a machine learning algorithm is trained on outdated data, it may not be able to accurately predict outcomes in the current environment, leading to a less reliable model.
4.7.3. Data Label Bias Impact Machine Learning Algorithms
- Inaccurate Predictions: If the training data used to train the machine learning algorithm contain biased labels, then the model will learn to make inaccurate predictions. For example, if a dataset for facial recognition algorithms contains mostly images of lighter-skinned individuals, then the model may struggle to accurately recognize individuals with darker skin tones.
- Unfairness: Data label bias can also lead to unfairness in machine learning algorithms. For instance, if a dataset used to train a hiring algorithm is biased towards hiring men candidates, then the algorithm may continue to discriminate against women candidates during the hiring process.
- Over generalization: If the training data used to train a machine learning algorithm contain biased labels, then the model may overgeneralize the learned patterns. For example, if a model is trained on biased data that suggest that all dogs are small, then the model may struggle to accurately recognize larger dog breeds.
- Lack of Diversity: If the training data used to train a machine learning algorithm are biased, then the model may not be able to accurately predict outcomes for under-represented groups. This can lead to a lack of diversity in the predictions made by the model. To mitigate data label bias, it is important to use diverse and representative data in the training of machine learning algorithms. This can be achieved by using diverse data sources, carefully selecting the data used in the training set, and implementing bias detection and correction techniques. Additionally, it is important to periodically review the model and data to ensure that they remain accurate and fair over time.
4.7.4. Some Techniques for Mitigating Data Label Bias
- 1.
- Collecting more diverse data: One way to reduce bias is to ensure that the data used to train a machine learning model are diverse and representative of the population they aim to serve. This can be achieved by collecting data from a variety of sources and ensuring that they include individuals from different backgrounds and experiences [77].
- 2.
- Using multiple annotators: Using multiple annotators to label data can help mitigate label bias by reducing the impact of individual biases. By aggregating the labels provided by multiple annotators, the final label is less likely to be influenced by the bias of any single individual [79].
- 3.
- Training with unbiased examples: Another approach is to augment the training data with examples that are known to be unbiased. This can help to balance the training set and reduce the impact of label bias [82].
- 4.
- Using debiasing techniques: Debiasing techniques can be used to adjust the labels assigned to data to make them more accurate and less biased [85]. This can be achieved through techniques such as re-weighting, which adjusts the importance of different data points based on their perceived bias, or adversarial training, which trains the model to be robust to different types of bias [86].
- 5.
- Regularly monitoring and auditing the data: It is important to regularly monitor and audit the data to ensure that it remains unbiased over time. This can involve reevaluating the labeling process and correcting any biases that are identified.
- 6.
- Exploring alternative labels: It may be helpful to explore alternative labeling schemes that could provide more accurate and less biased labels. This could involve working with domain experts or consulting with the intended users of the machine learning model to identify more appropriate labels. When training a model to recognize emotions in text, alternative labels such as “positive”, “negative”, and “neutral” could be used instead of more subjective labels like “happy” or “sad”. This can help to mitigate bias and improve the model’s accuracy.
4.7.5. Detect Data Label Bias in Labeled Data
- Data analysis: You can perform statistical analysis on the labeled data to identify any patterns or imbalances in the distribution of the labels. Look for any categories that are over-represented or under-represented in the data.
- Human review: Have human reviewers examine the labeled data to determine whether there are any inconsistencies or inaccuracies in the labeling. This can be performed through manual inspection or crowd-sourced reviews.
- Evaluation metrics: You can measure the performance of your machine learning model using different evaluation metrics for each label. If you notice that the model performs significantly better on some labels than others, it could be an indication of bias in the labeled data.
- A/B testing: You can test the model’s performance with two different sets of labeled data, and compare the results. This can help you identify any differences in the model’s accuracy or performance based on the labeled data it was trained on.
- Bias detection algorithms: There are algorithms designed specifically to detect bias in labeled data. These algorithms can help identify any inconsistencies or imbalances in the labeled data that may lead to biased machine learning models.
4.7.6. Evaluate the Accuracy and Quality of Labeled Data
- Inter-annotator agreement: You can calculate the agreement between multiple annotators who labeled the same data. This measure can help you identify any inconsistencies in the labeling and assess the quality of the labels. Common agreement metrics include Cohen’s kappa and Fleiss’ kappa.
- Error analysis: You can analyze the labeling errors to identify patterns or common mistakes made by the annotators. This can help you identify specific areas of the data that need further clarification or guidelines for better labeling.
- Domain expertise: Consult with subject matter experts who have knowledge of the domain and the data to evaluate the quality and accuracy of the labeled data. They can provide valuable insights into the nuances and complexities of the data, which can help identify any potential labeling errors.
- Gold-standard data: Create a subset of the data with manually verified and accurate labels as a gold standard. You can then compare the automated labels against the gold standard to measure the accuracy of the automated labeling process.
- Performance evaluation: Train a machine learning model on the labeled data and evaluate its performance using standard metrics such as precision, recall, and F1 score. The model’s performance can give you an indication of the quality and accuracy of the labeled data.
4.7.7. Limitation
5. AI Trojan Attacks
Significance of AI Trojan Attacks
6. Overview of Mitigation
6.1. Mitigation Techniques Types
- Pre-processing: This involves preparing the data before it is fed into the machine learning algorithm. This stage can include several tasks, such as data cleaning, data normalization, and feature engineering. Pre-processing can help reduce the impact of biased data on the machine learning algorithm [90]. Pre-processing involves cleaning, normalizing, and transforming data before they are used to train a machine learning model [8].For example, if a dataset contains biased data, such as gender- or race-based disparities, pre-processing techniques can be used to remove these biases or balance the dataset. Some examples of pre-processing techniques include the following:
- Oversampling or undersampling: This involves adding more examples of under-represented groups or removing from over-represented groups to balance the dataset [91].
For example, imagine we are trying to predict whether or not a customer will default on a loan. We have a dataset with 10,000 examples, but only 100 of them are defaults. This means that the dataset is imbalanced, with the default class being the minority class. If we were to train a machine learning model on this dataset without addressing the class imbalance, the model might be biased towards predicting the majority class since it has more examples to learn from. This could result in poor performance in the minority class. To address this imbalance, we can use oversampling or under-sampling techniques [92].- Feature scaling: This involves scaling the features to a similar range so that they can be easily interpreted by the machine learning algorithm [93].
Suppose [94] that we have a dataset of house prices with features like the number of bedrooms, the square footage of the house, and the distance from the city center. The number of bedrooms ranges from 1 to 5, the square footage ranges from 500 to 5000 square feet, and the distance from the city center ranges from 1 to 20 miles. If we were to apply a machine learning algorithm to this dataset without scaling the features, then the square footage feature would have a much larger range of values than the number of bedrooms or the distance from the city center. This means that the square footage feature would have a much larger impact on the output of the model compared to the other features. To address this issue, we can apply feature scaling to normalize the range of values for each feature. There are several ways to scale features, but one common method is to use normalization, which scales the values to a range between 0 and 1. We can normalize the square footage feature by subtracting the minimum value of the feature and dividing it by the range of values. Similarly, we can normalize the other features in the dataset as well [95]. This will ensure that all features have a similar range of values and that no one feature dominates over the others. Once we have scaled the features, we can apply a machine learning algorithm to the dataset to predict house prices. The algorithm will be able to learn from all features equally and make more accurate predictions, as a result [96].- One-hot encoding: This involves encoding categorical variables into a binary format to make them easily digestible by the machine learning algorithm. One-hot encoding is a technique used to represent categorical data as numerical data in machine learning models [97]. It is commonly used when the categorical data have no inherent order or hierarchy [98]. Here is an example case where one-hot encoding might be useful:
Suppose [99] that we have a dataset of customer information for an online retailer, and one of the features is “product category”, which can have values like “electronics”, “clothing”, “home goods”, and “books”. If we were to apply a machine learning algorithm to this dataset without one-hot encoding, the algorithm would not be able to interpret the categorical data as numerical data. This means that it would not be able to learn from the “product category” feature and would likely ignore it. To address this issue, we can use one-hot encoding to represent the “product category” feature as a set of binary features, where each feature represents a possible value of the original categorical feature. For example, we might create four binary features for the “product category”: “electronics”, “clothing”, “home goods”, and “books”. If the original “product category” feature for a customer is “electronics”, the corresponding binary feature would be set to 1 and all other binary features would be set to 0. If the original “product category” feature is “clothing”, then corresponding binary feature would be set to 1 and all other binary features would be set to 0, and so on. Once we have performed one-hot encoding on the “product category” feature, we can apply a machine learning algorithm to the dataset to predict customer behavior or preferences [99]. The algorithm will be able to interpret the “product category” feature as numerical data and use it to make more accurate predictions [97].In short, To mitigate data collection bias, it is important to ensure that data are collected in a representative and unbiased way. This can involve using random sampling techniques, collecting data from multiple sources, and carefully defining the population being studied. To mitigate feature selection bias, it is important to carefully consider the criteria used for feature selection and to avoid making assumptions based on stereotypes or incomplete information. - In-processing bias: In-processing involves verifying that the machine learning model is unbiased during training. It involves monitoring the machine learning algorithm’s performance during training to detect and correct for bias. This stage includes techniques such as adversarial training, which involves training the algorithm to recognize and correct for biased input data. In-processing techniques aim to reduce bias during the training process, rather than after the model has been trained [8]. By monitoring the algorithm’s performance, in-processing techniques can help detect and correct for bias as it arises. Some examples of in-processing techniques include:
- Adversarial training: This involves training the algorithm to recognize and correct for biased input data by generating adversarial examples that challenge the primary model and test its robustness [100]. Adversarial training is a technique used to improve the robustness of machine learning models against adversarial attacks. Adversarial attacks are when an attacker intentionally manipulates input data to cause a machine learning model to make a mistake. Here is an example case scenario where adversarial training might be useful:For example, we have a machine learning model that is trained to identify images of traffic signs. The model has high accuracy when tested on normal images of traffic signs, but when tested on adversarial images, which have been specifically designed to fool the model, the accuracy drops significantly. To improve the model’s robustness against adversarial attacks, we can use adversarial training. Adversarial training involves training the model on both normal and adversarial examples. The adversarial examples are generated by adding small, carefully crafted perturbations to the normal images, which are imperceptible to humans but can cause the model to misclassify the image.
- Batch normalization: This involves normalizing the output of each layer in the neural network to reduce the impact of biased data on the training process [95]. Batch normalization is a technique used in deep neural networks to improve their training speed and stability. It works by normalizing the activations of each layer in the network based on the statistics of the current mini-batch of data. Here is how batch normalization might work in this example:
- (a)
- Compute mean and variance: For each mini-batch of data during training, we compute the mean and variance of the activations for each layer in the network.
- (b)
- Normalize activations: We normalize the activations of each layer by subtracting the mean and dividing by the square root of the variance. This has the effect of centering and scaling the activations, making them more consistent across the mini-batch.
- (c)
- Scale and shift: We then scale and shift the normalized activations using learnable parameters, which allows the network to learn the optimal scaling and shifting for each layer.
- (d)
- Train the network: We then train the network using back propagation with the batch-normalized activation.
By applying batch normalization to the neural network, we can significantly improve the training speed and stability, which leads to faster convergence and higher accuracy on the test set. This technique can be applied to other types of deep neural networks and other types of data as well to improve their training speed and stability.
In short, to mitigate algorithmic bias, it is important to carefully select and evaluate machine learning algorithms to ensure that they are unbiased and do not perpetuate existing biases. To mitigate sampling bias, it is important to carefully select and curate training datasets to ensure that they are representative and do not exclude important subgroups. - Post-processing bias: Post-processing involves modifying the output of the machine learning model to make it more equitable [101]. Post-processing involves modifying the output of the machine learning algorithm to make it more equitable. This stage includes techniques such as re-calibrating the model’s predictions to reduce bias or using techniques such as demographic parity to ensure fairness [2]. Post-processing techniques aim to reduce bias in the model’s output after it has been trained. By modifying the output, post-processing techniques can help correct any bias that may have been introduced during the training process. Some examples of post-processing techniques include:
- Re-calibration: This involves adjusting the probabilities assigned by the model to different outcomes to ensure that they are fair and unbiased.
- Demographic parity: This involves ensuring that the model’s predictions do not unfairly favor one group over another by setting a threshold that is consistent across all groups of people.
In summary, pre-processing, in-processing, and post-processing are three stages of the machine learning pipeline that can be used to address bias. By using a combination of these techniques, machine learning practitioners can help reduce the impact of bias on their models and ensure fairness for all groups of people.
6.2. Mitigation Techniques
- Data Collection: The cornerstone of the whole machine learning process is the collection of impartial data. This is significant because a machine learning model’s behavior is greatly influenced by the data used to train it. To guarantee that the model is trained on broad data, which increases its robustness and applicability to real-world events, it is essential to use reputable sources and ensure that the whole population is represented. Important elements in this process include avoiding biased samples, using stratified sampling approaches, and making sure that features are represented in a diversified manner. The foundation for developing moral, open, and equitable machine learning models that can be trusted to be used in a variety of settings and applications is unbiased data collecting.
- Data pre-processing: By using methods like resampling, data augmentation, or feature engineering to make the dataset more representative of the complete population, bias can be removed from the data. As a sculptor, data pre-processing transforms unprocessed data into a shape that enables machine learning models to train efficiently and provide well-informed predictions in a variety of settings.
- Model Training: By using methods like adversarial training, regularization, or model interpretability, we can create models that are more resistant to bias. To avoid prejudice, the model must be tested and improved using a variety of datasets. It is also essential to test and refine the model on a range of datasets. This guarantees that the model performs fairly in all circumstances and for all demographic groups.
- Model Evaluation: Model evaluation is a critical step in assessing bias and ensuring fair and unbiased performance. By comparing the model’s performance across different datasets or by using metrics that are more accurate in describing the issue can check the model for bias. Additionally, choosing a fair success metric and testing the model using various datasets are required.
- Model Deployment: Model deployment is a crucial stage where monitoring and addressing bias. Use methods like debiasing or retraining to track and modify the model’s performance after it has been deployed. This entails testing the model for bias in the deployed environment and making sure that it is used in a manner. See Table 18.
6.3. Approaches of Mitigation
- Environmental mitigation approaches: The actions are to lessen the negative effects of human activity on the environment. For instance, conserving habitats, decreasing the use of non-renewable resources, and lowering greenhouse gas emissions to slow climate change.
- Disaster mitigation approaches: These actions are being performed to lessen the effects of calamities like earthquakes, hurricanes, and floods. Creating emergency response plans, constructing disaster-resistant structures, and practicing regularly are a few examples.
- Cybersecurity mitigation approaches: These actions are being performed to lessen the effects of data breaches and cyberattacks. Regular software updates, the use of strong passwords, and the encoding of private data are a few examples.
- Health mitigation approaches: These actions are being performed in order to lessen the effects of pandemics and other public health emergencies. Vaccination campaigns, mask use, hand hygiene, and social segregation strategies are a few instances.
7. Discuss the Future Possibility of Bias
8. Conclusions
9. Our Contribution
Author Contributions
Funding
Data Availability Statement
Conflicts of Interest
References
- van Giffen, B.; Herhausen, D.; Fahse, T. Overcoming the pitfalls and perils of algorithms: A classification of machine learning biases and mitigation methods. J. Bus. Res. 2022, 144, 93–106. [Google Scholar] [CrossRef]
- Abay, A.; Zhou, Y.; Baracaldo, N.; Rajamoni, S.; Chuba, E.; Ludwig, H. Mitigating bias in federated learning. arXiv 2020, arXiv:2012.02447. [Google Scholar]
- Hort, M.; Chen, Z.; Zhang, J.M.; Sarro, F.; Harman, M. Bia mitigation for machine learning classifiers: A comprehensive survey. arXiv 2022, arXiv:2207.07068. [Google Scholar] [CrossRef]
- Krco, N.; Laugel, T.; Loubes, J.M.; Detyniecki, M. When Mitigating Bias is Unfair: A Comprehensive Study on the Impact of Bias Mitigation Algorithms. arXiv 2023, arXiv:2302.07185. [Google Scholar]
- Dietterich, T.G.; Kong, E.B. Machine Learning Bias, Statistical Bias, and Statistical Variance of Decision Tree Algorithms; Technical Report; Citeseer: University Park, PA, USA, 1995. [Google Scholar]
- Wang, Z.; Zhou, Y.; Qiu, M.; Haque, I.; Brown, L.; He, Y.; Wang, J.; Lo, D.; Zhang, W. Towards Fair Machine Learning Software: Understanding and Addressing Model Bias Through Counterfactual Thinking. arXiv 2023, arXiv:2302.08018. [Google Scholar]
- Kraus, S.; Breier, M.; Dasí-Rodríguez, S. The art of crafting a systematic literature review in entrepreneurship research. Int. Entrep. Manag. J. 2020, 16, 1023–1042. [Google Scholar] [CrossRef]
- Chen, Z.; Zhang, J.M.; Sarro, F.; Harman, M. A comprehensive empirical study of bias mitigation methods for software fairness. arXiv 2022, arXiv:2207.03277. [Google Scholar]
- Dixon, L.; Li, J.; Sorensen, J.; Thain, N.; Vasserman, L. Measuring and mitigating unintended bias in text classification. In Proceedings of the 2018 AAAI/ACM Conference on AI, Ethics, and Society, New Orleans, LA, USA, 2–3 February 2018; pp. 67–73. [Google Scholar]
- Pan, Z.; Mishra, P. Design of AI trojans for evading machine learning-based detection of hardware trojans. In Proceedings of the 2022 Design, Automation & Test in Europe Conference & Exhibition (DATE), Antwerp, Belgium, 14–23 March 2022; pp. 682–687. [Google Scholar]
- Zhang, K.; Khosravi, B.; Vahdati, S.; Faghani, S.; Nugen, F.; Rassoulinejad-Mousavi, S.M.; Moassefi, M.; Jagtap, J.M.M.; Singh, Y.; Rouzrokh, P.; et al. Mitigating bias in radiology machine learning: 2. Model development. Radiol. Artif. Intell. 2022, 4, e220010. [Google Scholar] [CrossRef]
- Pagano, T.P.; Loureiro, R.B.; Lisboa, F.V.; Peixoto, R.M.; Guimarães, G.A.; Cruz, G.O.; Araujo, M.M.; Santos, L.L.; Cruz, M.A.; Oliveira, E.L.; et al. Bias and Unfairness in Machine Learning Models: A Systematic Review on Datasets, Tools, Fairness Metrics, and Identification and Mitigation Methods. Big Data Cogn. Comput. 2023, 7, 15. [Google Scholar] [CrossRef]
- Ranjit, J.; Wang, T.; Ray, B.; Ordonez, V. Variation of Gender Biases in Visual Recognition Models Before and After Finetuning. arXiv 2023, arXiv:2303.07615. [Google Scholar]
- Kim, S.J.; Singh, P.; Jung, S.Y. A machine learning-based concentration-encoded molecular communication system. Nano Commun. Netw. 2023, 35, 100433. [Google Scholar] [CrossRef]
- Chakrabortty, R.; Pal, S.C.; Ruidas, D.; Roy, P.; Saha, A.; Chowdhuri, I. Living with Floods Using State-of-the-Art and Geospatial Techniques: Flood Mitigation Alternatives, Management Measures, and Policy Recommendations. Water 2023, 15, 558. [Google Scholar] [CrossRef]
- Zhang, G.; Li, Y.; Wu, Y.; Zhang, L.; Lin, C.; Geng, J.; Wang, S.; Fu, J. CORGI-PM: A Chinese Corpus For Gender Bias Probing and Mitigation. arXiv 2023, arXiv:2301.00395. [Google Scholar]
- Saxena, N.A.; Zhang, W.; Shahabi, C. Unveiling and Mitigating Bias in Ride-Hailing Pricing for Equitable Policy Making. arXiv 2023, arXiv:2301.03489. [Google Scholar]
- Shrestha, R.; Kafle, K.; Kanan, C. An investigation of critical issues in bias mitigation techniques. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Waikoloa, HI, USA, 3–8 January 2022; pp. 1943–1954. [Google Scholar]
- Pastaltzidis, I.; Dimitriou, N.; Quezada-Tavarez, K.; Aidinlis, S.; Marquenie, T.; Gurzawska, A.; Tzovaras, D. Data augmentation for fairness-aware machine learning: Preventing algorithmic bias in law enforcement systems. In Proceedings of the 2022 ACM Conference on Fairness, Accountability, and Transparency, Seoul, Republic of Korea, 21–24 June 2022; pp. 2302–2314. [Google Scholar]
- Guo, Y.; Yang, Y.; Abbasi, A. Auto-debias: Debiasing masked language models with automated biased prompts. In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), Dublin, Ireland, 22–27 May 2022; Association for Computational Linguistics: Kerrville, TX, USA, 2022; pp. 1012–1023. [Google Scholar]
- Vokinger, K.N.; Feuerriegel, S.; Kesselheim, A.S. Mitigating bias in machine learning for medicine. Commun. Med. 2021, 1, 25. [Google Scholar] [CrossRef] [PubMed]
- Islam, R.; Keya, K.N.; Zeng, Z.; Pan, S.; Foulds, J. Debiasing career recommendations with neural fair collaborative filtering. In Proceedings of the Web Conference 2021, Ljubljana, Slovenia, 19–23 April 2021; pp. 3779–3790. [Google Scholar]
- Miron, M.; Tolan, S.; Gómez, E.; Castillo, C. Evaluating causes of algorithmic bias in juvenile criminal recidivism. Artif. Intell. Law 2021, 29, 111–147. [Google Scholar] [CrossRef]
- Ashokan, A.; Haas, C. Fairness metrics and bias mitigation strategies for rating predictions. Inf. Process. Manag. 2021, 58, 102646. [Google Scholar] [CrossRef]
- Alelyani, S. Detection and evaluation of machine learning bias. Appl. Sci. 2021, 11, 6271. [Google Scholar] [CrossRef]
- Zhang, Y.; Yang, Y.; Huang, Q.; Zhang, Q.; Li, M.; Wu, Y. The effectiveness of lifestyle interventions for diabetes remission on patients with type 2 diabetes mellitus: A systematic review and meta-analysis. Worldviews-Evid.-Based Nurs. 2023, 20, 64–78. [Google Scholar] [CrossRef]
- Straw, I.; Wu, H. Investigating for bias in healthcare algorithms: A sex-stratified analysis of supervised machine learning models in liver disease prediction. BMJ Health Care Inform. 2022, 29, e100457. [Google Scholar] [CrossRef]
- Yousefi-Nooraie, R.; Shakiba, B.; Mortaz-Hejri, S. Country development and manuscript selection bias: A review of published studies. BMC Med. Res. Methodol. 2006, 6, 37. [Google Scholar] [CrossRef] [PubMed]
- Ellenberg, J.H. Selection bias in observational and experimental studies. Stat. Med. 1994, 13, 557–567. [Google Scholar] [CrossRef] [PubMed]
- Heckman, J.J. Sample selection bias as a specification error. Econom. J. Econom. Soc. 1979, 47, 153–161. [Google Scholar] [CrossRef]
- Andringa, S.; Godfroid, A. Sampling bias and the problem of generalizability in applied linguistics. Annu. Rev. Appl. Linguist. 2020, 40, 134–142. [Google Scholar] [CrossRef]
- Garg, T.; Masud, S.; Suresh, T.; Chakraborty, T. Handling bias in toxic speech detection: A survey. ACM Comput. Surv. 2022, 55, 1–32. [Google Scholar] [CrossRef]
- McCullagh, P. Sampling bias and logistic models. J. R. Stat. Soc. Ser. (Stat. Methodol.) 2008, 70, 643–677. [Google Scholar] [CrossRef]
- Vella, F. Estimating models with sample selection bias: A survey. J. Hum. Resour. 1998, 33, 127–169. [Google Scholar] [CrossRef]
- Panzeri, S.; Senatore, R.; Montemurro, M.A.; Petersen, R.S. Correcting for the sampling bias problem in spike train information measures. J. Neurophysiol. 2007, 98, 1064–1072. [Google Scholar] [CrossRef]
- Baker, J.E. Reducing bias and inefficiency in the selection algorithm. In Proceedings of the Second International Conference on Genetic Algorithms, Cambridge, MA, USA, 28–31 July 1987; L. Erlbaum Associates Inc.: Broadway Hillsdale, NJ, USA, 1987; Volume 206, pp. 14–21. [Google Scholar]
- Edlund, M.J.; Craig, T.J.; Richardson, M.A. Informed consent as a form of volunteer bias. Am. J. Psychiatry 1985, 142, 624–627. [Google Scholar]
- Tripepi, G.; Jager, K.J.; Dekker, F.W.; Zoccali, C. Selection bias and information bias in clinical research. Nephron Clin. Pract. 2010, 115, c94–c99. [Google Scholar] [CrossRef]
- Brayne, C.; Moffitt, T.E. The limitations of large-scale volunteer databases to address inequalities and global challenges in health and aging. Nat. Aging 2022, 2, 775–783. [Google Scholar] [CrossRef] [PubMed]
- Buckley, J.P.; Keil, A.P.; McGrath, L.J.; Edwards, J.K. Evolving methods for inference in the presence of healthy worker survivor bias. Epidemiology 2015, 26, 204–212. [Google Scholar] [CrossRef] [PubMed]
- Ho, A.M.H.; Dion, P.W.; Yeung, J.H.; Holcomb, J.B.; Critchley, L.A.; Ng, C.S.; Karmakar, M.K.; Cheung, C.W.; Rainer, T.H.; Warner, D.S. Prevalence of survivor bias in observational studies on fresh frozen plasma: Erythrocyte ratios in trauma requiring massive transfusion. J. Am. Soc. Anesthesiol. 2012, 116, 716–728. [Google Scholar] [CrossRef] [PubMed]
- Prather, L.; Bertin, W.J.; Henker, T. Mutual fund characteristics, managerial attributes, and fund performance. Rev. Financ. Econ. 2004, 13, 305–326. [Google Scholar] [CrossRef]
- Van der Sluis, J.; Van Praag, M.; Vijverberg, W. Education and entrepreneurship selection and performance: A review of the empirical literature. J. Econ. Surv. 2008, 22, 795–841. [Google Scholar] [CrossRef]
- Thornton, A.; Lee, P. Publication bias in meta-analysis: Its causes and consequences. J. Clin. Epidemiol. 2000, 53, 207–216. [Google Scholar] [CrossRef] [PubMed]
- Yadav, K.; Lewis, R.J. Immortal time bias in observational studies. JAMA 2021, 325, 686–687. [Google Scholar] [CrossRef] [PubMed]
- Lesage, P. Characterization of frequency stability: Bias due to the juxtaposition of time-interval measurements. IEEE Trans. Instrum. Meas. 1983, 32, 204–207. [Google Scholar] [CrossRef]
- Giobbie-Hurder, A.; Gelber, R.D.; Regan, M.M. Challenges of guarantee-time bias. J. Clin. Oncol. 2013, 31, 2963. [Google Scholar] [CrossRef]
- Schwartzbaum, J.; Ahlbom, A.; Feychting, M. Berkson’s bias reviewed. Eur. J. Epidemiol. 2003, 18, 1109–1112. [Google Scholar] [CrossRef]
- Westreich, D. Berkson’s bias, selection bias, and missing data. Epidemiology 2012, 23, 159. [Google Scholar] [CrossRef] [PubMed]
- Feinstein, A.R.; Walter, S.D.; Horwitz, R.I. An analysis of Berkson’s bias in case-control studies. J. Chronic Dis. 1986, 39, 495–504. [Google Scholar] [CrossRef] [PubMed]
- Pearce, N.; Richiardi, L. Commentary: Three worlds collide: Berkson’s bias, selection bias and collider bias. Int. J. Epidemiol. 2014, 43, 521–524. [Google Scholar] [CrossRef] [PubMed]
- Conn, H.O.; Snyder, N.; Atterbury, C.E. The Berkson bias in action. Yale J. Biol. Med. 1979, 52, 141. [Google Scholar]
- Freeman, J.; Hutchison, G.B. Prevalence, incidence and duration. Am. J. Epidemiol. 1980, 112, 707–723. [Google Scholar] [CrossRef] [PubMed]
- Younossi, Z.M.; Koenig, A.B.; Abdelatif, D.; Fazel, Y.; Henry, L.; Wymer, M. Global epidemiology of nonalcoholic fatty liver disease—Meta-analytic assessment of prevalence, incidence, and outcomes. Hepatology 2016, 64, 73–84. [Google Scholar] [CrossRef] [PubMed]
- Quaade, A.S.; Simonsen, A.B.; Halling, A.S.; Thyssen, J.P.; Johansen, J.D. Prevalence, incidence, and severity of hand eczema in the general population—A systematic review and meta-analysis. Contact Dermat. 2021, 84, 361–374. [Google Scholar] [CrossRef]
- Qiang, J.; Wu, D.; Du, H.; Zhu, H.; Chen, S.; Pan, H. Review on Facial-Recognition-Based Applications in Disease Diagnosis. Bioengineering 2022, 9, 273. [Google Scholar] [CrossRef]
- Khalil, A.; Ahmed, S.G.; Khattak, A.M.; Al-Qirim, N. Investigating bias in facial analysis systems: A systematic review. IEEE Access 2020, 8, 130751–130761. [Google Scholar] [CrossRef]
- Langenkamp, M.; Costa, A.; Cheung, C. Hiring fairly in the age of algorithms. arXiv 2020, arXiv:2004.07132. [Google Scholar] [CrossRef]
- Alikhademi, K.; Drobina, E.; Prioleau, D.; Richardson, B.; Purves, D.; Gilbert, J.E. A review of predictive policing from the perspective of fairness. Artif. Intell. Law 2022, 30, 1–17. [Google Scholar] [CrossRef]
- Berk, R.A. Artificial intelligence, predictive policing, and risk assessment for law enforcement. Annu. Rev. Criminol. 2021, 4, 209–237. [Google Scholar] [CrossRef]
- Abd-Alrazaq, A.A.; Rababeh, A.; Alajlani, M.; Bewick, B.M.; Househ, M. Effectiveness and safety of using chatbots to improve mental health: Systematic review and meta-analysis. J. Med. Internet Res. 2020, 22, e16021. [Google Scholar] [CrossRef] [PubMed]
- Caldarini, G.; Jaf, S.; McGarry, K. A literature survey of recent advances in chatbots. Information 2022, 13, 41. [Google Scholar] [CrossRef]
- FitzGerald, C.; Hurst, S. Implicit bias in healthcare professionals: A systematic review. BMC Med. Ethics 2017, 18, 19. [Google Scholar] [CrossRef] [PubMed]
- Kunz, R.; Vist, G.E.; Oxman, A.D. Randomisation to protect against selection bias in healthcare trials. Cochrane Database Syst. Rev. 2007, 18, MR000012. [Google Scholar]
- Oswald, M.E.; Grosjean, S. Confirmation bias. In Cognitive Illusions: A Handbook on Fallacies and Biases in Thinking, Judgement and Memory; Ghent University Library: Ghent, Belgium, 2004; Volume 79. [Google Scholar]
- Nickerson, R.S. Confirmation bias: A ubiquitous phenomenon in many guises. Rev. Gen. Psychol. 1998, 2, 175–220. [Google Scholar] [CrossRef]
- Peters, U. What is the function of confirmation bias? Erkenntnis 2022, 87, 1351–1376. [Google Scholar] [CrossRef]
- Klayman, J. Varieties of confirmation bias. Psychol. Learn. Motiv. 1995, 32, 385–418. [Google Scholar]
- Obermeyer, Z.; Powers, B.; Vogeli, C.; Mullainathan, S. Dissecting racial bias in an algorithm used to manage the health of populations. Science 2019, 366, 447–453. [Google Scholar] [CrossRef]
- Tang, W.; Sanville, E.; Henkelman, G. A grid-based Bader analysis algorithm without lattice bias. J. Phys. Condens. Matter 2009, 21, 084204. [Google Scholar] [CrossRef] [PubMed]
- Kordzadeh, N.; Ghasemaghaei, M. Algorithmic bias: Review, synthesis, and future research directions. Eur. J. Inf. Syst. 2022, 31, 388–409. [Google Scholar] [CrossRef]
- Jak, S.; Oort, F.J.; Dolan, C.V. Measurement bias in multilevel data. Struct. Equ. Model. Multidiscip. J. 2014, 21, 31–39. [Google Scholar] [CrossRef]
- Mansoury, M.; Abdollahpouri, H.; Pechenizkiy, M.; Mobasher, B.; Burke, R. Feedback loop and bias amplification in recommender systems. In Proceedings of the 29th ACM International Conference on Information & Knowledge Management, Virtual Event, Ireland, 19–23 October 2020; pp. 2145–2148. [Google Scholar]
- Puttagunta, M.K.; Ravi, S.; Nelson Kennedy Babu, C. Adversarial examples: Attacks and defences on medical deep learning systems. Multimed. Tools Appl. 2023, 82, 33773–33809. [Google Scholar] [CrossRef]
- Dee, D.P. Bias and data assimilation. Q. J. R. Meteorol. Soc. J. Atmos. Sci. Appl. Meteorol. Phys. Oceanogr. 2005, 131, 3323–3343. [Google Scholar] [CrossRef]
- Ntoutsi, E.; Fafalios, P.; Gadiraju, U.; Iosifidis, V.; Nejdl, W.; Vidal, M.E.; Ruggieri, S.; Turini, F.; Papadopoulos, S.; Krasanakis, E.; et al. Bias in data-driven artificial intelligence systems—An introductory survey. Wiley Interdiscip. Rev. Data Min. Knowl. Discov. 2020, 10, e1356. [Google Scholar] [CrossRef]
- Stasaski, K.; Yang, G.H.; Hearst, M.A. More diverse dialogue datasets via diversity-informed data collection. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, Online, 5–10 July 2020; pp. 4958–4968. [Google Scholar]
- Li, T.; Higgins, J.P.; Deeks, J.J. Collecting data. In Cochrane Handbook for Systematic Reviews of Interventions; Wiley Online Library: Hoboken, NJ, USA, 2019; pp. 109–141. [Google Scholar]
- Rodrigues, F.; Pereira, F.; Ribeiro, B. Sequence labeling with multiple annotators. Mach. Learn. 2014, 95, 165–181. [Google Scholar] [CrossRef]
- Metallinou, A.; Narayanan, S. Annotation and processing of continuous emotional attributes: Challenges and opportunities. In Proceedings of the 2013 10th IEEE International Conference and Workshops on Automatic Face and Gesture Recognition (FG), Shanghai, China, 22–26 April 2013; pp. 1–8. [Google Scholar]
- Donmez, P.; Carbonell, J.; Schneider, J. A probabilistic framework to learn from multiple annotators with time-varying accuracy. In Proceedings of the 2010 SIAM International Conference on Data Mining, Columbus, OH, USA, 29 April–1 May 2010; pp. 826–837. [Google Scholar]
- Prokhorenkova, L.; Gusev, G.; Vorobev, A.; Dorogush, A.V.; Gulin, A. CatBoost: Unbiased boosting with categorical features. In Proceedings of the 32nd International Conference on Neural Information Processing Systems, Montréal, QC, Canada, 3–8 December 2018. [Google Scholar]
- Tang, K.; Niu, Y.; Huang, J.; Shi, J.; Zhang, H. Unbiased scene graph generation from biased training. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 3716–3725. [Google Scholar]
- Krishnan, R.; Rajpurkar, P.; Topol, E.J. Self-supervised learning in medicine and healthcare. Nat. Biomed. Eng. 2022, 6, 1346–1352. [Google Scholar] [CrossRef]
- Jugnandan, S.; Willows, G.D. Towards an integrated debiasing framework for consumer financial decisions: A reflection on debiasing research. Int. J. Consum. Stud. 2023, 47, 1544–1560. [Google Scholar] [CrossRef]
- Arkes, H.R. Costs and benefits of judgment errors: Implications for debiasing. Psychol. Bull. 1991, 110, 486. [Google Scholar] [CrossRef]
- Parraga, O.; More, M.D.; Oliveira, C.M.; Gavenski, N.S.; Kupssinskü, L.S.; Medronha, A.; Moura, L.V.; Simões, G.S.; Barros, R.C. Debiasing Methods for Fairer Neural Models in Vision and Language Research: A Survey. arXiv 2022, arXiv:2211.05617. [Google Scholar]
- Xu, X.; Wang, Q.; Li, H.; Borisov, N.; Gunter, C.A.; Li, B. Detecting ai trojans using meta neural analysis. In Proceedings of the 2021 IEEE Symposium on Security and Privacy (SP), San Francisco, CA, USA, 24–27 May 2021; pp. 103–120. [Google Scholar]
- Wang, L.; Wu, Z.; Chen, L.; Sun, Y.; Lin, W.; Li, G. iBEAT V2. 0: A multisite-applicable, deep learning-based pipeline for infant cerebral cortical surface reconstruction. Nat. Protoc. 2023, 18, 1488–1509. [Google Scholar] [CrossRef] [PubMed]
- Werner de Vargas, V.; Schneider Aranda, J.A.; dos Santos Costa, R.; da Silva Pereira, P.R.; Victória Barbosa, J.L. Imbalanced data preprocessing techniques for machine learning: A systematic mapping study. Knowl. Inf. Syst. 2023, 65, 31–57. [Google Scholar] [CrossRef] [PubMed]
- Tarawneh, A.S.; Hassanat, A.B.; Altarawneh, G.A.; Almuhaimeed, A. Stop oversampling for class imbalance learning: A review. IEEE Access 2022, 10, 47643–47660. [Google Scholar] [CrossRef]
- Sharma, S.; Gosain, A.; Jain, S. A review of the oversampling techniques in class imbalance problem. In International Conference on Innovative Computing and Communications: Proceedings of ICICC 2021; Springer: Singapore, 2022; Volume 1, pp. 459–472. [Google Scholar]
- Sharma, S.; Mandal, P.K. A comprehensive report on machine learning-based early detection of alzheimer’s disease using multi-modal neuroimaging data. ACM Comput. Surv. (CSUR) 2022, 55, 1–44. [Google Scholar] [CrossRef]
- Mondal, B.; Banerjee, A.; Gupta, S. A review of SQLI detection strategies using machine learning. Mach. Learn. 2022, 6, 9664–9677. [Google Scholar] [CrossRef]
- Huang, L.; Qin, J.; Zhou, Y.; Zhu, F.; Liu, L.; Shao, L. Normalization techniques in training dnns: Methodology, analysis and application. IEEE Trans. Pattern Anal. Mach. Intell. 2023, 45, 10173–10196. [Google Scholar] [CrossRef] [PubMed]
- Zahedi, R.; hasan Ghodusinejad, M.; Aslani, A.; Hachem-Vermette, C. Modelling community-scale renewable energy and electric vehicle management for cold-climate regions using machine learning. Energy Strategy Rev. 2022, 43, 100930. [Google Scholar] [CrossRef]
- Kosaraju, N.; Sankepally, S.R.; Mallikharjuna Rao, K. Categorical Data: Need, Encoding, Selection of Encoding Method and Its Emergence in Machine Learning Models—A Practical Review Study on Heart Disease Prediction Dataset Using Pearson Correlation. In Proceedings of International Conference on Data Science and Applications: ICDSA 2022; Springer: Singapore, 2023; Volume 1, pp. 369–382. [Google Scholar]
- Chebolu, S.U.S.; Rosso, P.; Kar, S.; Solorio, T. Survey on aspect category detection. ACM Comput. Surv. 2022, 55, 1–37. [Google Scholar] [CrossRef]
- Passos, D.; Mishra, P. A tutorial on automatic hyperparameter tuning of deep spectral modelling for regression and classification tasks. Chemom. Intell. Lab. Syst. 2022, 223, 104520. [Google Scholar] [CrossRef]
- Sajeeda, A.; Hossain, B.M. Exploring generative adversarial networks and adversarial training. Int. J. Cogn. Comput. Eng. 2022, 3, 78–89. [Google Scholar] [CrossRef]
- Zhang, Z.; Wang, S.; Meng, G. A Review on Pre-processing Methods for Fairness in Machine Learning. In Advances in Natural Computation, Fuzzy Systems and Knowledge Discovery: Proceedings of the ICNC-FSKD 2022, Fuzhou, China, 30 July–1 August 2022; Springer International Publishing: Cham, Switzerland, 2023; pp. 1185–1191. [Google Scholar]
Search Keyword | No. of Documents |
---|---|
“Machine learning bias” AND “mitigation” | 30 |
“algorithmic bias ” OR “fairness in machine learning” | 1530 |
“Unfairness model” OR “unintentional bias ” | 228 |
“Ethical machine learning” AND “algorithmic transparency” | 3 |
“Discrimination in machine learning” AND “counterfactual fairness” | 4 |
“Bias in natural language processing” OR “fairness in text classification” | 133 |
“Bias in computer vision” OR “fairness in image recognition” | 20 |
Total | 1948 |
Source | No. of Documents |
---|---|
Advances in Intelligent Systems and Computing (conference Proceedings) | 546 |
Lecture Notes in Computer Science (conference Proceedings) | 6 |
IEEE Access (conference Proceedings) | 1 |
ACM International Conference Proceedings Series (survey) | 138 |
International Journal of Production Research (Review article) | 49 |
Computers and Industrial Engineering | 20 |
CEUR Workshop Proceedings | 1 |
IEEE Transactions on AI and Ethics | 82 |
Machine Learning: A Multidisciplinary Approach (Springer) | 79 |
Total | 922 |
Source | No. of Occurrences |
---|---|
Bias | 45 |
Machine Learning | 32 |
Fairness | 9 |
Mitigation measures | 4 |
Algorithmic bias | 5 |
bias mitigation | 5 |
cognitive bias | 3 |
Debiasing techniques | 2 |
sampling bias | 5 |
decision making process | 8 |
Paper | Year | Contribution | Dataset | Limitation | Methods | Accuracy |
---|---|---|---|---|---|---|
Abay et al. [2] | 2020 | Framework both FL and fairness | Adult dataset, COMPAS | Focused only on binary classification but did not address biased data in FL on training data. | Local reweighing, global reweighing with privacy and federated bias removal. To mitigate bias pre-processing and in-processing methods. | Used 11.5% for fairness metrics and others use (FP, FN, curve) |
Additional exploration of advanced techniques for bias detection and quantification, and impact of different types of biases on FL. | ||||||
Zhenpeng et al. [8] | 2022 | Study (341 publications) on software engineering | Benchmark dataset, matrices, and benchmark | A limited set of bias mitigation techniques | Pre-processing, in-processing, and post-processing techniques | Revealing different methods and mitigating bias to predictive performance. |
Future offerings, investigate methods and datasets, as well as exploring performance and fairness | ||||||
Hort et al. [3] | 2022 | A systematic and extensive survey of various bias mitigation techniques. | 81 unique datasets. | Pre-processing, in-processing, and post-processing methods. | ML classifier | in-processing 212. |
Future offerings, on bias mitigation techniques of three factors: metrics, benchmarking, and statistics. | ||||||
Benjamin et al. [1] | 2022 | Comprehensive classification of machine learning biases and mitigation methods, highlighting potential pitfalls and perils of using algorithms. | Benchmarked | Focuses only business context | Six model-based methods used (bias removal approach, adversarial debiasing maximizes accuracy, Naive Bayes Classifiers, variable model, interpretable models, splitting and resampling | ML biases can be identified, avoided, and mitigated the CRISP-DM development process |
Future offering: conversion of human biases into machine learning biases. | ||||||
Lucas et al. [9] | 2022 | Measuring bias in text classification. | Hate Speech and Offensive Language dataset and Wikipedia Personal Attacks dataset. | Not suitable for limited diversity or representativeness | Statistical analysis and novel algorithm | Personal Attacks dataset, achieved 30% reduction maintain 90% predictive accuracy. Hate Speech and Offensive Language dataset, achieved a 13% reduction while maintaining 94% predictive accuracy. |
Future offering: approach on larger and diverse datasets. | ||||||
Zhixin et al. [10] | 2022 | Introduce a new attack model “neural Trojans”. Also proposed a new detection method on feature squeezing, to detect neural Trojans. | Does not use a dataset. | Not be effective against other types of AI Trojans | Demonstrate generative adversarial networks (GANs) | Demonstrates the feasibility of designing AI Trojans using GANs also propose a new detection method based on feature squeezing that are neural Trojans with high accuracy. |
Future offering: detecting different types of AI Trojans. | ||||||
Zhang et al. [11] | 2022 | Unbiased ML models in radiology by addressing the sources of bias | Dataset of chest X-rays | Lack of transparency regarding the dataset used in the case study. | Convolutional neural network (CNN) | AUC-ROC of 0.909 on the validation |
Future offering: expanding the framework to other types of medical imaging or healthcare applications. | ||||||
Korco et al. [4] | 2023 | Study of several bias mitigation approaches | Adult, Dutch, Compas, Bank, Credit | Not used on real-world datasets. Focused on a limited number of bias mitigation algorithms | Partial Least Squares Discriminant Analysis (PLS-DA) | Demographic parties and Equalized odds (Adult, Bank, COMPAS, Credit, Dutch-Biased 85.18, 73.74, 81.08, 83) |
Future offering: developing more robust and fair bias mitigation algorithms. |
Reference | Topic | Limitation with Issues | Contribution |
---|---|---|---|
[12] | Bias and Unfairness in Machine Learning Models | Exploring multi-class and multi-metric characteristics is limited to binary approaches. | The work examines ML model fairness and bias reduction, emphasizing attribute concerns and transparency for fairer algorithms. |
[13] | Variation of Gender Biases in Visual Recognition Models Before and After Fine-tuning | Pre-trained model biases and dataset size are poorly understood, according to the research. | Introduces image-based bias assessment for ML models. After extensive data fine-tuning, biases remain. To reduce downstream effects, it addressing fine-tuning biases. Bigger datasets may introduce and transmit bias. |
[14] | A machine learning-based concentration-encoded molecular communication system | Numerical simulations are used without experimental validation. | Suggested technology reduces molecular communication bias and interference over current techniques. Longer transmission and 4-ary CPSK improve simulations. |
[15] | Living with Floods Using State-of-the-Art and Geospatial Techniques | Lack of sufficient data, the Nonlinear relationship between causative factors and risk | A research employed ML to forecast floods and identify significant components. The best subtropical river basin model is ANN. A useful map for planning, prevention, and enforcement. |
[16] | A Chinese Corpus For Gender Bias Probing and Mitigation | The use of highly educated gender-bias annotators may cause cognitive bias. Only a few Chinese language models were tested. Work should diversify annotators and investigate other models and solutions for their issues. | Created CORGI-PM, a 32.9 k-sentence dataset revealing Chinese gender prejudice. AI was challenged to identify and correct gender bias using several language models. This helps researchers discover and mitigate gender bias in Chinese literature. |
[17] | Unveiling and Mitigating Bias in Ride-Hailing Pricing for Equitable Policy Making | Its focus on one city (Chicago) and the assumption that the government would subsidize discounts to make rides more affordable for disadvantaged residents. | It addresses ride-hailing price equity. It proposes fairness metrics pricing mechanisms and government subsidies. Practical experiments support ride-hailing policy improvements that promote fairness. |
Reference | Topic | limitation with Issues | Contribution |
---|---|---|---|
[18] | An Investigation of Critical Issues in Bias Mitigation Techniques | The study evaluates 7 algorithms, image tasks, and biases. Identifies bias mitigation difficulties, pushing for wider examination and hidden biases. | Evaluation techniques are improved, bias reduction recommendations are proposed, and diverse architectures and structured concept-based predictions are suggested. |
[19] | Data augmentation for fairness-aware machine learning | Focused on racial bias, overlooking gender and socioeconomic disparities. Data augmentation’s effectiveness varies with dataset variability, and detecting certain motions remains challenging. | The paper suggests fairness-aware ML for impartial law enforcement software. Data augmentation rebalances training data by race to reduce bias concerns without affecting data integrity. Real-world data experiments show balanced datasets may reduce law enforcement prejudice. |
[20] | Auto-Debias: Debiasing Masked Language Models with Automated Biased Prompts | Proposed method solely debiases PLMs. Because big pre-trained language models are often used in real-world applications, reducing their human-like biases and societal preconceptions is crucial. | Auto-Debias reduces NLP bias efficiently and objectively. It reduces gender and racial biases in PLMs by automatically recognizing biased prompts, outperforming previous methods. This innovation advances NLP practical debiasing strategies. |
Reference | Topic | limitation with Issues | Contribution |
---|---|---|---|
[21] | Mitigating bias in machine learning for medicine | Limited medical machine learning bias and patient outcome studies. Addressing bias essential for fair healthcare, advocating diverse data, robust model development, and careful clinical deployment. | Its integrative approach to building machine learning algorithms for medical applications provides concrete bias-reduction solutions. To provide equitable healthcare results, prejudice must be mitigated. |
[22] | Debiasing Career Recommendations with Neural Fair Collaborative Filtering | Gender bias in career suggestions might dissuade women from STEM disciplines, hurting economic competitiveness and gender pay inequalities. NFCF reduces bias just for gender, according to the research. | NFCF combats gender bias in career recommender systems, according to the research. After pre-training and bias correction, it beats current models on MovieLens and Facebook datasets, potentially reducing gender salary gaps and STEM career discrepancies. |
[23] | Evaluating causes of algorithmic bias in juvenile criminal recidivism | The research only included Catalan youth offenders, limiting its usefulness. Crime prediction using AI presents justice problems, especially for disadvantaged populations. Prediction accuracy and fairness are difficult to balance. | The research shows ML outperforms professional tools but favors certain populations. It pinpoints biases and offers solutions. Additionally, certain fairness approaches may have unintended bad effects. |
[24] | Fairness metrics and bias mitigation strategies for rating predictions | The bias reduction strategy is designed for rating-based recommender systems and may not work for others. The evaluation is limited to two datasets, requiring more testing on other datasets for relevance. | The work innovates recommender system bias reduction and aligns fairness measures across domains. Recent findings emphasize bias reduction and fairness in algorithmic decision-making, particularly in recommenders. It helps reduce biases and improve fairness. |
[25] | Detection and Evaluation of Bias in Machine Learning Models Using an Alternation Function | A single dataset with few gender and racial biases tests the study’s method. It needs further testing on varied datasets with different features. | The alternation function approach examines machine learning model bias for transparency and fairness. It focuses on detecting and quantifying human biases’ influence on ML through attribute value swaps. Practical model understanding and usefulness improve with this technique. |
Bias Type | Definition | Example | Mitigation Strategies |
---|---|---|---|
Data Bias | Bias that arises from the quality, quantity, or representativeness of the training data | suppose a machine learning model is to predict whether a customer is likely to purchase a product based on their demographic and purchase history. We train the model on a dataset that consists of only men customers. The model performs well on this dataset, but when you apply it to a test dataset that includes women customers, it performs poorly. Then, the system is data in a bias phase. | Data augmentation, collecting diverse datasets, representativeness of training data |
Feature Bias | Feature bias can arise from various sources such as incomplete data, selection bias, or pre-existing social biases. | For example, a company is building a machine learning model to predict whether a person is likely to default on a loan. They train the model on a dataset that includes features such as income, credit score, and employment history. However, they fail to include other relevant factors such as race, gender, or zip code, which can lead to biased predictions. In this case, feature bias arises. | Feature selection and pre-processing, careful consideration of feature choice, feature engineering to remove discriminatory features, dimensionality reduction |
Model Bias | Model bias can arise from a variety of sources, including biased training data, biased algorithms, or biased human decisions that influence the design or implementation of the model. | Candidate’s height as a factor in hiring decisions can be an example of model bias. If height is used as a selection criteria, then it may disproportionately disadvantage shorter candidates, even if height is not a relevant factor for job performance. This bias can be perpetuated if the hiring team or the machine learning model has been trained on data that includes the height of successful candidates in the past. | Model Transparency and interpretability, regularized model training, model val- ideation and testing on diverse datasets |
Algorithm Bias | Arises from the specific algorithm or optimization method used to train the machine learning model | Predicting flower species using ML could occur if the training dataset used to train the algorithm is biased towards certain types of flowers. if the training dataset contains mostly white flowers, the algorithm may have difficulty accurately predicting the species of flowers with different colors or shapes. This can result in the algorithm being biased towards certain types of flowers, which can lead to inaccurate predictions or misclassification of flowers. | Algorithmic auditing, hyperparameter tuning, and ensemble methods, fairness-aware algorithms, or optimization methods. |
Operational Bias | Bias that arises from the deployment, usage, or interpretation of the machine learning model in real-world settings | Loan decisions can occur when the loan approval process relies too heavily on automated systems or algorithms that are not properly designed or tested. If the algorithm is not designed to account for all relevant factors, such as the credit history or employment status of the applicant, it may lead to inaccurate or biased loan decisions. To address operational bias in loan decisions, it is important to ensure that the loan application process is designed to minimize the potential for bias, and that loan officers and algorithms are trained and monitored to ensure they are making fair and objective decisions. | Regularly monitoring model performance and bias, introducing feedback mechanisms, ensuring ethical and responsible use of the model. |
Method | Details |
---|---|
Random sampling | Random sampling helps to ensure that every member of the population has an equal chance of being selected for the sample. This helps to prevent bias that can be introduced through non-random sampling techniques. |
An increased sample size | larger sample size can help to reduce the impact of outliers and other anomalies that can introduce bias into the sample. However, it is important to ensure that the sample size is appropriate for the population being studied. |
Use stratified sampling | Stratified sampling involves dividing the population into subgroups and then selecting a random sample from each subgroup. This can help to ensure that the sample includes representation from all subgroups, preventing bias that can be introduced through non-representative subgroups. |
Be mindful of selection bias | Selection bias occurs when the sample is not representative of the population being studied. This can occur when the sample is self-selected or when the researcher selects participants based on certain characteristics. To minimize selection bias, it is important to use random sampling techniques and to ensure that the sample is diverse and representative of the population being studied. |
Use a diverse sample | A diverse sample helps to ensure that the sample is representative of the population being studied. This can include diversity in terms of age, gender, race/ethnicity, education level, and other characteristics that may impact the results of the study. |
Monitor participation rates | It is important to monitor participation rates to ensure that the sample is not biased towards those who are more likely to participate. This can help to prevent bias that can be introduced through non-response bias. |
Method | Details |
---|---|
Random Selection | Select participants at random from the population to minimize the risk of bias. |
Incentives | Offer incentives such as payment or a chance to win a prize to motivate more people to participate and reduce volunteer bias. |
Masked Studies | Use a blinded or masked study design where participants do not know the true nature of the study to reduce self-selection bias. |
Broad Advertising | Use of a variety of media outlets and advertising channels to reach a more diverse population and reduce volunteer bias. |
Community Partnerships | Partner with community organizations to recruit participants and ensure that the study is representative of the community as a whole. |
Method | Description |
---|---|
Use complete data | Collect and analyze data from all participants, including those who drop out or are lost to follow-up. |
Impute missing data | Estimate missing data values based on patterns in the observed data, using methods such as mean imputation, regression imputation, or multiple imputation. |
Sensitivity analysis | Assess of the robustness of the results by varying the assumptions about missing data, such as the degree of missing or the imputation method used. |
Weighted analysis | Assign weights to each observation based on the probability of being observed, to adjust for differential attrition rates between groups. |
Inverse probability weighting | Assign weights to each observation based on the inverse probability of being observed, to adjust for differential attrition rates between groups and the probability of missing. |
Propensity score analysis | Use propensity scores to match or stratify participants based on their likelihood of being observed, to adjust for confounding factors and selection bias. |
Method | Description |
---|---|
Prospective study design | where outcomes are measured at the same time or at regular intervals over a specified period. |
Matching outcome time intervals | Measuring outcomes to ensure that exposure status is accurately captured during the study period. |
Statistical adjustment | Use statistical techniques such as time-dependent covariant analysis or survival analysis to adjust for changes in exposure status over time. |
Sensitivity analysis | Perform sensitivity analyses to assess the impact of time interval bias on study results by varying the time intervals used to measure exposure and outcome. |
Stratification | Stratify study participants based on exposure status and measure the outcome at specific time intervals for each group to identify any potential differences in the effect of exposure on the outcome over time. |
Method | Description |
---|---|
Use a population-based sample | Select a sample of individuals from the general population, rather than from a specific group, such as those who are hospitalized or enrolled in a health program. |
Control for confounding factors | Identify and control for other factors that could influence the relationship between the disease and the factor of interest. This can be performed through statistical techniques, such as multivariable regression analysis. |
Use a randomized controlled trial | Randomly assign individuals to a treatment group or a control group, which can help minimize selection bias and other types of bias. |
Use multiple recruitment sources | Recruit participants from multiple sources to avoid over-representation of individuals with a specific disease or condition. |
Use clear inclusion and exclusion criteria | Develop clear criteria for inclusion and exclusion in the study to avoid selecting participants based on their disease status and other factors. |
Method | Description |
---|---|
Defining the study population | Clearly define the study population and excluding individuals who have a long duration of the disease or condition being studied. This can help ensure that the denominator used to calculate incidence is accurate and not artificially inflated. |
Identifying the time period | Clearly define the time period over which incidence is being measured. This can help to avoid including individuals who have had the disease or condition for a long time, which can inflate the prevalence and lead to an overestimation of incidence. |
Adjusting for disease duration | Use statistical techniques to adjust for the duration of the disease or condition when estimating incidence. This can involve calculating incidence rates over shorter time intervals or using survival analysis techniques to adjust for the duration of the disease. |
Conducting prospective studies | conducting prospective studies, which measure incidence over time, can help to minimize prevalence–incidence bias. In these studies, individuals are followed over time, and new cases of disease or condition are identified as they occur. |
Method | Description |
---|---|
Use random sampling | Random sampling ensures that every member of the population has an equal chance of being selected for the study. This helps to ensure that the sample is representative of the population. |
Use stratified sampling | Stratified sampling involves dividing the population into subgroups based on relevant characteristics, such as age, gender, or income level. Then, participants are randomly selected from each subgroup in proportion to their size in the population. This helps to ensure that the sample is representative of the population with respect to these important characteristics. |
Use appropriate recruitment methods | To avoid self-selection bias need appropriate recruitment methods. For example, if the study aims to investigate the prevalence of a certain disease, researchers should not recruit participants from a hospital, as this will bias the sample towards those who are already sick. |
Consider missing data | Missing data lead to selection bias. Estimate missing data or exclude participants with missing data only after examining whether their exclusion biases the sample. |
Analyze and report sample characteristics | Analyze and report the characteristics of the sample including demographics and other relevant information. This helps to understand the representativeness of the sample and potential sources of bias. |
Method | Description |
---|---|
Improving the quality and diversity of training data | One of the main causes of algorithm bias is flawed training data. By ensuring that the data used to train algorithms are representative of the real-world population and include diverse perspectives, biases can be minimized. |
Using transparent and explainable algorithms | Algorithms that are transparent and explainable can help to identify and address bias more effectively. By providing clear insights into how the algorithm arrived at its results, it is easier to understand and address any biases that may be present. |
Implementing regular audits and testing | Regular audits and testing can help to identify any biases in algorithms and ensure that they are producing fair and accurate results. This can involve analyzing the impact of the algorithm on different demographic groups and making adjustments as necessary. |
Involving diverse stakeholders in the development process | Involving a diverse range of stakeholders in the development process, including those from under-represented groups, can help to identify and address potential biases before the algorithm is deployed. |
Providing ongoing training and education | Providing ongoing training and education to those involved in the development and deployment of algorithms can help to raise awareness of algorithm bias and provide the skills and knowledge needed to prevent and mitigate it. |
Algorithm | Type | Description | Strengths | Weaknesses | Example |
---|---|---|---|---|---|
FGSM | Gradient-based | Computes the gradient of the loss function with respect to the input data and perturbs the data in the direction of the gradient. | Simple and fast; can generate effective adversarial examples. | May not generate robust adversarial examples that are resistant to other perturbation techniques. | Changing the label of an image from “dog” to “cat” by adding a small amount of noise. |
PGD | Gradient-based | An iterative version of FGSM that applies small perturbations to the input data and projects the perturbed point onto a valid region of the input space. | Can generate more robust adversarial examples than FGSM; can be used to create multiple adversarial examples with varying degrees of distortion. | More computationally expensive than FGSM. | Generating an adversarial example that is misclassified as a stop sign instead of a yield sign. |
CW attack | Optimization-based | Minimizes a custom loss function that encourages the perturbed input data point to be classified as the target label while minimizing the amount of perturbation. | Can generate adversarial examples that are difficult for a range of models to classify correctly; can generate adversarial examples with minimal perturbation. | Computationally expensive; may not always generate effective adversarial examples. | Generating an adversarial example that is misclassified as a “panda” instead of a “gibbon”. |
DeepFool | Optimization-based | Iteratively finds the closest decision boundary to the input data point and perturbs the point in the direction of the boundary. | Can generate small perturbations that cause a misclassification; can generate adversarial examples that are difficult for the model to detect. | May not generate robust adversarial examples that are resistant to other perturbation techniques. | Changing the label of an image from “car” to “truck” by adding a small amount of noise. |
JSMA | Gradient-based | Identifies the most important features of the input data point and perturbs them in a way that maximizes the model’s prediction error. | Can generate adversarial examples that are difficult for the model to classify correctly while minimizing the overall amount of perturbation. | May not generate robust adversarial examples that are resistant to other perturbation techniques. | Changing the label of an image from “bird” to “airplane” by perturbing the wings and beak of the bird. |
Technique | Description | Example |
---|---|---|
Data augmentation | Generate new data points by adding existing dataset. | Synthesize new data points for under-represented groups to increase dataset. |
Balancing the dataset | Oversample the minority class or undersample the majority class to make the dataset more balanced | Undersample the majority group to balance the proportion of defaulters between the groups. |
Feature engineering | Select and transform the input features used in model | Remove features that introduce bias, such as zip code, and add features that increase fairness, such as education level. |
Regularization | Constrain the model’s parameters to prevent overfitting and encourage it to generalize better to new data | Add regularization terms to the loss function that encourage the model to use a wider range of features. |
Counterfactual analysis | Identify scenarios and use them to test the model’s fairness | Simulate the effect of changing an applicant’s race or gender on the model’s prediction. |
Fairness constraints | Constraints model ensure that it treats all groups fairly | Add a constraint that limits the difference in the model’s false positive rate between different racial groups. |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2023 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Siddique, S.; Haque, M.A.; George, R.; Gupta, K.D.; Gupta, D.; Faruk, M.J.H. Survey on Machine Learning Biases and Mitigation Techniques. Digital 2024, 4, 1-68. https://doi.org/10.3390/digital4010001
Siddique S, Haque MA, George R, Gupta KD, Gupta D, Faruk MJH. Survey on Machine Learning Biases and Mitigation Techniques. Digital. 2024; 4(1):1-68. https://doi.org/10.3390/digital4010001
Chicago/Turabian StyleSiddique, Sunzida, Mohd Ariful Haque, Roy George, Kishor Datta Gupta, Debashis Gupta, and Md Jobair Hossain Faruk. 2024. "Survey on Machine Learning Biases and Mitigation Techniques" Digital 4, no. 1: 1-68. https://doi.org/10.3390/digital4010001
APA StyleSiddique, S., Haque, M. A., George, R., Gupta, K. D., Gupta, D., & Faruk, M. J. H. (2024). Survey on Machine Learning Biases and Mitigation Techniques. Digital, 4(1), 1-68. https://doi.org/10.3390/digital4010001