Article

AI-Series (3): Detect and Avoid Bias at an Early Stage

If the data quality or data usage is incorrect, the result can be bias. These systematic errors in predictive models, for example, create the risk of discriminatory AI results. There are proven concepts for avoiding, recognizing and correcting bias.

Detect and Avoid Bias at an Early Stage

No items found.

An example from New Zealand shows how explosive such errors can be: When identifying individuals at border controls, the focus clearly needs to be on the passport photos. However, an AI passport recognition system rejected Asians because the system interpreted asiatic eye shapes as closed. Systematic errors in predictive models can thus not only lead to incorrect results, they can also be unintentionally discriminatory. Here we present a method for avoiding, detecting and correcting such biases, which is already being used successfully in many projects.

Causes for bias

Systematic errors in predictive models, for example, are mainly due to poor data quality or incorrect use of data.

Lack of data quality

Non-representative samples and discriminatory data

The training data samples for an AI algorithm must be representative, otherwise subgroups with certain characteristics may be represented more frequently than in the actual total. For example, data collected online comes much more frequently from young people than from seniors – a distribution that does not, however, correspond to the real proportions in the population as a whole. The result is a disproportionate weighting of the characteristics of young people, with the risk of distorting the results for other age groups. The same is true if the training data is from discriminated groups. For example, if minorities were disadvantaged in old lending decisions, then the legacy lending data set is biased to the disadvantage of those minorities. Even when the discriminatory behavior has long since ended, without countermeasures the AI model persists and perpetuates the disparaging patterns preserved in the old training data. Moreover, it is irrelevant whether these groups are intentionally or unintentionally disadvantaged. Without countermeasures, this bias will negatively affect the predictive performance of an AI model.

Incorrect Data Usage

Incorrect handling of otherwise error-free data can also lead to bias. The four main causes are:

  • Subjective assessments
    • Manual data enrichment of an AI model is not uncommon, but it carries the risk of subjective annotations resulting in bias in the model output – for example, when biases, cultural perceptions, and semantics are unconsciously incorporated into the data and target variables. Thus, individuals with little contact with same-sex couples are less likely to classify such two-person constellations in photographs as relationships.
  • Methodological errors
    • Methodological errors can result from weaknesses in improving data quality. If, for example, missing values for male height are supplemented with the average height of all persons during data preparation, the bias has already occurred. The higher male average value should have been used here. This error will later be reflected in the model results.
  • Intentional manipulations
    • In addition, data and models can also be intentionally manipulated to provide advantages or disadvantages to certain groups. This is extremely difficult to detect because the chain from data collection to data use is very long and there are many places where manipulative interventions are possible.
  • Incorrect use of AI results
    • Even if an AI model produces results without bias, this is still no guarantee that it will be used that way. BCG Platinion has experienced in various projects that call center employees only use an AI recommendation for a customer if it coincides with their personal assessment. At least as problematic is the use of algorithms for purposes or target groups other than those originally intended. For example, an AI model for determining the insolvency risk for restaurants may produce systematically incorrect results if used in a different industry.

Integrated Solution Approach

Given the multiple causes, a process to effectively prevent bias must consist of several components. This is the only way to ensure that all error sources are taken into account. For this purpose, we recommend an integrated three-part approach that has proven its worth many times:

1. Avoid bias

Prevention is better than cure. Staff training and awareness-raising among those involved in development and operations are therefore effective levers for avoiding bias in the first place. Also useful are training concepts that convey special methodological, ethical and organizational content. It makes sense to impart this knowledge not only to the company’s own employees. Suppliers should also be appropriately screened for knowledge of anti-bias.

The creation of standards also counteracts bias, including the standardization of all development steps and control of their compliance. Comprehensive versioning, documentation, access controls, and logging of all AI input variables (features), AI model codes, and AI results, as well as strict data governance are all important. They have all been proven to avoid bias due to methodological errors, ignorance, or intentional incorporation of errors.

Finally, logging the assumptions and intended use of the final AI model prevents incorrect use of the AI results.

2. Detect bias

Experts are able to detect a systematic error in an AI model via the qualitative evaluation of input variables and results. The expected distribution values of a bias-free population are methodically compared with the input variables of the model in question. Abnormalities indicate data collection and processing errors.

Another key requirement for any plausibility check is transparency. The influence of individual variables on the model result as a whole must therefore be comprehensible. In this respect, there are significant differences between the algorithms. Accordingly, simpler models such as logistic regression are very transparent. For more complex models, indirect methods such as SHAP analyses (SHapley Additive exPlanations) are used to test individual influencing factors. Fairness measures can also be used to examine whether model outcome quality is similar across representations.

3. Correct bias

If bias occurs due to poor data quality, it can be eliminated by data set weighting, stratified random sampling (stratifying), or sampling strategies. The goal is always to obtain a weighting of the variables as in the data without bias. An example would be a male-female ratio that is expected to be 50/50. However, if a data set contains 33 percent males and 67 percent females, double weighting each male data set would be appropriate as a correction to restore the expected relationship. Potentially bias-causing input variables such as migration background or religion are best excluded right at the start of development.

On the other hand, however, this data may reveal that minorities are overrepresented. If the AI model is compromised in this respect, the additional inclusion of such a variable may even prevent bias. It is also important to ensure that features excluded due to bias are not included again in the results through the back door. To stick with the male-female example, it makes little sense to delete the gender variable but to train the algorithm with gender-specific first names. If subjective statements in manual data preparation are the cause of a bias problem, multiple surveys and additional test questions can remedy the situation.

Conclusion

Bias is an intrinsic problem of all human work, since there is no such thing as a perfectly and absolutely objective human being. AI has the advantage of being objective if it has been trained appropriately and only receives bias-free data. To achieve this, developers and users must always be aware of the causes of systematic errors. With the presented solution kit, they can succeed in minimizing bias to get realistic results from their AI model.


More to Explore

No items found.
No items found.
No items found.
Artificial Intelligence