GilHanbit1
ShinYejin2
KwakJoon Ho2
WooSungmin1*
-
( School of Electrical, Electronics and Communication Engineering, Koreatech / 1600,
Chungjeol-ro, Byeongcheon-myeon, Dongnam-gu, Cheonan-si, Chungcheongnam-do 31253,
Korea {hb35686, innosm}@koreatech.ac.kr)
-
( AI Trustworthiness Policy & Research Team, TTA, Korea {yepp1252, pentarous}@tta.or.kr)
Copyright © The Institute of Electronics and Information Engineers(IEIE)
Keywords
Fairness, Bias, Diversity, Discrimination, Equality
1. Introduction
The release of OpenAI's ChatGPT has sparked significant interest in the performance
and proliferation of artificial intelligence. ChatGPT, a representative hyperscale
artificial neural network, interacts with users conversationally, answers follow-up
questions, and occasionally acknowledges its own mistakes, making interactions feel
remarkably human-like. Moreover, it can quickly learn programming languages and write
essays at a human-like level.
However, concerns have emerged, particularly in the United States, about ChatGPT's
potential impact on education and academia, leading the New York Department of Education
to announce an official ban on its access in December 2022. Research papers published
by OpenAI, the creators of ChatGPT and the underlying GPT-3 model, have highlighted
issues related to potential misuse, bias, and fairness. These concerns include the
possibility of using the model for spreading false information, spam, phishing, and
writing fake essays, as well as its exploitation by malicious actors. Bias in the
training data can lead to the generation of text with stereotypes and prejudices.
For instance, GPT-3's authors reported that 83% of the 388 occupations generated by
the model were male-dominated, with Asian races receiving mainly positive scores while
negative scores were higher for black individuals. Additionally, violent and terrorism-related
words were more frequently associated with Islam than with other religions.
In March 2024, the European Union enacted the world's first artificial intelligence
regulation law, marking a significant milestone as the first enforceable legislation
of its kind. This development underscores the growing awareness and accountability
among academia, international organizations, and nations worldwide regarding the societal
risks associated with AI. However, while these AI regulations are essential, they
may potentially slow down AI development and affect competitive standing. Therefore,
it is imperative to analyze and ensure trustworthy AI systems, which encompass factors
such as fairness, transparency, safety, and other factors to effectively balance innovation
and regulation.
In this paper, we investigate the definitions of fairness and bias in the standards
and regulations of international organizations. We review academic approaches and
verification tools for evaluating the fairness of artificial intelligence, particularly
in the field of machine learning. Our aim is to predict the current level of fairness
in machine learning, understand the progression of standardization and regulation,
and provide guidance on the direction of fairness technology for industry experts.
Table 1. Summary of bias definitions from international standards organizations and
regulations.
Organization
|
Definition
|
ISO
(ISO/IEC TR24027) [2]
|
Common categories of negative impacts that might be perceived as “unfair” are listed
below. Unfair allocation / Unfair quality of service / Stereotyping / Denigration
/ “Over" or "under" representation
|
European Commission
(Ethics guidelines for trustworthy AI) [8]
|
The substantive dimension implies a commitment to: ensuring equal and just distribution
of both benefits and costs, and ensuring that individuals and groups are free from
unfair bias, discrimination and stigmatization. The procedural dimension of fairness
entails the ability to contest and seek effective redress against decisions made by
AI systems and by the humans operating them.
|
OECD
(Recommendation of the Council on Artificial Intelligence) [4]
|
AI actors should respect the rule of law, human rights and democratic values, throughout
the AI system lifecycle.
|
UNESCO
(Recommendation on the ethics of artificial intelligence) [5]
|
AI actors should promote social justice and safeguard fairness and non-discrimination
of any kind in compliance with international law.
This implies an inclusive approach to ensuring that the benefits of AI technologies
are available and accessible to all, taking into consideration the specific needs
of different age groups, cultural systems, different language groups, persons with
disabilities, girls and women, and disadvantaged, marginalized and vulnerable people
or people in vulnerable situations.
|
NIST
(AI Risk Management Framework: AI RMF 1.0) [6]
|
Absence of harmful bias is a necessary condition for fairness.
|
The Alan Turing Institute in UK (Understanding artificial intelligence ethics and
safety) [7]
|
All Al systems that process social or demographic data pertaining to features of human
subjects must be designed to meet a minimum threshold of discriminatory non-harm.
|
Australia
(Australia's AI Ethics Principles) [9]
|
AI systems should be inclusive and accessible, and should not involve or result in
unfair discrimination against individuals, communities or groups.
|
Ministry of Science and ICT
(Strategy to realize trustworthy artificial intelligence) [10]
|
Functionality that prevents the generation of conclusions that include discrimination
or bias towards specific groups during the processing of data by artificial intelligence
|
Major corporations
(Google, IBM, Microsoft, Facebook, NAVER, Kakao)
|
Google: Avoid creating or reinforcing unfair bias [11].
IBM: Fairness refers to the equitable treatment of individuals, or groups of individuals,
by an AI system [12].
Microsoft: AI systems should treat all people fairly [13].
Facebook: At Facebook, we believe that our products should treat everyone fairly and
work equally well for all people [14].
Naver: Naver will develop and utilize AI while considering the value of diversity,
in order to avoid any unfair discrimination towards all individuals including users
[15].
Kakao: We are cautious to ensure that intentional social discrimination does not occur
in algorithmic results [16].
|
2. Definition of fairness and bias
2.1 Definition of fairness and bias in standards and regulations
Standardization bodies such as ISO/IEC and international organizations such as OECD
and European Commission provide documents and guidelines on the trustworthiness of
artificial intelligence. ISO/IEC TR 24028:2020 [1] includes transparency, controllability, robustness, recoverability, fairness, stability,
personal data protection, and security as considerations necessary for ensuring trustworthiness,
and definitions and related content regarding fairness can be found in ISO/IEC TR
24028:2020, ISO/IEC TR 24027 [2], and ISO/IEC 23894:2023 [3].
In [1], bias is described as one of the factors that can undermine the trustworthiness of
artificial intelligence systems, defined as ``Favoritism towards some things, people
or groups over others.'' [2] explains the causes of unfairness as follows: ``Bias in objective functions, imbalanced
data sets, and human biases in training data and in providing feedback to systems,
bias issue in the product concept, the problem formulation or choices about when and
where to deploy AI systems.'' [3] defines bias in AI systems as ``a systematic difference in treatment of certain objects,
people, or groups in comparison to others'', and describes fairness as ``a concept
that is distinct from, but related to bias.'' When referring to the fairness of artificial
intelligence systems, OECD's ``Recommendation of the Council on Artificial Intelligence''
[4] and UNESCO's ``Recommendation on the ethics of artificial intelligence [5]'' emphasize the need to ensure fairness of the system by striving to prevent discriminatory
biased practices and outcomes throughout the entire life cycle.
The US National Institute of Standards and Technology (NIST) explains fairness as
one of the fundamental principles related to the risks of artificial intelligence
as follows: ``Fairness in AI includes concerns for equality and equity by addressing
issues such as harmful bias and discrimination. Standards of fairness can be complex
and difficult to define because perceptions of fairness differ among cultures and
may shift depending on application [6].'' The Alan Turing Institute in the UK has provided guidelines for applying AI ethics
and safety principles to the design and implementation of algorithmic systems in the
public sector in their publication ``Understanding artificial intelligence ethics
and safety.'' Here, fairness is described as a principle for responsible design and
use of AI technology, stating that ``All AI systems that process social or demographic
data pertaining to features of human subjects must be designed to meet a minimum threshold
of discriminatory non-harm [7].''
2.2 Fairness and bias in academia
2.2.1 Bias in artificial intelligent systems
Most artificial intelligence systems and algorithms are based on data and operate
by training on this data. Therefore, data is closely related to the operation of artificial
intelligence systems. If there is bias in the data used for training, the algorithm
learned from that data can generate biased results or amplify the bias in the data.
Additionally, even when the data is unbiased, a biased behavior can be exhibited due
to specific algorithmic design choices. The results of biased algorithms can affect
user decisions and ultimately lead to a vicious cycle of generating biased data for
future algorithm training. For example, with web search engines that place certain
results at the top of the list, users tend to interact more with the top results and
pay less attention to the results at the bottom of the list. User-item interactions
collected by the web search engine are then used to make decisions about how to display
information based on popularity and user interest. This can create bias towards the
top results not because of the characteristics of the results themselves, but due
to biased interactions and list placement. The types of bias in [17] are divided into three categories: data, algorithm, and user interaction, defined
as interacting loops. Fig. 1. illustrates the cause-and-effect relationships among these factors and categorizes
them accordingly. Table Table 2 summarizes the types and causes of biases in a table. For more detailed information,
refer to [17].
Fig. 1. Inherent bias in the feedback loop among data, algorithm, and users.
Table 2. Summary of three types and causes of biases.
Categories
|
Bias
|
Causes of occurrence
|
Data to Algorithm
|
Measurement
|
Occurs in the method of selecting, utilizing, and measuring features
|
Omitted variable Bias
|
When one or more important variables are excluded from the model
|
Representation Bias
|
Occurs in the method of sampling from a population
|
Aggregation bias
|
when drawing incorrect conclusions about individuals by observing the entire population
|
Simpson's Paradox
|
A type of aggregate bias that occurs in the analysis of heterogeneous data
|
Sampling Bias
|
Similar to selection bias, it occurs due to non-random sampling of subgroups
|
Longitudinal Data Fallacy
|
In cross-sectional analysis, combining diverse groups at a single point in time can
lead to bias, as heterogeneous groups can lead to different conclusions
|
Linking Bias
|
when the social connection properties obtained from user connections, activities,
or interactions differ from actual user behavior.
|
Algorithm to User
|
Algorithmic Bias
|
When bias is purely added by the algorithm and not present in the input data
|
User Interaction Bias
|
biased behavior or interactions based on user interfaces and user choices
|
Presentation Bias
|
Bias due to the way information is represented. For example, on the web, users can
only click on the visible content, so clicks occur on the displayed content and not
on any other content
|
Ranking Bias
|
Top-ranking results are thought to be the most relevant and important, so they induce
more clicks. This bias affects search engines and crowdsourcing applications.
|
Popularity Bias
|
More popular items tend to be displayed more frequently. Popularity measures can be
manipulated. This type of bias can be seen in recommendation systems, search engines,
or other systems where popular items are presented more to the public.
|
Emergent Bias
|
This refers to bias that emerges as a result of interaction with actual users, and
typically occurs after a certain period of time has elapsed following the completion
of the design, due to changes in population/cultural values or social knowledge.
|
Evaluation Bias
|
Bias that occurs during model evaluation, such as the use of imbalanced benchmarks.
|
User to Data
|
Historical Bias
|
an existing bias that is a socio-technological issue, where search results for "female
CEO" are influenced by the fact that only 5% of Fortune 500 CEOs are women.
|
Population Bias
|
when statistics, population distribution, representative characteristics, or user
characteristics differ from the platform's user distribution.
|
Self-Selection Bias
|
when research subjects are self-selected, such as passionate supporters being more
likely to participate in a poll.
|
Social Bias
|
when others' behavior influences our judgment, such as when we want to rate or review
an item poorly but are influenced by others' high ratings.
|
Behavioral Bias
|
when user behavior differs depending on the platform, context, or dataset.
|
Temporal Bias
|
Temporal bias arises from differences in population and behavior over time, such as
when people start using hashtags to attract attention to a topic and then continue
discussing the event without using hashtags.
|
Content Protection Bias
|
structural, lexical, semantic, and syntactical differences in user-generated content,
such as differences in language use based on gender, age, country, and population.
|
2.2.2 Academic definition of fairness
The term ``fairness'' in machine learning denotes the multiple efforts made to rectify
any form of algorithmic bias in automated decision-making processes that rely on machine
learning models. Algorithmic fairness has begun to garner the attention of researchers
in the artificial intelligence, software engineering, and legal communities, with
more than twenty different notions of fairness proposed in the past few years. There
is no clear consensus on which definition to apply in each situation, and the detailed
differences between multiple definitions are difficult to comprehend. However, most
statistical metrics in machine learning rely on the concepts of true positive (TP),
true negative (TN), false positive (FP), and false negative (FN) based on predicted
and actual class values commonly used in machine learning. [90] collected and theoretically interpreted definitions of fairness for algorithmic classification
problems, and explained them through a unified case study. [91] organized different academic definitions of fairness from the perspectives of independence,
separation, and sufficiency. Table Table 3 below summarizes the academic definitions of fairness.
Table 3. Summary of academic definitions of fairness.
Fairness definitions
|
Explanations
|
Equalized Odds
|
A predictor $\hat{Y}$ satisfies equalized odds with respect to protected attribute
${A}$ and outcome ${Y}$, if $\hat{Y}$ and ${A}$ are independent conditional on ${Y}$.
$
P(\hat{Y}=1\mid A=0,~Y=y)\\
=PA(\hat{Y} =1 \mid A=1, ~Y=y),~y\in \{0,~1\}
$
|
Equal opportunity
|
A binary predictor $\hat{Y}$satisfies equal opportunity with respect to ${A}$ and
${Y}$ if
$\begin{aligned}
P(\hat{Y}=1\mid A=0,~Y=1)\\
=P(\hat{Y}=1 \mid A=1,~Y=1).
\end{aligned}$
This means that the probability of a person in the positive class being classified
as positive should be the same for both the protected and unprotected groups. In other
words, both groups should have the same true positive rates.
|
Demographic Parity
(statistical parity)
|
A predictor $\hat{Y}$ satisfies demographic parity if $P(\hat{Y}\mid A=0)=P(\hat{Y}
\mid A=1)$ . It means that the likelihood of a positive outcome should be the same
regardless of membership in a protected group (e.g. female).
|
Fairness Through Awareness
|
n algorithm is fair if it gives similar predictions to similar individuals. If there
is no difference in similarity measurement (such as inverse distance) for a specific
task, the results should also be similar.
|
Fairness Through Unawareness
|
An algorithm is fair as long as any protected attributes A are not explicitly used
in the decision-making process. Fairness is achieved if the protected attribute (e.g.
race in loan decisions) is not explicitly used in the decision-making process.
|
Treatment Equality
|
Treatment equality is achieved when the ratio of false negatives and false positives
is the same for both protected group categories. the false negative and false positive
rates should be the same for both protected groups (male and female).
|
Test Fairness
|
$S = S(x)$ is test fair (well-calibrated) if it reflects the same likelihood of recidivism
irrespective of the individual's group membership, $R$. That is, if for all values
of $s$,
$\begin{aligned}
P(\hat{Y}=1\mid S=s,~ R=b)\\
=P(\hat{Y}=1\mid S=s,~ R=w).
\end{aligned}$
For all risk scores, the probability of individuals from protected and unprotected
groups being classified as positive class should be the same.
|
Counterfactual Fairness
|
A predictor $\hat{Y}$ is counterfactually fair if under any context $X=x$ and $A=a$
$\begin{aligned}
P({\hat{Y}}_{A\leftarrow a(U)}=y\mid X=x,~A=a)\\
=P({\hat{Y}}_{A\leftarrow a^{'}(U)}=y\mid X=x,~A=a)
\end{aligned}$
for all $y$ and for any value attainable by A. Individuals should have the same outcomes
in the real world and the counterfactual world (involving membership in a different
racial group).
|
Fairness in Relational Domains
|
A fairness concept that captures the relational structure of a domain by considering
not only individual attributes but also social, organizational, and other connections
between individuals.
|
Conditional Statistical Parity
|
For a set of legitimate factors $L$, predictor $\hat{Y}$ satisfies conditional statistical
parity, if
$\begin{aligned}
P(\hat{Y}\mid L=1,~A=0)=P(\hat{Y}\mid L=1,~ A=1).\end{aligned}$
Under controlled legal risk factors, an equal proportion of defendants from each racial
group should be detained. For example, Black and White defendants who receive the
same number of guilty verdicts should be detained at the same rate.
|
As seen in the table, satisfying multiple fairness simultaneously seems impossible
without specific constraints. Therefore, when considering fairness, it is important
to consider the context and situation where the fairness definition is used and apply
it accordingly, while also analyzing the potential impact on individuals or groups
over time. Furthermore, when addressing fairness-related issues, it is necessary to
consider the causes and types of bias and satisfy the fairness criteria mentioned
above by taking them into account.
3. Bias mitigation techniques
In machine learning, there are various methods to mitigate bias, and they differ depending
on individual academic fields. However, these methods can be categorized into pre-processing
methods that offset bias in the data, correction methods that are integrated with
individual learning algorithms, and post-processing methods.
3.1 Pre-processing
The creation of all datasets is the result of various design decisions made by the
data creator (curator). These decisions have an impact on the fairness of the dataset
and ultimately affect the results of the algorithm. Best practices have been proposed
to mitigate data bias, such as creating a datasheet containing information on the
dataset's creation method, features, motivations, and skewed distribution of the data
[18,19]. Labeling, such as nutrition labels on food, has also been suggested for better data
classification [20]. Studies have been conducted to propose methods for testing Simpson's paradox [21] or for automatically detecting bias [22,23]. Causal models and causal graphs are also used for tasks that involve detecting direct
discrimination in the data [24] and preventing techniques to modify the data [25]. Furthermore, research has been proposed to prevent discrimination in data mining
[26], including messaging [27], preferential sampling [28,29], and disparate impact removal [30], which are other preprocessing approaches aimed at removing bias from data. [31] suggests a fair data envelopment analysis method, incorporating an additional constraint
to mitigate disparate impact, thereby addressing the scoring parity between privileged
and unprivileged groups. [32] formulates fairness as an optimization problem, obscuring information about protected
group members to find a good representation for group and individual fairness during
data encoding. [33] presents optimized preprocessing, which uses a probabilistic transformation to modify
data features and labels with fairness and accuracy constraints for group fairness,
individual distortion, and data fidelity. [34] performs a theoretical study on the trade-off relationship between accuracy and discrimination
and proposes four methods using non-discriminatory constraints in classifier design.
[35] applies a linear transformation on the non the non-sensitive feature columns to eliminate
their correlation with the sensitive feature columns while preserving the maximum
amount of information possible.
3.2 In-Processing
Classification is a common task in machine learning used in various domains where
human contact is not feasible, and techniques have been proposed to minimize cost
while meeting certain definitions of fairness to achieve fair classification [36]. These techniques include adding constraints for fairness to the cost function through
regularization [37,38,39], optimizing the estimation probability of sensitive and non-sensitive features to
be equal [40], modifying the Naive Bayes classifier [41], maximizing the average accuracy of each group [42], mitigating dependency on sensitive features using Wasserstein distance measure [43], using Preferential Sampling (PS) for discrimination-free training data [44], and incorporating attention mechanism for interpretable and bias-mitigated classification
[45]. The ART classifier is used as a classifier to satisfy fairness such as demographic
parity and equalized odds [46]. [47] is a meta-algorithm that modifies the classification threshold of the base classifier
to maximize a fairness metric while maintaining a certain level of accuracy. Adversarial
debiasing [48] trains a classifier to improve prediction accuracy while simultaneously minimizing
the adversary's ability to infer the protected attribute from the predictions. [49] proposes a group fairness classifier aiming to avoid the susceptibility of intentional
or inadvertent ``fairness gerrymandering'' that can occur with traditional statistical
definitions of fairness that fix a small collection of pre-defined groups. Exponentiated
Gradient Reduction [50] reduce fair classification to a sequence of cost-sensitive classification problems,
resulting in a randomized classifier with the lowest empirical error while adhering
to fair classification constraints.
In community detection, users with low connectivity are often ignored, which can lead
to improper classification into communities. However, CLAN [51] first identifies important communities and then uses a method of classifying a small
number of users into important communities, thus improving the community classification
performance of these minority users. In addition, methods for community detection
using graph embedding [52] or clustering [53,54] have also been proposed.
In regression problems, [55] proposed to define the loss (cost) function by dividing individuals, groups, and
individuals/groups rather than using regression methods for the entire dataset in
order to minimize it. Methods for modeling demographic and group-specific fairness
as loss functions [56], and using decision trees that can be applied to classification and regression problems
have also been proposed [57]. [58] presented an algorithm-independent lower bound by characterizing the trade-off between
the inherent fairness and accuracy in regression problems. [59] proposed a two-step approach where an unconstrained regression is first estimated
from labeled data and then recalibrated with unlabeled data. [60] examines the problem of achieving demographic parity in real-valued function learning
by establishing a connection between fair regression and optimal transport theory,
and proposes a post-processing algorithm to achieve fairness with established guarantees.
Fair PCA [61] was proposed to address the issue of amplifying the reconstruction error of one group
in original PCA when multiple groups of equal size exist in the data, and it aims
to achieve similar representation across different groups. To avoid the problem of
inadequate representation of group data other than the lowest principal component
during dimensionality reduction of the entire dataset, this method defines a loss
function for each group and only performs PCA for the group with the highest loss
function. [62] formulated the maximum mean discrepancy (MMD)-based fair PCA as a constrained optimization
over the Stiefel manifold and solved it using Riemannian Exact Penalty Method via
Smoothing.
In addition to classification and regression, studies aimed at mitigating bias have
been proposed in the literature. Prejudice remover [63] modifies the learning objective by adding a regularization term that is sensitive
to discrimination. [64] exploits adversarial learning for mitigating unwanted biases in work embeddings.
Causal models are used to eliminate unwanted dependencies on sensitive attributes
such as gender or race by using causal graphs to confirm causal relationships between
variables [65,66]. In [67], a method is proposed for detecting direct/indirect discrimination from historical
data using causal graphs and removing discrimination effects before predictive analysis.
First, path-specific effects (direct/indirect) are defined as below and discrimination
thresholds for protected groups are set. For example, the UK law on gender discrimination
in 1975 allowed a 5% difference. Similarly, thresholds for indirect discrimination
for Redlining properties can be set. From this, direct and indirect discrimination
for all paths in the causal graph are determined, and a objective function is created
and constraints for mitigating discrimination are set to modify the causal graph through
quadratic programming.
There is research on autoencoders that learn fair representations and avoid interference
of sensitive attributes [68]. In this research, the goal is to obtain fair representations by removing information
about sensitive variables. A maximum mean discrepancy regularizer is used to ensure
that the posterior distribution of the latent variables is the same. In [69], the optimization objective function is designed to include the mutual information
between the encoding and sensitive variables, which helps to remove biases in the
representation. Debiased variational autoencoder (DB-VAE) is used to reduce the accuracy
gap by training the model to learn sensitive attribute variables such as skin color
and gender [70]. This method modifies the traditional variational autoencoder (VAE) by incorporating
the learning of classes corresponding to the samples and representing the frequency
of occurrence of specific classes as a histogram. The sample occurrence probability
is defined to be inversely proportional to the frequency, ensuring that during training,
the samples do not converge to a single class. Adaptive re-sampling is employed to
prevent the network from biasing towards a particular class as the learning progresses.
Fig. 2. Debiasing variational autoencoder (DB-VAE) employs resampling during training,
where the sample selection is inversely proportional to its frequency of appearance,
to mitigate data bias [70].
To mitigate biases from learned models with fixed conceptual associations, adversarial
debiasing methods were proposed. The goal here is to enhance prediction accuracy while
simultaneously reducing the predictions made by an adversary regarding protected and
sensitive variables [71]. Instead of removing dependency on discriminatory data, FairGAN [72] proposed a method to generate data that is similar to real data but without discrimination.
The architecture of FairGAN consists of one generator and two discriminators. The
generator generates fake data from protective attributes and noise, while the discriminator
plays a role in distinguishing real data from fake data. The third component is for
the fairness constraint, determining whether generated (fake) data is created from
a protected attribute group.
In the research of word embedding which represents a word to a vector for natural
language processing, [73] proposed the bias relationships such as ``man to computer programmer'' and ``woman
to homemaker.'' This method defined the gender-specific and gender-neutral words,
and re-embedded the latter so that gender relationship in the pair is removed. Recently,
[74] proposed a method to analyze and mitigate the biases of contextualized vectors.
Recent advancements in reinforcement learning have enabled automated online decision-making
in communication network design and optimization problems. However, most reinforcement
learning approaches prioritize maximizing agent rewards without considering fairness.
Recent research proposes introducing a fairness term in actor-critic reinforcement
learning, adjusting rewards multiplicatively [75]. In this study, weighted multiplicative-adjusted rewards are given to the Actor instead
of traditional rewards, addressing fairness concerns in network utility optimization.
3.3 Post-Processing
Post-processing techniques represent a subset of methods for mitigating fairness issues
in machine learning. These methods are applied subsequent to the model making predictions,
with the goal of adjusting the model's outputs to ensure fairness through various
means.
The ``Equalized Odds Postprocessing'' technique modifies predicted probabilities for
distinct groups to meet the equalized odds criterion [76,77]. Its objective is to guarantee that the odds of accurate classification are equivalent
across different demographic or sensitive attribute groups. Expanding on the equalized
odds concept, the ``Calibrated Equality Odds'' method involves calibrating the model's
predicted probabilities and subsequently applying equalized odds postprocessing [77]. The aim is to concurrently achieve calibration and fairness. The ``Reject Option
Classification'' approach [78] introduces a ``reject'' option, where the model refrains from making predictions
in situations of high uncertainty. This strategy seeks to alleviate biases by yielding
more favorable outcomes to underprivileged groups and less favorable outcomes to privileged
groups within a confidence band around the decision boundary [79]. In a binary classification context, [79] proposed a systematic approach for achieving fairness. The core idea is to reduce
fair classification to a series of cost-sensitive classification problems. The solutions
obtained yield a randomized classifier with the lowest empirical error while adhering
to specified constraints. For fair regression, [80] introduced a general scheme wherein the prediction is required to be statistically
independent of the protected attribute, and the prediction error for any protected
group must remain below a predetermined level.
4. Tools for fairness evaluation and mitigation
Leading companies in the field of artificial intelligence are at the forefront of
providing tools for assessing fairness and alleviating biases and discrimination within
machine learning models. These tools are instrumental in evaluating and fine-tuning
the outcomes of machine learning algorithms, leveraging the discrimination mitigation
techniques discussed earlier.
IBM's AI Fairness 360 Toolkit summarized in Table Table 4 encompasses a broad range of features, including various fairness metrics and mitigation
algorithms [81,82]. These tools collectively support the evaluation and mitigation of fairness issues
across diverse models and datasets, aiding in the identification and resolution of
biases that may manifest in model predictions as illustrated in Fig. 3. AIF360 offers more than 70 fairness metrics and 14 bias mitigation algorithms designed
to measure and address the fairness of artificial intelligence algorithms. AIF360
supports bias mitigation algorithms in pre-processing, in-processing, and post-processing.
Fig. 3. The fairness pipeline of AIF360 [81].
Table 4. Bias Mitigation Algorithm supported by IBM's AIF360.
Category
|
Algorithm
|
Pre-processing
|
Disparate Impact Remover
Learning Fair Representation
Optimized Preprocessing
Reweighing
|
In-processing
|
Adversarial Debiasing
ART Classifier
Gerry Fair Classifier
Meta Fair Classifier
Prejudice Remover
Exponentiated Gradient Reduction
Grid Search Reduction
|
Post-processing
|
Calibrated Equality of Odds
Equality Odd
Reject Option Classification
|
Fairlearn by Microsoft [83,84,85] is an open-source toolkit providing functions to assess fairness and algorithms to
mitigate unfairness in machine learning models. For fairness evaluation, Fairlearn
offers library functions based on various metrics, allowing users to select sensitive
attributes and examine evaluation results based on performance and prediction differences.
Additionally, a dashboard makes it easy to check whether a model is fair. Fairlearn's
unfairness mitigation algorithms are categorized into pre-processing, post-processing,
and in-processing. Pre-processing algorithms mitigate unfairness by transforming the
dataset, while post-processing adjusts the predictions to align with specified equalized
constraints. In-processing algorithms solve an optimization problem by replacing fairness-related
terms with Lagrange multipliers as constraints, effectively converting the problem
into a general machine learning task. As of the current version (January 4, 2024),
Fairlearn is at v0.9.0 and can be considered pre-release.
Google has developed Fairness Indicators [86], which facilitate the easy calculation of commonly identified fairness metrics for
binary and multiclass classifiers. Unlike existing fairness evaluation tools that
may not perform well on large datasets and models, Google's Fairness Indicators allow
the evaluation of fairness metrics for all scales of use cases.
KAIST's AI Fairness Research Center has developed the `MSIT AI FAIR 2022 (MAF 2022)'
framework for analyzing, detecting, mitigating, and eliminating bias in AI models
and training data [87]. AF 2022 boasts a total of 19 algorithms, making it more versatile than other fairness
tools for various situations as summarized in Table Table 5.
Table 5. Bias Mitigation Algorithm supported by KAIST' MSIT AI FAIR 2022.
Category
|
Algorithm
|
Pre-processing
|
Disparate Impact Remover
Learning Fair Representation
Optimized Preprocessing
Reweighing
|
In-processing
|
Adversarial Debiasing
ART Classifier
Gerry Fair Classifier
Meta Fair Classifier
Prejudice Remover
Exponentiated Gradient Reduction
Grid Search Reduction
Kernel density estimation
Fairness VAE
Learning from fairness
Fair feature distillation
FairBatch
|
Post-processing
|
Calibrated Equalized Odds Postprocessing
Equalized Odds Postprocessing
Reject Option Classification
|
Aequitas, developed by the Center for Data Science and Public Policy at the University
of Chicago, is an audit tool for bias and fairness [88,89]. It enables understanding various types of bias and utilizing them in development.
Aequitas guides users through the process of uploading data, selecting a protected
group, choosing fairness metrics, and selecting thresholds to provide a bias report.
5. Conclusion
In this paper, we examined international standards, national definitions, and definitions
from leading companies related to fairness. Additionally, we explored extensive research
on the increasing importance of fairness in artificial intelligence. We compared the
inherent biases in machine learning and their mitigation methods from various research
papers. The study delved into the vast research on fairness in artificial intelligence,
considering aspects of data preprocessing, in-processing from a processing perspective,
and post-processing when data and algorithms are inaccessible. Furthermore, we provided
a comparative analysis of AI bias measurement methods and mitigation approaches offered
by universities and companies, presenting guidelines for research on fairness in artificial
intelligence. Based on this comprehensive survey, we conclude several promising directions
for future research. First, it is essential to develop more sophisticated and context-aware
bias mitigation techniques that can adapt to various cultural norms and application
contexts. Current bias mitigation techniques are limited to addressing specific known
biases or constraints, making it challenging to find universally applicable tools.
Second, since biases can occur in data, algorithms, and user interactions, careful
attention is required at each stage of AI system development. Enhancing the transparency
and interpretability of AI models will help in more effectively identifying and mitigating
biases. Finally, collaboration between academia and industry can lead to the creation
of standardized guidelines and tools that are robust and universally applicable. By
integrating newly developed bias mitigation tools with existing ones, we can continuously
enhance scalability and effectiveness.
ACKNOWLEDGMENTS
This work was supported by the Korean MSIT (Ministry of Science and ICT) as Establishing
the foundation of AI Trustworthiness (TTA) and by the Education and Research Promotion
Program of KOREATECH in 2023.
REFERENCES
ISO/IEC TR 24028:2020, ``Information technology -- Artificial intelligence -- Overview
of trustworthiness in artificial intelligence,'' 2020.

ISO/IEC TR 24027:2021, ``Information technology -- Artificial intelligence (AI) --
Bias in AI systems and AI aided decision making,'' 2021.

ISO/IEC 23894:2023, ``Information technology -- Artificial intelligence -- Guidance
on risk management,'' 2023.

OECD, ``Artificial Intelligence in Society,'' June 2019.

UNESCO, ``Recommendation on the Ethics of Artificial Intelligence,'' November 2021.

National Institute of Standards and Technology, ``Artificial Intelligence Risk Management
Framework (AI RMF 1.0),'' January 2023.

D. Leslie, ``Understanding artificial intelligence ethics and safety: A guide for
the responsible design and implementation of AI systems in the public sector,'' Zenodo,
June 2019.

European Commission, Directorate-General for Communications Networks, Content and
Technology, ``Ethics guidelines for trustworthy AI,'' Publications Office, November
2019.

Australia, ``Australia's AI Ethics Principles,'' https://www.industry.gov.au/publications/australias-artificial-intelligence-ethics-framework/australias-ai-ethics-principles

Ministry of Science and ICT, ``Strategy to realize trustworthy artificial intelligence,''
May. 2021, https://www.msit.go.kr/eng/bbs/view.do?sCode=eng&mId=4&mPid=2&pageIndex=&bbsSeqNo=42&nttSeqNo=509&searchOpt=ALL&searchTxt

Google, ``Objectives for AI applications,'' https://ai.google/responsibility/principles

IBM, ``Our fundamental properties for trustworthy AI,'' https://www.ibm.com/artificial-intelligence/ai-ethics-focus-areas

Microsoft, ``Microsoft responsible AI principles,'' https://www.microsoft.com/en-us/ai/our-approach?activetab=pivot1%3aprimaryr5

J. Pesenti, ``Facebook's five pillars of Responsible AI,'' June 2021, https://ai.facebook.com/blog/facebooks-five-pillars-of-responsible-ai/

Naver, ``AI Ethics Principles,'' https://www.navercorp.com/value/aiCodeEthics

Kakao, ``AI Ethics,'' https://www.kakaocorp.com/page/responsible/detail/algorithm

N. Mehrabi, F. Morstatter, N. Saxena, K. Lerman, and A. Galdyan, ``A survey on bias
and fairness in machine learning,'' ACM Computing Surveys (CSUR), vol. 54, no. 6,
pp. 1-35, 2021.

M. Benjamin, P. Gagnon, N Rostamzadeh, C. Pan, Y. Bengio, and A. Shee, ``Towards standardization
of data licenses: The montreal data license,'' arXiv preprint arXiv:1903.12262, 2019.

T. Gebru, J. Morgenstern, B. Vecchione, J. W. Vaughan, H. Wallach, H. D. Ill, and
K. Crawford, ``Datasheets for datasets,'' Communications of the ACM, vol. 64, no.
12, pp. 86-92, 2021

S. Holland, A. Hosny, S. Newman, J. Joseph, and K. Chmielinski, ``The dataset nutrition
label,'' Data Protection and Privacy, vol. 12, 2020.

R. A. Kievit, W. E. Frankenhuis, L. J. Waldorp, and D. Borsboom, ``Simpson's paradox
in psychological science: a practical guide,'' Frontiers in Psychology, vol. 4, 513,
2013.

N. Alipourfard, P. G. Fennell, and K. Lerman, ``Can you trust the trend?'' Proc. of
the 11th ACM International Conference on Web Search and Data Mining, pp. 19-27, February
2018.

N. Alipourfard, P. G. Fennell, and K. Lerman, ``Using Simpson's paradox to discover
interesting patterns in behavioral data,'' Proc. of 12th International AAAI Conference
on Web and Social Media, June 2018.

L. Zhang, Y. Wu, and X. Wu, ``Achieving non-discrimination in data release,'' Proc.
of the 23rd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining,
pp. 1335-1344, August 2017.

L. Zhang, Y. Wu, and X. Wu, ``Causal modeling-based discrimination discovery and removal:
Criteria, bounds, and algorithms,'' IEEE Transactions on Knowledge and Data Engineering,
vol. 31, no. 11, pp. 2035-2050, 2018.

S. Hajian and J. Domingo-Ferrer, ``A methodology for direct and indirect discrimination
prevention in data mining,'' IEEE transactions on Knowledge and Data Engineering,
vol. 25, no. 7, pp. 1445-1459, July 2013.

F. Kamiran and T. Calders, ``Classifying without discriminating,'' Proc. of 2009 2nd
International Conference on Computer, Control and Communication, pp. 1-6, February
2009.

F. Kamiran and T. Calders, ``Classification with no discrimination by preferential
sampling,'' Proc. of 19th Machine Learning Conference in Belgium and The Netherlands,
vol. 1, no. 6, May 2010.

F. Kamiran and T. Calders, ``Data preprocessing techniques for classification without
discrimination,'' Knowledge and Information Systems, vol. 33, no. 1, pp. 1-33, 2012.

M. Feldman, S. A. Friedler, J. Moeller, C. Scheidegger, and S. Venkatasubramanian,
``Certifying and removing disparate impact,'' Proc. of the 21th ACM SIGKDD International
Conference on Knowledge Discovery and Data Mining, pp. 259-268, August 2015.

S. Radovanovi´c, G. Savi´c, B. Deliba˘si´c, and M. Suknovi´c, ``FairDEA-removing
disparate impact from efficiency scores,'' European Journal of Operational Research,
vol. 301, no. 3, pp. 1088-1098, 2022.

R. Zemel, Y. Wu, K. Swersky, T. Pitassi, and C. Dwork, ``Learning fair representations,''
Proc. of International Conference on Machine Learning, PMLR, pp. 325-333, 2013.

F. P. Calmon, D. Wei, B. Vinzamuri, K. N. Ramamurthy, and K. R. Varshney, ``Optimized
pre-processing for discrimination prevention,'' Proc. of Conference on Neural Information
Processing Systems, 2017.

F. Kamiran and T. Calders, ``Data preprocessing techniques for classification without
discrimination,'' Knowledge and Information Systems, 2012.

https://github.com/fairlearn/fairlearn/blob/main/fairlearn/preprocessing/_correlation_remover.py

A. K. Menon and R. C. Williamson, ``The cost of fairness in binary classification,''
Proc. of Conference on Fairness, Accountability and Transparency, PMLR, pp. 107-118,
January 2018.

T. Kamishima, S. Akaho, H. Asoh, and J. Sakuma, ``Fairness-aware classifier with prejudice
remover regularizer,'' Machine Learning and Knowledge Discovery in Databases, Springer,
Berlin, Heidelberg, pp. 35-50, 2012.

L. Oneto, M. Doninini, A. Elders, and M. Pontil, ``Taking advantage of multitask learning
for fair classification,'' Proc. of Joint European Conference on Machine Learning
and Knowledge Discovery in Databases, Springer, Berlin, Heidelberg, pp. 35-50, January
2019.

S. Jung, T. Park, S. Chun, and T. Moon, ``Re-weighting based group fairness regularization
via classwise robust optimization,'' arXiv preprint arXiv:2303.00442, 2023.

E. Krasanakis, E. Spyromitros-Xioufis, S. Papadopoulos, and Y. Kompatsiaris, ``Adaptive
sensitive reweighting to mitigate bias in fairness-aware classification,'' Proc. of
the 2018 World Wide Web Conference, pp. 853-862, April 2018.

T. Calders and S. Verwer, ``Three naive Bayes approaches for discrimination-free classification,''
Data Mining and Knowledge Discovery, vol. 21, no. 2, pp. 277-292, July 2010.

C. Dwork, N. Cynthia, A. T. Kalai, and M. Leiserson, ``Decoupled classifiers for group-fair
and efficient machine learning,'' Proc. of Conference on Fairness, Accountability
and Transparency, PMLR, pp. 119-133, January 2018.

R. Jiang, A. Pacchiano, T. Stepleton, H. Jiang, and S. Chiappa, ``Wasserstein fair
classification,'' Uncertainty in Artificial Intelligence, PMLR, pp. 862-872, August
2019.

F. Kamiran and T. Calders, ``Classification with no discrimination by preferential
sampling,'' Proc. of 19th Machine Learning Conference in Belgium and The Netherlands,
vol. 1. no. 6, May 2010.

N. Mehrabi, U. Gupta, F. Morstatter, G. V. Steeg, and A. Galstyan, ``Attributing fair
decisions with attention interventions,'' arXiv preprint arXiv:2109.03952, 2021.

https://github.com/Trusted-AI/AIF360/blob/master/aif360/algorithms/inprocessing/art_classifier.py

L. E. Celis, L. Huang, V. Keswani, and N. K. Vishnoi, ``Classification with fairness
constraints: A meta-algorithm with provable guarantees,'' Proc. of the Conference
on Fairness, Accountability, and Transparency, pp. 319-328, January 2019.

B. H. Zhang, B. Lemoine, and M. Mitchell, ``Mitigating unwanted biases with adversarial
learning,'' Proc. of of the 2018 AAAI/ACM Conference on AI, Ethics, and Society, New
Orleans, USA, February 2018.

M. Kearns, S. Neel, A. Roth, and Z. S. Wu, ``Preventing fairness gerrymandering: Auditing
and learning for subgroup fairness,'' Proc. of International Conference on Machine
Learning, PMLR, pp. 2564-2572, July 218.

A. Agarwal, A. Beygelzimer, M. Dudik, J. Langford, and H. Wallach, ``A reductions
approach to fair classification,'' Proc. of International Conference on Machine Learning,
2018.

N. Mehrabi, F. Morstatter, N. Peng, and A. Galstyan, ``Debiasing community detection,''
Proc. of the 2019 IEEE/ACM International Conference on Advances in Social Networks
Analysis and Mining, ACM, August 2019.

A. J. Bose and W. L. Hamilton, ``Compositional fairness constraints for graph embeddings,''
Proc. of International Conference on Machine Learning, PMLR, pp. 715-724, May. 2019.

A. Backurs, P. Indyk, K. Onak, B. Schieber, A. Vakilian, and T. Wagner, ``Scalable
fair clustering,'' Proc. of International Conference on Machine Learning, PMLR, pp.
405-413, May 2019.

X. Chen, B. Fain, L. Lyu, and K. Munagala, ``Proportionally fair clustering,'' Proc.
of International Conference on Machine Learning, PMLR, pp. 1032-1041, May 2019.

R. Berk, H. Heidari, S. Jabbari, M. Joseph, M. Kearns, J. Morhenstern, S. Neel, and
A. Roth, ``A convex framework for fair regression,'' arXiv preprint arXiv:1706.02409,
2017.

A. Agarwal, M. Dudík, and Z. S. Wu, ``Fair regression: Quantitative definitions and
reduction-based algorithms,'' Proc. of International Conference on Machine Learning,
PMLR, pp. 120-129, May 2019.

S. Aghaei, M. J. Azizi, and P. Vayanos, ``Learning optimal and fair decision trees
for non-discriminative decision-making,'' Proc. of the AAAI Conference on Artificial
Intelligence, Association for the Advancement of Artificial Intelligence (AAAI), vol.
33, pp. 1418-1426, July 2019.

H. Zhao, ``Costs and benefits of fair regression,'' arXiv preprint arXiv:2106.08812,
2021.

E. Chzhen, C. Denis, M. Hebiri, L. Oneto, and M. Pontil, ``Fair regression via plug-in
estimator and recalibration with statistical guarantees,'' Advances in Neural Information
Processing Systems, vol. 33, pp. 19137-19148, 2020.

E. Chzhen, C. Denis, M. Hebiri, L. Oneto, and M. Pontil, ``Fair regression with wasserstein
barycenters,'' Advances in Neural Information Processing Systems, vol. 33, pp. 7321-7331,
2020.

S. Samadi, U. Tantipongpipat, J. Morgenstern, M. Singh, and S. Vempala, ``The price
of fair PCA: One extra dimension,'' Advances in Neural Information Processing Systems,
vol. 31, 2018.

J. Lee, G. Kim, M. Olfat, M. Hasegawa-Johnson, and C. D. Yoo, ``Fast and efficient
MMD-based fair PCA via optimization over Stiefel manifold,'' Proc. of the AAAI Conference
on Artificial Intelligence, vol. 36, no. 7, pp. 7363-7371, June 2022.

T. Kamishima, S. Akaho, H. Asoh, and J. Sakuma, ``Fairness-aware classifier with prejudice
remover regularizer,'' Machine Learning and Knowledge Discovery in Databases, pp.
35-50, 2012.

D. Kenna, ``Using adversarial debiasing to remove bias from word embeddings,'' arXiv
preprint arXiv:2107.10251, 2021.

S. Chiappa and W. S. Isaac, ``A causal Bayesian networks viewpoint on fairness,''
Privacy and Identity Management: Fairness, Accountability, and Transparency in the
Age of Big Data, Springer International Publishing, pp. 3-20, 2019.

J. R. Loftus, C. Russell, M. J. Kusner, and R. Silva, ``Causal reasoning for algorithmic
fairness,'' arXiv preprint arXiv:1805.05859, 2018.

L. Zhang, Y. Wu, and X. Wu, ``A causal framework for discovering and removing direct
and indirect discrimination,'' arXiv preprint arXiv:1611.07509, 2016.

C. Louizos, K. Swersky, Y. Li, M. Welling, and R. Zemel, ``The variational fair autoencoder,''
arXiv preprint arXiv:1511.00830, 2015.

D. Moyer, S. Gao, R. Brekelmans, G. V. Steeg, and A. Galstyan, ``Invariant representations
without adversarial training,'' Advances in Neural Information Processing Systems,
vol. 31, 2018.

A. Amini, A. P. Soleimany, W. Schwarting, S. N. Bhatia, and D. Rus, ``Uncovering and
mitigating algorithmic bias through learned latent structure,'' Proc. of the 2019
AAAI/ACM Conference on AI, Ethics, and Society, ACM, January 2019.

B. H. Zhang, B. Lemoine, and M. Mitchell, ``Mitigating unwanted biases with adversarial
learning,'' Proc. of the 2018 AAAI/ACM Conference on AI, Ethics, and Society, ACM,
December 2018.

D. Xu, S. Yuan, L. Zhang, and X. Wu, ``FairGAN: Fairness-aware generative adversarial
networks,'' Proc. of 2018 IEEE International Conference on Big Data (Big Data), IEEE,
December 2018.

T. Bolukbasi, K.-W. Chang, J. Zou, V. Saligrama, and A. Kalai, ``Man is to computer
programmer as woman is to homemaker? Debiasing word embeddings,'' Advances in Neural
Information Processing Systems, vol. 29, 2016.

J. Zhao, T. Wang, M. Yatskar, R. Cotterell, V. Ordonez, and K.-W. Chang, ``Gender
bias in contextualized word embeddings,'' arXiv preprint arXiv:1904.03310, 2019.

J. Chen, Y. Wang, and T. Lan, ``Bringing fairness to actor-critic reinforcement learning
for network utility optimization,'' Proc. of IEEE INFOCOM 2021 - IEEE Conference on
Computer Communications, IEEE, May. 2021.

M. Hardt, E. Price, and N. Srebro, ``Equality of opportunity in supervised learning,''
Proc. of Conference on Neural Information Processing Systems, 2016.

G. Pleiss, M. Raghavan, F. Wu, J. Kleinberg, and K. Q. Weinberger, ``On fairness and
calibration,'' Proc. of Conference on Neural Information Processing Systems, 2017

F. Kamiran, A. Karim, and X. Zhang, ``Decision theory for discrimination-aware classification,''
Proc. of IEEE International Conference on Data Mining, 2012.

A. Agarwal, A. Beygelzimer, M. Dudík, J. Langford, and H. M. Wallach, ``A reductions
approach to fair classification,'' Proc. of Machine Learning Research, PMLR, vol.
80, pp. 60-69, 2018.

A. Agarwal, M. Dudík, and Z. S. Wu,, ``Fair regression: Quantitative definitions and
reduction-based algorithms,'' Proc. of Machine Learning Research, PMLR, vol. 97, pp.
120-129, 2019.

IBM Trusted-AI, ``AI Fairness 360 (AIF360),'' https://github.com/Trusted-AI/AIF360

R. K. E. BellamyK. Dey, M. Hind, et al., ``AI fairness 360: An extensible toolkit
for detecting, understanding, and mitigating unwanted algorithmic bias,'' arXiv preprint
arXiv.1810.01943, 2018.

S. Bird, M. Dudík, R. Edgar, B. Horn, R. Luts, V. Milan, M. Sameki, H. Wallach, and
K. Walker, ``Fairlearn: A toolkit for assessing and improving fairness in AI,'' Microsoft,
Tech. Rep. MSR-TR-2020-32, May 2020.

Microsoft, ``Fairlearn,'' https://github.com/fairlearn/fairlearn

https://fairlearn.org/

Google Tensorflow, ``Fairness Indicators,'' https://github.com/tensorflow/fairness-indicators

http://www.aitimes.com/news/articleView.html?idxno=143283

P. Saleiro, B. Kuester, A. Stevens, A. Anisfeld, L. Hinkson, J. London, and R. Ghani,
``Aequitas: A bias and fairness audit toolkit,'' arXiv preprint arXiv:1811.05577,
2018.

Data Science for Social Good, ``Aequitas,'' https://github.com/dssg/aequitas

S. Verma and J. Rubin, ``Fairness definitions explained,'' Proc. of the International
Workshop on Software Fairness, 2018.

S. Barocas, M. Hardt, and A. Narayanan, Fairness and Machine Learning: Limitations
and Opportunities, MIT Press, 2023.

Author
Hanbit Gil is a student at Korea University of Technology and Education, pursuing
a Master's degree in information and communication technology. Her current research
interests revolve around computer vision and machine learning.
Yejin Shin is a full-time principal research fellow in the AI Trustworthiness Policy
& Research Team at Telecommunications Technology Association (TTA). She earned her
doctorate in Information and Communication Engineering from the Korea University of
Technology and Education in 2020. She was a visiting researcher at the University
of Edinburgh from 2017 to 2018. She is currently working on research including AI
trustworthiness and ethics.
Joon Ho Kwak received a master's degree in electronic engineering from Seoul National
University, Seoul, Korea. He is a team manager with the AI Trustworthiness Policy
& Research Team, Telecommunications Technology Association, and is specialized in
testing and certification of software systems, software safety engineering. He is
currently involved in the research project of developing requirements and verification/validation
methodology for trustworthy AI, contributing to a risk-based framework of assuring
trustworthiness of AI especially.
Sungmin Woo received his B.S. degree in electrical engineering from Stony Brook
University, Stony Brook, NY, USA, in 2006, an M.S. degree from the Pohang University
of Science and Technology, Korea, in 2008, and a Ph.D. degree from Korea University,
Korea, in 2020. From 2008 to 2020, he participated in research and development on
mobile camera systems at the LG Electronics' Mobile Communication Division. He is
currently working as an Assistant Professor with the School of Electrical, Electronics
and Communication Engineering, Korea University of Technology and Education. His current
research interests include trustworthy AI, color constancy, image & video processing,
computer vision, and machine learning.