Mobile QR Code QR CODE

2024

Acceptance Ratio

21%


  1. ( School of Electrical, Electronics and Communication Engineering, Koreatech / 1600, Chungjeol-ro, Byeongcheon-myeon, Dongnam-gu, Cheonan-si, Chungcheongnam-do 31253, Korea {hb35686, innosm}@koreatech.ac.kr)
  2. ( AI Trustworthiness Policy & Research Team, TTA, Korea {yepp1252, pentarous}@tta.or.kr)



Fairness, Bias, Diversity, Discrimination, Equality

1. Introduction

The release of OpenAI's ChatGPT has sparked significant interest in the performance and proliferation of artificial intelligence. ChatGPT, a representative hyperscale artificial neural network, interacts with users conversationally, answers follow-up questions, and occasionally acknowledges its own mistakes, making interactions feel remarkably human-like. Moreover, it can quickly learn programming languages and write essays at a human-like level.

However, concerns have emerged, particularly in the United States, about ChatGPT's potential impact on education and academia, leading the New York Department of Education to announce an official ban on its access in December 2022. Research papers published by OpenAI, the creators of ChatGPT and the underlying GPT-3 model, have highlighted issues related to potential misuse, bias, and fairness. These concerns include the possibility of using the model for spreading false information, spam, phishing, and writing fake essays, as well as its exploitation by malicious actors. Bias in the training data can lead to the generation of text with stereotypes and prejudices. For instance, GPT-3's authors reported that 83% of the 388 occupations generated by the model were male-dominated, with Asian races receiving mainly positive scores while negative scores were higher for black individuals. Additionally, violent and terrorism-related words were more frequently associated with Islam than with other religions.

In March 2024, the European Union enacted the world's first artificial intelligence regulation law, marking a significant milestone as the first enforceable legislation of its kind. This development underscores the growing awareness and accountability among academia, international organizations, and nations worldwide regarding the societal risks associated with AI. However, while these AI regulations are essential, they may potentially slow down AI development and affect competitive standing. Therefore, it is imperative to analyze and ensure trustworthy AI systems, which encompass factors such as fairness, transparency, safety, and other factors to effectively balance innovation and regulation.

In this paper, we investigate the definitions of fairness and bias in the standards and regulations of international organizations. We review academic approaches and verification tools for evaluating the fairness of artificial intelligence, particularly in the field of machine learning. Our aim is to predict the current level of fairness in machine learning, understand the progression of standardization and regulation, and provide guidance on the direction of fairness technology for industry experts.

Table 1. Summary of bias definitions from international standards organizations and regulations.

Organization

Definition

ISO

(ISO/IEC TR24027) [2]

Common categories of negative impacts that might be perceived as “unfair” are listed below. Unfair allocation / Unfair quality of service / Stereotyping / Denigration / “Over" or "under" representation

European Commission

(Ethics guidelines for trustworthy AI) [8]

The substantive dimension implies a commitment to: ensuring equal and just distribution of both benefits and costs, and ensuring that individuals and groups are free from unfair bias, discrimination and stigmatization. The procedural dimension of fairness entails the ability to contest and seek effective redress against decisions made by AI systems and by the humans operating them.

OECD

(Recommendation of the Council on Artificial Intelligence) [4]

AI actors should respect the rule of law, human rights and democratic values, throughout the AI system lifecycle.

UNESCO

(Recommendation on the ethics of artificial intelligence) [5]

AI actors should promote social justice and safeguard fairness and non-discrimination of any kind in compliance with international law.

This implies an inclusive approach to ensuring that the benefits of AI technologies are available and accessible to all, taking into consideration the specific needs of different age groups, cultural systems, different language groups, persons with disabilities, girls and women, and disadvantaged, marginalized and vulnerable people or people in vulnerable situations.

NIST

(AI Risk Management Framework: AI RMF 1.0) [6]

Absence of harmful bias is a necessary condition for fairness.

The Alan Turing Institute in UK (Understanding artificial intelligence ethics and safety) [7]

All Al systems that process social or demographic data pertaining to features of human subjects must be designed to meet a minimum threshold of discriminatory non-harm.

Australia

(Australia's AI Ethics Principles) [9]

AI systems should be inclusive and accessible, and should not involve or result in unfair discrimination against individuals, communities or groups.

Ministry of Science and ICT

(Strategy to realize trustworthy artificial intelligence) [10]

Functionality that prevents the generation of conclusions that include discrimination or bias towards specific groups during the processing of data by artificial intelligence

Major corporations

(Google, IBM, Microsoft, Facebook, NAVER, Kakao)

Google: Avoid creating or reinforcing unfair bias [11].

IBM: Fairness refers to the equitable treatment of individuals, or groups of individuals, by an AI system [12].

Microsoft: AI systems should treat all people fairly [13].

Facebook: At Facebook, we believe that our products should treat everyone fairly and work equally well for all people [14].

Naver: Naver will develop and utilize AI while considering the value of diversity, in order to avoid any unfair discrimination towards all individuals including users [15].

Kakao: We are cautious to ensure that intentional social discrimination does not occur in algorithmic results [16].

2. Definition of fairness and bias

2.1 Definition of fairness and bias in standards and regulations

Standardization bodies such as ISO/IEC and international organizations such as OECD and European Commission provide documents and guidelines on the trustworthiness of artificial intelligence. ISO/IEC TR 24028:2020 [1] includes transparency, controllability, robustness, recoverability, fairness, stability, personal data protection, and security as considerations necessary for ensuring trustworthiness, and definitions and related content regarding fairness can be found in ISO/IEC TR 24028:2020, ISO/IEC TR 24027 [2], and ISO/IEC 23894:2023 [3].

In [1], bias is described as one of the factors that can undermine the trustworthiness of artificial intelligence systems, defined as ``Favoritism towards some things, people or groups over others.'' [2] explains the causes of unfairness as follows: ``Bias in objective functions, imbalanced data sets, and human biases in training data and in providing feedback to systems, bias issue in the product concept, the problem formulation or choices about when and where to deploy AI systems.'' [3] defines bias in AI systems as ``a systematic difference in treatment of certain objects, people, or groups in comparison to others'', and describes fairness as ``a concept that is distinct from, but related to bias.'' When referring to the fairness of artificial intelligence systems, OECD's ``Recommendation of the Council on Artificial Intelligence'' [4] and UNESCO's ``Recommendation on the ethics of artificial intelligence [5]'' emphasize the need to ensure fairness of the system by striving to prevent discriminatory biased practices and outcomes throughout the entire life cycle.

The US National Institute of Standards and Technology (NIST) explains fairness as one of the fundamental principles related to the risks of artificial intelligence as follows: ``Fairness in AI includes concerns for equality and equity by addressing issues such as harmful bias and discrimination. Standards of fairness can be complex and difficult to define because perceptions of fairness differ among cultures and may shift depending on application [6].'' The Alan Turing Institute in the UK has provided guidelines for applying AI ethics and safety principles to the design and implementation of algorithmic systems in the public sector in their publication ``Understanding artificial intelligence ethics and safety.'' Here, fairness is described as a principle for responsible design and use of AI technology, stating that ``All AI systems that process social or demographic data pertaining to features of human subjects must be designed to meet a minimum threshold of discriminatory non-harm [7].''

2.2 Fairness and bias in academia

2.2.1 Bias in artificial intelligent systems

Most artificial intelligence systems and algorithms are based on data and operate by training on this data. Therefore, data is closely related to the operation of artificial intelligence systems. If there is bias in the data used for training, the algorithm learned from that data can generate biased results or amplify the bias in the data. Additionally, even when the data is unbiased, a biased behavior can be exhibited due to specific algorithmic design choices. The results of biased algorithms can affect user decisions and ultimately lead to a vicious cycle of generating biased data for future algorithm training. For example, with web search engines that place certain results at the top of the list, users tend to interact more with the top results and pay less attention to the results at the bottom of the list. User-item interactions collected by the web search engine are then used to make decisions about how to display information based on popularity and user interest. This can create bias towards the top results not because of the characteristics of the results themselves, but due to biased interactions and list placement. The types of bias in [17] are divided into three categories: data, algorithm, and user interaction, defined as interacting loops. Fig. 1. illustrates the cause-and-effect relationships among these factors and categorizes them accordingly. Table Table 2 summarizes the types and causes of biases in a table. For more detailed information, refer to [17].

Fig. 1. Inherent bias in the feedback loop among data, algorithm, and users.

../../Resources/ieie/IEIESPC.2025.14.3.352/image1.png

Table 2. Summary of three types and causes of biases.

Categories

Bias

Causes of occurrence

Data to Algorithm

Measurement

Occurs in the method of selecting, utilizing, and measuring features

Omitted variable Bias

When one or more important variables are excluded from the model

Representation Bias

Occurs in the method of sampling from a population

Aggregation bias

when drawing incorrect conclusions about individuals by observing the entire population

Simpson's Paradox

A type of aggregate bias that occurs in the analysis of heterogeneous data

Sampling Bias

Similar to selection bias, it occurs due to non-random sampling of subgroups

Longitudinal Data Fallacy

In cross-sectional analysis, combining diverse groups at a single point in time can lead to bias, as heterogeneous groups can lead to different conclusions

Linking Bias

when the social connection properties obtained from user connections, activities, or interactions differ from actual user behavior.

Algorithm to User

Algorithmic Bias

When bias is purely added by the algorithm and not present in the input data

User Interaction Bias

biased behavior or interactions based on user interfaces and user choices

Presentation Bias

Bias due to the way information is represented. For example, on the web, users can only click on the visible content, so clicks occur on the displayed content and not on any other content

Ranking Bias

Top-ranking results are thought to be the most relevant and important, so they induce more clicks. This bias affects search engines and crowdsourcing applications.

Popularity Bias

More popular items tend to be displayed more frequently. Popularity measures can be manipulated. This type of bias can be seen in recommendation systems, search engines, or other systems where popular items are presented more to the public.

Emergent Bias

This refers to bias that emerges as a result of interaction with actual users, and typically occurs after a certain period of time has elapsed following the completion of the design, due to changes in population/cultural values or social knowledge.

Evaluation Bias

Bias that occurs during model evaluation, such as the use of imbalanced benchmarks.

User to Data

Historical Bias

an existing bias that is a socio-technological issue, where search results for "female CEO" are influenced by the fact that only 5% of Fortune 500 CEOs are women.

Population Bias

when statistics, population distribution, representative characteristics, or user characteristics differ from the platform's user distribution.

Self-Selection Bias

when research subjects are self-selected, such as passionate supporters being more likely to participate in a poll.

Social Bias

when others' behavior influences our judgment, such as when we want to rate or review an item poorly but are influenced by others' high ratings.

Behavioral Bias

when user behavior differs depending on the platform, context, or dataset.

Temporal Bias

Temporal bias arises from differences in population and behavior over time, such as when people start using hashtags to attract attention to a topic and then continue discussing the event without using hashtags.

Content Protection Bias

structural, lexical, semantic, and syntactical differences in user-generated content, such as differences in language use based on gender, age, country, and population.

2.2.2 Academic definition of fairness

The term ``fairness'' in machine learning denotes the multiple efforts made to rectify any form of algorithmic bias in automated decision-making processes that rely on machine learning models. Algorithmic fairness has begun to garner the attention of researchers in the artificial intelligence, software engineering, and legal communities, with more than twenty different notions of fairness proposed in the past few years. There is no clear consensus on which definition to apply in each situation, and the detailed differences between multiple definitions are difficult to comprehend. However, most statistical metrics in machine learning rely on the concepts of true positive (TP), true negative (TN), false positive (FP), and false negative (FN) based on predicted and actual class values commonly used in machine learning. [90] collected and theoretically interpreted definitions of fairness for algorithmic classification problems, and explained them through a unified case study. [91] organized different academic definitions of fairness from the perspectives of independence, separation, and sufficiency. Table Table 3 below summarizes the academic definitions of fairness.

Table 3. Summary of academic definitions of fairness.

Fairness definitions

Explanations

Equalized Odds

A predictor $\hat{Y}$ satisfies equalized odds with respect to protected attribute ${A}$ and outcome ${Y}$, if $\hat{Y}$ and ${A}$ are independent conditional on ${Y}$.

$ P(\hat{Y}=1\mid A=0,~Y=y)\\ =PA(\hat{Y} =1 \mid A=1, ~Y=y),~y\in \{0,~1\} $

Equal opportunity

A binary predictor $\hat{Y}$satisfies equal opportunity with respect to ${A}$ and ${Y}$ if

$\begin{aligned} P(\hat{Y}=1\mid A=0,~Y=1)\\ =P(\hat{Y}=1 \mid A=1,~Y=1). \end{aligned}$

This means that the probability of a person in the positive class being classified as positive should be the same for both the protected and unprotected groups. In other words, both groups should have the same true positive rates.

Demographic Parity

(statistical parity)

A predictor $\hat{Y}$ satisfies demographic parity if $P(\hat{Y}\mid A=0)=P(\hat{Y} \mid A=1)$ . It means that the likelihood of a positive outcome should be the same regardless of membership in a protected group (e.g. female).

Fairness Through Awareness

n algorithm is fair if it gives similar predictions to similar individuals. If there is no difference in similarity measurement (such as inverse distance) for a specific task, the results should also be similar.

Fairness Through Unawareness

An algorithm is fair as long as any protected attributes A are not explicitly used in the decision-making process. Fairness is achieved if the protected attribute (e.g. race in loan decisions) is not explicitly used in the decision-making process.

Treatment Equality

Treatment equality is achieved when the ratio of false negatives and false positives is the same for both protected group categories. the false negative and false positive rates should be the same for both protected groups (male and female).

Test Fairness

$S = S(x)$ is test fair (well-calibrated) if it reflects the same likelihood of recidivism irrespective of the individual's group membership, $R$. That is, if for all values of $s$,

$\begin{aligned} P(\hat{Y}=1\mid S=s,~ R=b)\\ =P(\hat{Y}=1\mid S=s,~ R=w). \end{aligned}$

For all risk scores, the probability of individuals from protected and unprotected groups being classified as positive class should be the same.

Counterfactual Fairness

A predictor $\hat{Y}$ is counterfactually fair if under any context $X=x$ and $A=a$

$\begin{aligned} P({\hat{Y}}_{A\leftarrow a(U)}=y\mid X=x,~A=a)\\ =P({\hat{Y}}_{A\leftarrow a^{'}(U)}=y\mid X=x,~A=a) \end{aligned}$

for all $y$ and for any value attainable by A. Individuals should have the same outcomes in the real world and the counterfactual world (involving membership in a different racial group).

Fairness in Relational Domains

A fairness concept that captures the relational structure of a domain by considering not only individual attributes but also social, organizational, and other connections between individuals.

Conditional Statistical Parity

For a set of legitimate factors $L$, predictor $\hat{Y}$ satisfies conditional statistical parity, if

$\begin{aligned} P(\hat{Y}\mid L=1,~A=0)=P(\hat{Y}\mid L=1,~ A=1).\end{aligned}$

Under controlled legal risk factors, an equal proportion of defendants from each racial group should be detained. For example, Black and White defendants who receive the same number of guilty verdicts should be detained at the same rate.

As seen in the table, satisfying multiple fairness simultaneously seems impossible without specific constraints. Therefore, when considering fairness, it is important to consider the context and situation where the fairness definition is used and apply it accordingly, while also analyzing the potential impact on individuals or groups over time. Furthermore, when addressing fairness-related issues, it is necessary to consider the causes and types of bias and satisfy the fairness criteria mentioned above by taking them into account.

3. Bias mitigation techniques

In machine learning, there are various methods to mitigate bias, and they differ depending on individual academic fields. However, these methods can be categorized into pre-processing methods that offset bias in the data, correction methods that are integrated with individual learning algorithms, and post-processing methods.

3.1 Pre-processing

The creation of all datasets is the result of various design decisions made by the data creator (curator). These decisions have an impact on the fairness of the dataset and ultimately affect the results of the algorithm. Best practices have been proposed to mitigate data bias, such as creating a datasheet containing information on the dataset's creation method, features, motivations, and skewed distribution of the data [18,19]. Labeling, such as nutrition labels on food, has also been suggested for better data classification [20]. Studies have been conducted to propose methods for testing Simpson's paradox [21] or for automatically detecting bias [22,23]. Causal models and causal graphs are also used for tasks that involve detecting direct discrimination in the data [24] and preventing techniques to modify the data [25]. Furthermore, research has been proposed to prevent discrimination in data mining [26], including messaging [27], preferential sampling [28,29], and disparate impact removal [30], which are other preprocessing approaches aimed at removing bias from data. [31] suggests a fair data envelopment analysis method, incorporating an additional constraint to mitigate disparate impact, thereby addressing the scoring parity between privileged and unprivileged groups. [32] formulates fairness as an optimization problem, obscuring information about protected group members to find a good representation for group and individual fairness during data encoding. [33] presents optimized preprocessing, which uses a probabilistic transformation to modify data features and labels with fairness and accuracy constraints for group fairness, individual distortion, and data fidelity. [34] performs a theoretical study on the trade-off relationship between accuracy and discrimination and proposes four methods using non-discriminatory constraints in classifier design. [35] applies a linear transformation on the non the non-sensitive feature columns to eliminate their correlation with the sensitive feature columns while preserving the maximum amount of information possible.

3.2 In-Processing

Classification is a common task in machine learning used in various domains where human contact is not feasible, and techniques have been proposed to minimize cost while meeting certain definitions of fairness to achieve fair classification [36]. These techniques include adding constraints for fairness to the cost function through regularization [37,38,39], optimizing the estimation probability of sensitive and non-sensitive features to be equal [40], modifying the Naive Bayes classifier [41], maximizing the average accuracy of each group [42], mitigating dependency on sensitive features using Wasserstein distance measure [43], using Preferential Sampling (PS) for discrimination-free training data [44], and incorporating attention mechanism for interpretable and bias-mitigated classification [45]. The ART classifier is used as a classifier to satisfy fairness such as demographic parity and equalized odds [46]. [47] is a meta-algorithm that modifies the classification threshold of the base classifier to maximize a fairness metric while maintaining a certain level of accuracy. Adversarial debiasing [48] trains a classifier to improve prediction accuracy while simultaneously minimizing the adversary's ability to infer the protected attribute from the predictions. [49] proposes a group fairness classifier aiming to avoid the susceptibility of intentional or inadvertent ``fairness gerrymandering'' that can occur with traditional statistical definitions of fairness that fix a small collection of pre-defined groups. Exponentiated Gradient Reduction [50] reduce fair classification to a sequence of cost-sensitive classification problems, resulting in a randomized classifier with the lowest empirical error while adhering to fair classification constraints.

In community detection, users with low connectivity are often ignored, which can lead to improper classification into communities. However, CLAN [51] first identifies important communities and then uses a method of classifying a small number of users into important communities, thus improving the community classification performance of these minority users. In addition, methods for community detection using graph embedding [52] or clustering [53,54] have also been proposed.

In regression problems, [55] proposed to define the loss (cost) function by dividing individuals, groups, and individuals/groups rather than using regression methods for the entire dataset in order to minimize it. Methods for modeling demographic and group-specific fairness as loss functions [56], and using decision trees that can be applied to classification and regression problems have also been proposed [57]. [58] presented an algorithm-independent lower bound by characterizing the trade-off between the inherent fairness and accuracy in regression problems. [59] proposed a two-step approach where an unconstrained regression is first estimated from labeled data and then recalibrated with unlabeled data. [60] examines the problem of achieving demographic parity in real-valued function learning by establishing a connection between fair regression and optimal transport theory, and proposes a post-processing algorithm to achieve fairness with established guarantees.

Fair PCA [61] was proposed to address the issue of amplifying the reconstruction error of one group in original PCA when multiple groups of equal size exist in the data, and it aims to achieve similar representation across different groups. To avoid the problem of inadequate representation of group data other than the lowest principal component during dimensionality reduction of the entire dataset, this method defines a loss function for each group and only performs PCA for the group with the highest loss function. [62] formulated the maximum mean discrepancy (MMD)-based fair PCA as a constrained optimization over the Stiefel manifold and solved it using Riemannian Exact Penalty Method via Smoothing.

In addition to classification and regression, studies aimed at mitigating bias have been proposed in the literature. Prejudice remover [63] modifies the learning objective by adding a regularization term that is sensitive to discrimination. [64] exploits adversarial learning for mitigating unwanted biases in work embeddings.

Causal models are used to eliminate unwanted dependencies on sensitive attributes such as gender or race by using causal graphs to confirm causal relationships between variables [65,66]. In [67], a method is proposed for detecting direct/indirect discrimination from historical data using causal graphs and removing discrimination effects before predictive analysis. First, path-specific effects (direct/indirect) are defined as below and discrimination thresholds for protected groups are set. For example, the UK law on gender discrimination in 1975 allowed a 5% difference. Similarly, thresholds for indirect discrimination for Redlining properties can be set. From this, direct and indirect discrimination for all paths in the causal graph are determined, and a objective function is created and constraints for mitigating discrimination are set to modify the causal graph through quadratic programming.

There is research on autoencoders that learn fair representations and avoid interference of sensitive attributes [68]. In this research, the goal is to obtain fair representations by removing information about sensitive variables. A maximum mean discrepancy regularizer is used to ensure that the posterior distribution of the latent variables is the same. In [69], the optimization objective function is designed to include the mutual information between the encoding and sensitive variables, which helps to remove biases in the representation. Debiased variational autoencoder (DB-VAE) is used to reduce the accuracy gap by training the model to learn sensitive attribute variables such as skin color and gender [70]. This method modifies the traditional variational autoencoder (VAE) by incorporating the learning of classes corresponding to the samples and representing the frequency of occurrence of specific classes as a histogram. The sample occurrence probability is defined to be inversely proportional to the frequency, ensuring that during training, the samples do not converge to a single class. Adaptive re-sampling is employed to prevent the network from biasing towards a particular class as the learning progresses.

Fig. 2. Debiasing variational autoencoder (DB-VAE) employs resampling during training, where the sample selection is inversely proportional to its frequency of appearance, to mitigate data bias [70].

../../Resources/ieie/IEIESPC.2025.14.3.352/image2.png

To mitigate biases from learned models with fixed conceptual associations, adversarial debiasing methods were proposed. The goal here is to enhance prediction accuracy while simultaneously reducing the predictions made by an adversary regarding protected and sensitive variables [71]. Instead of removing dependency on discriminatory data, FairGAN [72] proposed a method to generate data that is similar to real data but without discrimination. The architecture of FairGAN consists of one generator and two discriminators. The generator generates fake data from protective attributes and noise, while the discriminator plays a role in distinguishing real data from fake data. The third component is for the fairness constraint, determining whether generated (fake) data is created from a protected attribute group.

In the research of word embedding which represents a word to a vector for natural language processing, [73] proposed the bias relationships such as ``man to computer programmer'' and ``woman to homemaker.'' This method defined the gender-specific and gender-neutral words, and re-embedded the latter so that gender relationship in the pair is removed. Recently, [74] proposed a method to analyze and mitigate the biases of contextualized vectors.

Recent advancements in reinforcement learning have enabled automated online decision-making in communication network design and optimization problems. However, most reinforcement learning approaches prioritize maximizing agent rewards without considering fairness. Recent research proposes introducing a fairness term in actor-critic reinforcement learning, adjusting rewards multiplicatively [75]. In this study, weighted multiplicative-adjusted rewards are given to the Actor instead of traditional rewards, addressing fairness concerns in network utility optimization.

3.3 Post-Processing

Post-processing techniques represent a subset of methods for mitigating fairness issues in machine learning. These methods are applied subsequent to the model making predictions, with the goal of adjusting the model's outputs to ensure fairness through various means.

The ``Equalized Odds Postprocessing'' technique modifies predicted probabilities for distinct groups to meet the equalized odds criterion [76,77]. Its objective is to guarantee that the odds of accurate classification are equivalent across different demographic or sensitive attribute groups. Expanding on the equalized odds concept, the ``Calibrated Equality Odds'' method involves calibrating the model's predicted probabilities and subsequently applying equalized odds postprocessing [77]. The aim is to concurrently achieve calibration and fairness. The ``Reject Option Classification'' approach [78] introduces a ``reject'' option, where the model refrains from making predictions in situations of high uncertainty. This strategy seeks to alleviate biases by yielding more favorable outcomes to underprivileged groups and less favorable outcomes to privileged groups within a confidence band around the decision boundary [79]. In a binary classification context, [79] proposed a systematic approach for achieving fairness. The core idea is to reduce fair classification to a series of cost-sensitive classification problems. The solutions obtained yield a randomized classifier with the lowest empirical error while adhering to specified constraints. For fair regression, [80] introduced a general scheme wherein the prediction is required to be statistically independent of the protected attribute, and the prediction error for any protected group must remain below a predetermined level.

4. Tools for fairness evaluation and mitigation

Leading companies in the field of artificial intelligence are at the forefront of providing tools for assessing fairness and alleviating biases and discrimination within machine learning models. These tools are instrumental in evaluating and fine-tuning the outcomes of machine learning algorithms, leveraging the discrimination mitigation techniques discussed earlier.

IBM's AI Fairness 360 Toolkit summarized in Table Table 4 encompasses a broad range of features, including various fairness metrics and mitigation algorithms [81,82]. These tools collectively support the evaluation and mitigation of fairness issues across diverse models and datasets, aiding in the identification and resolution of biases that may manifest in model predictions as illustrated in Fig. 3. AIF360 offers more than 70 fairness metrics and 14 bias mitigation algorithms designed to measure and address the fairness of artificial intelligence algorithms. AIF360 supports bias mitigation algorithms in pre-processing, in-processing, and post-processing.

Fig. 3. The fairness pipeline of AIF360 [81].

../../Resources/ieie/IEIESPC.2025.14.3.352/image3.png

Table 4. Bias Mitigation Algorithm supported by IBM's AIF360.

Category

Algorithm

Pre-processing

Disparate Impact Remover

Learning Fair Representation

Optimized Preprocessing

Reweighing

In-processing

Adversarial Debiasing

ART Classifier

Gerry Fair Classifier

Meta Fair Classifier

Prejudice Remover

Exponentiated Gradient Reduction

Grid Search Reduction

Post-processing

Calibrated Equality of Odds

Equality Odd

Reject Option Classification

Fairlearn by Microsoft [83,84,85] is an open-source toolkit providing functions to assess fairness and algorithms to mitigate unfairness in machine learning models. For fairness evaluation, Fairlearn offers library functions based on various metrics, allowing users to select sensitive attributes and examine evaluation results based on performance and prediction differences. Additionally, a dashboard makes it easy to check whether a model is fair. Fairlearn's unfairness mitigation algorithms are categorized into pre-processing, post-processing, and in-processing. Pre-processing algorithms mitigate unfairness by transforming the dataset, while post-processing adjusts the predictions to align with specified equalized constraints. In-processing algorithms solve an optimization problem by replacing fairness-related terms with Lagrange multipliers as constraints, effectively converting the problem into a general machine learning task. As of the current version (January 4, 2024), Fairlearn is at v0.9.0 and can be considered pre-release.

Google has developed Fairness Indicators [86], which facilitate the easy calculation of commonly identified fairness metrics for binary and multiclass classifiers. Unlike existing fairness evaluation tools that may not perform well on large datasets and models, Google's Fairness Indicators allow the evaluation of fairness metrics for all scales of use cases.

KAIST's AI Fairness Research Center has developed the `MSIT AI FAIR 2022 (MAF 2022)' framework for analyzing, detecting, mitigating, and eliminating bias in AI models and training data [87]. AF 2022 boasts a total of 19 algorithms, making it more versatile than other fairness tools for various situations as summarized in Table Table 5.

Table 5. Bias Mitigation Algorithm supported by KAIST' MSIT AI FAIR 2022.

Category

Algorithm

Pre-processing

Disparate Impact Remover

Learning Fair Representation

Optimized Preprocessing

Reweighing

In-processing

Adversarial Debiasing

ART Classifier

Gerry Fair Classifier

Meta Fair Classifier

Prejudice Remover

Exponentiated Gradient Reduction

Grid Search Reduction

Kernel density estimation

Fairness VAE

Learning from fairness

Fair feature distillation

FairBatch

Post-processing

Calibrated Equalized Odds Postprocessing

Equalized Odds Postprocessing

Reject Option Classification

Aequitas, developed by the Center for Data Science and Public Policy at the University of Chicago, is an audit tool for bias and fairness [88,89]. It enables understanding various types of bias and utilizing them in development. Aequitas guides users through the process of uploading data, selecting a protected group, choosing fairness metrics, and selecting thresholds to provide a bias report.

5. Conclusion

In this paper, we examined international standards, national definitions, and definitions from leading companies related to fairness. Additionally, we explored extensive research on the increasing importance of fairness in artificial intelligence. We compared the inherent biases in machine learning and their mitigation methods from various research papers. The study delved into the vast research on fairness in artificial intelligence, considering aspects of data preprocessing, in-processing from a processing perspective, and post-processing when data and algorithms are inaccessible. Furthermore, we provided a comparative analysis of AI bias measurement methods and mitigation approaches offered by universities and companies, presenting guidelines for research on fairness in artificial intelligence. Based on this comprehensive survey, we conclude several promising directions for future research. First, it is essential to develop more sophisticated and context-aware bias mitigation techniques that can adapt to various cultural norms and application contexts. Current bias mitigation techniques are limited to addressing specific known biases or constraints, making it challenging to find universally applicable tools. Second, since biases can occur in data, algorithms, and user interactions, careful attention is required at each stage of AI system development. Enhancing the transparency and interpretability of AI models will help in more effectively identifying and mitigating biases. Finally, collaboration between academia and industry can lead to the creation of standardized guidelines and tools that are robust and universally applicable. By integrating newly developed bias mitigation tools with existing ones, we can continuously enhance scalability and effectiveness.

ACKNOWLEDGMENTS

This work was supported by the Korean MSIT (Ministry of Science and ICT) as Establishing the foundation of AI Trustworthiness (TTA) and by the Education and Research Promotion Program of KOREATECH in 2023.

REFERENCES

1 
ISO/IEC TR 24028:2020, ``Information technology -- Artificial intelligence -- Overview of trustworthiness in artificial intelligence,'' 2020.URL
2 
ISO/IEC TR 24027:2021, ``Information technology -- Artificial intelligence (AI) -- Bias in AI systems and AI aided decision making,'' 2021.URL
3 
ISO/IEC 23894:2023, ``Information technology -- Artificial intelligence -- Guidance on risk management,'' 2023.URL
4 
OECD, ``Artificial Intelligence in Society,'' June 2019.URL
5 
UNESCO, ``Recommendation on the Ethics of Artificial Intelligence,'' November 2021.URL
6 
National Institute of Standards and Technology, ``Artificial Intelligence Risk Management Framework (AI RMF 1.0),'' January 2023.URL
7 
D. Leslie, ``Understanding artificial intelligence ethics and safety: A guide for the responsible design and implementation of AI systems in the public sector,'' Zenodo, June 2019.DOI
8 
European Commission, Directorate-General for Communications Networks, Content and Technology, ``Ethics guidelines for trustworthy AI,'' Publications Office, November 2019.URL
9 
Australia, ``Australia's AI Ethics Principles,'' https://www.industry.gov.au/publications/australias-artificial-intelligence-ethics-framework/australias-ai-ethics-principlesURL
10 
Ministry of Science and ICT, ``Strategy to realize trustworthy artificial intelligence,'' May. 2021, https://www.msit.go.kr/eng/bbs/view.do?sCode=eng&mId=4&mPid=2&pageIndex=&bbsSeqNo=42&nttSeqNo=509&searchOpt=ALL&searchTxtURL
11 
Google, ``Objectives for AI applications,'' https://ai.google/responsibility/principlesURL
12 
IBM, ``Our fundamental properties for trustworthy AI,'' https://www.ibm.com/artificial-intelligence/ai-ethics-focus-areasURL
13 
Microsoft, ``Microsoft responsible AI principles,'' https://www.microsoft.com/en-us/ai/our-approach?activetab=pivot1%3aprimaryr5URL
14 
J. Pesenti, ``Facebook's five pillars of Responsible AI,'' June 2021, https://ai.facebook.com/blog/facebooks-five-pillars-of-responsible-ai/URL
15 
Naver, ``AI Ethics Principles,'' https://www.navercorp.com/value/aiCodeEthicsURL
16 
Kakao, ``AI Ethics,'' https://www.kakaocorp.com/page/responsible/detail/algorithmURL
17 
N. Mehrabi, F. Morstatter, N. Saxena, K. Lerman, and A. Galdyan, ``A survey on bias and fairness in machine learning,'' ACM Computing Surveys (CSUR), vol. 54, no. 6, pp. 1-35, 2021.DOI
18 
M. Benjamin, P. Gagnon, N Rostamzadeh, C. Pan, Y. Bengio, and A. Shee, ``Towards standardization of data licenses: The montreal data license,'' arXiv preprint arXiv:1903.12262, 2019.DOI
19 
T. Gebru, J. Morgenstern, B. Vecchione, J. W. Vaughan, H. Wallach, H. D. Ill, and K. Crawford, ``Datasheets for datasets,'' Communications of the ACM, vol. 64, no. 12, pp. 86-92, 2021DOI
20 
S. Holland, A. Hosny, S. Newman, J. Joseph, and K. Chmielinski, ``The dataset nutrition label,'' Data Protection and Privacy, vol. 12, 2020.DOI
21 
R. A. Kievit, W. E. Frankenhuis, L. J. Waldorp, and D. Borsboom, ``Simpson's paradox in psychological science: a practical guide,'' Frontiers in Psychology, vol. 4, 513, 2013.DOI
22 
N. Alipourfard, P. G. Fennell, and K. Lerman, ``Can you trust the trend?'' Proc. of the 11th ACM International Conference on Web Search and Data Mining, pp. 19-27, February 2018.DOI
23 
N. Alipourfard, P. G. Fennell, and K. Lerman, ``Using Simpson's paradox to discover interesting patterns in behavioral data,'' Proc. of 12th International AAAI Conference on Web and Social Media, June 2018.DOI
24 
L. Zhang, Y. Wu, and X. Wu, ``Achieving non-discrimination in data release,'' Proc. of the 23rd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, pp. 1335-1344, August 2017.DOI
25 
L. Zhang, Y. Wu, and X. Wu, ``Causal modeling-based discrimination discovery and removal: Criteria, bounds, and algorithms,'' IEEE Transactions on Knowledge and Data Engineering, vol. 31, no. 11, pp. 2035-2050, 2018.DOI
26 
S. Hajian and J. Domingo-Ferrer, ``A methodology for direct and indirect discrimination prevention in data mining,'' IEEE transactions on Knowledge and Data Engineering, vol. 25, no. 7, pp. 1445-1459, July 2013.DOI
27 
F. Kamiran and T. Calders, ``Classifying without discriminating,'' Proc. of 2009 2nd International Conference on Computer, Control and Communication, pp. 1-6, February 2009.DOI
28 
F. Kamiran and T. Calders, ``Classification with no discrimination by preferential sampling,'' Proc. of 19th Machine Learning Conference in Belgium and The Netherlands, vol. 1, no. 6, May 2010.URL
29 
F. Kamiran and T. Calders, ``Data preprocessing techniques for classification without discrimination,'' Knowledge and Information Systems, vol. 33, no. 1, pp. 1-33, 2012.DOI
30 
M. Feldman, S. A. Friedler, J. Moeller, C. Scheidegger, and S. Venkatasubramanian, ``Certifying and removing disparate impact,'' Proc. of the 21th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, pp. 259-268, August 2015.DOI
31 
S. Radovanovi´c, G. Savi´c, B. Deliba˘si´c, and M. Suknovi´c, ``FairDEA-removing disparate impact from efficiency scores,'' European Journal of Operational Research, vol. 301, no. 3, pp. 1088-1098, 2022.DOI
32 
R. Zemel, Y. Wu, K. Swersky, T. Pitassi, and C. Dwork, ``Learning fair representations,'' Proc. of International Conference on Machine Learning, PMLR, pp. 325-333, 2013.URL
33 
F. P. Calmon, D. Wei, B. Vinzamuri, K. N. Ramamurthy, and K. R. Varshney, ``Optimized pre-processing for discrimination prevention,'' Proc. of Conference on Neural Information Processing Systems, 2017.DOI
34 
F. Kamiran and T. Calders, ``Data preprocessing techniques for classification without discrimination,'' Knowledge and Information Systems, 2012.DOI
35 
https://github.com/fairlearn/fairlearn/blob/main/fairlearn/preprocessing/_correlation_remover.pyURL
36 
A. K. Menon and R. C. Williamson, ``The cost of fairness in binary classification,'' Proc. of Conference on Fairness, Accountability and Transparency, PMLR, pp. 107-118, January 2018.DOI
37 
T. Kamishima, S. Akaho, H. Asoh, and J. Sakuma, ``Fairness-aware classifier with prejudice remover regularizer,'' Machine Learning and Knowledge Discovery in Databases, Springer, Berlin, Heidelberg, pp. 35-50, 2012.DOI
38 
L. Oneto, M. Doninini, A. Elders, and M. Pontil, ``Taking advantage of multitask learning for fair classification,'' Proc. of Joint European Conference on Machine Learning and Knowledge Discovery in Databases, Springer, Berlin, Heidelberg, pp. 35-50, January 2019.DOI
39 
S. Jung, T. Park, S. Chun, and T. Moon, ``Re-weighting based group fairness regularization via classwise robust optimization,'' arXiv preprint arXiv:2303.00442, 2023.DOI
40 
E. Krasanakis, E. Spyromitros-Xioufis, S. Papadopoulos, and Y. Kompatsiaris, ``Adaptive sensitive reweighting to mitigate bias in fairness-aware classification,'' Proc. of the 2018 World Wide Web Conference, pp. 853-862, April 2018.DOI
41 
T. Calders and S. Verwer, ``Three naive Bayes approaches for discrimination-free classification,'' Data Mining and Knowledge Discovery, vol. 21, no. 2, pp. 277-292, July 2010.DOI
42 
C. Dwork, N. Cynthia, A. T. Kalai, and M. Leiserson, ``Decoupled classifiers for group-fair and efficient machine learning,'' Proc. of Conference on Fairness, Accountability and Transparency, PMLR, pp. 119-133, January 2018.DOI
43 
R. Jiang, A. Pacchiano, T. Stepleton, H. Jiang, and S. Chiappa, ``Wasserstein fair classification,'' Uncertainty in Artificial Intelligence, PMLR, pp. 862-872, August 2019.URL
44 
F. Kamiran and T. Calders, ``Classification with no discrimination by preferential sampling,'' Proc. of 19th Machine Learning Conference in Belgium and The Netherlands, vol. 1. no. 6, May 2010.URL
45 
N. Mehrabi, U. Gupta, F. Morstatter, G. V. Steeg, and A. Galstyan, ``Attributing fair decisions with attention interventions,'' arXiv preprint arXiv:2109.03952, 2021.DOI
46 
https://github.com/Trusted-AI/AIF360/blob/master/aif360/algorithms/inprocessing/art_classifier.pyURL
47 
L. E. Celis, L. Huang, V. Keswani, and N. K. Vishnoi, ``Classification with fairness constraints: A meta-algorithm with provable guarantees,'' Proc. of the Conference on Fairness, Accountability, and Transparency, pp. 319-328, January 2019.DOI
48 
B. H. Zhang, B. Lemoine, and M. Mitchell, ``Mitigating unwanted biases with adversarial learning,'' Proc. of of the 2018 AAAI/ACM Conference on AI, Ethics, and Society, New Orleans, USA, February 2018.DOI
49 
M. Kearns, S. Neel, A. Roth, and Z. S. Wu, ``Preventing fairness gerrymandering: Auditing and learning for subgroup fairness,'' Proc. of International Conference on Machine Learning, PMLR, pp. 2564-2572, July 218.DOI
50 
A. Agarwal, A. Beygelzimer, M. Dudik, J. Langford, and H. Wallach, ``A reductions approach to fair classification,'' Proc. of International Conference on Machine Learning, 2018.DOI
51 
N. Mehrabi, F. Morstatter, N. Peng, and A. Galstyan, ``Debiasing community detection,'' Proc. of the 2019 IEEE/ACM International Conference on Advances in Social Networks Analysis and Mining, ACM, August 2019.DOI
52 
A. J. Bose and W. L. Hamilton, ``Compositional fairness constraints for graph embeddings,'' Proc. of International Conference on Machine Learning, PMLR, pp. 715-724, May. 2019.DOI
53 
A. Backurs, P. Indyk, K. Onak, B. Schieber, A. Vakilian, and T. Wagner, ``Scalable fair clustering,'' Proc. of International Conference on Machine Learning, PMLR, pp. 405-413, May 2019.DOI
54 
X. Chen, B. Fain, L. Lyu, and K. Munagala, ``Proportionally fair clustering,'' Proc. of International Conference on Machine Learning, PMLR, pp. 1032-1041, May 2019.DOI
55 
R. Berk, H. Heidari, S. Jabbari, M. Joseph, M. Kearns, J. Morhenstern, S. Neel, and A. Roth, ``A convex framework for fair regression,'' arXiv preprint arXiv:1706.02409, 2017.DOI
56 
A. Agarwal, M. Dudík, and Z. S. Wu, ``Fair regression: Quantitative definitions and reduction-based algorithms,'' Proc. of International Conference on Machine Learning, PMLR, pp. 120-129, May 2019.DOI
57 
S. Aghaei, M. J. Azizi, and P. Vayanos, ``Learning optimal and fair decision trees for non-discriminative decision-making,'' Proc. of the AAAI Conference on Artificial Intelligence, Association for the Advancement of Artificial Intelligence (AAAI), vol. 33, pp. 1418-1426, July 2019.DOI
58 
H. Zhao, ``Costs and benefits of fair regression,'' arXiv preprint arXiv:2106.08812, 2021.DOI
59 
E. Chzhen, C. Denis, M. Hebiri, L. Oneto, and M. Pontil, ``Fair regression via plug-in estimator and recalibration with statistical guarantees,'' Advances in Neural Information Processing Systems, vol. 33, pp. 19137-19148, 2020.URL
60 
E. Chzhen, C. Denis, M. Hebiri, L. Oneto, and M. Pontil, ``Fair regression with wasserstein barycenters,'' Advances in Neural Information Processing Systems, vol. 33, pp. 7321-7331, 2020.DOI
61 
S. Samadi, U. Tantipongpipat, J. Morgenstern, M. Singh, and S. Vempala, ``The price of fair PCA: One extra dimension,'' Advances in Neural Information Processing Systems, vol. 31, 2018.DOI
62 
J. Lee, G. Kim, M. Olfat, M. Hasegawa-Johnson, and C. D. Yoo, ``Fast and efficient MMD-based fair PCA via optimization over Stiefel manifold,'' Proc. of the AAAI Conference on Artificial Intelligence, vol. 36, no. 7, pp. 7363-7371, June 2022.DOI
63 
T. Kamishima, S. Akaho, H. Asoh, and J. Sakuma, ``Fairness-aware classifier with prejudice remover regularizer,'' Machine Learning and Knowledge Discovery in Databases, pp. 35-50, 2012.DOI
64 
D. Kenna, ``Using adversarial debiasing to remove bias from word embeddings,'' arXiv preprint arXiv:2107.10251, 2021.DOI
65 
S. Chiappa and W. S. Isaac, ``A causal Bayesian networks viewpoint on fairness,'' Privacy and Identity Management: Fairness, Accountability, and Transparency in the Age of Big Data, Springer International Publishing, pp. 3-20, 2019.DOI
66 
J. R. Loftus, C. Russell, M. J. Kusner, and R. Silva, ``Causal reasoning for algorithmic fairness,'' arXiv preprint arXiv:1805.05859, 2018.DOI
67 
L. Zhang, Y. Wu, and X. Wu, ``A causal framework for discovering and removing direct and indirect discrimination,'' arXiv preprint arXiv:1611.07509, 2016.DOI
68 
C. Louizos, K. Swersky, Y. Li, M. Welling, and R. Zemel, ``The variational fair autoencoder,'' arXiv preprint arXiv:1511.00830, 2015.DOI
69 
D. Moyer, S. Gao, R. Brekelmans, G. V. Steeg, and A. Galstyan, ``Invariant representations without adversarial training,'' Advances in Neural Information Processing Systems, vol. 31, 2018.DOI
70 
A. Amini, A. P. Soleimany, W. Schwarting, S. N. Bhatia, and D. Rus, ``Uncovering and mitigating algorithmic bias through learned latent structure,'' Proc. of the 2019 AAAI/ACM Conference on AI, Ethics, and Society, ACM, January 2019.DOI
71 
B. H. Zhang, B. Lemoine, and M. Mitchell, ``Mitigating unwanted biases with adversarial learning,'' Proc. of the 2018 AAAI/ACM Conference on AI, Ethics, and Society, ACM, December 2018.DOI
72 
D. Xu, S. Yuan, L. Zhang, and X. Wu, ``FairGAN: Fairness-aware generative adversarial networks,'' Proc. of 2018 IEEE International Conference on Big Data (Big Data), IEEE, December 2018.DOI
73 
T. Bolukbasi, K.-W. Chang, J. Zou, V. Saligrama, and A. Kalai, ``Man is to computer programmer as woman is to homemaker? Debiasing word embeddings,'' Advances in Neural Information Processing Systems, vol. 29, 2016.DOI
74 
J. Zhao, T. Wang, M. Yatskar, R. Cotterell, V. Ordonez, and K.-W. Chang, ``Gender bias in contextualized word embeddings,'' arXiv preprint arXiv:1904.03310, 2019.DOI
75 
J. Chen, Y. Wang, and T. Lan, ``Bringing fairness to actor-critic reinforcement learning for network utility optimization,'' Proc. of IEEE INFOCOM 2021 - IEEE Conference on Computer Communications, IEEE, May. 2021.DOI
76 
M. Hardt, E. Price, and N. Srebro, ``Equality of opportunity in supervised learning,'' Proc. of Conference on Neural Information Processing Systems, 2016.DOI
77 
G. Pleiss, M. Raghavan, F. Wu, J. Kleinberg, and K. Q. Weinberger, ``On fairness and calibration,'' Proc. of Conference on Neural Information Processing Systems, 2017DOI
78 
F. Kamiran, A. Karim, and X. Zhang, ``Decision theory for discrimination-aware classification,'' Proc. of IEEE International Conference on Data Mining, 2012.DOI
79 
A. Agarwal, A. Beygelzimer, M. Dudík, J. Langford, and H. M. Wallach, ``A reductions approach to fair classification,'' Proc. of Machine Learning Research, PMLR, vol. 80, pp. 60-69, 2018.DOI
80 
A. Agarwal, M. Dudík, and Z. S. Wu,, ``Fair regression: Quantitative definitions and reduction-based algorithms,'' Proc. of Machine Learning Research, PMLR, vol. 97, pp. 120-129, 2019.DOI
81 
IBM Trusted-AI, ``AI Fairness 360 (AIF360),'' https://github.com/Trusted-AI/AIF360URL
82 
R. K. E. BellamyK. Dey, M. Hind, et al., ``AI fairness 360: An extensible toolkit for detecting, understanding, and mitigating unwanted algorithmic bias,'' arXiv preprint arXiv.1810.01943, 2018.DOI
83 
S. Bird, M. Dudík, R. Edgar, B. Horn, R. Luts, V. Milan, M. Sameki, H. Wallach, and K. Walker, ``Fairlearn: A toolkit for assessing and improving fairness in AI,'' Microsoft, Tech. Rep. MSR-TR-2020-32, May 2020.URL
84 
Microsoft, ``Fairlearn,'' https://github.com/fairlearn/fairlearnURL
85 
https://fairlearn.org/URL
86 
Google Tensorflow, ``Fairness Indicators,'' https://github.com/tensorflow/fairness-indicatorsURL
87 
http://www.aitimes.com/news/articleView.html?idxno=143283URL
88 
P. Saleiro, B. Kuester, A. Stevens, A. Anisfeld, L. Hinkson, J. London, and R. Ghani, ``Aequitas: A bias and fairness audit toolkit,'' arXiv preprint arXiv:1811.05577, 2018.DOI
89 
Data Science for Social Good, ``Aequitas,'' https://github.com/dssg/aequitasURL
90 
S. Verma and J. Rubin, ``Fairness definitions explained,'' Proc. of the International Workshop on Software Fairness, 2018.DOI
91 
S. Barocas, M. Hardt, and A. Narayanan, Fairness and Machine Learning: Limitations and Opportunities, MIT Press, 2023.URL

Author

Hanbit Gil
../../Resources/ieie/IEIESPC.2025.14.3.352/author1.png

Hanbit Gil is a student at Korea University of Technology and Education, pursuing a Master's degree in information and communication technology. Her current research interests revolve around computer vision and machine learning.

Yejin Shin
../../Resources/ieie/IEIESPC.2025.14.3.352/author2.png

Yejin Shin is a full-time principal research fellow in the AI Trustworthiness Policy & Research Team at Telecommunications Technology Association (TTA). She earned her doctorate in Information and Communication Engineering from the Korea University of Technology and Education in 2020. She was a visiting researcher at the University of Edinburgh from 2017 to 2018. She is currently working on research including AI trustworthiness and ethics.

Joon Ho Kwak
../../Resources/ieie/IEIESPC.2025.14.3.352/author3.png

Joon Ho Kwak received a master's degree in electronic engineering from Seoul National University, Seoul, Korea. He is a team manager with the AI Trustworthiness Policy & Research Team, Telecommunications Technology Association, and is specialized in testing and certification of software systems, software safety engineering. He is currently involved in the research project of developing requirements and verification/validation methodology for trustworthy AI, contributing to a risk-based framework of assuring trustworthiness of AI especially.

Sungmin Woo
../../Resources/ieie/IEIESPC.2025.14.3.352/author4.png

Sungmin Woo received his B.S. degree in electrical engineering from Stony Brook University, Stony Brook, NY, USA, in 2006, an M.S. degree from the Pohang University of Science and Technology, Korea, in 2008, and a Ph.D. degree from Korea University, Korea, in 2020. From 2008 to 2020, he participated in research and development on mobile camera systems at the LG Electronics' Mobile Communication Division. He is currently working as an Assistant Professor with the School of Electrical, Electronics and Communication Engineering, Korea University of Technology and Education. His current research interests include trustworthy AI, color constancy, image & video processing, computer vision, and machine learning.