|
International Journal of Computer Applications
Foundation of Computer Science (FCS), NY, USA
|
| Volume 187 - Issue 64 |
| Published: December 2025 |
| Authors: Syed Imtiazul Sami, Mohammad Rasel Mahmud, Khaled Bin Showkot Tanim, Mohammad Zubair Hussain, Mahpara Khan Orpa |
10.5120/ijca2025926064
|
Syed Imtiazul Sami, Mohammad Rasel Mahmud, Khaled Bin Showkot Tanim, Mohammad Zubair Hussain, Mahpara Khan Orpa . Advancing Fairness in Multimodal Machine Learning for Internet‑Scale Video Data: Comprehensive Bias Mitigation and Evaluation Framework. International Journal of Computer Applications. 187, 64 (December 2025), 1-9. DOI=10.5120/ijca2025926064
@article{ 10.5120/ijca2025926064,
author = { Syed Imtiazul Sami,Mohammad Rasel Mahmud,Khaled Bin Showkot Tanim,Mohammad Zubair Hussain,Mahpara Khan Orpa },
title = { Advancing Fairness in Multimodal Machine Learning for Internet‑Scale Video Data: Comprehensive Bias Mitigation and Evaluation Framework },
journal = { International Journal of Computer Applications },
year = { 2025 },
volume = { 187 },
number = { 64 },
pages = { 1-9 },
doi = { 10.5120/ijca2025926064 },
publisher = { Foundation of Computer Science (FCS), NY, USA }
}
%0 Journal Article
%D 2025
%A Syed Imtiazul Sami
%A Mohammad Rasel Mahmud
%A Khaled Bin Showkot Tanim
%A Mohammad Zubair Hussain
%A Mahpara Khan Orpa
%T Advancing Fairness in Multimodal Machine Learning for Internet‑Scale Video Data: Comprehensive Bias Mitigation and Evaluation Framework%T
%J International Journal of Computer Applications
%V 187
%N 64
%P 1-9
%R 10.5120/ijca2025926064
%I Foundation of Computer Science (FCS), NY, USA
The progression of multimodal machine learning (MML), a pivotal aspect of the artificial intelligence (AI) revolution, greatly enhances the analysis and comprehension of video data by providing insights across several modalities, including text, audio, and visual formats. MML models have extensive applications in entertainment, healthcare, and autonomous systems; nevertheless, when trained on expansive video datasets that encompass a diverse cultural, ethnic, and linguistic spectrum, they encounter significant challenges related to fairness and prejudice. This work presents a comprehensive investigation of bias reduction and fairness in MML, addressing issues arising from the intricacies of large-scale video data. This study (1) identifies the origins and mechanisms of bias in MML systems, (2) introduces advanced methodologies to enhance model fairness, and (3) offers a thorough framework for assessing fairness in large-scale video datasets. We offer a framework that integrates bias-aware pre-processing, fairness-aware modeling across multimodal settings, and scalable assessment metrics. Specifically, we employ balanced sampling, GANs for synthetic data augmentation, and adversarial debiasing to provide equitable representation and prediction. We validate our methods on extensive benchmark datasets (YouTube-8M, Activity Net, and Ego4D), demonstrating substantial enhancements in performance and fairness. The experimental findings indicate that the multimodal model surpasses both the modal models and the state-of-the-art techniques, achieving an accuracy of 90.5% and an F1 score of 91.0%. Ultimately, we enhanced fairness measurements, specifically differential impact and equalized chances, by 32.3% and 17.9%, respectively, demonstrating the efficacy of our bias mitigation strategies. However, comparative assessments reveal that our technique delivers state-of-the-art performance on the trade-off between predictive accuracy and ethical fairness, making it a feasible option for real-life contexts where equity is critical. Qualitative study corroborates the alleviation of demographic bias in model predictions, especially on sensitive tasks such as emotion detection and demographic classification. Our research enhances the ethical use of MML systems, guaranteeing that these models are resilient and equitable among diverse population subgroups while establishing a foundation for future advancements in multimodal fusion methodologies and task-specific fairness metrics.