International Journal of Computer Applications
Foundation of Computer Science (FCS), NY, USA
|
Volume 187 - Issue 22 |
Published: July 2025 |
Authors: Mehek Richharia, Aryan Gupta |
![]() |
Mehek Richharia, Aryan Gupta . Text-to-Image Synthesis with Stable Diffusion: Evaluation and Performance Analysis. International Journal of Computer Applications. 187, 22 (July 2025), 23-30. DOI=10.5120/ijca2025925353
@article{ 10.5120/ijca2025925353, author = { Mehek Richharia,Aryan Gupta }, title = { Text-to-Image Synthesis with Stable Diffusion: Evaluation and Performance Analysis }, journal = { International Journal of Computer Applications }, year = { 2025 }, volume = { 187 }, number = { 22 }, pages = { 23-30 }, doi = { 10.5120/ijca2025925353 }, publisher = { Foundation of Computer Science (FCS), NY, USA } }
%0 Journal Article %D 2025 %A Mehek Richharia %A Aryan Gupta %T Text-to-Image Synthesis with Stable Diffusion: Evaluation and Performance Analysis%T %J International Journal of Computer Applications %V 187 %N 22 %P 23-30 %R 10.5120/ijca2025925353 %I Foundation of Computer Science (FCS), NY, USA
Recent progress in machine learning, especially in imaging, has led to success in generating high-quality images from text descriptions. Among these advances, the widespread adoption of the face stands out for enhancing the model's strength, flexibility, and ability to produce realistic and diverse images. Unlike traditional generative models such as Generative Adversarial Networks (GANs) and Variational Autoencoders (VAEs), which often face issues like training instability and mode collapse, diffusion models offer a more stable structure for image generation. These models benefit from principles of diffusion processes, which iteratively transform random noise into coherent images, resulting in improved performance and reliability. This paper provides a comprehensive review of the latest version of Stable Diffusion, focusing on its revolutionary architecture, core principles, and practical applications. The study compares Stable Diffusion with other leading generative models in terms of image quality, stability, and computational efficiency. It also highlights Hugging Face's role in democratizing AI-driven image generation by making Stable Diffusion accessible through open-source platforms, enabling researchers, developers, and enthusiasts to customize and enhance the model for a wide range of innovative and practical applications. The overview further considers the broader implications of diffusion models in AI-driven creativity, especially in fields such as art, design, advertising, and entertainment. By analyzing the strengths and limitations of Stable Diffusion, the paper aims to offer insights into its potential to influence the future of image generation technology. Additionally, it addresses existing challenges, including the need for greater diversity in generated images and reductions in computational costs. This paper serves as a valuable resource for researchers and practitioners interested in the evolving landscape of text-to-image synthesis and the transformative potential of diffusion models in artificial intelligence.