|
ShodhKosh: Journal of Visual and Performing ArtsISSN (Online): 2582-7472
AI-Based Noise Reduction in Artistic Photography Damanjeet Aulakh 1 1 Centre
of Research Impact and Outcome, Chitkara University, Rajpura- 140417, Punjab,
India 2 Assistant
Professor, Department of Computer Science and IT, ARKA JAIN University
Jamshedpur, Jharkhand, India 3 Associate Professor, School of Business Management, Noida International
University, India 4 Department of Information Technology, Vishwakarma Institute of
Technology, Pune, Maharashtra, 411037, India 5 Associate Professor, Department of Computer Science and Engineering, Aarupadai Veedu Institute of Technology, Vinayaka Mission’s
Research Foundation (DU), Tamil Nadu, India 6 Assistant Professor, Department of Centre for Internet of Things,
Siksha 'O' Anusandhan (Deemed to be University),
Bhubaneswar, Odisha, India
1. INTRODUCTION Artistic photography has a special niche in the greater framework of visual imaging, in which the technical aspects of an image compete with subjective beauty, emotional appeal, and intentionality. In art photography, unlike the traditional approach of photography often being focused on clarity and realism, artistic photography allows the variances in tone, grain, texture, and light to be used to convey a mood or narrative richness. But the prolific nature of noise, either due to the low-light situation, high ISO, the weakness of sensor, the aging of the film or the environmental constraints, may make or break the artistic value of an image. Although grain has been deliberately used in some genres as a form of expression, noise that cannot be controlled may blur fine details and break tonal coherence and cause drawbacks in downstream creative processes like digital painting or composites or other forms of style adjustment Zhang et al. (2024). This means that noise reduction is still a crucial aspect of the modern artistic pipelines of photography. Conventional methods of noise reduction spatial filters, frequency-domain transforms, and model-based denoising have long been important in improving images, but have all basic constraints. Such methods as bilateral filtering, non-local means, and wavelet shrinkage frequently fail to determine the difference between texture that is of artistic significance and noise that is undesirable. Over smoothing often leads to the formation of plastic surfaces, loss of microstructure and suppression of emotional expression and is not appropriate in creative photography where expression rich in detail is of primary importance Chen et al. (2024). Moreover, classical approaches are not adaptive; they fail to learn data and hence fail to learn the trends of style like the tonal variations of skin, distribution of film grains, or structure of brushstroke as in paintings. With the ever-growing variety of artistic photography styles, such as fine-art portraits and abstracts, to movie frames and blends of digital and analog, the need to create a smarter and more versatile noisy reducing technologies is more apparent Gano et al. (2024). The new developments of artificial intelligence have changed this picture. Algorithms based on deep learning Noise reduction, which can be enabled via such architecture as Denoising Autoencoders (DAEs), U-Nets, Generative Adversarial Networks (GANs), and Transformer-hybrid models, constitute a strong alternative to traditional methods. Such models learn the latent structure of clean artistic pictures and the statistical characteristics of different types of noise and hence can better distinguish meaningful detail and noise. Multi-scale feature extractors, attention, perceptual loss training and adversarial training allow the retention of the tiny artistic details, such as texture gradients, tonal changes, and stylistic features, that classical denoisers tend to blur Agrawal et al. (2022). Significantly, such AI models can fit various artistic styles, which is possible by training on curated datasets including fine-art photography, digital portraits, stylized imagery, and go as far as recreating film scans. There are serious implications of the introduction of AI-based noise reduction in creative workflows. To the restoration professionals, it presents an unparalleled option to salvage decayed or damaged old and film photographs without losing their historical or artistic value Rao (2020). 2. Background Work Noise mitigation has always been a burning issue of interest in the research on imaging (especially in artistic and fine-art photography) because the visual quality has a direct bearing on the emotional and aesthetic interpretation. Initial background work was based on classical filtering methods and statistical modeling to reduce undesired noise and in the process, trying to maintain structural information. Some of the most popular methods were bilateral filtering, anisotropic diffusion, wavelet shrinkage, and non-local means (NLM). Through these techniques, significant concepts were brought on board like edge-sensitive smoothing, patch similarity and multi-resolution analysis Vijayakumar and Vairavasundaram (2024). Their nature however was that they were inherently dependent on handcrafted assumptions making them less able to differentiate between artistic texture such as film grain, brushstroke-like patterns, or subtle tonal variations, and unwanted noise so over-smoothing or erased expressive information. Later working was done on more advanced probabilistic and transform-based frameworks. Variational models, sparse coding and BM3D gained impact as they enhanced recovery of texture and provided a stronger noise modeling. In particular, BM3D established a high standard of classical denoising because of its multi-user filtering scheme and block-matching scheme Wang et al. (2025). However, even these more sophisticated techniques still failed in complex noise profiles in real-world artistic photography, such mixed noise, noise that varies with ISO, and noise that is sensor-specific. In addition, they had no capacity to respond to large scale visual cues, and this was what restricted their flexibility to various artistic genres and signature styles. The revolution of deep learning was a major transformation. The first attempts at denoising Autoencoders (DAEs) were made by showing that neural networks could be trained to learn nonlinear noise-to-clean mappings. U-Net-based designs also contribute to the progress of the sphere as they allow to extract features at the multi-scale and achieve a skip connection, which lead to the higher structural preservation Emek et al. (2023). Generative Adversarial Networks (GANs) offered perceptual realism where models are trained to maintain fine artistic textures and tonal features by training them against discriminators that are sensitive to artistic features. Table 1 presents the previous research that indicates AI methods to remove noise in photographs. Transformer-based hybrid models have since endeavored to even further with the help of attention mechanisms to model long range dependencies, and are thus especially useful at maintaining overall coherence in artistic content Liu et al. (2024). Table 1
3. Methodology 3.1. Dataset collection The process of data collection involving the artificial intelligence (AI)-based noise reduction in artistic photography is a matter of close curation to guarantee stylistic heterogeneity as well as the authenticity of noise. Examples of artistic photography include portraiture in the fine-art, conceptual, film-inspired, and abstract, low-light creative, and conceptual painting-photography hybrids. Thus, the dataset should include artistic images of high quality obtained through the sources of public repositories (e.g., DPED, DIV2K, Flickr Creative Commons artistic sets, MIT FiveK), film photography scans, and collections owned by artists Ma et al. (2024). In order to be rich, the images are chosen depending on the color depth, tonal variation, artistic grain, lighting complexity, richness of texture, and distinctiveness of style. Supervised learning has noise-free (or slightly noisy) reference images which are used as ground truth. But even actual artistic images are bound to have some grain or artistic flaws, so the use of noise simulation methods that simulate the degradation of reality must be taken. Techniques used to employ noise commonly used in texture-smoothing, such as Gaussian noise with dynamic standard deviation, poisson noise to simulate photon restrictions in dark-creative photography, and speckle noise or salt-and-pepper noise in scans of older films Gui et al. (2024). 3.2. Preprocessing and Noise Modeling Preprocessing is considered a crucial step in terms of training artistic images using AI-based denoising models and providing stability in terms of quality, noise modeling, and style preservation. It starts with color-space normalization (usually sRGB or linear RGB), adjustment of the dynamic range, and correction of the white-balance to make sure that there are identical tonal properties in the input pictures. In order to preserve the granularity of details that are important to artistic imagery, high-resolution image patches (e.g., 256 x 256 or 512 x 512) are cropped. Further preprocessing operations are histogram equalization, gamma correction and edge-enhancement analysis to bring out the structural motifs e.g. brushstroke-like patterns, grain transitions or textured surfaces. Figure 1 |
||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||
|
Table 2 Objective Quality Metrics Comparison Across Models |
|||||
|
Metric / Model |
Traditional (BM3D) |
DAE |
U-Net |
GAN |
Transformer Hybrid |
|
PSNR (dB) |
27.4 |
29.1 |
31.8 |
32.6 |
33.4 |
|
SSIM |
0.842 |
0.883 |
0.914 |
0.927 |
0.941 |
|
LPIPS |
0.189 |
0.152 |
0.118 |
0.094 |
0.072 |
|
ΔE Color
Error |
5.84 |
4.67 |
3.92 |
3.41 |
2.98 |
|
Edge Preservation (%) |
72.5 |
78.3 |
84.6 |
86.9 |
89.4 |
Table 2 shows an evident development of performance between traditional and advanced AI-based models. The model BM3D gives the lowest PSNR of 27.4 dB and SSIM of 0.842, indicating that it has low capabilities to preserve structural and tonal information in artistic images. Multimetric comparison of BM3D, DAE, U-Net, GAN and Transformer methods is presented in Figure 3.
Figure 3

Figure 3 Multimetric Evaluation of BM3D, DAE,
U-Net, GAN, and Transformer Hybrid Methods
The Denoising Autoencoder attains 29.1 dB PSNR and 0.883 SSIM, but it still has issues with subtle texture. U-Net generates a large leap reaching 31.8 dB PSNR and 0.914 SSIM with lower LPIPS (0.118) and 4.92 8E error which indicates enhanced perceptual and color fidelity.
6.2. Visual Comparison on Texture, Tone, and Artistic Details
Visual comparisons also draw attention to the strong points that the AI-based denoising possesses the ability to preserve the artistic richness, tonal smoothness, and small texture details. Traditional techniques can lead to over-smoothing of the skin, flattening of skin texture or blurring of edges or film-like texture. In particular, AIs such as U-Net and GAN models can recreate subtle gradients of shadows, maintain the richness of the midtone, and capture micro-textures that are vital in the process of expressiveness. Transformer hybrids have been found to exhibit extraordinary tone continuity and spatial coherence that is perfect when capturing a portrait or a fine art composition.
Table 3
|
Table 3 Texture, Tone, and Detail Preservation Scores (%) |
||||
|
Evaluation Aspect |
BM3D |
DAE |
U-Net |
GAN |
|
Texture Fidelity (%) |
61.2 |
68.7 |
82.5 |
88.4 |
|
Tone Continuity (%) |
67.9 |
74.1 |
85.3 |
89.5 |
|
Artistic Grain Retention (%) |
54.6 |
63.2 |
79.8 |
86.7 |
|
Shadow Gradient Smoothness
(%) |
58.1 |
72.4 |
83.6 |
88.2 |
Table 3 shows how AI-based models have brought significant enhancement to texture, tone, and other stylistic features, which are critical to artistic photography. BM3D offers the poorest performance on all the aspects tested with 61.2 texture fidelity, 67.9 tone continuity, 54.6 preservation of grain and 58.1 smoothness of a shadow graduate. Bar comparison of texture, tone, grain and shadow quality has been presented in Fig. 4.
Figure 4

Figure 4
Comparative Bar
Chart of Texture, Tone, Grain, and Shadow Quality Across Denoising Models
These values indicate that it has a preference to smooth out the images and the expressive details are lost. The Denoising AutoEncoder has a moderate improvement in both texture fidelity (68.7 percent) and tone continuity (74.1 percent) but also has difficulties with the subtle textures of creative images.
6.3. Impact of AI Models on Artistic Expression Preservation
The impact of AI in noise reduction can have a positive effect on artistic expression since it can be used to improve the clarity of the image without diminishing the stylistic meaning. Then with noise to detail relationship models can memorize expressive textures, movie grain and tonal breaks that help an image have an emotive appeal. GAN and Transformer-based models are specifically useful when it comes to preserving artistic features like soft gradient, mood of shadows, and composition. This will guarantee that even after the use of aggressive noise removal, the creative vision of the photographer will not be affected.
Table 4
|
Table 4 Artistic Expression Preservation Metrics |
||||
|
Artistic Metric |
BM3D |
DAE |
U-Net |
GAN |
|
Stylistic Integrity (%) |
62.8 |
71.4 |
84.2 |
89.3 |
|
Emotion Preservation Score
(%) |
61 |
69 |
82 |
88 |
|
Gradient Naturalness (%) |
65.2 |
73.6 |
86.1 |
90.4 |
|
Texture–Noise Separation
Accuracy (%) |
59.7 |
68.3 |
82.4 |
88.7 |
Table 4 illustrates the efficacy of various denoising models to retain the expression of art, a paramount need in creative photography whereby, mood, tone and stylistic delicacies ought to be preserved. Figure 5 illustrates performance curves of the artistic quality indicators of the denoising models.
Figure 5

Figure 5 Performance Curve of Artistic Quality Indicators for
BM3D, DAE, U-Net, and GAN
BM3D is least performing at 62.8% stylistic preservation and 61% emotion preservation, which is also an indication that normal filtering tends to squash expressive content whilst trying to eliminate noise. The fact that its gradient naturalness is 65.2% and the ability to separate texture and noise is 59.7% is another indicator of its shortcomings of separating artistic detail and undesirable artifacts. The Denoising Autoencoder gives significantly higher scores with 71.4% stylistic integrity and 69% emotion preservation.
7. Conclusion
The concept of noise reduction through AI has emerged as a critical new technology in the artistic photography sector, providing a complex tool between the technical improvement and maintenance of the creative intent. Compared to more conventional methods of denoising where it is common to reduce textures or sacrifice important aspects of style, modern AI models, in the form of Denoising Autoencoders, U-Nets, GANs and Transformer-based hybrids, have been shown to be able to discriminate between significant artistic detail and harmful noise. It is especially important in the imaginative image in which texture, grain, tonal gradients, and slight flaws play an essential role in the visual story. By undertaking a vast amount of experiments and testing on objective metrics in the form of PSNR, SSIM, LPIPS, and Delta-E, AI-based techniques are generally able to provide the best performance, which identifies their capability to produce clean and at the same time expressive photographic results. Their effect is further confirmed by visual comparison showing enhanced tonal continuity, stronger structures, and maintained artistic textures in old, digital and stylized photography images. The research, however, also highlights some drawbacks, such as the inabilities to preserve the stylistic faithfulness, the bias in data, and the inability to capture the noise variations in the real world.
CONFLICT OF INTERESTS
None.
ACKNOWLEDGMENTS
None.
REFERENCES
Agrawal, S., Panda, R., Mishro, P. K., and Abraham, A. (2022). A Novel Joint Histogram Equalization Based Image Contrast Enhancement. Journal of King Saud University – Computer and Information Sciences, 34, 1172–1182. https://doi.org/10.1016/j.jksuci.2019.05.010
Archana, R., and Jeevaraj, P. E. (2024). Deep Learning Models for Digital Image Processing: A Review. Artificial Intelligence Review, 57, Article 11. https://doi.org/10.1007/s10462-023-10631-z
Badjie, B., Cecílio, J., and Casimiro, A. (2024). Adversarial Attacks and Countermeasures on Image Classification-Based Deep Learning Models in Autonomous Driving Systems: A Systematic Review. ACM Computing Surveys, 57(1), 1–52. https://doi.org/10.1145/3691625
Cai, Y., Zhang, W., Chen, H., and Cheng, K. T. (2025). Medianomaly: A Comparative Study of Anomaly Detection in Medical Images. Medical Image Analysis, 102, Article 103500. https://doi.org/10.1016/j.media.2025.103500
Chen, W., Feng, S., Yin, W., Li, Y., Qian, J., Chen, Q., and Zuo, C. (2024). Deep-Learning-Enabled Temporally Super-Resolved Multiplexed Fringe Projection Profilometry: High-Speed kHz 3D Imaging with Low-Speed Camera. PhotoniX, 5, Article 25. https://doi.org/10.1186/s43074-024-00139-2
Emek Soylu, B., Guzel, M. S., Bostanci, G. E., Ekinci, F., Asuroglu, T., and Acici, K. (2023). Deep-Learning-Based Approaches for Semantic Segmentation of Natural Scene Images: A Review. Electronics, 12(12), Article 2730. https://doi.org/10.3390/electronics12122730
Gano, B., Bhadra, S., Vilbig, J. M., Ahmed, N., Sagan, V., and Shakoor, N. (2024). Drone-Based Imaging Sensors, Techniques, and Applications in Plant Phenotyping for Crop Breeding: A Comprehensive Review. Plant Phenome Journal, 7, Article e20100. https://doi.org/10.1002/ppj2.20100
Gui, S., Song, S., Qin, R., and Tang, Y. (2024). Remote Sensing Object Detection in the Deep Learning Era: A Review. Remote Sensing, 16(2), Article 327. https://doi.org/10.3390/rs16020327
Lepcha, D. C., Goyal, B., Dogra, A., Sharma, K. P., and Gupta, D. N. (2023). A Deep Journey into Image Enhancement: A Survey of Current and Emerging Trends. Information Fusion, 93, 36–76. https://doi.org/10.1016/j.inffus.2022.12.012
Liu, Y., Bai, X., Wang, J., Li, G., Li, J., and Lv, Z. (2024). Image Semantic Segmentation Approach Based on DeepLabV3+ Network with an Attention Mechanism. Engineering Applications of Artificial Intelligence, 127, Article 107260. https://doi.org/10.1016/j.engappai.2023.107260
Ma, J., He, Y., Li, F., Han, L., You, C., and Wang, B. (2024). Segment Anything in Medical Images. Nature Communications, 15, Article 654. https://doi.org/10.1038/s41467-024-44824-z
Rao, B. S. (2020). Dynamic Histogram Equalization for Contrast Enhancement for Digital Images. Applied Soft Computing, 89, Article 106114. https://doi.org/10.1016/j.asoc.2020.106114
Vijayakumar, A., and Vairavasundaram, S. (2024). YOLO-Based Object Detection Models: A Review and its Applications. Multimedia Tools and Applications, 83, 83535–83574. https://doi.org/10.1007/s11042-024-18872-y
Wang, A., Chen, H., Liu, L., Chen, K., Lin, Z., and Han, J. (2025). YOLOv10: Real-Time End-to-End Object Detection. Advances in Neural Information Processing Systems, 37, 107984–108011. https://doi.org/10.52202/079017-3429
Zhang, P., Zhou, F., Wang, X., Wang, S., and Song, Z. (2024). Omnidirectional Imaging Sensor Based on Conical Mirror for Pipelines. Optics and Lasers in Engineering, 175, Article 108003. https://doi.org/10.1016/j.optlaseng.2023.108003
Zhao, T., Guo, P., and Wei, Y. (2024). Road Friction Estimation Based on Vision for Safe Autonomous Driving. Mechanical Systems and Signal Processing, 208, Article 111019. https://doi.org/10.1016/j.ymssp.2023.111019
|
|
This work is licensed under a: Creative Commons Attribution 4.0 International License
© ShodhKosh 2024. All Rights Reserved.