Academic literature on the topic 'Low-light enhancement and denoising'

Create a spot-on reference in APA, MLA, Chicago, Harvard, and other styles

Select a source type:

Consult the lists of relevant articles, books, theses, conference reports, and other scholarly sources on the topic 'Low-light enhancement and denoising.'

Next to every source in the list of references, there is an 'Add to bibliography' button. Press on it, and we will generate automatically the bibliographic reference to the chosen work in the citation style you need: APA, MLA, Harvard, Chicago, Vancouver, etc.

You can also download the full text of the academic publication as pdf and read online its abstract whenever available in the metadata.

Journal articles on the topic "Low-light enhancement and denoising"

1

Carré, Maxime, and Michel Jourlin. "Extending Camera’s Capabilities in Low Light Conditions Based on LIP Enhancement Coupled with CNN Denoising." Sensors 21, no. 23 (2021): 7906. http://dx.doi.org/10.3390/s21237906.

Full text
Abstract:
Using a sensor in variable lighting conditions, especially very low-light conditions, requires the application of image enhancement followed by denoising to retrieve correct information. The limits of such a process are explored in the present paper, with the objective of preserving the quality of enhanced images. The LIP (Logarithmic Image Processing) framework was initially created to process images acquired in transmission. The compatibility of this framework with the human visual system makes possible its application to images acquired in reflection. Previous works have established the ability of the LIP laws to perform a precise simulation of exposure time variation. Such a simulation permits the enhancement of low-light images, but a denoising step is required, realized by using a CNN (Convolutional Neural Network). A main contribution of the paper consists of using rigorous tools (metrics) to estimate the enhancement reliability in terms of noise reduction, visual image quality, and color preservation. Thanks to these tools, it has been established that the standard exposure time can be significantly reduced, which considerably enlarges the use of a given sensor. Moreover, the contribution of the LIP enhancement and denoising step are evaluated separately.
APA, Harvard, Vancouver, ISO, and other styles
2

Zhang, Jialiang, Ruiwen Ji, Jingwen Wang, Hongcheng Sun, and Mingye Ju. "DEGAN: Decompose-Enhance-GAN Network for Simultaneous Low-Light Image Lightening and Denoising." Electronics 12, no. 14 (2023): 3038. http://dx.doi.org/10.3390/electronics12143038.

Full text
Abstract:
Images taken in low-light situations frequently have a significant quality reduction. Taking care of these degradation problems in low-light images is essential for raising their visual quality and enhancing high-level visual task performance. However, because of the inherent information loss in dark images, conventional Retinex-based approaches for low-light image enhancement frequently fail to accomplish real denoising. This research introduces DEGANet, a revolutionary deep-learning framework created particularly for improving and denoising low-light images. To overcome these restrictions, DEGANet makes use of the strength of a Generative Adversarial Network (GAN). The Decom-Net, Enhance-Net, and an Adversarial Generative Network (GAN) are three linked subnets that make up our novel Retinex-based DEGANet architecture. The Decom-Net is in charge of separating the reflectance and illumination components from the input low-light image. This decomposition enables Enhance-Net to effectively enhance the illumination component, thereby improving the overall image quality. Due to the complicated noise patterns, fluctuating intensities, and intrinsic information loss in low-light images, denoising them presents a significant challenge. By incorporating a GAN into our architecture, DEGANet is able to effectively denoise and smooth the enhanced image as well as retrieve the original data and fill in the gaps, producing an output that is aesthetically beautiful while maintaining key features. Through a comprehensive set of studies, we demonstrate that DEGANet exceeds current state-of-the-art methods in both terms of image enhancement and denoising quality.
APA, Harvard, Vancouver, ISO, and other styles
3

Kim, Minjae, Dubok Park, David Han, and Hanseok Ko. "A novel approach for denoising and enhancement of extremely low-light video." IEEE Transactions on Consumer Electronics 61, no. 1 (2015): 72–80. http://dx.doi.org/10.1109/tce.2015.7064113.

Full text
APA, Harvard, Vancouver, ISO, and other styles
4

Malik, Sameer, and Rajiv Soundararajan. "A low light natural image statistical model for joint contrast enhancement and denoising." Signal Processing: Image Communication 99 (November 2021): 116433. http://dx.doi.org/10.1016/j.image.2021.116433.

Full text
APA, Harvard, Vancouver, ISO, and other styles
5

Das Mou, Trisha, Saadia Binte Alam, Md Hasibur Rahman, Gautam Srivastava, Mahady Hasan, and Mohammad Faisal Uddin. "Multi-Range Sequential Learning Based Dark Image Enhancement with Color Upgradation." Applied Sciences 13, no. 2 (2023): 1034. http://dx.doi.org/10.3390/app13021034.

Full text
Abstract:
Images under low-light conditions suffer from noise, blurring, and low contrast, thus limiting the precise detection of objects. For this purpose, a novel method is introduced based on convolutional neural network (CNN) dual attention unit (DAU) and selective kernel feature synthesis (SKFS) that merges with the Retinex theory-based model for the enhancement of dark images under low-light conditions. The model mentioned in this paper is a multi-scale residual block made up of several essential components equivalent to an onward convolutional neural network with a VGG16 architecture and various Gaussian convolution kernels. In addition, backpropagation optimizes most of the parameters in this model, whereas the values in conventional models depend on an artificial environment. The model was constructed using simultaneous multi-resolution convolution and dual attention processes. We performed our experiment in the Tesla T4 GPU of Google Colab using the Customized Raw Image Dataset, College Image Dataset (CID), Extreme low-light denoising dataset (ELD), and ExDark dataset. In this approach, an extended set of features is set up to learn from several scales to incorporate contextual data. An extensive performance evaluation on the four above-mentioned standard image datasets showed that MSR-MIRNeT produced standard image enhancement and denoising results with a precision of 97.33%; additionally, the PSNR/SSIM result is 29.73/0.963 which is better than previously established models (MSR, MIRNet, etc.). Furthermore, the output of the proposed model (MSR-MIRNet) shows that this model can be implemented in medical image processing, such as detecting fine scars on pelvic bone segmentation imaging, enhancing contrast for tuberculosis analysis, and being beneficial for robotic visualization in dark environments.
APA, Harvard, Vancouver, ISO, and other styles
6

Han, Guang, Yingfan Wang, Jixin Liu, and Fanyu Zeng. "Low-light images enhancement and denoising network based on unsupervised learning multi-stream feature modeling." Journal of Visual Communication and Image Representation 96 (October 2023): 103932. http://dx.doi.org/10.1016/j.jvcir.2023.103932.

Full text
APA, Harvard, Vancouver, ISO, and other styles
7

Hu, Linshu, Mengjiao Qin, Feng Zhang, Zhenhong Du, and Renyi Liu. "RSCNN: A CNN-Based Method to Enhance Low-Light Remote-Sensing Images." Remote Sensing 13, no. 1 (2020): 62. http://dx.doi.org/10.3390/rs13010062.

Full text
Abstract:
Image enhancement (IE) technology can help enhance the brightness of remote-sensing images to obtain better interpretation and visualization effects. Convolutional neural networks (CNN), such as the Low-light CNN (LLCNN) and Super-resolution CNN (SRCNN), have achieved great success in image enhancement, image super resolution, and other image-processing applications. Therefore, we adopt CNN to propose a new neural network architecture with end-to-end strategy for low-light remote-sensing IE, named remote-sensing CNN (RSCNN). In RSCNN, an upsampling operator is adopted to help learn more multi-scaled features. With respect to the lack of labeled training data in remote-sensing image datasets for IE, we use real natural image patches to train firstly and then perform fine-tuning operations with simulated remote-sensing image pairs. Reasonably designed experiments are carried out, and the results quantitatively show the superiority of RSCNN in terms of structural similarity index (SSIM) and peak signal-to-noise ratio (PSNR) over conventional techniques for low-light remote-sensing IE. Furthermore, the results of our method have obvious qualitative advantages in denoising and maintaining the authenticity of colors and textures.
APA, Harvard, Vancouver, ISO, and other styles
8

Zhao, Meng Ling, and Min Xia Jiang. "Research on Enhanced of Mine-Underground Picture." Advanced Materials Research 490-495 (March 2012): 548–52. http://dx.doi.org/10.4028/www.scientific.net/amr.490-495.548.

Full text
Abstract:
Because of the based on S3C6410 Field information recorder mine- underground non-uniform illumination and mine- underground non-uniform illumination that a large of noise collected and transferred,image is low contrast ,dim and dark. Based on the theory of Donoho's wavelet threshold denoising, several typical wavelet threshold denoising methods are compared.the best denoising effect of peak signal to noise ratio is obtained. The image enhancement method that combination of the adaptive thresholding denoising and histogram equalization is proposed. The experiment result shows that the method has a good denoising performance, which removed the readout noise of CCD Camera,at the same time, image quality is improved .So the wavelet enhancement in image processing of mine- underground can improve image quality.
APA, Harvard, Vancouver, ISO, and other styles
9

Wu, Zeju, Yang Ji, Lijun Song, and Jianyuan Sun. "Underwater Image Enhancement Based on Color Correction and Detail Enhancement." Journal of Marine Science and Engineering 10, no. 10 (2022): 1513. http://dx.doi.org/10.3390/jmse10101513.

Full text
Abstract:
To solve the problems of underwater image color deviation, low contrast, and blurred details, an algorithm based on color correction and detail enhancement is proposed. First, the improved nonlocal means denoising algorithm is used to denoise the underwater image. The combination of Gaussian weighted spatial distance and Gaussian weighted Euclidean distance is used as the index of nonlocal means denoising algorithm to measure the similarity of structural blocks. The improved algorithm can retain more edge features and texture information while maintaining noise reduction ability. Then, the improved U-Net is used for color correction. Introducing residual structure and attention mechanism into U-Net can effectively enhance feature extraction ability and prevent network degradation. Finally, a sharpening algorithm based on maximum a posteriori is proposed to enhance the image after color correction, which can increase the detailed information of the image without expanding the noise. The experimental results show that the proposed algorithm has a remarkable effect on underwater image enhancement.
APA, Harvard, Vancouver, ISO, and other styles
10

Xu, Xiaogang, Ruixing Wang, Chi-Wing Fu, and Jiaya Jia. "Deep Parametric 3D Filters for Joint Video Denoising and Illumination Enhancement in Video Super Resolution." Proceedings of the AAAI Conference on Artificial Intelligence 37, no. 3 (2023): 3054–62. http://dx.doi.org/10.1609/aaai.v37i3.25409.

Full text
Abstract:
Despite the quality improvement brought by the recent methods, video super-resolution (SR) is still very challenging, especially for videos that are low-light and noisy. The current best solution is to subsequently employ best models of video SR, denoising, and illumination enhancement, but doing so often lowers the image quality, due to the inconsistency between the models. This paper presents a new parametric representation called the Deep Parametric 3D Filters (DP3DF), which incorporates local spatiotemporal information to enable simultaneous denoising, illumination enhancement, and SR efficiently in a single encoder-and-decoder network. Also, a dynamic residual frame is jointly learned with the DP3DF via a shared backbone to further boost the SR quality. We performed extensive experiments, including a large-scale user study, to show our method's effectiveness. Our method consistently surpasses the best state-of-the-art methods on all the challenging real datasets with top PSNR and user ratings, yet having a very fast run time. The code is available at https://github.com/xiaogang00/DP3DF.
APA, Harvard, Vancouver, ISO, and other styles
More sources
We offer discounts on all premium plans for authors whose works are included in thematic literature selections. Contact us to get a unique promo code!

To the bibliography