A novel 8-connected Pixel Identity GAN with Neutrosophic (ECP-IGANN) for missing imputation.

Sci Rep

Computer Science Department, Faculty of Computers and Informatics, Zagazig University, 44511, Zagazig, Egypt.

Published: October 2024


Category Ranking

98%

Total Visits

921

Avg Visit Duration

2 minutes

Citations

20

Article Abstract

Missing pixel imputation presents a critical challenge in image processing and computer vision, particularly in applications such as image restoration and inpainting. The primary objective of this paper is to accurately estimate and reconstruct missing pixel values to restore complete visual information. This paper introduces a novel model called the Enhanced Connected Pixel Identity GAN with Neutrosophic (ECP-IGANN), which is designed to address two fundamental issues inherent in existing GAN architectures for missing pixel generation: (1) mode collapse, which leads to a lack of diversity in generated pixels, and (2) the preservation of pixel integrity within the reconstructed images. ECP-IGANN incorporates two key innovations to improve missing pixel imputation. First, an identity block is integrated into the generation process to facilitate the retention of existing pixel values and ensure consistency. Second, the model calculates the values of the 8-connected neighbouring pixels around each missing pixel, thereby enhancing the coherence and integrity of the imputed pixels. The efficacy of ECP-IGANN was rigorously evaluated through extensive experimentation across five diverse datasets: BigGAN-ImageNet, the 2024 Medical Imaging Challenge Dataset, the Autonomous Vehicles Dataset, the 2024 Satellite Imagery Dataset, and the Fashion and Apparel Dataset 2024. These experiments assessed the model's performance in terms of diversity, pixel imputation accuracy, and mode collapse mitigation, with results demonstrating significant improvements in the Inception Score (IS) and Fréchet Inception Distance (FID). ECP-IGANN markedly enhanced image segmentation performance in the validation phase across all datasets. Key metrics, such as Dice Score, Accuracy, Precision, and Recall, were improved substantially for various segmentation models, including Spatial Attention U-Net, Dense U-Net, and Residual Attention U-Net. For example, in the 2024 Medical Imaging Challenge Dataset, the Residual Attention U-Net's Dice Score increased from 0.84 to 0.90, while accuracy improved from 0.88 to 0.93 following the application of ECP-IGANN. Similar performance enhancements were observed with the other datasets, highlighting the model's robust generalizability across diverse imaging domains.

Download full-text PDF

Source
http://www.ncbi.nlm.nih.gov/pmc/articles/PMC11471839PMC
http://dx.doi.org/10.1038/s41598-024-73976-7DOI Listing

Publication Analysis

Top Keywords

missing pixel
20
pixel imputation
12
pixel
10
pixel identity
8
identity gan
8
gan neutrosophic
8
neutrosophic ecp-igann
8
pixel values
8
mode collapse
8
2024 medical
8

Similar Publications

Multi-modal data fusion plays a critical role in enhancing the accuracy and robustness of perception systems for autonomous driving, especially for the detection of small objects. However, small object detection remains particularly challenging due to sparse LiDAR points and low-resolution image features, which often lead to missed or imprecise detections. Currently, many methods process LiDAR point clouds and visible-light camera images separately, and then fuse them in the detection head.

View Article and Find Full Text PDF

Portraits or selfie images taken from a close distance typically suffer from perspective distortion. In this paper, we propose an end-to-end deep learning-based rectification pipeline to mitigate the effects of perspective distortion. We learn to predict the facial depth by training a deep CNN.

View Article and Find Full Text PDF

This study presents an innovative unmanned aerial vehicle (UAV)-based intelligent detection method utilizing an improved Faster Region-based Convolutional Neural Network (Faster R-CNN) architecture to address the inefficiency and inaccuracy inherent in manual wheat spike counting. We systematically collected a high-resolution image dataset (2000 images, 4096 × 3072 pixels) covering key growth stages (heading, grain filling, and maturity) of winter wheat ( L.) during 2022-2023 using a DJI M300 RTK equipped with multispectral sensors.

View Article and Find Full Text PDF

Timely detection and effective management of postoperative flap crises are critical for improving flap salvage rates. Flap crisis often stems from impaired blood circulation, leading to changes in the flap's color, texture, and temperature. Therefore, we analyzed flap crises using anatomical and colorimetric parameters and designed pixel curve features using a biologically derived foundation model.

View Article and Find Full Text PDF

Objective: To enhance the automatic detection precision of diabetic retinopathy (DR) lesions, this study introduces an improved YOLOv8 model specifically designed for the precise identification of DR lesions.

Method: This study integrated two attention mechanisms, convolutional exponential moving average (convEMA) and convolutional simple attention module (convSimAM), into the backbone of the YOLOv8 model. A dataset consisting of 3,388 ultra-widefield (UWF) fundus images obtained from patients with DR, each with a resolution of 2,600 × 2048 pixels, was utilized for both training and testing purposes.

View Article and Find Full Text PDF