Category Ranking

98%

Total Visits

921

Avg Visit Duration

2 minutes

Citations

20

Article Abstract

Deep learning-based methods have achieved notable progress in removing blocking artifacts caused by lossy JPEG compression on images. However, most deep learning-based methods handle this task by designing black-box network architectures to directly learn the relationships between the compressed images and their clean versions. These network architectures are always lack of sufficient interpretability, which limits their further improvements in deblocking performance. To address this issue, in this article, we propose a model-driven deep unfolding method for JPEG artifacts removal, with interpretable network structures. First, we build a maximum posterior (MAP) model for deblocking using convolutional dictionary learning and design an iterative optimization algorithm using proximal operators. Second, we unfold this iterative algorithm into a learnable deep network structure, where each module corresponds to a specific operation of the iterative algorithm. In this way, our network inherits the benefits of both the powerful model ability of data-driven deep learning method and the interpretability of traditional model-driven method. By training the proposed network in an end-to-end manner, all learnable modules can be automatically explored to well characterize the representations of both JPEG artifacts and image content. Experiments on synthetic and real-world datasets show that our method is able to generate competitive or even better deblocking results, compared with state-of-the-art methods both quantitatively and qualitatively.

Download full-text PDF

Source
http://dx.doi.org/10.1109/TNNLS.2021.3083504DOI Listing

Publication Analysis

Top Keywords

jpeg artifacts
12
model-driven deep
8
deep unfolding
8
unfolding method
8
method jpeg
8
artifacts removal
8
deep learning-based
8
learning-based methods
8
network architectures
8
iterative algorithm
8

Similar Publications

A spatial-frequency hybrid restoration network for JPEG compressed image deblurring.

Neural Netw

September 2025

organization=Chongqing Key Laboratory of Computer Network and Communication Technology, School of Computer Science and Technology (National Exemplary Software School), Chongqing University of Posts and Telecommunications, city=Chongqing, postcode=400065, country=China. Electronic address: tianh519@1

Image deblurring and compression-artifact removal are both ill-posed inverse problems in low-level vision tasks. So far, although numerous image deblurring and compression-artifact removal methods have been proposed respectively, the research for explicit handling blur and compression-artifact coexisting degradation image (BCDI) is rare. In the BCDI, image contents will be damaged more seriously, especially for edges and texture details.

View Article and Find Full Text PDF

Transformer is leading a trend in the field of image processing. While existing lightweight image processing transformers have achieved notable success, they primarily focus on reducing FLOPs (floating-point operations) or the number of parameters, rather than on practical inference acceleration. In this paper, we present a latency-aware image processing transformer, termed LIPT.

View Article and Find Full Text PDF

Denoising diffusion models produce high-fidelity image samples by capturing the image distribution in a progressive manner while initializing with a simple distribution and compounding the distribution complexity. Although these models have unlocked new applicabilities, the sampling mechanism of diffusion does not offer means to extract image-specific semantic representation, which is inherently provided by auto-encoders. The encoding component of auto-encoders enables mapping between a specific image and its latent space, thereby offering explicit means of enforcing structures in the latent space.

View Article and Find Full Text PDF

Infrared line-scanning images have high redundancy and large file sizes. In JPEG2000 compression, the MQ arithmetic encoder's complexity slows down processing. Huffman coding can achieve O(1) complexity based on a code table, but its integer-bit encoding mechanism and ignorance of the continuity of symbol distribution result in suboptimal compression performance.

View Article and Find Full Text PDF

JPEG artifacts produced during the JPEG compression may obscure forgery artifacts, impairing the efficiency of regular forgery detection and localization approaches. To address the issue, we introduce a Multi-Scale Contextual Spatial-Channel Correlation Network, which has been designed for detecting and locating forgeries. Our MSCSCC-Net uses multi-scale mechanisms, which improves forgery detection and localization performance by better handling the scale variation of the forged areas and further helps to remove JPEG artifacts from forgery artifacts at different scales.

View Article and Find Full Text PDF