Category Ranking

98%

Total Visits

921

Avg Visit Duration

2 minutes

Citations

20

Article Abstract

3D scene stylization refers to generating stylized images of the scene at arbitrary novel view angles following a given set of style images while ensuring consistency when rendered from different views. Recently, several 3D style transfer methods leveraging the scene reconstruction capabilities of pre-trained neural radiance fields (NeRF) have been proposed. To successfully stylize a scene this way, one must first reconstruct a photo-realistic radiance field from collected images of the scene. However, when only sparse input views are available, pre-trained few-shot NeRFs often suffer from high-frequency artifacts, which are generated as a by-product of high-frequency details for improving reconstruction quality. Is it possible to generate more faithful stylized scenes from sparse inputs by directly optimizing encoding-based scene representation with target style? In this paper, we consider the stylization of sparse-view scenes in terms of disentangling content semantics and style textures. We propose a coarse-to-fine sparse-view scene stylization framework, where a novel hierarchical encoding-based neural representation is designed to generate high-quality stylized scenes directly from implicit scene representations. We also propose a new optimization strategy with content strength annealing to achieve realistic stylization and better content preservation. Extensive experiments demonstrate that our method can achieve high-quality stylization of sparse-view scenes and outperforms fine-tuning-based baselines in terms of stylization quality and efficiency.

Download full-text PDF

Source
http://dx.doi.org/10.1109/TVCG.2025.3558468DOI Listing

Publication Analysis

Top Keywords

sparse-view scenes
12
neural representation
8
scene
8
scene stylization
8
images scene
8
stylized scenes
8
stylization sparse-view
8
stylization
6
scenes
5
stylizing sparse-view
4

Similar Publications

Radiance fields represented by 3D Gaussians excel at synthesizing novel views, offering both high training efficiency and fast rendering. However, with sparse input views, the lack of multi-view consistency constraints results in poorly initialized Gaussians and unreliable heuristics for optimization, leading to suboptimal performance. Existing methods often incorporate depth priors from dense estimation networks but overlook the inherent multi-view consistency in input images.

View Article and Find Full Text PDF

Neural implicit functions including signed distance functions (SDFs) and unsigned distance functions (UDFs) have shown powerful ability in fitting the shape geometry. However, inferring continuous distance fields from discrete unoriented point clouds still remains a challenge. The neural network typically fits the shape with a rough surface and omits fine-grained geometric details such as shape edges and corners.

View Article and Find Full Text PDF

Objectives: Computed tomography (CT) provides high spatial-resolution visualization of 3D structures for various applications. Traditional analytical/iterative CT reconstruction algorithms require hundreds of angular samplings, a condition may not be met practically for physical and mechanical limitations. Sparse view CT reconstruction has been proposed using constrained optimization and machine learning methods with varying success, less so for ultra-sparse view reconstruction.

View Article and Find Full Text PDF

SREGS: Sparse-view Gaussian radiance fields with geometric regularization and region exploration.

Neural Netw

November 2025

Shandong Key Laboratory of Technologies and Systems for Intelligent Construction Equipment, Shandong Jiaotong University, Jina, 250357, Shandong, China; School of Information Science and Electrical Engineering, Shandong Jiaotong University, Jina, 250357, Shandong, China. Electronic address: 23208037

Recent advances in few-shot novel-view synthesis based on 3D Gaussian Splatting (3DGS) have shown remarkable progress. Existing methods usually rely on carefully designed geometric regularizers to reinforce geometric supervision; however, applying multiple regularizers consistently across scenes is hard to tune and often degrades robustness. Consequently, generating reliable geometry from extremely sparse viewpoints remains a key challenge.

View Article and Find Full Text PDF

Dynamic three-dimensional (4D) reconstruction from two-dimensional X-ray coronary angiography (CA) remains a significant clinical problem. Existing CA reconstruction methods often require extensive user interaction or large training datasets. Recently, Neural Radiance Field (NeRF) has successfully reconstructed high-fidelity scenes in natural and medical contexts without these requirements.

View Article and Find Full Text PDF