PBR-Inspired Controllable Diffusion for Image Generation

Loading...
Thumbnail Image
Date
2026
Journal Title
Journal ISSN
Volume Title
Publisher
The Eurographics Association and John Wiley & Sons Ltd.
Abstract
Despite recent advances in text-to-image generation, controlling geometric layout and PBR material properties in synthesized scenes remains challenging. We present a pipeline that first produces a G-buffer (albedo, normals, depth, roughness, shading, and metallic) from a text prompt and then renders a final image through a PBR-inspired branch network. This intermediate representation enables fine-grained control: users can copy and paste within specific G-buffer channels to insert or reposition objects, or apply masks to the irradiance channel to adjust lighting locally. As a result, real objects can be seamlessly integrated into virtual scenes. By separating user-friendly scene description from image rendering, our method offers a practical balance between detailed post-generation control and efficient text-driven synthesis. We demonstrate its effectiveness through quantitative evaluations and a user study with 156 participants, showing consistent human preference over strong baselines and confirming that G-buffer control extends the flexibility of text-guided image generation.
Description

CCS Concepts: Computing methodologies → Reflectance modeling; Image-based rendering;

        
@article{
10.1111:cgf.70329
, journal = {Computer Graphics Forum}, title = {{
PBR-Inspired Controllable Diffusion for Image Generation
}}, author = {
Xue, Bowen
and
Guarnera, Giuseppe Claudio
and
Zhao, Shuang
and
Montazeri, Zahra
}, year = {
2026
}, publisher = {
The Eurographics Association and John Wiley & Sons Ltd.
}, ISSN = {
1467-8659
}, DOI = {
10.1111/cgf.70329
} }
Citation