Smarter screen space shading

Eurographics DL Repository

Show simple item record

dc.contributor.author Nalbach, Oliver
dc.date.accessioned 2018-01-15T13:26:15Z
dc.date.available 2018-01-15T13:26:15Z
dc.date.issued 2017-11-10
dc.identifier.uri https://diglib.eg.org:443/handle/10.2312/2632112
dc.description.abstract This dissertation introduces a range of new methods to produce images of virtual scenes in a matter of milliseconds. Imposing as few constraints as possible on the set of scenes that can be handled, e.g., regarding geometric changes over time or lighting conditions, precludes pre-computations and makes this a particularly difficult problem. We first present a general approach, called deep screen space, using which a variety of light transport aspects can be simulated within the aforementioned setting. This approach is then further extended to additionally handle scenes containing participating media like clouds. We also show how to improve the correctness of deep screen space and related algorithms by accounting for mutual visibility of points in a scene. After that, we take a completely different point of view on image generation using a learning-based approach to approximate a rendering function. We show that neural networks can hallucinate shading effects which otherwise have to be computed using costly analytic computations. Finally, we contribute a holistic framework to deal with phosphorescent materials in computer graphics, covering all aspects from acquisition of real materials, to easy editing, to image synthesis. en_US
dc.language.iso en en_US
dc.title Smarter screen space shading en_US
dc.type Thesis en_US


Files in this item

Item/paper (currently) not available via TIB Hannover.

This item appears in the following Collection(s)

Show simple item record

Search Eurographics DL


Browse

My Account