Deep learning in volume rendering
Loading...
Date
2024-11-11
Authors
Engel, Dominik
Journal Title
Journal ISSN
Volume Title
Publication Type
Published in
Abstract
A variety of scientific fields commonly acquire volumetric data, for example by using computed tomography (CT) or magnetic resonance imaging (MRI) in medicine. Such volumetric data is often complex and requires visualization to gain understanding. However, rendering volumetric data entails several challenges, such as classifying the data to reveal the structures of interest. Meaningful classification of 3D data is difficult to implement in 2D graphical user interfaces. Furthermore, rendering of volumetric data is generally compute intensive, especially when considering volumetric shading. Lastly, volume rendering needs to be interactive and achieve reasonable frame rates in order to support fully exploring the 3D data from different views, while adapting the classification. To combat these challenges, this work explores how volume rendering and its individual aspects can be assisted by means of deep neural networks. Deep neural networks have recently proven very competent in many disciplines, like natural language processing, vision, but also computer graphics. They excel in the approximation of complex functions and can learn relevant features and representations when trained with sufficient data. In the scope of this dissertation, we show how these capabilities can be used throughout the volume rendering pipeline. This pipeline consists of the classification of structures of interest, shading of those structures and the composition of the light transmitted through the volume. In the classification step, we leverage the strong representations learned by self-supervised neural networks to enable an interactive click-to-select workflow that segments structures annotated by users within slice views. For shading, we propose a volume-to-volume network to predict volumetric ambient occlusion that respects how the volume is classified. Lastly, we employ a deep neural network to invert the composition step, separating different structures in an already composited semi-transparent volume rendered image into a modifiable layered representation.
In reference to IEEE copyrighted material which is used with permission in this thesis, the IEEE does not endorse any of [university/educational entity's name goes here]'s products or services. Internal or personal use of this material is permitted. If interested in reprinting/republishing IEEE copyrighted material for advertising or promotional purposes or for creating new collective works for resale or redistribution, please go to http://www.ieee.org/publications_standards/publications/rights/rights_link.html to learn how to obtain a License from RightsLink.
Description
Faculties
Fakultät für Ingenieurwissenschaften, Informatik und Psychologie
Institutions
Citation
DFG Project uulm
EU Project THU
Other projects THU
License
Lizenz A
Is version of
Has version
Supplement to
Supplemented by
Has erratum
Erratum to
Has Part
D. Engel and T. Ropinski, "Deep Volumetric Ambient Occlusion," in IEEE Transactions on Visualization and Computer Graphics, vol. 27, no. 2, pp. 1268-1278, Feb. 2021, doi: https://doi.org/10.1109/TVCG.2020.3030344
D. Engel, S. Hartwig and T. Ropinski, "Monocular Depth Decomposition of Semi-Transparent Volume Renderings," in IEEE Transactions on Visualization and Computer Graphics, vol. 30, no. 7, pp. 3981-3994, July 2024, doi: https://doi.org/10.1109/TVCG.2023.3245305
D. Engel, L. Sick and T. Ropinski, "Leveraging Self-Supervised Vision Transformers for Segmentation-based Transfer Function Design," in IEEE Transactions on Visualization and Computer Graphics, doi: https://doi.org/10.1109/TVCG.2024.3401755
D. Engel, S. Hartwig and T. Ropinski, "Monocular Depth Decomposition of Semi-Transparent Volume Renderings," in IEEE Transactions on Visualization and Computer Graphics, vol. 30, no. 7, pp. 3981-3994, July 2024, doi: https://doi.org/10.1109/TVCG.2023.3245305
D. Engel, L. Sick and T. Ropinski, "Leveraging Self-Supervised Vision Transformers for Segmentation-based Transfer Function Design," in IEEE Transactions on Visualization and Computer Graphics, doi: https://doi.org/10.1109/TVCG.2024.3401755
