Departments  

Book Series (78) 
1097

Medienwissenschaften 
10

Humanities 
2027

Natural Sciences 
5127

Mathematics  216 
Informatics  291 
Physics  957 
Chemistry  1296 
Geosciences  124 
Human medicine  225 
Stomatology  9 
Veterinary medicine  82 
Pharmacy  141 
Biology  791 
Biochemistry, molecular biology, gene technology  107 
Biophysics  23 
Domestic and nutritional science  40 
Agricultural science  967 
Forest science  197 
Horticultural science  15 
Environmental research, ecology and landscape conservation  131 
Engineering 
1567

Common 
82

Leitlinien Unfallchirurgie
5. Auflage bestellen 
Extract, PDF (680 KB)
Table of Contents, PDF (520 KB)
ISBN13 (Hard Copy)  9783736999978 
ISBN13 (eBook)  9783736989979 
Language  English 
Page Number  124 
Lamination of Cover  matt 
Edition  1. 
Publication Place  Göttingen 
Place of Dissertation  Braunschweig 
Publication Date  20190415 
General Categorization  Dissertation 
Departments 
Mathematics
Applied mathematics 
Keywords  Angewandte Mathematik, Optimierung, Variationsmethoden, mathematische Bildverarbeitung, Deflektometrie, PrimalDual Algorithmus, Totalvariation, verallgemeinerte Totalvariation, Regularisierungen, Dualitätslücke, applied mathematics, optimization, variational methods, mathematical imaging, denoising, deflectometry, primaldual algorithm, total variation, total generalized variation, regularization, primaldual gaps 
URL to External Homepage  https://www.tubraunschweig.de/iaa/personal/komander 
Consider manufactured parts, such as screws, car doors, lenses, or mirrors for lasers, for example. All these manufactured parts have to go through quality inspections checking if there are unwanted bumps or scratches that should not be there. There are different methods to measure the manufactured parts. The goal is to describe the measured object exactly by the data. This is one example for a socalled inverse problem.
The application that we consider in the first part of this thesis is a data fusion process. The given dataset is a result of a deflectometric measurement process. These processes deal with object measurements of specular objects, such as lenses or mirrors. The aim is to calculate a dataset that describes the measured object exactly. The output data consists of two sets of separately measured types of data which suffer from different accuracy due to different sensitivity to noise. We resolve this issue with a data fusion process by solving a minimization problem which uses the more accurat data as a reference value and by duing so increases the other.
Taking the gained insights we are able to develop new theories for image denoising. We formulate certain minimization problems in which use suitable reference values. In image denoising the reference value we want to use is an approximation of the image gradient vectors. Consequently, our approaches calculate such an approximation and use it as a reference value. Hence, our approaches are twostage methods. Another approach to prevent the staircasing effect is to go to higher orders of differentiation within the regularization term. One approach was proposed in 2010 namely the total generalized variation (TGV) functional. We propose different kinds of combinations of these functionals. In this way we are able to formulate different minimization problems that are in some sense equivalent to the TGV problem. One advantage of some of these problems lies in the easy parameter choice rules that perform equally well as the TGV problem. Additionally, the duality gaps of these new problems are finite instead of infinite as it is usually the case in the primaldual gap for the TGV problem. Hence, these can be used to create a reasonable stopping criterion for the optimization process. An additional advantage is the decreased runtime of the twostage methods, since the problem is divided into two smaller problems.