Digital Art of Bas
Transcription
Digital Art of Bas
Max-Planck-Institut für Informatik Computer Graphics Group Saarbrücken, Germany Digital Art of Bas-Relief Sculpting Master Thesis in Computer Science Computer Science Department University of Saarland Jens Kerber Supervisor: Advisor: Prof. Dr. Hans-Peter Seidel (1st reviewer) Prof. Dr. Alexander Belyaev (2nd reviewer) Max-Planck-Institut für Informatik Computer Graphics Group Saarbrücken, Germany Statement Hereby I confirm that this thesis is my own work and that I have documented all sources used. Saarbrücken, July 31, 2007 Jens Kerber Declaration of Consent Herewith I agree that my thesis will be made available through the library of the Computer Science Department. Saarbrücken, July 31, 2007 Jens Kerber Abstract In this thesis, we present a semi-automatic method for the generation of bas-reliefs from a given shape. A Bas-reliefs is an artistic sculptural representation of a three dimensional scene which possesses a negligible depth. The main idea behind our method is to work with a range image, whose depth interval size is compressed in a feature preserving way. Our approach operates in the gradient domain, whereas we also present an extension which works on the Laplacian. The algorithm relies on the achievements of High-Dynamic-RangeCompression and adapts necessary elements to our purpose. We manipulate the partial derivatives of the range image rather than the image data itself. These derivatives are compressed by applying an attenuation function which leads to a relative convergence of the entries. By a feature enhancing technique we boost small details in order to keep them perceivable in the result. In the end, the compressed and enhanced partial derivatives are recombined to the final bas-relief. The results exhibit a very small depth range but still contain perceptually important details. A user can specify the desired compression ratio and steer the relevance of small features in the outcome. The approach is intuitive, fast and works very well for high compression ratios as experiments have shown. Possible applications are of artistic nature like sculpting, embossment, engraving or carving. i ii ABSTRACT Acknowledgement First, I would like to thank Alexander Belyaev for assigning this interesting project to me and for his great supervision. Special thanks to Hans-Peter Seidel for providing an excellent working environment in his research group. I am also grateful to the members of the computer graphics department at the MPII who were at the right place to the right time and gave me helpful suggestions when I got stuck; especially Zhao Dong, Dorotea Dudas, Wenhao Song, Carsten Stoll, Hitoshi Yamauchi and Wenxiang Ying (in alphabetical order). I would like to thank Mathias Pohl and Werner Schwinn who have proofread this thesis. Finally, I am indebted to my family who gave me the opportunity to study and always supported and encouraged me, although they suffered under my absence. iii iv ACKNOWLEDGEMENT Contents Abstract i Acknowledgement iii 1 Introduction 1 1.1 Motivation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1 1.2 Goal . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4 1.3 Fundamentals . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4 1.4 Thesis Outline . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5 2 Related Work 7 2.1 High-Dynamic-Range-Compression . . . . . . . . . . . . . . . . 7 2.2 Bas-Reliefs . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9 2.3 Virtual Shape Decoration . . . . . . . . . . . . . . . . . . . . . . 11 3 Gradient Domain Approach 3.1 13 Algorithm . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13 3.1.1 Overview . . . . . . . . . . . . . . . . . . . . . . . . . . 13 3.1.2 Background Detection . . . . . . . . . . . . . . . . . . . 14 3.1.3 Decomposition . . . . . . . . . . . . . . . . . . . . . . . 15 3.1.4 Thresholding . . . . . . . . . . . . . . . . . . . . . . . . 17 3.1.5 Attenuation . . . . . . . . . . . . . . . . . . . . . . . . . 18 3.1.6 Unsharp Masking . . . . . . . . . . . . . . . . . . . . . . 20 3.1.7 Poisson Problem . . . . . . . . . . . . . . . . . . . . . . 22 3.1.8 Post Processing . . . . . . . . . . . . . . . . . . . . . . . 23 v CONTENTS vi 3.2 Results . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3.3 3.2.1 Parameters . . . . . . . . . . . . . . . . . . . . . . . . . 28 3.2.2 Performance . . . . . . . . . . . . . . . . . . . . . . . . 31 Discussion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 31 4 Laplacian Domain Approach 26 35 4.1 Results . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4.1.1 Performance . . . . . . . . . . . . . . . . . . . . . . . . 38 41 4.2 Discussion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 41 5 Discussion & Conclusion 5.1 Discussion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5.2 Conclusion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 43 43 44 5.3 Future Work . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 44 List of Figures 1.1 Ancient artistic examples . . . . . . . . . . . . . . . . . . . . . . 1 1.2 Modern artistic examples . . . . . . . . . . . . . . . . . . . . . . 2 1.3 Linear rescaling . . . . . . . . . . . . . . . . . . . . . . . . . . . 3 2.1 High-Dynamic-Range-Compression . . . . . . . . . . . . . . . . 8 2.2 Related Work Results . . . . . . . . . . . . . . . . . . . . . . . . 11 3.1 Algorithm workflow . . . . . . . . . . . . . . . . . . . . . . . . 14 3.2 Input data . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 15 3.3 1D Gradient‘ . . . . . . . . . . . . . . . . . . . . . . . . . . . . 16 3.4 2D Gradient images . . . . . . . . . . . . . . . . . . . . . . . . . 16 3.5 Modified gradient . . . . . . . . . . . . . . . . . . . . . . . . . . 18 3.6 Attenuation function . . . . . . . . . . . . . . . . . . . . . . . . 19 3.7 Unsharp masking results . . . . . . . . . . . . . . . . . . . . . . 22 3.8 Bas-relief . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 25 3.9 Distortion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 25 3.10 Armadillo bas-relief details . . . . . . . . . . . . . . . . . . . . . 26 3.11 More gradient domain results I . . . . . . . . . . . . . . . . . . . 27 3.12 More gradient domain results II . . . . . . . . . . . . . . . . . . 28 3.13 Thresholding influence . . . . . . . . . . . . . . . . . . . . . . . 29 3.14 Unsharp masking influence . . . . . . . . . . . . . . . . . . . . . 30 3.15 Low-high frequency relation . . . . . . . . . . . . . . . . . . . . 30 4.1 Laplacian and threshold . . . . . . . . . . . . . . . . . . . . . . . 36 4.2 Attenuation and enhancement . . . . . . . . . . . . . . . . . . . . 37 vii LIST OF FIGURES viii 4.3 Results . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 38 4.4 4.5 Different perspectives . . . . . . . . . . . . . . . . . . . . . . . . More results . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 39 40 List of Tables 3.1 Runtime table gradient approach . . . . . . . . . . . . . . . . . . 31 4.1 Runtime table Laplacian approach . . . . . . . . . . . . . . . . . 41 ix x LIST OF TABLES Chapter 1 Introduction This thesis begins with definitions and an explanation of the problem setting. Then we describe the contribution of our approach. After that we introduce and explain some basics and nomenclature. At the end of this introduction we present a prospect of how this thesis is further structured. 1.1 Motivation Figure 1.1: a Persian relief portraying a hunt (left); a corridor wall of the Buddhistic temple Borobudur in Java, Indonesia (right); both images courtesy of [29] A relief is a sculptured artwork where a modeled form projects out from a flat background [29]. There are two different kinds of reliefs. One is the so called high-relief or alto-relievo which allows that objects may stand out significantly 1 2 CHAPTER 1. INTRODUCTION from the background or are even completely detached from it. This kind of relief is not further considered here. A bas-relief or basso-relievo is carved or embossed into a flat surface and only hardly sticks out of the background. Bas-reliefs are of a small height and give the impression of a three dimensional scene if they are viewed from an orthogonal vantage point. They have been used by artists for centuries to decorate surfaces like stone monuments, coins or vases. Figure 1.1 shows some artwork examples from different cultures. Today, in the time of 3D printers, automatic milling devices and laser carvers, bas-reliefs are used in many different areas like the decoration of dishes and bottles, the production of dies or seals in order to mark products with a company logo and they are still applied in coinage and the creation of modern pieces of art. Figure 1.2 contains some more recent examples. Figure 1.2: modern coin showing Max Planck (left); huge stone carving in Stone Mountain, Georgia, USA (right); both images courtesy of [29] Let us suppose that we are given an arbitrary synthetic model and want to create an embossment of it on a metallic plate. The problem is that in most cases the depth of the model exceeds the thickness of the material. Hence, in order to make this embossment possible, it is necessary to compress the spatial extend of the object to a fractional amount of its initial size. At first glance, this seems to be quite simple, because one could apply global linear rescaling which serves the needs. However, the feature preservation of this methods is very poor. Figure 1.3 shows the Stanford armadillo model [15] and a 1.1. MOTIVATION 3 Chinese liondog [1] as well as their linearly rescaled version and the corresponding results of our approach. In the linearly rescaled case, the viewer can only see the outline and estimate the global shape, which appears to be very flat. Due to the lack of visually important details in the outcome, this naive approach is not applicable. Whereas, our result still contains small features which give a precise impression of the constitution of the object’s surface. Figure 1.3: In the left column the initial model is shown; the middle column contains the results after using the naive linear rescaling approach for a compression to 2% of its former extend; the outcome of our method is shown on the right, note that they are compressed to the same amount; the details in the middle images are almost not perceivable, whereas the reliefs generated by our algorithm exhibit many small features like the different muscle parts and the mail structure for the armadillo or the hair and the claws of the liondog CHAPTER 1. INTRODUCTION 4 This example shows that the solution to the problem of compressing the depth of a model in a feature preserving way is not straightforward. Nevertheless, we present a simple method to overcome this challenge. We do not compress the height field immediately but we change its gradient. A rescaling of the gradient components would not improve the feature preservation either. Therefore we attenuate them first, and boosts their high frequencies afterwards. This enhancement helps us to keep small visually important details perceivable in the result. In the end, the manipulated gradient components are recombined to the final bas-relief. 1.2 Goal Our aim is to provide a semi automatic system which supports an artist in producing bas-reliefs. Given a shape, we want to flatten it without losing the impression as long as it is seen from the same vantage point. Of course, any kind of compression modifies the geometry of a model, but our goal is to create a version of smaller spatial extent which still has a similar appearance. That’s why we have to keep the perceptually relevant details up through the compression process. Our approach can be applied in any area of synthetic or real world shape decoration. Examples are 3D printing, milling, embossment, sculpting, carving or engraving. Nowadays, it is still the case that mints design stamps for the production of coins or medals on a CAD machine by hand from scratch. A use of synthetic data as input is not possible yet, or it is rescaled linearly like it was described above. Our method contributes to bridge this gap by generating virtual prototypes. 1.3 Fundamentals Range images also called height fields or depth maps are a special class of digital images which cover height respectively distance information based on a regular two dimensional grid z = I(x, y). Due to the fact that this information describes shapes in a 3D scene it is often denoted as a 2.5D representation. 1.4. THESIS OUTLINE 5 A range image can be achieved in different ways. One example is ray casting a 3D scene, which means that for every pixel in the projection plane a ray is shot and the distance to the first intersection with the scene is measured. If such an intersection does not exist, the corresponding pixel will be set to a default background value. Z-buffering is a related technique. Here a given scene is rendered with the near and far clipping plane chosen in a way that they tightly enclose the scene, after that the z-buffer, which now contains the distance information, is readout. These approaches only work for virtual scenes. In contrast to that, a 3D scanner can produce depth maps of a real world object from different viewpoints . An important property of height fields is that they consist of a foreground part, containing one or more objects, and a background part which is filled with a default value usually very different from the foreground data. In this thesis we exclusively describe depth compression, which means the reduction of a model’s depth extent. We use the words compression and compressed only in this context. In no case, storage compression of range images, that uses different representations in order to consume less resources, like in [6] is meant. In the following, the term compression ratio describes the relation between the length of the depth interval after and before the processing: compression ratio = M axresult − M inresult M axoriginal − M inoriginal (1.1) Here, M axresult and M inresult represent the upper and lower boundaries of the entries at the object pixels after the depth compression (only foreground pixels count here). M axoriginal and M inoriginal stand for the corresponding extrema of the initial shape. 1.4 Thesis Outline The next chapter contains a survey of the actual state of affairs in the area of basrelief generation and we study the influence of related research fields. Furthermore the importance for different kinds of applications is investigated. In Chapter 3, we describe the different phases of our gradient domain approach and the effects they 6 CHAPTER 1. INTRODUCTION have on the intermediate results. We present several bas-reliefs which have been obtained with the help of this method and analyze how the user can influence the results. After that, the algorithm is discussed in detail and compared to other existing bas-relief generation techniques. Chapter 4 describes how the gradient domain approach is raised to the Laplacian domain. This extension and its differences to the method in Chapter 3 are investigated after the corresponding results are presented. In Chapter 5 we wrap up and give an outlook on future research. Chapter 2 Related Work In this chapter we describe what has been done in the area of bas-relief generation and related research fields so far. In the end we explain how our method can support existing applications for digital shape decoration. 2.1 High-Dynamic-Range-Compression The problem of depth compression for shapes is closely related to High-DynamicRange-Compression (HDR-Compression) which is a hot topic in the area of digital photography. HDR-Images contain luminance values distributed in a very large interval, e.g. a contrast relation in the order of 250.000:1. The goal of HDRCompression, also called tone mapping, is to diminish this interval in order to make it possible to display the corresponding Low-Dynamic-Range Images on regular monitors, which require a contrast relation of at most 1000:1, in the case of an up-to-date TFT-device. It is necessary that visually important details are preserved during the compression process. In the last years this area has been intensively studied, see [9], [10], [27] and [18] for example. Figure 2.1 motivates the problem setting. It shows the HDR images of the Stanford Memorial Church and its corresponding Low-Dynamic-Range results which were obtained by different methods. 7 8 CHAPTER 2. RELATED WORK Figure 2.1: Original HDR-image (courtesy of Paul Debevec) (left); result after applying the method of [27] (middle); outcome of the approach of [11] (right); the details in the windows are very well reproduced in both cases, but the floor and the steps in the right image look more natural; moreover the contrast in the middle left and upper right part of the images is better enhanced by the method of [11] Since Bas-relief generation has to ”squeeze”a shape into the available depth range without destroying perceptually relevant features, it can be regarded as a geometric pendant to tone mapping. Our method mainly relies on the pioneering work of [11]. Their main idea is to work in the gradient domain and manipulate the partial derivatives in a way that large magnitudes are attenuated stronger than low ones. Therefore, they use a multi-resolution approach. In the end, they recombine the gradient components to the new low dynamic range image. Nevertheless, an extension of this approach to range data is not straightforward because there are significant differences between features in digital images and features on shapes. Moreover, HDR-images are more or less continuous, whereas depth maps consist of a foreground and a background part which leads to problems along an object’s silhouette. These discrepancies forced us to perform some radical changes in order to adapt the method to our needs. 2.2. BAS-RELIEFS 9 2.2 Bas-Reliefs The effect that bas-reliefs, viewed from an orthogonal vantage point, are almost indistinguishable from a full 3D scene is called the bas-relief ambiguity. This phenomenon is a matter of human perception which is studied in [3]. The authors proof that there exists a 3-parameter family of transformation from a shape to a corresponding bas-relief such that the shading and shadowing in both cases are identical. As long as a viewer only changes his perspective slightly around an orthogonal view, the impression of a full relief remains. Once a certain vantage point is exceeded, the flattening and also distortion are revealed. If an object is recorded from several angles under unknown lighting conditions, the bas-relief ambiguity will make it hard for a computer to resolve the shape of this object, because the solution is not unique in general, unless assumptions about the surface properties are made. This problem constitutes an active area of research in computer vision [2], [26]. In contrast to that, we exploit the existence of the bas-relief ambiguity because we want to achieve exactly the opposite. Our aim is to create a shape which is different from an original but exhibits the same appearance. The bas-relief ambiguity justifies our motivation that for each model such corresponding shapes do exist. Currently, there are four works which directly address the challenge of generating reliefs with the help of computers. In [7] this problem is studied for the first time. The authors present simple algorithms to produce bas- and high-reliefs. They introduce the idea to work with a view dependent height field from a given 3D scene. First, they generate the range image by z-buffering and then a perspective transformation (reciprocally proportional to the depth) is applied such that objects in the background are mapped to a smaller z-range than those closer to the foreground. This results in a kind of perspective foreshortening. After that, a linear scaling is applied in order to receive the appropriate range in the result. This is the very first work in this field and the results are great for high-reliefs but unfortunately for bas-reliefs and especially for high compression ratios it does hardly better than linear rescaling in terms of CHAPTER 2. RELATED WORK 10 feature preservation. Then, for a long time this research area has completely been out of focus until three works have appeared in the same year. In [23] the authors describe a method to compute a bas-relief of a given mesh. Therefore, they represent the shape in appropriate differential coordinates, then they use a saliency measure [16] and combine it with a feature enhancing technique. In the end, the shape is rescaled and the bas-relief is reconstructed from the corresponding differential coordinates. Inspired by the results of [23], we developed an algorithm for range image compression in a feature preserving way which is presented in [14]. It adapts the main ideas of [11] to our shape processing purpose and operates in the gradient domain. In this work, we have mainly been interested in keeping the surface structure of an object perceivable in the generated depth compressed version. The information about the surface structure is covered in the high frequencies of the partial derivatives. So, we first split the image into its gradient components and extract their high frequencies, which are then rescaled before we recombine them to the depth compressed result. At this time, we have ignored the low frequencies. The results of this method, look quite good but in some sense they seem unnatural and exaggerated. This thesis is an extension of our earlier work. It addresses some open drawbacks (see Section 3.3) and drastically improves the quality of the outcomes.. Figure 2.2 contains a comparison. The work of [28] provides a semi-automatic tool which assists an artist with creating bas-reliefs. Their approach is also based on the ideas presented in [11] and it works in the gradient domain too. A logarithmic attenuation function is used in order to compress large gradients stronger than low ones. Then, they give a user the possibility to treat several frequency bands in the new gradient image individually. This requires the artist to adjust several weight-parameters. At first glance, their approach seems to be quite similar to ours but it differs a lot in some crucial points, as it will be discussed at the end of the next chapter. These current works are an indication for the growing interest in this very young research field. 2.3. VIRTUAL SHAPE DECORATION 11 Figure 2.2: Bas-relief of the armadillo and the lion vase model [1] achieved with the apporach proposed in [23] (left), obtained by the algorithm of [14] (middle) and the result of the method presented in this thesis (right); compression ratio is 2% in all cases 2.3 Virtual Shape Decoration The area of virtual shape decoration covers several techniques, e.g. embossment, engraving, carving or digital sculpting. Its aim is the generation of synthetic pieces of art which cover the same styles, effects and impressions like their real world pendants. In [24] and [25] the authors provide a set of tools and an interactive shape modeler which allows an artist to create synthetic embossments and wood cuttings. The drawback of this method is that the work is completely in the hand of a 12 CHAPTER 2. RELATED WORK user. This is time-consuming and the quality of the results largely depends on the skill of the artist. Another set of tools for synthetic carving is presented in [19]. Among interactive carving their approach allows two and three dimensional input to decorate shapes. The authors restrict themselves to implicit surfaces and explain that, among other techniques, ray casting is used to compute depth data from a given 3D scene but the problem of how to compress this data to an appropriate size for their purpose is not considered. A computer based sculpting system for the creation of digital characters for the entertainment industry is presented in [20]. They require 3D or range data as input, then a user has several editing options to manipulate the outcome. For their algorithm they use adaptively sampled distance fields [12] and exploit their properties. In the first case our algorithm can contribute by acting as a kind of preprocessor, such that range data of synthetic objects can be used to produce a template which can then be further modified, rather than working from scratch. This could improve the quality of the results and would lead to greater success for untrained users. If a model should be carved into another virtual object, then our method can be used to preserve more features in the outcome or even exaggerate them. Once range data is used as input, we can support digital sculpting by requiring less user interaction to produce similar or even better looking results. Chapter 3 Gradient Domain Approach In this chapter we describe the different phases of our gradient domain bas-relief generation method. First, we give a global survey of the algorithm and then we explain the purpose and the contribution of each step in detail. We illustrate those effects using the armadillo model as an example. Later in this chapter, we present several results which have been obtained with the help of our approach. We inspect the influence of the user adjusted parameters and investigate the performance. Finally, the algorithm is discussed and compared to other recent bas-relief generation methods. 3.1 Algorithm 3.1.1 Overview The main idea of our algorithm is to work in the gradient domain like it was proposed in [11]. The workflow shown in Figure 3.1 describes the interplay of the different steps. After having read a given range image I(x, y), it is split into its gradient components Ix and Iy . A problem which naturally arises when working with depth maps forces us to perform a thresholding in order to get rid of boundary artifacts; this leads to I¯x and I¯y . Then, an attenuation function is applied in order to compress the interval size, and we receive Iˆx and Iˆy . The following unsharp masking step allows a user to treat the high frequency band individually. So, an artist can 13 14 CHAPTER 3. GRADIENT DOMAIN APPROACH Figure 3.1: Survey of the different algorithm phases decide whether small details should be boosted or not present in the result at all. At this stage, the new gradient components Jx and Jy of the intermediate result are already computed. Solving a Poisson problem lets us reconstruct J with the help of its partial derivatives. Although this range image now exhibits a very small depth, its interval range is adapted by a global linear scaling in order to make it ˜ suit to the given purpose, and so we receive the final bas-relief J. 3.1.2 Background Detection An input file which has been generated with 3D scanners typically contains a binary background mask in addition to the height data I(x, y) itself. In other cases, e.g. ray casted or z-buffered shapes, such a binary mask B(x, y) is extracted after having read the range image data: B(i, j) = ( 0, if I(i, j) = δ 1, else (3.1) Here, δ represents the default background value of the depth map. It is important that background information does not influence the further computation of the foreground data. This mask helps us to distinguish which part a pixel belongs to and is used to normalize the result in the end. Figure 3.2 shows the initial range image and the extracted background mask for the armadillo model. 3.1. ALGORITHM 15 Figure 3.2: Initial color-coded shape (left); most salient parts are indicated by red and those further to the background are colored blue; the value range is 300 at that stage; corresponding binary background mask (right) 3.1.3 Decomposition In order to work in the gradient domain the partial derivatives Ix and Iy of the height field have to be computed. They are obtained with the help of a differential quotient in two dimensions: I(i + h, j) − I(i, j) h I(i, j + h) − I(i, j) Iy (i, j) ≈ h Ix (i, j) ≈ (3.2) (3.3) In our discrete case it holds that h = 1, and so the equations shown above collapse to a finite forward difference: Ix (i, j) ≈ I(i + 1, j) − I(i, j) (3.4) Iy (i, j) ≈ I(i, j + 1) − I(i, j) (3.5) The fact that the background value is usually very different from the foreground data leads to large peaks at the object’s boundary. In Figure 3.3 a 1D signal is used to illustrate the problem. 16 CHAPTER 3. GRADIENT DOMAIN APPROACH Figure 3.3: 1D signal with boundary and default background value (left) and corresponding first derivative obtained by finite difference (right) In the two dimensional case, which is shown in Figure 3.4, a viewer can only see a thin line (exactly one pixel of width) of large values along the silhouette and all other information inside is close to 0 relative to these discontinuities. Keeping such jumps results in a larger depth interval and the problem that small details will hardly be perceivable in the result. Furthermore, these peaks lead to artifacts during the high frequency computation. In general, this problem cannot be solved by adapting the background value in advance. So, we have to modify Ix and Iy in a way that the outliers are eliminated. Figure 3.4: Gradient components of the armadillo model; X-gradient (left) and Y-gradient (right); both exhibit a very large value range 3.1. ALGORITHM 17 3.1.4 Thresholding This step detaches high values form the gradient images. Therefore, a user defined parameter τ is introduced and all pixels whose absolute value is greater than this threshold are set to 0. First, we generate a binary threshold mask T : T (X, i, j) = ( 1, if |X(i, j)| ≤ τ 0, else (3.6) This mask is then used to cut out the corresponding entries. The -operator means component-wise multiplication in this case: I¯x = T (Ix ) Ix (3.7) I¯y = T (Iy ) Iy (3.8) Note, that this method also affects larger jumps on the object’s surface. If τ is chosen too high, then larger gradients will form the result and smaller details will hardly be visible. If it is chosen too small, then flat artifacts can arise during the reconstruction step, because important information is lost. So, this parameter gives an artist the opportunity to steer which kind of gradients should be removed (respectively remain), but conceals the risk of cutting off too much. The influence that τ has on the outcome is further discussed in the result section. Neither the pixels in the background nor the peaks along the silhouette and on the object’s surface may contribute to the further processing. Therefore we construct a binary combined mask C with the help of B and T for which a pixel is set to 1 if and only if it is equal to 1 in the background mask and in both threshold masks: C = B T (Ix ) T (Iy ) (3.9) Figure 3.5 illustrates the effect of this step on the armadillo model. It shows the thresholded gradient images I¯x and I¯y as well as the corresponding combined mask C. CHAPTER 3. GRADIENT DOMAIN APPROACH 18 Figure 3.5: (left and middle) Thresholded X- and Y-gradient of the armadillo model; in this example τ = 5; note that the details of the armadillo are now visible in contrast to Figure 3.4 and the interval size changes drastically according to τ ; (right) combined binary mask; the positions of detached pixels are marked in the mask, this holds especially for the jaw in this example 3.1.5 Attenuation Up to now, we have produced continuous gradient images. By definition, attenuation is the reduction in amplitude and intensity of a signal [29]. In practice, this means that values of larger magnitudes have to be diminished stronger than small entries. This relative convergence leads to a compression of the remaining interval size by keeping the appearance of the signal. There is a number of different ways in order to achieve attenuation. We have decided to apply the function proposed in [11]. Therefore, we first construct a weight matrix A: A(X, i, j) = 0, a |X(i,j)| · |X(i,j)| a b if X(i, j) = 0 , else (3.10) Then, these weights are applied by component-wise multiplication and we obtain the thresholded and attenuated gradient components Iˆx and Iˆy : Iˆx = A(I¯x ) I¯x Iˆy = A(I¯y ) I¯y (3.11) (3.12) Note, that the gradients in the background and those which were detached before remain unchanged, because zero entries are mapped to zero entries again. 3.1. ALGORITHM 19 This method needs two parameters a and b. It has the effect that absolute values above a are attenuated and those below are slightly enhanced. At first glance, this magnification seems to be counterproductive for the purpose of compression but it comes up with the benefit of preserving small entries by boosting them in that way. Figure 3.6 shows the graph of the attenuation function. Figure 3.6: The parameter ”a”of the attenuation function marks the position of the 1-intersection, it is 10 in this particular example; the parameter ”b”steers how fast the weights approach the x-axis, here it is 0.9 If a depends on the interval-constitution then the whole function will be adaptive because it returns different weights for the same value if it is contained in different intervals. For all images in the following, the reference value a is chosen 10 % of the average absolute value of all unmasked pixels in X. The parameter b which affects the compression ratio is fixed at 0.9 for all results shown in this thesis. So, basically this attenuation step does not need any user interaction. CHAPTER 3. GRADIENT DOMAIN APPROACH 20 3.1.6 Unsharp Masking If a user wants to achieve a very high compression ratio, then it is likely that fine details will hardly be perceivable in the result. The slight enhancement from the previous step is not sufficient to overcome this problem in general. The small features are covered in the high frequencies of the signal and that’s why the high frequency part has to be boosted in order to emphasise the fine details and keep them up through the procedure. Therefore, we use the concept of unsharp masking, which is a classical technique to sharpen images that was even applied in historical photography around the 1930’s [29]. Today unsharp masking is widely used in many areas throughout computer graphics e.g. to enhance features in digital images [17] or meshes [13] [8]. The main idea is to create a smoothed version of a signal and subtract it from the original in order to obtain the high frequent part, which is then linearly scaled by a user specified factor and added back to the low frequency component. If the background pixels and those which have been thresholded before contribute to the blurring, then some pixels in the low frequent part will be of a magnitude which is too small, because there is a number of undesired 0-valued entries in their neighbourhood. Moreover, some of the masked pixels would be given an unpredictable value different from 0, since the smoothing kernel reaches the foreground pixels. Both cases lead to peaks in the extracted high frequency part, because the smoothed version of the signal exhibits values which are either higher or lower than expected. Such discontinuities will cause artifacts in the result, so we have to address this problem. The most important step is the following discrete convolution. For the smoothing itself we use a 2D Gaussian filter Gσ of standard deviation σ and corresponding kernel size. Here, m and n are the indices of the filter in x respectively y direction and C represents the combined binary mask obtained from Equation 3.9: Blurσ (X, i, j) = 0,P if C(i, j) = 0 Gσ (m,n)C(i−m,j−n)X(i−m,j−n) m,n P , Gσ (m,n)C(i−m,j−n) m,n else (3.13) 3.1. ALGORITHM 21 The main idea of this function is that it only considers unmasked pixels in a neighbourhood for the weighted averaging, and like in the previous step it leaves the masked pixels untouched. The denominator is used for normalization, so that the weights which are actually used sum up to 1. If we use the above mentioned function in our special case, we can go on like it is intended by regular unsharp masking. The signals are split into two components Low and High, whereas we apply another slight smoothing to the high frequent part in order to prevent occurring noise. After that, the weight of the high frequent part is modified and it is added back to obtain the new partial derivatives of the depth compressed range image, called Jx and Jy : Low(X) = Blurσ1 (X) (3.14) High(X) = Blurσ2 (X − Low(X)) Jx = Low(Iˆx ) + α · High(Iˆx ) (3.15) Jy = Low(Iˆy ) + α · High(Iˆy ) (3.17) (3.16) The parameter σ1 decides how the signal is decomposed; in our examples the value ranges from 2 to 5. In contrast to that, σ2 only serves to diminish noise and is chosen 1 for all the results in this thesis. The new relation between the high and low frequencies is steered by the parameter α > 0 which it depends on the desired compression ratio or the intention of the artist. α = 1 leaves the balance unchanged, α > 1 enhances the high frequencies whereas α < 1 impairs them. The higher the compression, the higher α has to be set in order to keep the small details up. We want to accentuate that the second smoothing with σ2 is not essential. Strictly spoken, this slight modification even destroys the idea of unsharp masking because we only boost a part of the high frequencies since the highest ones are eliminated. Nevertheless, experiments have shown that the results exhibit some noise in regions with a small number of features and that the details appear to be unnaturally sharp if the second blurring is omitted. Figure 3.7 contains the enhanced gradient images Jx and Jy of the armadillo model. CHAPTER 3. GRADIENT DOMAIN APPROACH 22 Figure 3.7: Intermediate result after unsharp masking; new X-gradient (left) and Y-gradient (right); compare to Figure 3.5; here α = 6, σ1 = 4 and σ2 = 1 3.1.7 Poisson Problem The results of the last step represent the new gradient components of the bas-relief. So far, their discontinuities have been removed, they have been compressed and their high frequencies have been boosted. Now, we reconstruct the depth compressed range image J from its modified h i Jx gradient G = Jy . Since the partial derivatives have been manipulated, G is not integrable anymore, in general. Therefore, we use an optimization process in order to find a function whose derivative is closest G in a least square sense, like it was proposed in [11] and [28]: J(x, y) = arg min J Whereas, ∇ = ∂ ∂x ∂ ∂y Z Z ||∇J(x, y) − G(x, y)||2 dxdy (3.18) represents the vector of partial derivatives. Reformulating the equation shown above leads to the Poisson equation: ∆J = divG 2 2 ∂ ∂ Here, ∆ = ∇2 = ∂x 2 + ∂y 2 denotes the Laplace operator, and div = stands for the divergence operator. (3.19) ∂ ∂x + ∂ ∂y Adapted to our discrete case this means that the Laplacian of J is given by the 3.1. ALGORITHM 23 sum of the partial derivatives of G. Due to the discrete case which we are in, they are obtained by a finite difference again like in Chapter 3.1.3, but here it has to be the backward difffernce in order to produce a central difference like it is defined for the Laplacian: ∂Jx ∂Jy + ∂x ∂y = Jxx + Jyy ∆J = (3.20) (3.21) Jxx (i, j) ≈ Jx (i, j) − Jx (i − 1, j) (3.22) Jyy (i, j) ≈ Jy (i, j) − Jy (i, j − 1) (3.23) (3.24) Now, the task is to compute a function J whose Laplacian is equal to the right hand side of Equation 3.24. The solution to this Poisson equation is a well known and well studied problem and there are standard techniques to address it, so it should not be discussed in detail here. One has to solve a sparse system of linear equations and therefore the Laplace matrix itself and a boundary is required. In our case, the boundary is given by the default background value. As further reading we recommend [4]. This step lets us recombine both partial derivatives, and so it brings us back form the gradient domain to a height field (spatial domain). 3.1.8 Post Processing So far, we have produced a range image which exhibits a greatly reduced depth interval size. Small features are preserved because the relation between high and low frequencies has been adapted by unsharp masking. The above mentioned parameters τ , σ1 , σ2 and α lead to a compression ratio which is unknown in advance. In order to obtain the desired depth range, we introduce a global linear scaling factor λ > 0 which makes the bas-relief suit to 24 CHAPTER 3. GRADIENT DOMAIN APPROACH ˜ the specific purpose. This bas-relief is called J. desired ratio achieved ratio ˜ J = λ·J λ = (3.25) (3.26) Then, we normalize the result with the help of the binary background mask B, in a way that it ranges from 0 to a certain positive amount, whereas the background pixels are set to 0. Finally, we map our depth compressed height field to a regular triangulated grid of appropriate resolution in order to visualize it as a three dimensional surface. That means that the z-value of each vertex is set to the height of the corresponding pixel. The underlying connectivity of the vertices leads to a mesh representation of the bas-relief. The outcomes are shown in Figure 3.8. The careful reader might realize that the color distribution in the left part of Figure 3.8 is very different compared to the one in Figure 3.2. In the original the claws and the nose of the armadillo are almost on the same level whereas this relation is heavily distorted in the result because the nose is much more salient than the claws which, are even further to the background then the breast. In the right image of Figure 3.8 such an effect is not perceivable because, on the one hand it shows an almost orthogonal vantage point, and on the other hand these differences on the surface are only visible if a very small scaling is applied to the z-axis, since it is very flat. Another viewpoint and a different scaling reveal the distortion, see Figure 3.9. 3.1. ALGORITHM 25 Figure 3.8: (left) reconstructed color-coded depth map compressed to 2% of its initial size (6:300) and the corresponding 3D surface representation (right) Figure 3.9: side view of the original model (left); same perspective on the depthcompressed result with different scaling (right) 26 CHAPTER 3. GRADIENT DOMAIN APPROACH 3.2 Results In this section we present several results which demonstrate the potential of our algorithm. We inspect the influence which the user adjusted parameters have on the outcome and analyze the speed of this approach. In Figure 3.10 we present different views on the bas-relief which corresponds directly to the result of the last chapter. Note the preservation of the fine details and the very small height. More results for models of different size and complexity with varying compression ratios are shown in Figure 3.11 and 3.12. Figure 3.10: (top row) For the lower leg one can recognize the fine surface structure as well as the nails; the parts of the mail on the upper leg and the different muscles around the stomach are very well distinguishable; the claws and the details on the inner part are greatly preserved; (bottom row) side view of original range image and three different perspectives on the result after compressing it to 2% of its initial extent 3.2. RESULTS 27 Figure 3.11: (top row) original Stanford dragon model [15] and its compressed version (ratio of 1%); (bottom row) an ornament [1] and its corresponding basrelief compressed to 2% of its former depth, note the preservation of the order, a viewer can recognize the overlapping levels of the model and knows which part is above or behind the other one 28 CHAPTER 3. GRADIENT DOMAIN APPROACH Figure 3.12: a range image of a Buddhistic statue [5] (left) and the depth compressed model on the right (2%); note how well the low frequent features like eyelids nose or mouth and the very high frequent structure on the head are preserved at the same time 3.2.1 Parameters Figure 3.13 shows the influence of the threshold parameter τ . If the value for τ is too high, then discontinuities on the object’s surface will be preserved at the expense of the visibility of small details. A small value for τ can lead to large masked areas which make it hard to reconstruct the range image properly. Hence, a meaningful setting of the threshold is required in order to find a balance between the preservation of visually relevant details and continuity of the result. The following two figures demonstrate the contribution of the unsharp masking step. The bas-reliefs in Figure 3.14 are produced using different values for σ 1 which cause a different degree of smoothness in the outcome. Figure 3.15 demonstrates the influence of α which steers the new relation between the high- and low frequency part. 3.2. RESULTS 29 Figure 3.13: combined masks (top row) and corresponding results (bottom row) for τ = [75, 5, 1] from left to right (σ1 = 4, σ2 = 1, α = 6); in the left images one can see that all inner pixels are taken into account (no difference to background mask) which leads to an exaggeration of the jaw an the claws whereas the visibility of the fine structure is impaired; the small features in the right part are better preserved but the loss of too much information on the coarse structure leads to flat artifacts; this holds especially for the silhouette, the jaw and the claws here; τ = 5 leads to a meaningful relation 30 CHAPTER 3. GRADIENT DOMAIN APPROACH Figure 3.14: σ1 = 15 (left); σ1 = 10 (middle); σ1 = 2 (right); for large values of σ1 the high frequent features are still visible but they appear to be blurred; smaller values emphasise the details but the ridges are very sharp; compare the smoothness of the reflections along the transitions of the stomach and the breast part; in all cases τ = 5, σ2 = 1, α = 6 Figure 3.15: from left to right α = [20, 1, 0.2] (τ = 5, σ1 = 4, σ2 = 1); these images show that one can exaggerate the small details (high reflections around details) by choosing a high boosting factor; for the case of α = 1 it is neutral and the features are visible but very flat (almost no reflections), this corresponds to a result for which only thresholding and attenuation are applied; impairing the high frequencies leads to an even more schematic appearance of the whole model (right) 3.3. DISCUSSION 31 3.2.2 Performance The MATLAB® implementation of our algorithm turns out to be quite fast, although we are sure that it can be further accelerated because we have not paid special attention to efficiency yet. The desired compression ratio and different values for the parameters during the attenuation do not influence the speed. In contrast to that, larger values for the blurring parameter σ1 can cause a larger Gaussian kernel and so the discrete convolution in the unsharp masking step takes significantly longer. If the number of masked pixels, due to a small threshold, is high then the algorithm will be faster because more entries can be ignored during the processing. The reconstruction step is the most time consuming part. Solving the Poisson equation takes more time for larger depth maps, because the system of linear equation grows according to the resolution. For our experiments we use an Intel® Xeon™ 3.06 GHz dual core with 2GB main memory. Table 3.1 shows the runtime for different models excluding fileI/O. Name Angel statue Buddha statue Armadillo Lion Dog Ornament Happy Buddha Stanford Dragon Resolution 200x200 200x200 640x460 535x445 970x680 970x900 970x900 σ1 4 4 4 2 2 5 5 Elapsed Time 0.6 sec 0.6 sec 3.2 sec 2.4 sec 7.1 sec 12.8 sec 19.5 sec Table 3.1: Runtime overview for selected models; Happy Buddha and angel statue results are shown in the next chapter 3.3 Discussion Our gradient domain approach produces bas-reliefs from a given shape in a feature preserving way. The method is simple, fast and leads to a richness of detail in the results even if a very high compression ratio has been achieved. CHAPTER 3. GRADIENT DOMAIN APPROACH 32 The reader may ask why the work in the gradient domain is necessary since almost all operations used are linear. The crucial point is the thresholding which helps to ignore larger jumps on the objects surface. The result of attenuating, unsharp masking and linear rescaling in the spatial domain is absolutely identical to the outcome shown in 3.13 for the reference value τ = 75 where the visibility of the small features is impaired because the discontinuities around the jaw are kept. The distortion described in Section 3.1.8 is caused by the interplay of thresholding and the solution of the Poisson equation for reconstruction. It would arise even if no attenuation, unsharp masking or linear rescaling were performed. In our special case such a distortion is desired because it supports the correct impression of the model from an orthogonal vantage point like it is intended for bas-reliefs. In [28] the authors use a logarithmic weighting function for attenuation, which needs one parameter to steer the compression ratio. The problem with this approach is that the weight only depends on the absolute value of a pixel. This means it returns the same result for a specific magnitude, disregarding its relation to the other entries. Suppose we are given two very different models as input. The thresholding step can cause gradient intervals of the same size for both models although the distribution of values inside of them is very different. If we used logarithmic weights, then the values in both intervals would be treated exactly in the same way. This is why we apply the function proposed in [11] which individually adapts to different intervals. Another difference to the approach presented in [28] is that it gives a user the possibility to decompose the signal into an arbitrary number of different frequency bands, which can then be weighted individually. On the one hand this leads to more artistic control and even allows stop-band-filtering like it is described in [13] but on the other hand it requires the user to find meaningful weights by trial and error. Our method is limited to two different frequency bands, because we used the approach presented in [17] and [8]. This means that stop-band-filtering is not possible, but we can produce good looking results in a much more user friendly, simpler and faster way. As mentioned earlier, the method introduced in [7] is not applicable for high compression ratios because it scales down globally once a height field is gen- 3.3. DISCUSSION 33 erated. This means that, if the range image is compressed too much, then the preservation of small features will be similar to the one achieved by naive linear rescaling. In contrast to that, our algorithm preserves fine details by boosting the high frequencies and locally taking into account the constitution of each neighbourhood. This thesis is an extension of our earlier work [14] in which we only concentrated on the high frequencies and ignored the low ones. This has caused several problems in areas which exhibit only a small number of features, e.g. spherical parts. Depending on the parameters, these regions have either been noisy or completely flattened. Hence, we include the low frequencies now, and obtain results which look much more natural because the convexity in such areas is preserved. Moreover, in [14] we have used global blurring, which has led to exaggerated large and sharp ridges and other undesired artifacts, because it introduces discontinuities at locations where entries have been thresholded, as it is explained in Section 3.1.6. Currently we apply the discrete convolution function described in Equation 3.13, in order to prevent those peaks. 34 CHAPTER 3. GRADIENT DOMAIN APPROACH Chapter 4 Laplacian Domain Approach For the gradient domain approach it is necessary to compute the Laplacian of the range image in order to get back to the spatial domain. That’s why we have thought of manipulating the Laplacian directly. To the best of our knowledge, for the purpose of depth compression of shapes this has not been done before. The idea behind each step is the same like in Chapter 3, but the effects which they have are different because the Laplacian and the gradient represent different properties of the model. In order to not repeat what has been described before, we keep this chapter more technical and only demonstrate the differences to the gradient domain approach. Therefore we use the armadillo model here again. Let the height field I(x, y) be given and the binary background mask B(x, y) be extracted like described in Section 3.1.2. We begin with the computation of the Laplacian ∆I of the range image. By definition ∆I is the sum of the second derivatives in both dimensions. ∆I = Ixx + Iyy I(i + h, j) − 2I(i, j) + I(i − h, j) Ixx (i, j) ≈ h2 I(i, j + h) − 2I(i, j) + I(i, j − h) Iyy (i, j) ≈ h2 Since the range images are discrete it holds again that h = 1 in our case. 35 (4.1) (4.2) (4.3) CHAPTER 4. LAPLACIAN DOMAIN APPROACH 36 Reformulating the above mentioned equations under consideration of this fact leads to the following approximation: 0 1 0 ∆I ≈ 1 −4 1 ⊗ I 0 1 0 (4.4) This 3x3 filter and the large differences between foreground and background data lead to the fact that the peaks along the silhouette have a width of 2 pixels now. Like in Section 3.1.4 we use thresholding with reference value τ to eliminate those outliers and obtain ∆I. B(x, y) is extended to the combined mask C(x, y) accordingly. ∆I(i, j) = ( 1, if |∆I(i, j)| ≤ τ 0, else (4.5) Figure 4.1 shows the Laplacian of the armadillo depth map before and after thresholding. Figure 4.1: Initial Laplacian of the armadillo range image (left) and the corresponding thresholded version (right) c Then we apply the attenuation function described in Section 3.1.5 to get ∆I, whereas the parameter a depends on the signal constitution and b is chosen 0.9 again. 37 c j) = ∆I(i, 0, a |∆I(i,j)| · |∆I(i,j)| a b if ∆I(i, j) = 0 , else (4.6) This step is followed by unsharp masking which boosts the high frequencies and produces the new Laplacian ∆J for the bas-relief. Therefore, the discrete convolution Blurσ , which has been introduced in Section 3.1.6, is used. We omit the additional smoothing of the high frequency component here because the visibility of very fine details in the result is impaired even for a slight blurring. The noise reduction which could be achieved by the second smoothing is hardly perceivable as experiments have shown, and so it plays a minor role here. c = Blurσ (∆I) c Low(∆I) c = ∆I c − Low(∆I) c High(∆I) c + α · High(∆I) c ∆J = Low(∆I) (4.7) (4.8) (4.9) The intermediate results after attenuation and unsharp masking are presented in Figure 4.2. Figure 4.2: Thresholded and attenuated signal (left); corresponding image after unsharp masking (σ = 4, alpha = 100) CHAPTER 4. LAPLACIAN DOMAIN APPROACH 38 Due to the fact that we have produced a new Laplacian directly, the reconstruction of J can be done by solving the Poisson equation ∆J = f immediately. Figure 4.3 contains the reconstructed, rescaled and normalized height field in a color-coded way as well as the corresponding surface representation (see Section 3.1.8 for details). Figure 4.3: Color-coded bas-relief (left) and its surface representation (right) 4.1 Results The color distribution in the left part of Figure 4.3 shows that the generated bas relief exhibits a heavy bending. This is the counterpart to the distortion in the gradient domain approach. The bending is caused by the thresholding step and the solution of the Poisson equation for reconstruction. In our case this bending is undesired because it elevates the center of an object too far from the background. Figure 4.4 contains different views on the bas-relief. In Figure 4.5 we compare two more results of this approach with those obtained by the gradient domain method. The values for the threshold τ and the blurring parameter σ are similar to the ones for the gradient method (τ ≈ 5, σ ≈ 4 ), whereas the enhancement factor α is larger than 100 for the results presented in this chapter. 4.1. RESULTS 39 Figure 4.4: All three views show that the middle of the bas-relief is significantly elevated, whereas the ears and feet stand out only slightly from the background. Especially the tail which normally belongs to the background plane is affected by this undesired rising towards the object’s center 40 CHAPTER 4. LAPLACIAN DOMAIN APPROACH Figure 4.5: Happy Buddha model [15] compressed with gradient domain approach (left) and the corresponding bas-relief obtained by Laplacian method (middle); depth compressed version of a kneeling angel statue [5] achieved with gradient (top right) and Laplacian algortihm (bottom right); all models are compressed to 2% of their former spatial extend; note that the ends of the rope as well as the rope itself, the chain, the contours of the coat and the facial expression can be recognized for the Buddha model; the preservation of the fine structure at the wings and the hair are striking in the angel example; The Laplacian images cover all these details but the overall appearance of the gradient domain results is more plastic and natural 4.2. DISCUSSION 41 4.1.1 Performance Due to the fact that we only work on one image (Laplacian), we achieve an enormous increase in speed, compared to the work on two different components. Thresholding, attenuation and unsharp masking have to be applied only once, which saves almost half of the runtime. The computation of the Laplacian only requires one convolution instead of four finite differences. Table 4.1 shows the computation times for the results presented in this section (file-I/O excluded; Intel® Xeon™ 3.06 GHz dual core). Name Angel Armadillo Happy Buddha Resolution 200x200 640x460 970x900 σ 4 4 5 Elapsed Time 0.3 sec 2.0 sec 7.5 sec Table 4.1: Runtime overview for selected models 4.2 Discussion All in all, the feature preservation of this Laplacian algorithm is acceptable but the appearance of the model in the final bas-relief is not as good as the results obtained by the gradient domain approach. For high compression ratios the bending itself is not visible from an orthogonal vantage point, but if the compression is relatively low, then the reflections on the surface will reveal this deformation. This Laplacian domain approach is an extension of an algorithm which is pretty novel itself, hence the current state is only preliminary. A different method to overcome the boundary problem at the object’s silhouette is required in order to produce evenly flat results. We are still searching for an appropriate solution but we could not find a satisfying one until the deadline for this thesis. The virtual results in both of our approaches look very promising, but there will be two grave practical problems if these results should be ”brought to real life” with the help of a CNC machine, a milling device, or a laser carver. (1) The tools which are used e.g. by a milling device, have a certain spatial extend so that very small details either cannot be produced or will be destroyed again during the process. 42 CHAPTER 4. LAPLACIAN DOMAIN APPROACH (2) The crucial point is the fact that our generated height fields are discrete. A CNC machine needs to know the transition from one entry to the next. These can be e.g. linear, parabolic, cubic or a circular arc of given radius. So, if our depth compressed range images serve as input, one has to decide in advance which transition should be applied where; or a spline curve has to be fitted through all discrete points of a row for each column. But then, it cannot be guaranteed that the effects and details which are visible in the synthetic results also arise on the material in the outcome. The authors of [28] have produced a limestone sculpture but due to the above mentioned reasons, it exhibits only a small number of coarse features compared to their virtual results. Chapter 5 Discussion & Conclusion 5.1 Discussion The approach presented in [28] is currently the most flexible one for the generation of bas-reliefs. This method offers a number of possible artistic effects and the user is given a lot of freedom, e.g one can use an arbitrary number of frequency bands and change their relative weights. But this freedom requires either an experienced artist or a trial and error adjustment for the parameters in case of an untrained user. The gradient domain approach presented in this thesis is restricted in the number of possible artistic effects so far. Nevertheless, it produces bas-reliefs which can at least compete with the ones achieved by the algorithm presented in [28], it is reasonable fast and much more user friendly. The method presented in [14] is, in some sense, a subset of our current gradient domain algorithm. It focuses only on high frequencies which leads to either slightly noisy or exaggerated results. Depending on the model or the artists intention this method is absolutely sufficient. Its advantages are the simplicity and the higher speed. The Laplacian approach from Chapter 4 is still under development. All in all, the results are not yet satisfying but promising and they require only short computation times. The look of the outcomes of the gradient algorithm is still more natural than the one of the bas-reliefs generated by the Laplacian method. 43 CHAPTER 5. DISCUSSION & CONCLUSION 44 The pioneering work of [7] is not applicable if high compression ratios are needed because then the feature preservation is hardly better than the one of linear rescaling. Nevertheless their method is simpler and faster than any other approach in this area. 5.2 Conclusion In this thesis we have described a method which assists a user with the generation of bas-reliefs from a given shape. Furthermore, we have shown how this approach, which operates on the gradient, can be raised to the Laplacian domain. The artist has the possibility to steer the relative attenuation between coarse and fine details in order to control their presence in the outcome. For an orthogonal vantage point we can achieve arbitrary compression ratios and preserve visually important details at the same time. Due to the fact that this problem is related to High-Dynamic-RangeCompression we have adapted several ideas from tone mapping to our purpose. Our algorithm is simple, fast, easy to implement and independent of the model’s complexity. Possible applications are any kind of synthetic or real world shape decoration like embossment, engraving, carving or sculpting. This is a very young and interesting research field which currently receives more attention. The results in this area look very promising but there is still a number of possible extensions. 5.3 Future Work So far, the three dimensional scenes which are used as examples, only consist of one object. For a very complex or even panoramic scene with many objects, z-buffering from one perspective camera or orthogonal ray-casting can leads to distortion, bending and other undesired artifacts. For a sequence of two dimensional images this problem is e.g. studied in [21] and [22]. That’s why the use of multi-perspective techniques constitutes a direction for future research. We expect a huge difference between the results obtained by a regular one camera capturing 5.3. FUTURE WORK 45 of a scene and the outcomes achieved with a multi-view approach. It also offers the opportunity to make hidden objects visible by having a camera pointing behind an occluding object. Multiple viewpoints have been applied in paintings for a very long time e.g. by Pablo Picasso. In our case multi-perspective methods would give more respect to human perception on the one hand and provide many more possible artistic effects on the other hand. The design of a hybrid approach which keeps and enhances gradient details as well as properties of the Laplacian and combines both methods is another possible extension. Therefore, we want to improve the work in the Laplacian domain by applying a different silhouette treatment. Using and developing other techniques to extract and treat the frequency bands, e.g. by multi-resolution methods, would give the user even more control on the outcome and offer further artistic effects as well. In contrast to that, we also think of reducing the user intervention by developing a quality measure which takes into account the properties of the generated bas-reliefs and those of the original height field. This could help to introduce adaptive parameters for the functions used inside of our algorithm and hence make the process more automatic. 46 CHAPTER 5. DISCUSSION & CONCLUSION Bibliography [1] AIM@SHAPE. shape.net/. AIM@SHAPE shape repository. http://shapes.aim-at- [2] Neil Alldrin, Satya Mallick, and David Kriegman. Resolving the generalized bas-relief ambiguity by entropy minimization. In IEEE Conf. on Computer Vision and Pattern Recognition, June 2007. [3] Peter N. Belhumeur, David J. Kriegman, and Alan L. Yuille. The bas-relief ambiguity. Int. J. Comput. Vision, 35(1):33–44, 1999. [4] Dietrich Braess. Finite Elements: Theory, Fast Solvers, and Applications in Solid Mechanics. Cambridge University Press, 2002. second edition. [5] Richard J. Campbell and Patrick J. Flynn. A www-accessible 3D image and model database for computer vision research. In Empirical Evaluation Methods in Computer Vision, K.W. Bowyer and P.J. Phillips (eds.), pages 148–154, 1998. http://sampl.ece.ohio-state.edu/data/3DDB/index.htm. [6] Richard J. Campbell and Patrick J. Flynn. Experiments in transform-based range image compression. In ICPR ’02: Proceedings of the 16 th International Conference on Pattern Recognition (ICPR’02) Volume 3, page 30875, 2002. [7] Paolo Cignoni, Claudio Montani, and Roberto Scopigno. Automatic generation of bas- and high-reliefs. Journal of Graphics Tools, 2(3):15–28, 1997. [8] Paolo Cignoni, Roberto Scopigno, and Marco Tarini. A simple normal enhancement technique for interactive non-photorealistic renderings. Computer & Graphics, 29(1):125–133, 2005. [9] Paul Debevec and Erik Reinhard. High-dynamic-range imaging: Theory and applications. SIGGRAPH 2006 Course #5, 2006. http://www.siggraph.org/s2006/main.php?f=conference&p=courses&s=5. 47 48 BIBLIOGRAPHY [10] Frédo Durand and Julie Dorsey. Fast bilateral filtering for the display of high-dynamic-range images. In ACM Transactions on Graphics (Proc. SIGGRAPH 2002), pages 257–266, 2002. [11] Raanan Fattal, Dani Lischinski, and Michael Werman. Gradient domain high dynamic range compression. In ACM Transactions on Graphics (Proc. SIGGRAPH 2002), pages 249–256, 2002. [12] Sarah F. Frisken, Ronald N. Perry, Alyn P. Rockwood, and Thouis R. Jones. Adaptively sampled distance fields: a general representation of shape for computer graphics. In ACM Transactions on Graphics (Proc. SIGGRAPH 2000), pages 249–254, 2000. [13] Igor Guskov, Wim Sweldens, and Peter Schröder. Multiresolution signal processing for meshes. In SIGGRAPH ’99: Proceedings of the 26th annual conference on Computer graphics and interactive techniques, pages 325– 334, 1999. [14] Jens Kerber, Alexander Belyaev, and Hans-Peter Seidel. Feature Preserving Depth Compression of Range Images. In SCCG ’07: Proceedings of the 23rd spring conference on Computer graphics, April 2007. Winner 2nd best SCCG ’07 paper award. Post-conference proceedings are in print. [15] Stanford Computer Graphics Laboratory. Stanford 3D scanning repository. http://graphics.stanford.edu/data/3Dscanrep/. [16] Chang Ha Lee, Amitabh Varshney, and David W. Jacobs. Mesh saliency. In ACM Transactions on Graphics (Proc. SIGGRAPH 2005), pages 659–666, 2005. [17] Thomas Luft, Carsten Colditz, and Oliver Deussen. Image enhancement by unsharp masking the depth buffer. In ACM Transactions on Graphics (Proc. SIGGRAPH 2006), pages 1206–1213, 2006. [18] Rafał Mantiuk, Karol Myszkowski, and Hans-Peter Seidel. A perceptual framework for contrast processing of high dynamic range images. In APGV ’05: Proceedings of the 2nd symposium on Appied perception in graphics and visualization, pages 87–94, 2005. [19] Alexander A. Pasko, Vladimir Savchenko, and Alexei Sourin. Synthetic carving using implicit surface primitives. Computer-Aided Design, Elsevier, 33(5):379–388, 2001. BIBLIOGRAPHY 49 [20] Ronald N. Perry and Sarah F. Frisken. Kizamu: a system for sculpting digital characters. In ACM Transactions on Graphics (Proc. SIGGRAPH 2001), pages 47–56, 2001. [21] Augusto Román, Gaurav Garg, and Marc Levoy. Interactive design of multiperspective images for visualizing urban landscapes. In VIS ’04: Proceedings of the conference on Visualization ’04, pages 537–544, 2004. [22] Augusto Román and Hendrik P. A. Lensch. Automatic multiperspective images. In Rendering Techniques 2006: Eurographics Symposium on Rendering, pages 161–171, 2006. [23] Wenhao Song, Alexander Belyaev, and Hans-Peter Seidel. Automatic generation of bas-reliefs from 3D shapes. In SMI ’07: Proceedings of the IEEE International Conference on Shape Modeling and Applications 2007, pages 211–214, June 2007. [24] Alexei Sourin. Functionally based virtual computer art. In SI3D ’01: Proceedings of the 2001 symposium on Interactive 3D graphics, pages 77–84, 2001. [25] Alexei Sourin. Functionally based virtual embossing. The Visual Computer, Springer, 17(4):258–271, 2001. [26] Ping Tan, Satya P. Mallick, Long Quan, David Kriegman, and Todd Zickler. Isotropy, reciprocity and the generalized bas-relief ambiguity. In IEEE Conf. on Computer Vision and Pattern Recognition, June 2007. [27] Jack Tumblin and Greg Turk. Lcis: a boundary hierarchy for detailpreserving contrast reduction. In ACM Transactions on Graphics (Proc. SIGGRAPH 1999), pages 83–90, 1999. [28] Tim Weyrich, Jia Deng, Connelly Barnes, Szymon Rusinkiewicz, and Adam Finkelstein. Digital bas-relief from 3D scenes. To appear in ACM SIGGRAPH, August 2007. [29] Wikipedia. Wikipedia, the free encyclopedia, 2007. http://en.wikipedia.org.