Gao / Dai | View-based 3-D Object Retrieval | E-Book | sack.de
E-Book

E-Book, Englisch, 154 Seiten

Reihe: Computer Science Reviews and Trends

Gao / Dai View-based 3-D Object Retrieval


1. Auflage 2014
ISBN: 978-0-12-802623-6
Verlag: Elsevier Science & Techn.
Format: EPUB
Kopierschutz: 6 - ePub Watermark

E-Book, Englisch, 154 Seiten

Reihe: Computer Science Reviews and Trends

ISBN: 978-0-12-802623-6
Verlag: Elsevier Science & Techn.
Format: EPUB
Kopierschutz: 6 - ePub Watermark



Content-based 3-D object retrieval has attracted extensive attention recently and has applications in a variety of fields, such as, computer-aided design, tele-medicine,mobile multimedia, virtual reality, and entertainment. The development of efficient and effective content-based 3-D object retrieval techniques has enabled the use of fast 3-D reconstruction and model design. Recent technical progress, such as the development of camera technologies, has made it possible to capture the views of 3-D objects. As a result, view-based 3-D object retrieval has become an essential but challenging research topic. View-based 3-D Object Retrieval introduces and discusses the fundamental challenges in view-based 3-D object retrieval, proposes a collection of selected state-of-the-art methods for accomplishing this task developed by the authors, and summarizes recent achievements in view-based 3-D object retrieval. Part I presents an Introduction to View-based 3-D Object Retrieval, Part II discusses View Extraction, Selection, and Representation, Part III provides a deep dive into View-Based 3-D Object Comparison, and Part IV looks at future research and developments including Big Data application and geographical location-based applications. - Systematically introduces view-based 3-D object retrieval, including problem definitions and settings, methodologies, and benchmark testing beds - Discusses several key challenges in view-based 3-D object retrieval, and introduces the state-of-the-art solutions - Presents the progression from general image retrieval techniques to view-based 3-D object retrieval - Introduces future research efforts in the areas of Big Data, feature extraction, and geographical location-based applications

Yue Gao is with the Department of Automation, Tsinghua University. His recent research focuses on the areas of neuroimaging, multimedia and remote sensing. He is a senior member of IEEE.

Gao / Dai View-based 3-D Object Retrieval jetzt bestellen!

Autoren/Hrsg.


Weitere Infos & Material


1;Front Cover;1
2;View-Based 3-D Object Retrieval;4
3;Copyright;5
4;Contents;6
5;Acknowledgments;8
6;Preface;10
7;Part I: The Start;12
7.1;Chapter 1: Introduction;14
7.1.1;1.1 The Definition of 3DOR;14
7.1.2;1.2 Model-Based 3DOR Versus View-Based 3DOR;15
7.1.3;1.3 The Challenges of V3DOR;16
7.1.4;1.4 Summary of Our Work;17
7.1.4.1;1.4.1 View Extraction;18
7.1.4.2;1.4.2 Representative View Selection;18
7.1.4.3;1.4.3 Learning the Weights for Multiple Views;19
7.1.4.4;1.4.4 Distance Measures for Object Matching;20
7.1.4.5;1.4.5 Learning the Relevance Among 3-D Objects;21
7.1.5;1.5 Structure of This Book;23
7.1.6;1.6 Summary;25
7.1.7;References;25
7.2;Chapter 2: The Benchmark and Evaluation;28
7.2.1;2.1 Introduction;28
7.2.2;2.2 The Standard Benchmarks;28
7.2.3;2.3 The Shape Retrieval Contest;36
7.2.4;2.4 Evaluation Criteria in 3DOR;43
7.2.5;2.5 Summary;46
7.2.6;References;46
8;Part II View Extraction, Selection, and Representation;50
8.1;Chapter 3: View Extraction;52
8.1.1;3.1 Introduction;52
8.1.2;3.2 Dense Sampling Viewpoints;52
8.1.3;3.3 Predefined Camera Array;53
8.1.4;3.4 Generated View;56
8.1.5;3.5 Summary;61
8.1.6;References;61
8.2;Chapter 4: View Selection;62
8.2.1;4.1 Introduction;62
8.2.2;4.2 Unsupervised View Selection;63
8.2.3;4.3 Interactive View Selection;65
8.2.3.1;4.3.1 Multiview 3-D Object Matching;66
8.2.3.2;4.3.2 View Clustering;68
8.2.3.3;4.3.3 Initial Query View Selection;68
8.2.3.4;4.3.4 Interactive View Selection with User Relevance Feedback;70
8.2.3.5;4.3.5 Learning a Distance Metric;71
8.2.3.6;4.3.6 Multiple Query Views Linear Combination;73
8.2.3.7;4.3.7 The Computational Cost;75
8.2.4;4.4 Summary;75
8.2.5;References;76
8.3;Chapter 5: View Representation;78
8.3.1;5.1 Introduction;78
8.3.2;5.2 Shape Feature Extraction;79
8.3.2.1;5.2.1 Zernike Moments;79
8.3.2.2;5.2.2 Fourier Descriptor;80
8.3.3;5.3 The Bag-of-Visual-Features Method;81
8.3.3.1;5.3.1 The Bag-of-Visual-Words;81
8.3.3.2;5.3.2 The Bag-of-Region-Words;82
8.3.4;5.4 Learning the Weights for Multiple Views;85
8.3.4.1;5.4.1 K-Partite Graph Reinforcement;86
8.3.4.2;5.4.2 Weight Learning for Multiple Views Usingthe k-Partite Graph;90
8.3.5;5.5 Summary;92
8.3.6;References;92
9;Part III View-Based 3-D Object Comparison;96
9.1;Chapter 6: Multiple-View Distance Metric;98
9.1.1;6.1 Introduction;98
9.1.2;6.2 Fundamental Many-to-Many Distance Measures;99
9.1.3;6.3 Bipartite Graph Matching;101
9.1.3.1;6.3.1 View Selection and Weighting;102
9.1.3.2;6.3.2 Bipartite Graph Construction;104
9.1.3.3;6.3.3 Bipartite Graph Matching;104
9.1.4;6.4 Statistical Matching;104
9.1.4.1;6.4.1 Adaptive View Clustering;105
9.1.4.2;6.4.2 CCFV;105
9.1.4.2.1;6.4.2.1 View Clustering and Query Model Training;106
9.1.4.2.2;6.4.2.2 Positive and Negative Matching Models;107
9.1.4.2.3;6.4.2.3 Calculation of the Similarity Between Q and O S(Q,O);108
9.1.4.2.4;6.4.2.4 Analysis of Computational Cost;110
9.1.4.3;6.4.3 Markov Chain;111
9.1.4.4;6.4.4 Gaussian Mixture Model Formulation;112
9.1.4.4.1;6.4.4.1 Conventional GMM Training;112
9.1.4.4.2;6.4.4.2 Generative Adaptation of GMM;114
9.1.4.4.3;6.4.4.3 Discriminative Adaptation of GMM;115
9.1.4.4.4;6.4.4.4 Learning the Weights for Multiple GMMs;117
9.1.5;6.5 Summary;119
9.1.6;References;119
9.2;Chapter 7: Learning-Based 3-D Object Retrieval;122
9.2.1;7.1 Introduction;122
9.2.2;7.2 Learning Optimal Distance Metrics;123
9.2.2.1;7.2.1 Hausdorff Distance Learning;123
9.2.2.2;7.2.2 Learning Bipartite Graph Optimal Matching;128
9.2.3;7.3 3-D Object Relevance Estimation via Hypergraph Learning;132
9.2.3.1;7.3.1 Hypergraph and Its Applications;133
9.2.3.2;7.3.2 Learning on Single Hypergraph;135
9.2.3.3;7.3.3 Learning on Multiple Hypergraphs;139
9.2.3.4;7.3.4 Learning the Weights for Multiple Hypergraphs;141
9.2.3.5;7.3.5 Learning the Weights for Edges;143
9.2.4;7.4 Summary;145
9.2.5;References;145
10;Part IV Conclusions and Future Work;148
10.1;Chapter 8: Conclusions and Future Work;150
10.1.1;8.1 Summary of This Book;150
10.1.2;8.2 Future Work;151
10.1.2.1;8.2.1 The Issue of Big Data;151
10.1.2.2;8.2.2 Feature Extraction;151
10.1.2.3;8.2.3 Multiple-View Matching;152
10.1.2.4;8.2.4 Multimodal Data;152
10.1.2.5;8.2.5 Geographical Location-Based Applications;152
10.1.3;References;153


Chapter 2 The Benchmark and Evaluation
Abstract
In all research areas, standard benchmarks are important for justifying the performance of different methods. With the progress of recent 3-D technology, many 3-D benchmarks have been released. In this chapter, we introduce the widely employed 3-D benchmarks and the 3DOR contest, that is, the Shape Retrieval Contest (SHREC). In the last part of this chapter, we discuss widely used 3DOR retrieval performance evaluation methods. Keywords 3-D Benchmark Data example Data distribution Shape retrieval contest Evaluation criteria 2.1 Introduction
Given the array of available applications, 3-D shape representation and matching and retrieval algorithms have been extensively investigated. These methods require standard benchmarks to evaluate their performance. Standard benchmarks are important for justifying the effectiveness of different method, and many 3-D object benchmarks have been released in recent years. In this chapter, we review several popular 3-D benchmarks, including the Princeton Shape Benchmark (PSB) [1], the Engineering Shape Benchmark (ESB) [2], the ITI database [3], the Amsterdam Library of Object Images (ALOI) [4], the Eidgenössische Technische Hochschule (ETH) [5], and the National Taiwan University (NTU) [6] 3-D data sets. We also introduce the recent 3DOR contest, SHREC, on each year, respectively. Evaluation methods are important to judge the performance of 3DOR. In the last part of this chapter, we introduce several widely used 3DOR retrieval performance evaluation criteria. 2.2 The Standard Benchmarks
In this section, we introduce six standard benchmarks. • Princeton shape benchmark
PSB was collected by Princeton University. PSB contains 1814 polygonal 3-D models, which are collected from the World Wide Web from 293 different Web domains. All these models are manually classified based on functions and forms and multiple semantic labels are annotated for each 3-D model. These labels belong to a hierarchical structure, which can reflect the semantic in different levels. For example, a 3-D model with an annotation of “aircraft” can be further divided into subclasses, such as “winged_vehicle aircraft,” “balloon_vehicle aircraft,” and “helicopter aircraft.” PSB has been split into a training set and a testing set. The training set contains 907 models, and the testing set contains the other 907 models. The distribution of all the models in PSB is shown in Figures 2.1 and 2.2. Example 3-D models from PSB are shown in Figure 2.3. Figure 2.1 The distribution of different model categories in the PSB training set. Figure 2.2 The distribution of different model categories in the PSB testing set. Figure 2.3 Example 3-D models in PSB. • Engineering shape benchmark
ESB focuses on engineering shape representation. Computer-aided design product design is an iterative process, which is highly dependent on previous model designs. ESB aims to evaluate the shape representation to discriminate different shape forms for product design. ESB is composed of 867 3-D CAD models. For each 3-D model, there are CAD files, including two neutral formats: “.stl” and “.obj”, and one thumbnail image. These models have been classified into three higher-level categories: Flat-thin wall components, Rectangular-cubic prism, and Solids of revolution. As introduced in [2], these three higher-level categories are defined as follows: • Flat-thin wall components are the parts with thin-walled sections and shell-like components. • Rectangular-cubic prism are the parts whose envelopes are largely rectangular or cubic prism. • Solids of revolution are the parts whose envelopes are largely solids of revolution. These higher-level classes are further divided into 45 shape categories. The distribution of all the models in ESB is shown in Figure 2.4. Example 3-D models from ESB are shown in Figure 2.5. Figure 2.4 The distribution of different model categories in the ESB. Figure 2.5 Example 3-D models in ESB. • The ITI benchmark
The ITI data set is from the VICTORY 3-D search engineer [7]. The ITI data set contains 544 3-D models from 13 categories, including animals, spheroid objects, conventional airplanes, delta airplanes, helicopters, cars, motorcycles, tubes, couches, chairs, fish, humans, and some other models. Each 3-D model contains 5,080 vertices and 7,061 triangles on average. The distribution of all the models in the ITI data set is shown in Figure 2.6. Figure 2.6 The distribution of different model categories in the ITI data set. • Amsterdam Library of Object Images
ALOI is a color image collection containing 1,000 small objects. We note that this data set does not provide the virtual 3-D model data, but only a set of images for each object. To obtain these images, the camera view angle, the illumination angle, and the illumination color for each object are systematically varied. For the camera view angle, the frontal camera is rotated in the plane at 5° to record 72 views of each object. For the illumination angle, there are eight different light conditions employed in the image data collection procedure, and five lights are employed. Each object is recorded with only one light turned on (five kinds in total), with only the left or right lights turned on (two kinds), and with all the lights turned on (one kind). Three cameras are employed to capture the views, which leads to a total of 24 different illumination conditions for recording views. The voltage of the lamps is varied 12 times for the illumination color, which yields 12 different illumination colors. Wide-baseline stereo images for 750 different scenarios are also recorded. In total, there are 110,250 images in the ALOI data set. Example objects from the ALOI benchmark are shown in Figure 2.7. Figure 2.7 Example objects in the ALOI benchmark. • The Eidgenössische Technische Hochschule Zürich data set
The ETH benchmark contains 80 objects from eight categories and each category consists of 10 objects. These categories cover the following areas: fruits and vegetables, animals, small human-made objects, and large human-made objects. These eight categories include apple, pear, tomato, cow, dog, horse, cup, and car. In the ETH data set, a group of 41 views are recorded for each object. To capture these views, the object is placed on a table with blue chroma keying background and the views are obtained from directions spaced equally over the upper viewing hemisphere, subdividing an octahedron to the third recursion level. These views are recorded using a Sony DFW-X700 progressive scan digital camera at a resolution of 1024 × 768. For each view, the original color image and an associated high-quality segmentation mask are provided. Example objects from the ETH benchmark are shown in Figure 2.8. Figure 2.8 Example objects in the ETH benchmark. • The National Taiwan University data set
NTU contains two parts, the NTU 3-D model benchmark and the NTU 3-D model database. The NTU 3-D model benchmark consists of 1833 3-D models and the NTU 3-D model database consists of 10,911 3-D models. All these 3-D models are downloaded from the Internet in July 2002. These models are in Wavefront file format. For each model, a thumbnail image is provided in the data set. The NTU 3-D model benchmark contains 47 categories with 549 3-D models, while another 1284 models are regarded as miscellaneous. The distribution of all models in the NTU 3-D model benchmark is shown in Figure 2.9, and example objects from the NTU 3-D model benchmark are shown in Figure 2.10. Figure 2.9 The distribution of different model categories in the NTU benchmark. Figure 2.10 Example objects in the NTU benchmark. 2.3 The Shape Retrieval Contest
SHREC [8] is an annual contest that started from 2006. SHREC aims to evaluate the effectiveness of 3-D shape retrieval algorithms. Each year, different tracks are organized with various objectives and different testing benchmarks are employed or released. In this section, we briefly review the existing SHREC events, the employed benchmarks, and the challenging tasks. • SHREC2006
SHREC2006 is the first SHREC and there is only one task: retrieving a ranking list of similar 3-D objects given one...



Ihre Fragen, Wünsche oder Anmerkungen
Vorname*
Nachname*
Ihre E-Mail-Adresse*
Kundennr.
Ihre Nachricht*
Lediglich mit * gekennzeichnete Felder sind Pflichtfelder.
Wenn Sie die im Kontaktformular eingegebenen Daten durch Klick auf den nachfolgenden Button übersenden, erklären Sie sich damit einverstanden, dass wir Ihr Angaben für die Beantwortung Ihrer Anfrage verwenden. Selbstverständlich werden Ihre Daten vertraulich behandelt und nicht an Dritte weitergegeben. Sie können der Verwendung Ihrer Daten jederzeit widersprechen. Das Datenhandling bei Sack Fachmedien erklären wir Ihnen in unserer Datenschutzerklärung.