The Princeton Shape Benchmark Philip Shilane, Patrick Min, Michael Kazhdan, and Thomas Funkhouser Shape Retrieval Problem 3D Model Shape Descriptor Best Matches Model Database Example Shape Descriptors • • • • • • • D2 Shape Distributions Extended Gaussian Image Shape Histograms Spherical Extent Function Spherical Harmonic Descriptor Light Field Descriptor etc. Example Shape Descriptors • • • • • • • D2 Shape Distributions Extended Gaussian Image Shape Histograms Spherical Extent Function Spherical Harmonic Descriptor Light Field Descriptor etc. How do we know which is best? Typical Retrieval Experiment • Create a database of 3D models • Group the models into classes • For each model: • Rank other models by similarity • Measure how many models in the same class appear near the top of the ranked list • Present average results Typical Retrieval Experiment • Create a database of 3D models • Group the models into classes • For each model: • Rank other models by similarity • Measure how many models in the same class appear near the top of the ranked list • Present average results Typical Retrieval Experiment • Create a database of 3D models • Group the models into classes • For each model: • Rank other models by similarity • Measure how many models in the same class appear near the top of the ranked list • Present average results Typical Retrieval Experiment • Create a database of 3D models • Group the models into classes • For each model: • Rank other models by similarity • Measure how many models in the same class appear near the top of the ranked list • Present average results Query Typical Retrieval Experiment 1 precision • Create a database of 3D models 0.8 • Group the models into0.6 classes 0.4 • For each model: • Rank other models by similarity 0 • Measure how many models 0 0.2 in the same class appear near the top of the ranked list 0.2 • Present average results Query 0.4 recall 0.6 0.8 1 Typical Retrieval Experiment 1 precision • Create a database of 3D models 0.8 • Group the models into0.6 classes 0.4 • For each model: • Rank other models by similarity 0 • Measure how many models 0 0.2 in the same class appear near the top of the ranked list 0.2 • Present average results Query 0.4 recall 0.6 0.8 1 Typical Retrieval Experiment 1 precision • Create a database of 3D models 0.8 • Group the models into0.6classes 0.4 • For each model: • Rank other models by similarity 0 • Measure how many models 0 0.2 in the same class appear near the top of the ranked list 0.2 • Present average results 0.4 recall 0.6 0.8 1 Typical Retrieval Experiment 1 precision • Create a database of 3D models 0.8 • Group the models into0.6classes 0.4 • For each model: • Rank other models by similarity 0 0 0.2 • Measure how many models in the same class appear near the top of the ranked list 0.2 • Present average results 0.4 0.6 recall 0.8 1 Shape Retrieval Results Shape Descriptor Compare Time (µs) Storage Size (bytes) Norm. DCGain 1,300 4,700 +21.3% REXT 229 17,416 +13.3% SHD 27 2,148 +10.2% GEDT 450 32,776 +10.1% 8 552 +6.0% SECSHEL 451 32,776 +2.8% VOXEL 450 32,776 +2.4% SECTORS 14 552 -0.3% CEGI 27 2,056 -9.6% EGI 14 1,032 -10.9% D2 2 136 -18.2% SHELLS 2 136 -27.3% LFD EXT Outline • • • • • • • Introduction Related work Princeton Shape Benchmark Comparison of 12 descriptors Evaluation techniques Results Conclusion Typical Shape Databases Osada MPEG-7 Hilaga Technion Zaharia CCCC Utrecht Taiwan Viewpoint Num Models 133 1,300 230 1,068 1,300 1,841 684 1,833 1,890 Num Num Classes Classified 25 133 15 227 32 230 17 258 23 362 54 416 6 512 47 549 85 1,280 Largest Class 20% 15% 15% 10% 14% 13% 45% 12% 12% Typical Shape Databases Osada MPEG-7 Hilaga Technion Zaharia CCCC Utrecht Taiwan Viewpoint Num Models 133 1,300 230 1,068 1,300 1,841 684 1,833 1,890 Num Num Classes Classified 25 133 15 227 32 230 17 258 23 362 54 416 6 512 47 549 85 1,280 Largest Class 20% 15% 15% 10% 14% 13% 45% 12% 12% Typical Shape Databases Osada MPEG-7 Hilaga Technion Zaharia CCCC Utrecht Taiwan Viewpoint Num Num Num Models Classes Classified 133 25 133 1,300 15 227 230 32 230 1,068 17 258 Aerodynamic 23 1,300 362 1,841 54 416 684 6 512 1,833 47 549 1,890 85 1,280 Largest Class 20% 15% 15% 10% 14% 13% 45% 12% 12% Typical Shape Databases Osada MPEG-7 Hilaga Technion Zaharia CCCC Utrecht Taiwan Viewpoint Num Models 133 1,300 230 1,068 1,300 1,841 684 1,833 1,890 Num Num Classes Classified 25 133 15 227 32 230 17 258 Letter ‘C’ 23 362 54 416 6 512 47 549 85 1,280 Largest Class 20% 15% 15% 10% 14% 13% 45% 12% 12% Typical Shape Databases Vehicles Furniture Animals Plants Household Buildings Osada MPEG-7 Hilaga Zaharia CCCC Utrecht Taiwan Viewpoint 47% 12% 12% 35% 33% 100% 44% 0% 12% 0% 0% 0% 13% 0% 13% 42% 12% 14% 23% 7% 21% 0% 0% 1% 0% 13% 2% 7% 5% 0% 0% 0% 24% 0% 12% 11% 25% 0% 36% 50% 0% 7% 0% 0% 0% 0% 0% 0% Typical Shape Databases Vehicles Furniture Animals Plants Household Buildings Osada MPEG-7 Hilaga Zaharia CCCC Utrecht Taiwan Viewpoint 47% 12% 12% 35% 33% 100% 44% 0% 12% 0% 0% 0% 13% 0% 13% 42% 12% 14% 23% 7% 21% 0% 0% 1% 0% 13% 2% 7% 5% 0% 0% 0% 24% 0% 12% 11% 25% 0% 36% 50% 0% 7% 0% 0% 0% 0% 0% 0% Typical Shape Databases 13% 2% 7% 39 5% vases 0% 0% 0% Buildings 160% beds Household Plants Animals Furniture Vehicles room chairs Osada153 dining chairs 47%25 living 12% 12% MPEG-7 12% 0% 14% Hilaga 12% 0% 23% Zaharia 35% 0% 7% 28 bottles 21% CCCC 8 chests33% 13% Utrecht 100% 0% 0% Taiwan 44% 13% 0% Viewpoint 0% 42% 1% tables 24%12 dining0% 0% 7% 12% 0% 11% 0% tables 25%36 end0% 0% 0% 36% 0% 50% 0% Typical Shape Databases Vehicles Furniture Animals Plants Household Buildings Osada MPEG-7 Hilaga Zaharia CCCC Utrecht Taiwan Viewpoint 47% 12% 12% 35% 33% 100% 44% 0% 12% 0% 0% 0% 13% 0% 13% 42% 12% 14% 23% 7% 21% 0% 0% 1% 0% 13% 2% 7% 5% 0% 0% 0% 24% 0% 12% 11% 25% 0% 36% 50% 0% 7% 0% 0% 0% 0% 0% 0% Goal: Benchmark for 3D Shape Retrieval • • • • • • Large number of classified models Wide variety of class types Not too many or too few models in each class Standardized evaluation tools Ability to investigate properties of descriptors Freely available to researchers Princeton Shape Benchmark • Large shape database • 6,670 models • 1,814 classified models, 161 classes • Separate training and test sets • Standardized suite of tests • Multiple classifications • Targeted sets of queries • Standardized evaluation tools • Visualization software • Quantitative metrics Princeton Shape Benchmark 51 potted plants 33 faces 15 desk chairs 22 dining chairs 100 humans 28 biplanes 14 flying birds 11 ships Princeton Shape Benchmark (PSB) Num Models Osada Num Classes Num Classified Largest Class 133 25 133 20% 1,300 15 227 15% 230 32 230 15% Technion 1,068 17 258 10% Zaharia 1,300 23 362 14% CCCC 1,841 54 416 13% Utrecht 684 6 512 45% Taiwan 1,833 47 549 12% Viewpoint 1,890 85 1,280 12% PSB 6,670 161 1,814 6% MPEG-7 Hilaga Princeton Shape Benchmark (PSB) Vehicles Furniture Animals Plants Household Buildings Osada MPEG-7 Hilaga Zaharia CCCC Utrecht Taiwan Viewpoint PSB 47% 12% 12% 35% 33% 100% 44% 0% 26% 12% 0% 0% 0% 13% 0% 13% 42% 11% 12% 14% 23% 7% 21% 0% 0% 1% 16% 0% 13% 2% 7% 5% 0% 0% 0% 8% 24% 0% 12% 11% 25% 0% 36% 50% 22% 0% 7% 0% 0% 0% 0% 0% 0% 6% Outline • • • • • • • Introduction Related work Princeton Shape Benchmark Comparison of 12 descriptors Evaluation techniques Results Conclusion Comparison of Shape Descriptors • • • • • • • • • • • • Shape Histograms (Shells) Shape Histograms (Sectors) Shape Histograms (SecShells) D2 Shape Distributions Extended Gaussian Image (EGI) Complex Extended Gaussian Image (CEGI) Spherical Extent Function (EXT) Radialized Spherical Extent Function (REXT) Voxel Gaussian Euclidean Distance Transform (GEDT) Spherical Harmonic Descriptor (SHD) Light Field Descriptor (LFD) Comparison of Shape Descriptors Base (92) LFD 1 REXT SHD 0.8 precision GEDT EXT 0.6 SecShells l Voxel 0.4 Sectors CEGI 0.2 EGI D2 0 0 0.2 0.4 0.6 recall 0.8 1 Shells Evaluation Tools Visualization tools Precision/recall plot Best matches Distance image Tier image Quantitative metrics Nearest neighbor First and Second tier E-Measure Discounted Cumulative Gain (DCG) Evaluation Tools Visualization tools Precision/recall plot Best matches Distance image Tier image Quantitative metrics Nearest neighbor First and Second tier E-Measure Discounted Cumulative Gain (DCG) Evaluation Tools Query Visualization tools Precision/recall plot Best matches Distance image Tier image Quantitative metrics Nearest neighbor First and Second tier E-Measure Discounted Cumulative Gain (DCG) Wrong class Correct class Evaluation Tools Visualization tools Precision/recall plot Best matches Distance image Tier image Quantitative metrics Nearest neighbor First and Second tier E-Measure Discounted Cumulative Gain (DCG) Evaluation Tools Visualization tools Precision/recall plot Best matches Distance image Tier image Quantitative metrics Nearest neighbor First and Second tier E-Measure Discounted Cumulative Gain (DCG) Evaluation Tools Visualization tools Precision/recall plot Best matches Distance image Tier image Quantitative metrics Nearest neighbor First and Second tier E-Measure Discounted Cumulative Gain (DCG) Dining Chair Desk Chair Function vs. Shape Functional at the top levels of the hierarchy, shape based at the lower levels root Man-made Vehicle Furniture Table Rectangular table Chair Round table Natural Base Classification (92 classes) Man-made 1 precision 0.8 0.6 SHD Furniture EGI 0.4 Table 0.2 0 0 0.2 0.4 0.6 recall 0.8 1 Round table Coarse Classification (44 classes) Man-made 1 precision 0.8 0.6 SHD Furniture EGI 0.4 Table 0.2 0 0 0.2 0.4 0.6 recall 0.8 1 Round table Coarser Classification (6 classes) Man-made 1 precision 0.8 0.6 SHD Furniture EGI 0.4 Table 0.2 0 0 0.2 0.4 0.6 recall 0.8 1 Round table Coarsest Classification (2 classes) Man-made 1 precision 0.8 0.6 SHD Furniture EGI 0.4 Table 0.2 0 0 0.2 0.4 0.6 recall 0.8 1 Round table Granularity Comparison Base (92) Man-made vs. Natural (2) LFD 1 REXT 0.8 SHD precision GEDT 0.6 EXT SecShells 0.4 Voxel 0.2 Sectors CEGI 0 0 0.2 0.4 0.6 recall 0.8 1 EGI D2 Shells Rotationally Aligned Models (650) 1 precision 0.8 0.6 SHD GEDT 0.4 0.2 0 0 0.2 0.4 0.6 recall 0.8 1 All Models (907) 1 precision 0.8 0.6 SHD GEDT 0.4 0.2 0 0 0.2 0.4 0.6 recall 0.8 1 Complex Models (200) 1 precision 0.8 0.6 SHD GEDT 0.4 0.2 0 0 0.2 0.4 0.6 recall 0.8 1 Performance by Property Rotation Aligned Depth Complexity Base LFD 18.8 21.3 28.2 REXT 12.3 13.3 15.0 7.6 10.2 8.9 13.0 10.1 13.5 EXT 5.0 6.0 6.1 SecShells 5.2 2.8 2.2 Voxel 4.7 2.4 0.2 Sectors 2.0 -0.3 -1.6 CEGI -8.7 -9.6 -12.7 EGI -11.2 -10.9 -9.1 D2 -19.7 -18.2 -19.9 Shells -29.1 -27.3 -30.9 SHD GEDT Conclusion • Methodology to compare shape descriptors • Vary classifications • Query lists targeted at specific properties • Unexpected results • EGI: good at discriminating man-made vs. natural objects, though poor at fine-grained distinctions • LFD: good overall performance across tests • Freely available Princeton Shape Benchmark • 1,814 classified polygonal models • Source code for evaluation tools Future Work • Multi-classifiers • Evaluate statistical significance of results • Application of techniques to other domains • Text retrieval • Image retrieval • Protein classification Acknowledgements David Bengali partitioned thousands of models. Ming Ouhyoung and his students provided the light field descriptor. Dejan Vranic provided the CCCC and MPEG-7 databases. Viewpoint Data Labs donated the Viewpoint database. Remco Veltkamp and Hans Tangelder provided the Utrecht database. Funding: The National Science Foundation grants CCR-0093343 and 11S-0121446. The End http://shape.cs.princeton.edu/benchmark