Abstract
This paper describes a unified approach for two-dimensional (2-D) shape matching and similarity ranking of objects by means of a modal representation. In particular, we propose a new shape-similarity metric in the eigenshape space for object/image retrieval from a visual database via query-by-example. This differs from prior work which performed point correspondence determination and similarity ranking of shapes in separate steps. The proposed method employs selected boundary and/or contour points of an object as a coarse-to-fine shape representation, and does not require extraction of connected boundaries or silhouettes. It is rotation-, translation- and scale-invariant, and can handle mild deformations of objects (e.g. due to partial occlusions or pose variations). Results comparing the unified method with an earlier two-step approach using B-spline-based modal matching and Hausdorff distance ranking are presented on retail and museum catalog style still-image databases.
Original language | English |
---|---|
Pages (from-to) | 931-944 |
Number of pages | 14 |
Journal | Pattern Recognition |
Volume | 31 |
Issue number | 7 |
DOIs | |
Publication status | Published - 31 Jul 1998 |
Funding
This work is supported by a National Science Foundation SIUCRC grant and a New York State Science and Technology Foundation grant to the Center for Electronic Imaging Systems at the University of Rochester.
Funders | Funder number |
---|---|
New York State Science and Technology Foundation | |
National Science Foundation |
Keywords
- Content-based access
- Image databases
- Modal matching
- Object retrieval
- Shape similarity metrics