Shape matching and object recognition using shape contexts

Research output: Contribution to journalJournal articleResearchpeer-review

We present a novel approach to measuring similarity between shapes and exploit it for object recognition. In our framework, the measurement of similarity is preceded by 1) solving for correspondences between points on the two shapes, 2) using the correspondences to estimate an aligning transform. In order to solve the correspondence problem, we attach a descriptor, the shape context, to each point. The shape context at a reference point captures the distribution of the remaining points relative to it, thus offering a globally discriminative characterization. Corresponding points on two similar shapes will have similar shape contexts, enabling us to solve for correspondences as an optimal assignment problem. Given the point correspondences, we estimate the transformation that best aligns the two shapes; regularized thin-plate splines provide a flexible class of transformation maps for this purpose. The dissimilarity between the two shapes is computed as a sum of matching errors between corresponding points, together with a term measuring the magnitude of the aligning transform. We treat recognition in a nearest-neighbor classification framework as the problem of finding the stored prototype shape that is maximally similar to that in the image. Results are presented for silhouettes, trademarks, handwritten digits, and the COIL data set.

Original languageEnglish
JournalIEEE Transactions on Pattern Analysis and Machine Intelligence
Volume24
Issue number4
Pages (from-to)509-522
Number of pages14
ISSN0162-8828
DOIs
Publication statusPublished - Apr 2002
Externally publishedYes

Bibliographical note

Funding Information:
This research is supported by the Army Research Office (ARO) DAAH04-96-1-0341, the Digital Library Grant IRI-9411334, a US National Science Foundation Graduate Fellowship for S. Belongie, and the Germ an Research Foundation by DFG grant PU-165/1. Parts of this work have appeared in [3], [2]. The authors wish to thank H. Chui and A. Rangarajan for providing the synthetic testing data used in Section 4.2. We would also like to thank them and various members of the Berkeley computer vision group, particularly A. Berg, A. Efros, D. Forsyth, T. Leung, J. Shi, and Y. Weiss, for useful discussions. This work was carried out while the authors were with the Departm ent of Electrical Engineering and Computer Science Division, University of California at Berkeley.

    Research areas

  • Correspondence problem, Deformable templates, Digit recognition, Image registration, MPEG7, Object recognition, Shape

ID: 302058015