Abstract
We are interested in matching objects in photographs, paintings, sketches and so on; after all, humans have a remarkable ability to recognise objects in images, no matter how they are depicted. We conduct experiments in matching, and conclude that the key to robustness lies in object description. The existing literature consists of numerous feature descriptors that rely heavily on photometric properties such as colour and illumination to describe objects. Although these methods achieve high rates of accuracy in applications such as detection and retrieval of photographs, they fail to generalise datasets consisting of mixed depictions. Here, we propose a more general approach for describing objects invariant to depictive style. We use structure at a global level, which is combined with simple non-photometric descriptors at a local level. There is no need for any prior learning. Our descriptor achieves results on par with existing state of the art, when applied to object matching on a standard dataset consisting of photographs alone and outperforms the state of the art when applied to depiction-invariant object matching.
Original language | English |
---|---|
DOIs | |
Publication status | Published - 1 Jan 2012 |
Event | 2012 23rd British Machine Vision Conference, BMVC 2012 - Guildford, Surrey, UK United Kingdom Duration: 3 Sept 2012 → 7 Sept 2012 |
Conference
Conference | 2012 23rd British Machine Vision Conference, BMVC 2012 |
---|---|
Country/Territory | UK United Kingdom |
City | Guildford, Surrey |
Period | 3/09/12 → 7/09/12 |
ASJC Scopus subject areas
- Computer Vision and Pattern Recognition