Open-set low-shot classification leveraging the power of instance discrimination
Spiridon Kasapis,
Geng Zhang,
Jonathon M Smereka and
Nickolas Vlahopoulos
The Journal of Defense Modeling and Simulation, 2023, vol. 20, issue 4, 527-539
Abstract:
In search, exploration, and reconnaissance operations of autonomous ground vehicles, an image recognition capability is needed for specifically classifying targeted objects (relevant classes) and at the same time identifying as unknown (irrelevant and unseen) objects that do not belong to any known classes, as opposed to falsely classifying them in one of the relevant classes. This paper integrates an unsupervised learning feature extraction framework based on the Instance Discrimination method with an Open-Set Low-Shot (IDLS) classifier for creating the desired new capability. Unlabeled images from the vehicle’s operating environment are used for training the feature extractor while a modest number (less than 40) images for each relevant class and unlabeled irrelevant images are used for training the Open-Set Low-Shot (OSLS) classifier in a manner that enables recognition of images unseen during training as irrelevant. The value and the accuracy of the new IDLS approach are demonstrated through a thorough comparison with alternative unsupervised and fully supervised methods.
Keywords: Unsupervised training; open-set classification; low-shot learning (search for similar items in EconPapers)
Date: 2023
References: Add references at CitEc
Citations:
Downloads: (external link)
https://journals.sagepub.com/doi/10.1177/15485129221111172 (text/html)
Related works:
This item may be available elsewhere in EconPapers: Search for items with the same title.
Export reference: BibTeX
RIS (EndNote, ProCite, RefMan)
HTML/Text
Persistent link: https://EconPapers.repec.org/RePEc:sae:joudef:v:20:y:2023:i:4:p:527-539
DOI: 10.1177/15485129221111172
Access Statistics for this article
More articles in The Journal of Defense Modeling and Simulation
Bibliographic data for series maintained by SAGE Publications ().