Auto-Probabilistic Mining Method for Siamese Neural Network Training
Arseniy Mokin (),
Alexander Sheshkus and
Vladimir L. Arlazarov
Additional contact information
Arseniy Mokin: Smart Engines Service LLC, 117312 Moscow, Russia
Alexander Sheshkus: Smart Engines Service LLC, 117312 Moscow, Russia
Vladimir L. Arlazarov: Smart Engines Service LLC, 117312 Moscow, Russia
Mathematics, 2025, vol. 13, issue 8, 1-16
Abstract:
Training deep learning models for classification with limited data and computational resources remains a challenge when the number of classes is large. Metric learning offers an effective solution to this problem. However, it has its own shortcomings due to the known imperfections of widely used loss functions such as contrastive loss and triplet loss, as well as sample mining methods. This paper address these issues by proposing a novel mining method and metric loss function. Firstly, this paper presents an auto-probabilistic mining method designed to automatically select the most informative training samples for Siamese neural networks. Combined with a previously proposed auto-clustering technique, the method improves model training by optimizing the utilization of available data and reducing computational overhead. Secondly, this paper proposes the novel cluster-aware triplet-based metric loss function that addresses the limitations of contrastive and triplet loss, enhancing the overall training process. To evaluate the proposed methods, experiments were conducted with the optical character recognition task using the PHD08 and Omniglot datasets. The proposed loss function with the random-mining method achieved 82.6 % classification accuracy on the PHD08 dataset with full training on the Korean alphabet, surpassing the known baseline. The same experiment, using a reduced training alphabet, set a new baseline of 88.6 % on the PHD08 dataset. The application of the novel mining method further enhanced the accuracy to 90.6 % (+2.0%) and, combined with auto-clustering, achieved 92.3 % (+3.7%) compared with the new baseline. On the Omniglot dataset, the proposed mining method reached 92.32 % , rising to 93.17 % with auto-clustering. These findings highlight the potential effectiveness of the developed loss function and mining method in addressing a wide range of pattern recognition challenges.
Keywords: deep metric learning; optical character recognition; Siamese neural networks; pattern recognition (search for similar items in EconPapers)
JEL-codes: C (search for similar items in EconPapers)
Date: 2025
References: Add references at CitEc
Citations:
Downloads: (external link)
https://www.mdpi.com/2227-7390/13/8/1270/pdf (application/pdf)
https://www.mdpi.com/2227-7390/13/8/1270/ (text/html)
Related works:
This item may be available elsewhere in EconPapers: Search for items with the same title.
Export reference: BibTeX
RIS (EndNote, ProCite, RefMan)
HTML/Text
Persistent link: https://EconPapers.repec.org/RePEc:gam:jmathe:v:13:y:2025:i:8:p:1270-:d:1633319
Access Statistics for this article
Mathematics is currently edited by Ms. Emma He
More articles in Mathematics from MDPI
Bibliographic data for series maintained by MDPI Indexing Manager ().