EconPapers    
Economics at your fingertips  
 

LLM on a Budget: Active Knowledge Distillation for Efficient Classification of Large Text Corpora

Leland Crane, Xiaoyu Ge, Flora Haberkorn, Rithika Iyengar, Seung Jung Lee, Viviana Luccioli, Ryan Panley and Nitish R. Sinha
Additional contact information
Nitish R. Sinha: https://www.federalreserve.gov/econres/nitish-r-sinha.htm

No 2025-108, Finance and Economics Discussion Series from Board of Governors of the Federal Reserve System (U.S.)

Abstract: Large Language Models (LLMs) are highly accurate in classification tasks, however, substantial computational and financial costs hinder their large-scale deployment in dynamic environments. Knowledge Distillation (KD) where a LLM ""teacher"" trains a smaller and more efficient ""student"" model, offers a promising solution to this problem. However, the distillation process itself often remains costly for large datasets, since it requires the teacher to label a vast number of samples while incurring significant token consumption. To alleviate this challenge, in this work we explore the active learning (AL) as a way to create efficient student models at a fraction of the cost while preserving the LLM's performance. In particular, we introduce M-RARU (Multi-class Randomized Accept/Reject Uncertainty Sampling), a novel AL algorithm that significantly reduces training costs. M-RARU employs an innovative strategy combining uncertainty with a randomized accept-reject mechanism to select only the most informative data points for the LLM teacher. This focused approach significantly minimizes required API calls and data processing time. We evaluate M-RARU against random sampling across five diverse student models (SVM, LDA, RF, GBDT, and DistilBERT) on multiple benchmark datasets. Experiments demonstrate that our proposed method achieves up to 80\% reduction in sample requirements as compared to random sampling, substantially improving classification accuracy while reducing financial costs and overall training time.

Keywords: Machine learning; Sampling; Computational techniques (search for similar items in EconPapers)
JEL-codes: C38 C45 C55 (search for similar items in EconPapers)
Pages: 26 p.
Date: 2025-12-15
References: Add references at CitEc
Citations:

Downloads: (external link)
https://www.federalreserve.gov/econres/feds/files/2025108pap.pdf Full text (application/pdf)

Related works:
This item may be available elsewhere in EconPapers: Search for items with the same title.

Export reference: BibTeX RIS (EndNote, ProCite, RefMan) HTML/Text

Persistent link: https://EconPapers.repec.org/RePEc:fip:fedgfe:102367

DOI: 10.17016/FEDS.2025.108

Access Statistics for this paper

More papers in Finance and Economics Discussion Series from Board of Governors of the Federal Reserve System (U.S.) Contact information at EDIRC.
Bibliographic data for series maintained by Ryan Wolfslayer ; Keisha Fournillier ().

 
Page updated 2026-02-01
Handle: RePEc:fip:fedgfe:102367