EconPapers    
Economics at your fingertips  
 

Classification error in multiclass discrimination from Markov data

Sören Christensen (), Albrecht Irle () and Lars Willert ()
Additional contact information
Sören Christensen: Chalmers University of Technology and Göteborg University
Albrecht Irle: University of Kiel
Lars Willert: University of Kiel

Statistical Inference for Stochastic Processes, 2016, vol. 19, issue 3, No 3, 336 pages

Abstract: Abstract As a model for an on-line classification setting we consider a stochastic process $$(X_{-n},Y_{-n})_{n}$$ ( X - n , Y - n ) n , the present time-point being denoted by 0, with observables $$\ldots ,X_{-n},X_{-n+1}, \ldots , X_{-1}, X_0$$ … , X - n , X - n + 1 , … , X - 1 , X 0 from which the pattern $$Y_0$$ Y 0 is to be inferred. So in this classification setting, in addition to the present observation $$X_0$$ X 0 a number l of preceding observations may be used for classification, thus taking a possible dependence structure into account as it occurs e.g. in an ongoing classification of handwritten characters. We treat the question how the performance of classifiers is improved by using such additional information. For our analysis, a hidden Markov model is used. Letting $$R_l$$ R l denote the minimal risk of misclassification using l preceding observations we show that the difference $$\sup _k |R_l - R_{l+k}|$$ sup k | R l - R l + k | decreases exponentially fast as l increases. This suggests that a small l might already lead to a noticeable improvement. To follow this point we look at the use of past observations for kernel classification rules. Our practical findings in simulated hidden Markov models and in the classification of handwritten characters indicate that using $$l=1$$ l = 1 , i.e. just the last preceding observation in addition to $$X_0$$ X 0 , can lead to a substantial reduction of the risk of misclassification. So, in the presence of stochastic dependencies, we advocate to use $$ X_{-1},X_0$$ X - 1 , X 0 for finding the pattern $$Y_0$$ Y 0 instead of only $$X_0$$ X 0 as one would in the independent situation.

Keywords: Optimal classification; Asymptotic risk; Hidden Markov model (search for similar items in EconPapers)
Date: 2016
References: View references in EconPapers View complete reference list from CitEc
Citations:

Downloads: (external link)
http://link.springer.com/10.1007/s11203-015-9129-6 Abstract (text/html)
Access to the full text of the articles in this series is restricted.

Related works:
This item may be available elsewhere in EconPapers: Search for items with the same title.

Export reference: BibTeX RIS (EndNote, ProCite, RefMan) HTML/Text

Persistent link: https://EconPapers.repec.org/RePEc:spr:sistpr:v:19:y:2016:i:3:d:10.1007_s11203-015-9129-6

Ordering information: This journal article can be ordered from
http://www.springer. ... ty/journal/11203/PS2

DOI: 10.1007/s11203-015-9129-6

Access Statistics for this article

Statistical Inference for Stochastic Processes is currently edited by Denis Bosq, Yury A. Kutoyants and Marc Hallin

More articles in Statistical Inference for Stochastic Processes from Springer
Bibliographic data for series maintained by Sonal Shukla () and Springer Nature Abstracting and Indexing ().

 
Page updated 2025-03-20
Handle: RePEc:spr:sistpr:v:19:y:2016:i:3:d:10.1007_s11203-015-9129-6