Predictability of Classification Performance Measures with Meta-Learning

Aus SDQ-Institutsseminar
Version vom 2. April 2019, 10:48 Uhr von Jakob Bach (Diskussion | Beiträge) (Die Seite wurde neu angelegt: „{{Vortrag |vortragender=Huijie Wang | |vortragstyp=Proposal |betreuer=Jakob Bach |termin=Institutsseminar/2019-04-12 |kurzfassung=In…“)
(Unterschied) ← Nächstältere Version | Aktuelle Version (Unterschied) | Nächstjüngere Version → (Unterschied)
Vortragende(r) Huijie Wang
Vortragstyp Proposal
Betreuer(in) Jakob Bach
Termin Fr 12. April 2019
Kurzfassung In machine learning, classification is the problem of identifying to which of a set of categories a new instance belongs. Usually, we cannot tell how the model performs until it is trained. Meta-learning, which learns about the learning algorithms themselves, can predict the performance of a model without training it based on meta-features of datasets and performance measures of previous runs. Though there is a rich variety of meta-features and performance measures on meta-learning, existing works usually focus on which meta-features are likely to correlate with model performance using one particular measure. The effect of different types of performance measures remain unclear as it is hard to draw a comparison between results of existing works, which are based on different meta-data sets as well as meta-models. The goal of this thesis is to study if certain types of performance measures can be predicted better than other ones and how much does the choice of the meta-model matter, by constructing different meta-regression models on same meta-features and different performance measures. We will use an experimental approach to evaluate our study.