TY - GEN
T1 - Multi-stage transfer learning with an application to selection process
AU - Mendes, Andre
AU - Togelius, Julian
AU - Dos Santos Coelho, Leandro
N1 - Publisher Copyright:
© 2020 The authors and IOS Press.
PY - 2020/8/24
Y1 - 2020/8/24
N2 - In multi-stage processes, decisions happen in an ordered sequence of stages. Many of them have the structure of dual funnel problem: As the sample size decreases from one stage to the other, the information increases. A related example is a selection process, where applicants apply for a position, prize or grant. In each stage, more applicants are evaluated and filtered out and from the remaining ones, more information is collected. In the last stage, decision-makers use all available information to make their final decision. To train a classifier for each stage becomes impracticable as they can underfit due to the low dimensionality in early stages or overfit due to the small sample size in the latter stages. In this work, we proposed a Multi-StaGe Transfer Learning (MSGTL) approach that uses knowledge from simple classifiers trained in early stages to improve the performance of classifiers in the latter stages. By transferring weights from simpler neural networks trained in larger datasets, we able to fine-tune more complex neural networks in the latter stages without overfitting due to the small sample size. We show that is possible to control the trade-off between conserving knowledge and fine-tuning using a simple probabilistic map. Experiments using real-world data show the efficacy of our approach as it outperforms other state-of-the-art methods for transfer learning and regularization.
AB - In multi-stage processes, decisions happen in an ordered sequence of stages. Many of them have the structure of dual funnel problem: As the sample size decreases from one stage to the other, the information increases. A related example is a selection process, where applicants apply for a position, prize or grant. In each stage, more applicants are evaluated and filtered out and from the remaining ones, more information is collected. In the last stage, decision-makers use all available information to make their final decision. To train a classifier for each stage becomes impracticable as they can underfit due to the low dimensionality in early stages or overfit due to the small sample size in the latter stages. In this work, we proposed a Multi-StaGe Transfer Learning (MSGTL) approach that uses knowledge from simple classifiers trained in early stages to improve the performance of classifiers in the latter stages. By transferring weights from simpler neural networks trained in larger datasets, we able to fine-tune more complex neural networks in the latter stages without overfitting due to the small sample size. We show that is possible to control the trade-off between conserving knowledge and fine-tuning using a simple probabilistic map. Experiments using real-world data show the efficacy of our approach as it outperforms other state-of-the-art methods for transfer learning and regularization.
UR - http://www.scopus.com/inward/record.url?scp=85091801600&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85091801600&partnerID=8YFLogxK
U2 - 10.3233/FAIA200291
DO - 10.3233/FAIA200291
M3 - Conference contribution
AN - SCOPUS:85091801600
T3 - Frontiers in Artificial Intelligence and Applications
SP - 1770
EP - 1777
BT - ECAI 2020 - 24th European Conference on Artificial Intelligence, including 10th Conference on Prestigious Applications of Artificial Intelligence, PAIS 2020 - Proceedings
A2 - De Giacomo, Giuseppe
A2 - Catala, Alejandro
A2 - Dilkina, Bistra
A2 - Milano, Michela
A2 - Barro, Senen
A2 - Bugarin, Alberto
A2 - Lang, Jerome
PB - IOS Press BV
T2 - 24th European Conference on Artificial Intelligence, ECAI 2020, including 10th Conference on Prestigious Applications of Artificial Intelligence, PAIS 2020
Y2 - 29 August 2020 through 8 September 2020
ER -