TY - GEN
T1 - Adversarial encoder-multi-task-decoder for multi-stage processes
AU - Mendes, Andre
AU - Togelius, Julian
AU - dos Santos Coelho, Leandro
N1 - Publisher Copyright:
© 2020 IEEE
PY - 2020
Y1 - 2020
N2 - In multi-stage processes, decisions occur in an ordered sequence of stages. Early stages usually have more observations with general information (easier/cheaper to collect), while later stages have fewer observations but more specific data. This situation can be represented as a dual funnel structure, in which the sample size decreases from one stage to the other while the information available about each instance increases. Training classifiers in this scenario is challenging since information in the early stages may not contain distinct patterns to learn (underfitting). In contrast, the small sample size in later stages can cause overfitting. We address both cases by introducing a framework that combines adversarial autoencoders (AAE), multitask learning (MTL), and multi-label semi-supervised learning (MLSSL). We improve the decoder of the AAE with MTL so it can jointly reconstruct the original input and use feature nets to predict the features for the next stages. We also introduce a sequence constraint in the output of an MLSSL classifier to guarantee the sequential pattern in the predictions. Using different domains (selection process, medical diagnosis), we show that our approach outperforms other state-of-the-art methods.
AB - In multi-stage processes, decisions occur in an ordered sequence of stages. Early stages usually have more observations with general information (easier/cheaper to collect), while later stages have fewer observations but more specific data. This situation can be represented as a dual funnel structure, in which the sample size decreases from one stage to the other while the information available about each instance increases. Training classifiers in this scenario is challenging since information in the early stages may not contain distinct patterns to learn (underfitting). In contrast, the small sample size in later stages can cause overfitting. We address both cases by introducing a framework that combines adversarial autoencoders (AAE), multitask learning (MTL), and multi-label semi-supervised learning (MLSSL). We improve the decoder of the AAE with MTL so it can jointly reconstruct the original input and use feature nets to predict the features for the next stages. We also introduce a sequence constraint in the output of an MLSSL classifier to guarantee the sequential pattern in the predictions. Using different domains (selection process, medical diagnosis), we show that our approach outperforms other state-of-the-art methods.
KW - Adversarial autoencoder
KW - Multi-stage
KW - Multi-task
UR - http://www.scopus.com/inward/record.url?scp=85110463142&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85110463142&partnerID=8YFLogxK
U2 - 10.1109/ICPR48806.2021.9412044
DO - 10.1109/ICPR48806.2021.9412044
M3 - Conference contribution
AN - SCOPUS:85110463142
T3 - Proceedings - International Conference on Pattern Recognition
SP - 763
EP - 770
BT - Proceedings of ICPR 2020 - 25th International Conference on Pattern Recognition
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 25th International Conference on Pattern Recognition, ICPR 2020
Y2 - 10 January 2021 through 15 January 2021
ER -