TY - JOUR
T1 - Adaptively sharing multi-levels of distributed representations in multi-task learning
AU - Wang, Tianxin
AU - Zhuang, Fuzhen
AU - Sun, Ying
AU - Zhang, Xiangliang
AU - Lin, Leyu
AU - Xia, Feng
AU - He, Lei
AU - He, Qing
N1 - Publisher Copyright:
© 2022
PY - 2022/4
Y1 - 2022/4
N2 - In multi-task learning, the performance is often sensitive to the relationships between tasks. Thus it is important to study how to exploit the complex relationships across different tasks. One line of research captures the complex task relationships, by increasing the model capacity and thus requiring a large training dataset. However in many real-world applications, the amount of labeled data is limited. In this paper, we propose a light weight and specially designed architecture, which aims to model task relationships for small or middle-sized datasets. The proposed framework learns a task-specific ensemble of sub-networks in different depths, and is able to adapt the model architecture for the given data. The task-specific ensemble parameters are learned simultaneously with the weights of the network by optimizing a single loss function defined with respect to the end task. The hierarchical model structure is able to share both general and specific distributed representations to capture the inherent relationships between tasks. We validate our approach on various types of tasks, including synthetic task, article recommendation task and vision task. The results demonstrate the advantages of our model over several competitive baselines especially when the tasks are less-related.
AB - In multi-task learning, the performance is often sensitive to the relationships between tasks. Thus it is important to study how to exploit the complex relationships across different tasks. One line of research captures the complex task relationships, by increasing the model capacity and thus requiring a large training dataset. However in many real-world applications, the amount of labeled data is limited. In this paper, we propose a light weight and specially designed architecture, which aims to model task relationships for small or middle-sized datasets. The proposed framework learns a task-specific ensemble of sub-networks in different depths, and is able to adapt the model architecture for the given data. The task-specific ensemble parameters are learned simultaneously with the weights of the network by optimizing a single loss function defined with respect to the end task. The hierarchical model structure is able to share both general and specific distributed representations to capture the inherent relationships between tasks. We validate our approach on various types of tasks, including synthetic task, article recommendation task and vision task. The results demonstrate the advantages of our model over several competitive baselines especially when the tasks are less-related.
KW - Deep learning
KW - Machine learning
KW - Multi-task learning
UR - https://www.webofscience.com/wos/woscc/full-record/WOS:000770687400012
UR - https://openalex.org/W4206950106
UR - https://www.scopus.com/pages/publications/85123581293
M3 - Journal Article
SN - 0020-0255
VL - 591
SP - 226
EP - 234
JO - Information Sciences
JF - Information Sciences
ER -