Multitask learning (MTL) via a shared representation has been adopted to alleviate problems with sparsity of labeled data across different learning tasks. Active learning, on the …
Q Zhou, Q Zhao - IEEE transactions on pattern analysis and …, 2015 - ieeexplore.ieee.org
Multi-task learning (MTL) methods have shown promising performance by learning multiple relevant tasks simultaneously, which exploits to share useful information across relevant …
Y Zhang, Q Yang - IEEE transactions on knowledge and data …, 2021 - ieeexplore.ieee.org
Multi-Task Learning (MTL) is a learning paradigm in machine learning and its aim is to leverage useful information contained in multiple related tasks to help improve the …
This paper introduces two new frameworks, Doubly Supervised Latent Dirichlet Allocation (DSLDA) and its non-parametric variation (NP-DSLDA), that integrate two different types of …
J Xu, D Tang, T Jebara - arXiv preprint arXiv:2103.13420, 2021 - arxiv.org
The cost of annotating training data has traditionally been a bottleneck for supervised learning approaches. The problem is further exacerbated when supervised learning is …
L Han, Y Zhang - Proceedings of the AAAI Conference on Artificial …, 2015 - ojs.aaai.org
In multi-task learning (MTL), multiple related tasks are learned jointly by sharing information across them. Many MTL algorithms have been proposed to learn the underlying task groups …
Y Zhang, Q Yang - National Science Review, 2018 - academic.oup.com
As a promising area in machine learning, multi-task learning (MTL) aims to improve the performance of multiple related learning tasks by leveraging useful information among them …
D Gao, W Yang, H Zhou, Y Wei, Y Hu… - arXiv preprint arXiv …, 2021 - arxiv.org
The Multi-Task Learning (MTL) technique has been widely studied by word-wide researchers. The majority of current MTL studies adopt the hard parameter sharing structure …
In this paper, we propose an online multitask learning framework where the weight vectors are updated in an adaptive fashion based on inter-task relatedness. Our work is in contrast …