Learning Multi-Task Sparse Representation Based on Fisher Information

Authors

  • Yayu Zhang Shanxi University City University of Hong Kong
  • Yuhua Qian Shanxi University
  • Guoshuai Ma North University of China
  • Keyin Zheng Shanxi University
  • Guoqing Liu Shanxi University City University of Hong Kong
  • Qingfu Zhang City University of Hong Kong The City University of Hong Kong Shenzhen Research Institute

DOI:

https://doi.org/10.1609/aaai.v38i15.29632

Keywords:

ML: Transfer, Domain Adaptation, Multi-Task Learning

Abstract

Multi-task learning deals with multiple related tasks simultaneously by sharing knowledge. In a typical deep multi-task learning model, all tasks use the same feature space and share the latent knowledge. If the tasks are weakly correlated or some features are negatively correlated, sharing all knowledge often leads to negative knowledge transfer among. To overcome this issue, this paper proposes a Fisher sparse multi-task learning method. It can obtain a sparse sharing representation for each task. In such a way, tasks share features on a sparse subspace. Our method can ensure that the knowledge transferred among tasks is beneficial. Specifically, we first propose a sparse deep multi-task learning model, and then introduce Fisher sparse module into traditional deep multi-task learning to learn the sparse variables of task. By alternately updating the neural network parameters and sparse variables, a sparse sharing representation can be learned for each task. In addition, in order to reduce the computational overhead, an heuristic method is used to estimate the Fisher information of neural network parameters. Experimental results show that, comparing with other methods, our proposed method can improve the performance for all tasks, and has high sparsity in multi-task learning.

Published

2024-03-24

How to Cite

Zhang, Y., Qian, Y., Ma, G., Zheng, K., Liu, G., & Zhang, Q. (2024). Learning Multi-Task Sparse Representation Based on Fisher Information. Proceedings of the AAAI Conference on Artificial Intelligence, 38(15), 16899-16907. https://doi.org/10.1609/aaai.v38i15.29632

Issue

Section

AAAI Technical Track on Machine Learning VI