Efficient multi-task learning with adaptive temporal structure for progression prediction.

Menghui Zhou, Yu Zhang, Tong Liu, Yun Yang, Po Yang
Author Information
  1. Menghui Zhou: Department of Software, Yunnan University, Kunming, 674199 Yunnan Province China.
  2. Yu Zhang: Department of Computer Science, University of Sheffield, Sheffield, S10 2TT UK.
  3. Tong Liu: Department of Computer Science, University of Sheffield, Sheffield, S10 2TT UK.
  4. Yun Yang: Department of Software, Yunnan University, Kunming, 674199 Yunnan Province China.
  5. Po Yang: Department of Computer Science, University of Sheffield, Sheffield, S10 2TT UK. ORCID

Abstract

In this paper, we propose a novel efficient multi-task learning formulation for the class of progression problems in which its state will continuously change over time. To use the shared knowledge information between multiple tasks to improve performance, existing multi-task learning methods mainly focus on feature selection or optimizing the task relation structure. The feature selection methods usually fail to explore the complex relationship between tasks and thus have limited performance. The methods centring on optimizing the relation structure of tasks are not capable of selecting meaningful features and have a bi-convex objective function which results in high computation complexity of the associated optimization algorithm. Unlike these multi-task learning methods, motivated by a simple and direct idea that the state of a system at the current time point should be related to all previous time points, we first propose a novel relation structure, termed adaptive global temporal relation structure (AGTS). Then we integrate the widely used sparse group Lasso, fused Lasso with AGTS to propose a novel convex multi-task learning formulation that not only performs feature selection but also adaptively captures the global temporal task relatedness. Since the existence of three non-smooth penalties, the objective function is challenging to solve. We first design an optimization algorithm based on the alternating direction method of multipliers (ADMM). Considering that the worst-case convergence rate of ADMM is only sub-linear, we then devise an efficient algorithm based on the accelerated gradient method which has the optimal convergence rate among first-order methods. We show the proximal operator of several non-smooth penalties can be solved efficiently due to the special structure of our formulation. Experimental results on four real-world datasets demonstrate that our approach not only outperforms multiple baseline MTL methods in terms of effectiveness but also has high efficiency.

Keywords

References

  1. IEEE Trans Knowl Data Eng. 2022 Oct;34(10):4854-4873 [PMID: 37915376]
  2. KDD. 2012;2012:1095-1103 [PMID: 25309808]
  3. IEEE Trans Biomed Eng. 2010 Apr;57(4):884-93 [PMID: 19932995]
  4. BMJ. 2021 May 19;373:n1239 [PMID: 34011499]
  5. KDD. 2012 Aug 12;2012:895-903 [PMID: 24078896]
  6. Biometrika. 2009 Jun;96(2):339-355 [PMID: 20037673]
  7. J Mach Learn Res. 2016 Apr;17: [PMID: 28428735]
  8. Nat Hum Behav. 2020 Dec;4(12):1303-1312 [PMID: 33199859]
  9. J Magn Reson Imaging. 2008 Apr;27(4):685-91 [PMID: 18302232]
  10. Neuroimage. 2013 Sep;78:233-48 [PMID: 23583359]
  11. IEEE Trans Pattern Anal Mach Intell. 2017 Jan;39(1):102-114 [PMID: 26955018]
  12. ACM Trans Knowl Discov Data. 2012 Feb 1;5(4):22 [PMID: 24077658]
  13. IEEE Trans Pattern Anal Mach Intell. 2013 Sep;35(9):2104-16 [PMID: 23868773]
  14. Biometrics. 2015 Mar;71(1):53-62 [PMID: 25257196]

Word Cloud

Created with Highcharts 10.0.0structurelearningmethodsmulti-taskrelationtemporalproposenovelformulationtimetasksfeatureselectionalgorithmefficientprogressionstatemultipleperformanceoptimizingtaskobjectivefunctionresultshighoptimizationfirstadaptiveglobalAGTSLassoalsonon-smoothpenaltiesbasedmethodADMMconvergenceratepredictionpaperclassproblemswillcontinuouslychangeusesharedknowledgeinformationimproveexistingmainlyfocususuallyfailexplorecomplexrelationshipthuslimitedcentringcapableselectingmeaningfulfeaturesbi-convexcomputationcomplexityassociatedUnlikemotivatedsimpledirectideasystemcurrentpointrelatedpreviouspointstermedintegratewidelyusedsparsegroupfusedconvexperformsadaptivelycapturesrelatednessSinceexistencethreechallengingsolvedesignalternatingdirectionmultipliersConsideringworst-casesub-lineardeviseacceleratedgradientoptimalamongfirst-ordershowproximaloperatorseveralcansolvedefficientlyduespecialExperimentalfourreal-worlddatasetsdemonstrateapproachoutperformsbaselineMTLtermseffectivenessefficiencyEfficientAdaptiveMulti-taskProgression

Similar Articles

Cited By

No available data.