Abstract

State-of-the-art deep survival prediction approaches expand network parameters to accommodate performance over a fine discretization of output time. For medical applications where data are limited, the regression-based Deepsurv approach is more advantageous because its continuous output design limits unnecessary network parameters. Despite the practical advantage, the typical network lacks control over the feature distribution causing the network to be more prone to noisy information and occasional poor prediction performance. We propose a novel projection loss as a regularizing objective to improve the time-to-event Deepsurv model. The loss formulation maximizes the lower bound of the multiple-correlation coefficient between the network’s features and the desired hazard value. Reducing the loss also theoretically lowers the upper bound on the likelihood of discordant pair and improves C-index performance. We observe superior performances and robustness of regularized Deepsurv over many state-of-the-art approaches in our experiments with five public medical datasets and two cross-cohort validation tasks.

Highlights

  • INTRODUCTIONKnown as time-to-event analysis, is a crucial task in medical prognosis and risk assessment

  • Survival analysis, known as time-to-event analysis, is a crucial task in medical prognosis and risk assessment

  • Unlike the typical regression tasks, the model has to account for right-censored outcomes The problem has been widely investigated in the Statistics community [18]

Read more

Summary

INTRODUCTION

Known as time-to-event analysis, is a crucial task in medical prognosis and risk assessment. A prominent trend in developing deep survival approaches is discretizing possible survival outcomes for classification architecture in modeling hitting time distribution. DeepHit [12] is the state-of-the-art method that combines the discretized density training similar to that of the PMF with a multi-task network for the prediction Notice that these discrete-time network can adapt to new input forms by changing the input layer architecture (e.g., Convolutional Neural Network [3], [40].) The drawback of this approach is the requirement of hitting time discretization, which inconveniently introduces a trade-off between the number of parameters and the granularity of output time. Inspired by the advantages offered by the deep survival regression approach, we investigate the lack of feature distribution control during the regular Deepsurv training that exposes the network to suboptimal performance. We discuss the advantages of Deepsurv over the other baselines in further sections

PARAMETRIC SURVIVAL REGRESSION
DEEPSURV NETWORK
EXPERIMENTS
DATASETS
Findings
CONCLUSION
Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call