Implicit Sparse Regularization: The Impact of Depth and Early Stopping Academic Article uri icon

abstract

  • In this paper, we study the implicit bias of gradient descent for sparse regression. We extend results on regression with quadratic parametrization, which amounts to depth-2 diagonal linear networks, to more general depth-N networks, under more realistic settings of noise and correlated designs. We show that early stopping is crucial for gradient descent to converge to a sparse model, a phenomenon that we call implicit sparse regularization. This result is in sharp contrast to known results for noiseless and uncorrelated-design cases. We characterize the impact of depth and early stopping and show that for a general depth parameter N, gradient descent with early stopping achieves minimax optimal sparse recovery with sufficiently small initialization and step size. In particular, we show that increasing depth enlarges the scale of working initialization and the early-stopping window so that this implicit sparse regularization effect is more likely to take place.

author list (cited authors)

  • Li, J., Nguyen, T. V., Hegde, C., & Wong, R.

complete list of authors

  • Li, Jiangyuan||Nguyen, Thanh V||Hegde, Chinmay||Wong, Raymond KW

publication date

  • August 2021