Controlled learning of pointwise nonlinearities in neural-network-like architectures

We present a general variational framework for the training of freeform nonlinearities in layered computational architectures subject to some slope constraints. The regularization that we add to the traditional training loss penalizes the second-order total variation of each trainable activation. Th...

Full description

Saved in:
Bibliographic Details
Published inApplied and computational harmonic analysis Vol. 77; p. 101764
Main Authors Unser, Michael, Goujon, Alexis, Ducotterd, Stanislas
Format Journal Article
LanguageEnglish
Published Elsevier Inc 01.06.2025
Online AccessGet full text
ISSN1063-5203
DOI10.1016/j.acha.2025.101764

Cover

Loading…
More Information
Summary:We present a general variational framework for the training of freeform nonlinearities in layered computational architectures subject to some slope constraints. The regularization that we add to the traditional training loss penalizes the second-order total variation of each trainable activation. The slope constraints allow us to impose properties such as 1-Lipschitz stability, firm non-expansiveness, and monotonicity/invertibility. These properties are crucial to ensure the proper functioning of certain classes of signal-processing algorithms (e.g., plug-and-play schemes, unrolled proximal gradient, invertible flows). We prove that the global optimum of the stated constrained-optimization problem is achieved with nonlinearities that are adaptive nonuniform linear splines. We then show how to solve the resulting function-optimization problem numerically by representing the nonlinearities in a suitable (nonuniform) B-spline basis. Finally, we illustrate the use of our framework with the data-driven design of (weakly) convex regularizers for the denoising of images and the resolution of inverse problems.
ISSN:1063-5203
DOI:10.1016/j.acha.2025.101764