Püttschneider, JensJensPüttschneiderFaulwasser, TimmTimmFaulwasser2026-02-062026-02-062026-04-01Automatica 186: 112767 (2026)https://hdl.handle.net/11420/61335The training of ResNets and neural ODEs can be formulated and analyzed from the perspective of optimal control. This paper proposes a dissipative formulation of the training of ResNets and neural ODEs for classification problems. Specifically, we consider a variant of the cross-entropy (label smoothing) as a loss function and as a regularization in the stage cost. Based on our dissipative formulation of the training, we prove that the training OCPs for ResNets and neural ODEs alike exhibit the turnpike phenomenon. We illustrate this finding with numerical results for the two spirals and MNIST datasets. Crucially, our training formulation ensures that the transformation of the data from input to output is achieved in the first layers. In the following layers, which constitute the turnpike, the data remains at an equilibrium state and therefore these layers do not contribute to the transformation learned. In principle, these layers can be pruned after training, resulting in a network with only the necessary number of layers thus simplifying tuning of hyperparameters.en0005-1098Automatica2026Elsevierhttps://creativecommons.org/licenses/by/4.0/Deep learningDissipativityHuber lossLabel smoothingManifold turnpikeNeural networksOptimal controlComputer Science, Information and General Works::006: Special computer methodsNatural Sciences and Mathematics::515: AnalysisComputer Science, Information and General Works::004: Computer SciencesOn dissipativity of cross-entropy loss in training ResNets — A turnpike towards architecture searchJournal Articlehttps://doi.org/10.15480/882.1663710.1016/j.automatica.2025.11276710.15480/882.16637Journal Article