Abstract

In the last decade, we have witnessed the introduction of several novel deep neural network (DNN) architectures exhibiting ever-increasing performance across diverse tasks. Explaining the upward trend of their performance, however, remains difficult as different DNN architectures of comparable depth and width - common factors associated with their expressive power - may exhibit a drastically different performance even when trained on the same dataset. In this paper, we introduce the concept of the non-linearity signature of DNN, the first theoretically sound solution for approximately measuring the non-linearity of deep neural networks. Built upon a score derived from closed-form optimal transport mappings, this signature provides a better understanding of the inner workings of a wide range of DNN architectures and learning paradigms, with a particular emphasis on the computer vision task. We provide extensive experimental results that highlight the practical usefulness of the proposed non-linearity signature and its potential for long-reaching implications. The code for our work is available at https://github.com/qbouniot/AffScoreDeep.

Original languageEnglish
Title of host publication2025 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)
PublisherIEEE
Pages25250-25260
Number of pages11
ISBN (Electronic)979-8-3315-4364-8
DOIs
Publication statusPublished - Aug 2025
MoE publication typeA4 Conference publication
EventIEEE Conference on Computer Vision and Pattern Recognition - Nashville, TN, USA, Nashville, United States
Duration: 10 Jun 202517 Jun 2025

Publication series

NameProceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition
PublisherIEEE
ISSN (Print)1063-6919

Conference

ConferenceIEEE Conference on Computer Vision and Pattern Recognition
Abbreviated titleCVPR
Country/TerritoryUnited States
CityNashville
Period10/06/202517/06/2025

Funding

This work was supported by the EU Horizon project ELIAS (No. 101120237) and by the French National Research Agency (ANR) under the PEPR IA FOUNDRY (ANR-23-PEIA-0003), the LIMPID (ANR-20-CE23-0028) and the FAR-SEE (ANR-24-CE23-0921) projects. We also thank Florence d'Alche-Buc, Remi Flamary and Devis Tuia for providing feedback on early versions of the manuscript.

Keywords

  • activation functions
  • deep neural networks
  • optimal transport

Fingerprint

Dive into the research topics of 'From Alexnet to Transformers : Measuring the Non-linearity of Deep Neural Networks with Affine Optimal Transport'. Together they form a unique fingerprint.

Cite this