Stable fixed points of combinatorial threshold-linear networks

Adv Appl Math. 2024 Mar:154:102652. doi: 10.1016/j.aam.2023.102652. Epub 2023 Dec 13.

Abstract

Combinatorial threshold-linear networks (CTLNs) are a special class of recurrent neural networks whose dynamics are tightly controlled by an underlying directed graph. Recurrent networks have long been used as models for associative memory and pattern completion, with stable fixed points playing the role of stored memory patterns in the network. In prior work, we showed that target-free cliques of the graph correspond to stable fixed points of the dynamics, and we conjectured that these are the only stable fixed points possible [1, 2]. In this paper, we prove that the conjecture holds in a variety of special cases, including for networks with very strong inhibition and graphs of size n4. We also provide further evi-dence for the conjecture by showing that sparse graphs and graphs that are nearly cliques can never support stable fixed points. Finally, we translate some results from extremal com-binatorics to obtain an upper bound on the number of stable fixed points of CTLNs in cases where the conjecture holds.

Keywords: 15; 34; 92; Collatz-Wielandt formula; attractor neural networks; cliques; stable fixed points; threshold-linear networks.