{Reference Type}: Journal Article {Title}: A topological description of loss surfaces based on Betti Numbers. {Author}: Bucarelli MS;D'Inverno GA;Bianchini M;Scarselli F;Silvestri F; {Journal}: Neural Netw {Volume}: 178 {Issue}: 0 {Year}: 2024 Oct 14 {Factor}: 9.657 {DOI}: 10.1016/j.neunet.2024.106465 {Abstract}: In the context of deep learning models, attention has recently been paid to studying the surface of the loss function in order to better understand training with methods based on gradient descent. This search for an appropriate description, both analytical and topological, has led to numerous efforts in identifying spurious minima and characterize gradient dynamics. Our work aims to contribute to this field by providing a topological measure for evaluating loss complexity in the case of multilayer neural networks. We compare deep and shallow architectures with common sigmoidal activation functions by deriving upper and lower bounds for the complexity of their respective loss functions and revealing how that complexity is influenced by the number of hidden units, training models, and the activation function used. Additionally, we found that certain variations in the loss function or model architecture, such as adding an ℓ2 regularization term or implementing skip connections in a feedforward network, do not affect loss topology in specific cases.