Cargando…

Training a Two-Layer ReLU Network Analytically

Neural networks are usually trained with different variants of gradient descent-based optimization algorithms such as the stochastic gradient descent or the Adam optimizer. Recent theoretical work states that the critical points (where the gradient of the loss is zero) of two-layer ReLU networks wit...

Descripción completa

Detalles Bibliográficos
Autor principal: Barbu, Adrian
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2023
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10146164/
https://www.ncbi.nlm.nih.gov/pubmed/37112413
http://dx.doi.org/10.3390/s23084072