|
Boris Galerkin posted:I'm just not really sure how this is any different from newton's method or like basically any numerical method that minimizes an objective/loss function. Newton's method and other second-order techniques aren't really appropriate for NNs for a couple reasons, not least of which being that the Hessian is so enormous: even if you don't store it you still have to compute with it. Backpropagation is just a way of quickly computing the gradient of the objective function that works with the particular structure of NNs.
|
# ¿ Jun 15, 2022 19:51 |
|
|
# ¿ May 18, 2024 13:03 |