Hacker Read top | best | new | newcomments | leaders | about | bookmarklet login

isn't it just gradient descent?


sort by: page size:

Gradient descent is for non-linear problems where you can't directly invert or somehow linearize the problem.

does it mean we don't need gradient descent after all to achieve the same result?

I know what gradient descent is, thanks, I was referring to the rest of that mess.

"gradient descent" should be enough

Gradient descent in action.

Or gradient descent if you mentally negate the number in question. It's the same thing.

Thing is, gradient descent is not really a complex algorithm.

No. I think you should rework your gradient descent algorithm, it's bad.

Gradient descent is covered. We even show a pytorch implementation :)

AI is gradient descent?

Accelerated gradient descent is pretty common everywhere.

note that "gradient descent" isn't AI either. it's more computational linear algebra: a heuristic for numerical methods used to solve (usually to a local extrema) systems of equations without direct analytical solution(s).

People mostly use gradient descent to "solve" nonconvex problems.

They're all various optimization technique, variations on gradient descent.

Yeah, it's quite similar to "Learning to learn by gradient descent by gradient descent" and related works

What he calls back propagation is actually gradient descent.

You are not off base at all, thanks for clarify and sorry for the confusion, I did not mean to say it was using gradient descent. It's been a while. The term I was thinking of was multiple "simulated annealing".

Most likely, gradient descent with momentum.

So it doesn't use a neural network, but it is still optimized by gradient descent. Differentiability is the key!
next

Legal | privacy