roatienza / Deep-Learning-Experiments

Videos, notes and experiments to understand deep learning
MIT License
1.1k stars 760 forks source link

Optimization in Deep Learning #17

Closed rrmina closed 6 years ago

rrmina commented 6 years ago

image

Hello. Prof Atienza. You said here (Lecture 3, Slide 23) that we can only use gradient descent in optimizing deep learning models (particularly loss function). I understand that both closed-form solution and Newton's method are computationally painful to use, that non-iterative 1st order optimization is inefficient, and that convex optimization is also not an option since most deep learning models have multiple extremas. We are therefore left only with gradient descent.

Are there other approaches to optimization in deep learning?

roatienza commented 6 years ago

The whole idea of deep learning is based on gradient-based method optimization. There are areas of AI (eg Evolutionary Computing) that attempt to use gradient-free approaches.