Implemented methods:
- Batch Gradient Descent
- Adadelta
Function to approximate: F(w, x): w_1 + w_2x + w_3x^2
Loss function: MSE
Further details could be found in paper.pdf
numpy
matplotlib
- Matthew D. Zeiler, ADADELTA: AN ADAPTIVE LEARNING RATE METHOD.
- Sebastian Ruder, An overview of gradient descent optimization algorithms.