1 year ago
#228927

Skywear
Problem of convergence when trying to implement a basic batch gradient descent algortihm in python
I am working with this dataset and I want to plot the training error as a function of the number of iterations of gradient descent with square loss. Here is my attempt :
import numpy as np
import scipy.io
import matplotlib.pyplot as plt
mat = scipy.io.loadmat('data_orsay_2017.mat')
x=mat['Xtrain']
y=mat['ytrain']
def gradientdescent(x,y,n,alpha,max_iterations): # n is the sample size, alpha is the learning rate
d = x.shape[1] # dimension of the data
theta = np.random.random(d)
error = []
for j in range(max_iterations):
prediction = x.dot(theta)
cost = 1/(2*n)*sum((y[i,0]-prediction[i])**2 for i in range(n))
error.append(cost)
grad = (1/n) * sum((prediction[i] - y[i,0])*x[i] for i in range(n))
theta-=alpha*grad
return (theta,error)
plt.xlabel("iterations")
plt.ylabel("training error for square loss")
plt.plot(range(1000),gradientdescent(x,y,1000,1,1000)[1])
plt.show()
However it seems that the cost does not converge to zero (here is an example with 1000 iterations, it doesn't go below 0.27) Is there something wrong with my algorithm ?
python
gradient-descent
convergence
0 Answers
Your Answer