I'm trying to write mini-batch gradient descent for log regression.
Given numpy matrices X_batch
(of shape (n_samples, n_features)
) and y_batch
(of shape (n_samples,)
).
Naive way is to write loop:
def calc_loss_grad(self, X_batch, y_batch):
n_samples, n_features = X_batch.shape
loss_grad = np.zeros((n_features,))
for i in range(n_samples):
sigm = sigmoid(X_batch[i] @ self.weights)
loss_grad += - (y_batch[i] - sigm) * X_batch[i]
return loss_grad
But seems like using loop is a bad idea w.r.t. speed. Any better ways? Pure numpy without a loop? Rewrite expression for gradient somehow?
Note that this algorithm is memory bandwidth limited. If you optimize this in a larger context (real application) there is very likely a higher speedup possible.
Example