Is plain torch.optim.SGD
doing the same as gradient descent?
#1146
Unanswered
shenhai-ran
asked this question in
Q&A
Replies: 0 comments
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
Hi,
I am wondering about the implementation of
torch.optim.SGD
.If I don't use any batch size or dataloader, as the code snippet below from Chatper 1.
In this case, if I understand it right, there is no stochastic, but calculating native gradient descent for the whole dataset, and the
epoch
is basically the counter of steps in the optimization.Do I get it right?
Beta Was this translation helpful? Give feedback.
All reactions