I assumed the "stochastic" in Stochastic Gradient Descent came from the random selection of samples within each batch. But the articles I have read on the topic seem to indicate that SGD makes a small move (weight change) with every data point. How does Tensorflow implement it?
Does SGD in Tensorflow make a move with each data point?
290 views Asked by Mastiff At
1
There are 1 answers
Related Questions in TENSORFLOW
- A deterministic GPU implementation of fused batch-norm backprop, when training is disabled, is not currently available
- Keras similarity calculation. Enumerating distance between two tensors, which indicates as lists
- Does tensorflow have a way of calculating input importance for simple neural networks
- How to predict input parameters from target parameter in a machine learning model?
- Windows 10 TensorFlow cannot detect Nvidia GPU
- unable to use ignore_class in SparseCategoricalCrossentropy
- Why is this code not working? I've tried everything and everything seems to be fine, but no
- Why convert jpeg into tfrecords?
- ValueError: The shape of the target variable and the shape of the target value in `variable.assign(value)` must match
- The kernel appears to have died. It will restart automatically. whenever i try to run the plt.imshow() and plt.show() function in jupyter notebook
- Pneumonia detection, using transfer learning
- Cannot install tensorflow ver 2.3.0 (distribution not found)
- AttributeError: module 'keras._tf_keras.keras.layers' has no attribute 'experimental'
- Error while loading .keras model: Layer node index out of bounds
- prediction model with python tensorflow and keras, gives error when predicting
Related Questions in KERAS
- Keras similarity calculation. Enumerating distance between two tensors, which indicates as lists
- How to predict input parameters from target parameter in a machine learning model?
- What is the alternative to module: tf.keras.preprocessing?
- My MSE and MAE are low, but my R2 is not good, how to improve it?
- No module named 'keras.layers.core
- AttributeError: 'Sequential' object has no attribute 'predict_classes'. Did you mean: 'predict_step'?
- AttributeError: module 'keras._tf_keras.keras.layers' has no attribute 'experimental'
- Error while loading .keras model: Layer node index out of bounds
- prediction model with python tensorflow and keras, gives error when predicting
- Recommended way to use Gymnasium with neural networks to avoid overheads in model.fit and model.predict
- Keras OCR - Getting different results from Keras
- No gradients provided for any variable in R
- Error Encountered: InvalidArgumentError: Graph execution error using Keras and Transformers
- How to import logsumexp from keras.backend?
- Keras predict/predict_on_batch giving different answers than predict_step/__call__()
Related Questions in STOCHASTIC-GRADIENT
- unexpected output with stochastic gradient descent algorithm for linear regression
- How to get Stochastic Gradient Descent result in contour plot
- Gradient Descent Cost Function Blows up after locating minimum value
- Tensorflow gradient descent algorithm
- how to set the gradient for a network in pytorch
- How variable alpha changes SGDRegressor behavior for outlier?
- Visualize Stochastic Gradient Descent using Contour plot in Python
- Encountering a TypeError: can't multiply sequence by non-int of type 'float' when creating an SGD algorithm
- Why are the gradients not equivalent when using loss.backward() v.s torch.auto.grad?
- pytorch: implementing a custom optimizer
- SGD breaks down when encountering unseen values
- Does SGD in Tensorflow make a move with each data point?
- GAN - Generator loss decreasing but Discriminator fake loss increase after a initial drop, why?
- How can I get my neural net to correctly do linear regression?
- What is the problem with this SGD loss graph?
Popular Questions
- How do I undo the most recent local commits in Git?
- How can I remove a specific item from an array in JavaScript?
- How do I delete a Git branch locally and remotely?
- Find all files containing a specific text (string) on Linux?
- How do I revert a Git repository to a previous commit?
- How do I create an HTML button that acts like a link?
- How do I check out a remote Git branch?
- How do I force "git pull" to overwrite local files?
- How do I list all files of a directory?
- How to check whether a string contains a substring in JavaScript?
- How do I redirect to another webpage?
- How can I iterate over rows in a Pandas DataFrame?
- How do I convert a String to an int in Java?
- Does Python have a string 'contains' substring method?
- How do I check if a string contains a specific word?
Popular Tags
Trending Questions
- UIImageView Frame Doesn't Reflect Constraints
- Is it possible to use adb commands to click on a view by finding its ID?
- How to create a new web character symbol recognizable by html/javascript?
- Why isn't my CSS3 animation smooth in Google Chrome (but very smooth on other browsers)?
- Heap Gives Page Fault
- Connect ffmpeg to Visual Studio 2008
- Both Object- and ValueAnimator jumps when Duration is set above API LvL 24
- How to avoid default initialization of objects in std::vector?
- second argument of the command line arguments in a format other than char** argv or char* argv[]
- How to improve efficiency of algorithm which generates next lexicographic permutation?
- Navigating to the another actvity app getting crash in android
- How to read the particular message format in android and store in sqlite database?
- Resetting inventory status after order is cancelled
- Efficiently compute powers of X in SSE/AVX
- Insert into an external database using ajax and php : POST 500 (Internal Server Error)
Yes, SGD is indeed randomly sampled, but the point here is a little different.
SGD itself doesn't do the sampling. You do the sampling by batching and hopefully shuffling between each epoch.
GD means you generate gradients for each weight after forward propping the entire dataset (batchsize = cardinality, and steps per epoch = 1). If your batch size is less than the cardinality of the dataset, then you are the one doing sampling, and you are running SGD not GD.
The implementation is pretty simple, and something like