I was looking at tensorflow examples by Aymeric Damien (https://github.com/aymericdamien/TensorFlow-Examples/blob/master/examples/3_NeuralNetworks/multilayer_perceptron.py) and in multilayer_perceptron.py he uses a neural net to classify MNIST digits. I think he is using a neural network with 784inputs, with 2 hidden layers with 256 neurons each, and 10 outputs. Am I correct? How do matrix dimensions in weights and biases in multilayer_perceptron.py correspond with ANN "dimensions" (#inputs, #hidden layers, #output, #neurons in each hidden layer, etc. Thank you!
Dimensions of this neural network i.e (4 inputs, 2 hidden layer with X neurons each, etc)
2k views Asked by Richter Brzeski At
1
There are 1 answers
Related Questions in MACHINE-LEARNING
- How to cluster a set of strings?
- Enforcing that inputs sum to 1 and are contained in the unit interval in scikit-learn
- scikit-learn preperation
- Spark MLLib How to ignore features when training a classifier
- Increasing the efficiency of equipment using Amazon Machine Learning
- How to interpret scikit's learn confusion matrix and classification report?
- Amazon Machine Learning for sentiment analysis
- What Machine Learning algorithm would be appropriate?
- LDA generated topics
- Spectral clustering with Similarity matrix constructed by jaccard coefficient
Related Questions in NEURAL-NETWORK
- How to choose good SURF feature keypoints?
- How to avoid overfitting (Encog3 C#)?
- Run out of VRAM using Theano on Amazon cluster
- Calculating equation from image in Java
- Print output of a Theano network
- Torch Lua: Why is my gradient descent not optimizing the error?
- How can I train a neural (pattern recognition) network multiple times in matlab?
- Using Convolution Neural Net with Lasagne in Python error
- Random number of hidden units improves accuracy/F-score on test set
- Matlab example code for deep belief network for classification
Related Questions in TENSORFLOW
- (Tensorflow)Does the op assign change the gradient computation?
- Tensorflow Windows Accessing Folders Denied:"NewRandomAccessFile failed to Create/Open: Access is denied. ; Input/output error"
- Android App TensorFlow Google Cloud ML
- Convert Tensorflow model to Caffe model
- Google Tensorflow LSTMCell Variables Mapping to Hochreiter97_lstm.pdf paper
- additive Gaussian noise in Tensorflow
- TFlearn evaluate method results meaning
- Regularization losses Tensorflow - TRAINABLE_VARIABLES to Tensor Array
- feed picture to model tensorflow for training
- Fail to read the new format of tensorflow checkpoint?
Related Questions in ARTIFICIAL-INTELLIGENCE
- Developing a Checkers (Draughts) engine, how to begin?
- STRIPS representation of monkey in the lab
- Difference between weak AI and strong AI?
- Q-learning in game not working as expected
- How do I use a class as a template in C#?
- prolog rules as arguments
- How to evaluate a recurrent connection in an artificial neural network?
- AS3 AI barrier detection and movement
- How to simulate neural network by changing different parameters after training in MATLAB?
- Debugging Neural Network for (Natural Language) Tagging
Related Questions in MNIST
- What is the image type in MNIST dataset?
- Error in building mean image file(Caffe)
- Dimensions of this neural network i.e (4 inputs, 2 hidden layer with X neurons each, etc)
- Feeding image into tensorflow
- how to change softmaxlayer with regression in matconvnet
- How can I use Mnist dataset in MapReduce project?
- How to create a file like the MNIST dataset
- Failed to unpack MNIST data set using python3
- InvalidArgumentError while coding MNIST tutorial
- Logistic Regression implementation with MNIST - not converging?
Popular Questions
- How do I undo the most recent local commits in Git?
- How can I remove a specific item from an array in JavaScript?
- How do I delete a Git branch locally and remotely?
- Find all files containing a specific text (string) on Linux?
- How do I revert a Git repository to a previous commit?
- How do I create an HTML button that acts like a link?
- How do I check out a remote Git branch?
- How do I force "git pull" to overwrite local files?
- How do I list all files of a directory?
- How to check whether a string contains a substring in JavaScript?
- How do I redirect to another webpage?
- How can I iterate over rows in a Pandas DataFrame?
- How do I convert a String to an int in Java?
- Does Python have a string 'contains' substring method?
- How do I check if a string contains a specific word?
Popular Tags
Trending Questions
- UIImageView Frame Doesn't Reflect Constraints
- Is it possible to use adb commands to click on a view by finding its ID?
- How to create a new web character symbol recognizable by html/javascript?
- Why isn't my CSS3 animation smooth in Google Chrome (but very smooth on other browsers)?
- Heap Gives Page Fault
- Connect ffmpeg to Visual Studio 2008
- Both Object- and ValueAnimator jumps when Duration is set above API LvL 24
- How to avoid default initialization of objects in std::vector?
- second argument of the command line arguments in a format other than char** argv or char* argv[]
- How to improve efficiency of algorithm which generates next lexicographic permutation?
- Navigating to the another actvity app getting crash in android
- How to read the particular message format in android and store in sqlite database?
- Resetting inventory status after order is cancelled
- Efficiently compute powers of X in SSE/AVX
- Insert into an external database using ajax and php : POST 500 (Internal Server Error)
This is a 3-layer neural network (2 hidden layers and an output layer).
The connection between the inputs to the first hidden layer has 784 x 256 weights with 256 biases. This configuration is due to the fact that each of the 784 inputs is fully connected to the 256 hidden layer nodes, and each hidden layer node has 1 bias.
The connection between that first hidden layer to the second hidden layer has 256 x 256 weights due to full connectivity between the layers. The second layer's 256 nodes each has 1 bias.
The connection between the second hidden layer and the output layer is similar. There are 256 x 10 weights (for the second hidden layer's 256 nodes and the output layer's 10 nodes), and each output node has 1 bias.
There are thus 785*256 + 256*256 + 256*10 = 269,056 weights and 256 + 256 + 10 = 522 biases.
The figure below should explain it fully.