I've been reading a lot about transformers and self attention and have seen both BERT and GPT-2 are a newer version that only use an encoder transformer (BERT) and decoder transformer (GPT-2). I've been trying to build a decoder only model for myself for next sequence prediction but am confused by one thing. I'm using PyTorch and have looked at thereSeq2Seq tutorial and then looked into the Transformer Decoder Block which is made up of Transformer Decoder Layers. My confusion comes from the memory these need to be passed as well. In the documentation they say memory is the last layer of the encoder block which makes sense for a Seq2Seq model but I'm wanting to make a decoder only model. So my question is what do you pass a decoder only model like GPT-2 for memory if you do not have an encoder?
Related Questions in PYTHON
- How to store a date/time in sqlite (or something similar to a date)
- Instagrapi recently showing HTTPError and UnknownError
- How to Retrieve Data from an MySQL Database and Display it in a GUI?
- How to create a regular expression to partition a string that terminates in either ": 45" or ",", without the ": "
- Python Geopandas unable to convert latitude longitude to points
- Influence of Unused FFN on Model Accuracy in PyTorch
- Seeking Python Libraries for Removing Extraneous Characters and Spaces in Text
- Writes to child subprocess.Popen.stdin don't work from within process group?
- Conda has two different python binarys (python and python3) with the same version for a single environment. Why?
- Problem with add new attribute in table with BOTO3 on python
- Can't install packages in python conda environment
- Setting diagonal of a matrix to zero
- List of numbers converted to list of strings to iterate over it. But receiving TypeError messages
- Basic Python Question: Shortening If Statements
- Python and regex, can't understand why some words are left out of the match
Related Questions in PYTORCH
- Influence of Unused FFN on Model Accuracy in PyTorch
- Conda CMAKE CXX Compiler error while compiling Pytorch
- Which library can replace causal_conv1d in machine learning programming?
- yolo v5 export to torchscript: how to generate constants.pkl
- Pytorch distribute process across nodes and gpu
- My ICNN doesn't seem to work for any n_hidden
- a problem for save and load a pytorch model
- The meaning of an out_channel in nn.Conv2d pytorch
- config QConfig in pytorch QAT
- Can't load the saved model in PyTorch
- How can I convert a flax.linen.Module to a torch.nn.Module?
- Snuffle in PyTorch Dataloader
- Cuda out of Memory but I have no free space
- Can not load scripted model using torch::jit::load
- Should I train my model with a set of pictures as one input data or I need to crop to small one using Pytorch
Related Questions in DECODER
- Parlance's CTC prefix Beam search decoding logic insights
- PNG decode data from IDAT chunk
- Android media codec: Mismatch between encoder and decoder buffer sizes
- How to save model after adding attention mechanism?
- How do I use a std::map to change a vector of strings into another vector of characters?
- How to confirm that whether decoder support HLG or HDR format in Android?
- Strange names in Javascript code - how do I decode it?
- How to find frame length of stereo mp3
- Splitting Pytorch transformer into encoder & decoder
- How to hardware decode a Stream from an RTSP source
- How to solve "Error: Type 'DecoderCallback' not found." to run a Flutter app locally on an iPhone?
- How to decode and fmp4 + chunked live stream
- GStreamer Pipeline: Decoding H.264 from Appsrc to Appsink
- How to decode .twb file
- How to set the Microsoft WMF VP9 decoder to decode VP9 videos
Related Questions in TRANSFORMER-MODEL
- Understanding batching in pytorch models
- Using an upstream-downstream ML model, with the upstream being Wav2Vec 2.0 transformer and the downstream CNN. The model's accuracy is plateaued, why?
- How to obtain latent vectors from fine-tuned model with transformers
- What is the difference between PEFT and RAFT?
- Improving Train Punctuality Prediction Using a Transformer Model: Model Setup and Performance Issues
- How to remove layers in Huggingface's transformers GPT2 pre-trained models?
- NPL Keras transformers model not converging
- How to convert pretrained hugging face model to .pt and run it fully locally?
- LLaMA2 Workload Traces
- Inference question through LoRA in Whisper model
- is there any way to use RL for decoder only models
- What's the exact input size in MultiHead-Attention of BERT?
- How to solve this error "UnsupportedOperation: fileno"
- Transformers // Predicting next transaction based on sequence of previous transactions // Sequence2One task
- I was using colab: I want to run a .py file having argparse function to train a model
Related Questions in GPT-2
- Can't resolve KeyError in Pandas
- My API with OpenAI is not working properly
- How to use GPT2 as a Question-Answering System (What to put in context?)
- terminal keeps saying "The name tf.disable_v2_behavior is deprecated." even though i thought i changed it
- Struggling with Hugging Face PEFT
- How can i use two textual inputs for GPT2 based regression model effectively?
- Converting a GPT2 h5 model to torch for conversion to ggml - shape mismatches
- how can I feed GPT2 with prespecified embeddings?
- Issue with fine tuning GPT2 for IPA transcription model
- Error "Layer is not connected, no input to return" when loading pre-trained model
- loading gpt2 simple checkpoint getting OpError: /content/model.data-00000-of-00001; No such file or directory
- Handling GPT-2 Tokenization and Encoding in a Text Processing Pipeline
- gpt2 tokenizer issue ( AssertionError: Cannot handle batch sizes > 1 if no padding token is defined )
- Transformers fine-tuning script doesn't work with FSDP
- Aitextgen doesn't generate text when generate with prompt
Popular Questions
- How do I undo the most recent local commits in Git?
- How can I remove a specific item from an array in JavaScript?
- How do I delete a Git branch locally and remotely?
- Find all files containing a specific text (string) on Linux?
- How do I revert a Git repository to a previous commit?
- How do I create an HTML button that acts like a link?
- How do I check out a remote Git branch?
- How do I force "git pull" to overwrite local files?
- How do I list all files of a directory?
- How to check whether a string contains a substring in JavaScript?
- How do I redirect to another webpage?
- How can I iterate over rows in a Pandas DataFrame?
- How do I convert a String to an int in Java?
- Does Python have a string 'contains' substring method?
- How do I check if a string contains a specific word?
Trending Questions
- UIImageView Frame Doesn't Reflect Constraints
- Is it possible to use adb commands to click on a view by finding its ID?
- How to create a new web character symbol recognizable by html/javascript?
- Why isn't my CSS3 animation smooth in Google Chrome (but very smooth on other browsers)?
- Heap Gives Page Fault
- Connect ffmpeg to Visual Studio 2008
- Both Object- and ValueAnimator jumps when Duration is set above API LvL 24
- How to avoid default initialization of objects in std::vector?
- second argument of the command line arguments in a format other than char** argv or char* argv[]
- How to improve efficiency of algorithm which generates next lexicographic permutation?
- Navigating to the another actvity app getting crash in android
- How to read the particular message format in android and store in sqlite database?
- Resetting inventory status after order is cancelled
- Efficiently compute powers of X in SSE/AVX
- Insert into an external database using ajax and php : POST 500 (Internal Server Error)
After further investigation I believe I can now answer this myself. A decoder only transformer doesn't actually use any memory as there is no encoder-decoder self attention in it like there is in a encoder-decoder transformer. A decoder only transformer looks a lot like an encoder transformer only instead it uses a masked self attention layer over a self attention layer. In order to do this you can pass a square subsequent mask (upper triangle) so that the model cannot look forward to achieve a decoder only model like found in GPT-2/GPT-3.