I have a very complex LSTM based neural network model which I'm training on Quora Duplicate Question pairs. There are approximately 400 000 sentence pairs in the original dataset. It would take a lot of processing power and computation time to train on the entire (or 80%) dataset. Would it be unwise if I choose a random subset of the dataset (say 8000 pairs only) for training and 2000 for testing? Would it have a severe impact on the performance? Is always "more the data, better the model" true?
1
There are 1 answers
Related Questions in DEEP-LEARNING
- Influence of Unused FFN on Model Accuracy in PyTorch
- How to train a model with CSV files of multiple patients?
- Does tensorflow have a way of calculating input importance for simple neural networks
- What is the alternative to module: tf.keras.preprocessing?
- Which library can replace causal_conv1d in machine learning programming?
- My MSE and MAE are low, but my R2 is not good, how to improve it?
- Sketch Guided Text to Image Generation
- ValueError: The shape of the target variable and the shape of the target value in `variable.assign(value)` must match
- a problem for save and load a pytorch model
- Optuna Hyperband Algorithm Not Following Expected Model Training Scheme
- How can I resolve this error and work smoothly in deep learning?
- Difference between model.evaluate and metrics.accuracy_score
- Integrating Mesonet algorithm with a webUI for deepfake detection model
- How can i edit the "wake-word-detection notebook" on coursera so it fit my own word?
- PyTorch training on M2 GPU slower than Colab CPU
Related Questions in NLP
- Seeking Python Libraries for Removing Extraneous Characters and Spaces in Text
- Clarification on T5 Model Pre-training Objective and Denoising Process
- The training accuracy and the validation accuracy curves are almost parallel to each other. Is the model overfitting?
- Give Bert an input and ask him to predict. In this input, can Bert apply the first word prediction result to all subsequent predictions?
- Output of Cosine Similarity is not as expected
- Getting an error while using the open ai api to summarize news atricles
- SpanRuler on Retokenized tokens links back to original token text, not the token text with a split (space) introduced
- Should I use beam search on validation phase?
- Dialogflow failing to dectect the correct intent
- How to detect if two sentences are simmilar, not in meaning, but in syllables/words?
- Is BertForSequenceClassification using the CLS vector?
- Issue with memory when using spacy_universal_sentence_encoder for similarity detection
- Why does the Cloud Natural Language Model API return so many NULLs?
- Is there any OCR or technique that can recognize/identify radio buttons printed out in the form of pdf document?
- Model, lexicon to do fine grained emotions analysis on text in r
Related Questions in SEMANTIC-COMPARISON
- Should I be using whole available data for training my deep learning model ? What are the pros and cons of using only a subset?
- Use dkpro semantic similarity with uby
- NUnit: Track down difference between expected and actual in composite result using SemanticComparison .NET library
- Is it possible in AutoFixture SemanticComparison to set custom comparer for specified type
- How to compare two anonymous types or two collection of different types using SemanticComparison
- Why can't I create a likeness proxy when abstract class not exposing every constructor argument?
- Comparing nested object properties using SemanticComparison
- Verifying complete Mapping of an [unordered] Collection/Set of Items in a Unit Test
- Likeness - polishing and packaging
- Why doesn't Autofixture Likeness behave like I'd expect for one of these two tests?
- Trouble using Autofixture's CreateProxy to use Likeness, SemanticComparison features
- Why doesn't simple test pass using AutoFixture Freeze, SemanticComparison Likeness and CreateProxy?
- Applying [AutoFixture] SemanticComparison OfLikeness to sequences / collections / arrays / IEnumerable
Popular Questions
- How do I undo the most recent local commits in Git?
- How can I remove a specific item from an array in JavaScript?
- How do I delete a Git branch locally and remotely?
- Find all files containing a specific text (string) on Linux?
- How do I revert a Git repository to a previous commit?
- How do I create an HTML button that acts like a link?
- How do I check out a remote Git branch?
- How do I force "git pull" to overwrite local files?
- How do I list all files of a directory?
- How to check whether a string contains a substring in JavaScript?
- How do I redirect to another webpage?
- How can I iterate over rows in a Pandas DataFrame?
- How do I convert a String to an int in Java?
- Does Python have a string 'contains' substring method?
- How do I check if a string contains a specific word?
Popular Tags
Trending Questions
- UIImageView Frame Doesn't Reflect Constraints
- Is it possible to use adb commands to click on a view by finding its ID?
- How to create a new web character symbol recognizable by html/javascript?
- Why isn't my CSS3 animation smooth in Google Chrome (but very smooth on other browsers)?
- Heap Gives Page Fault
- Connect ffmpeg to Visual Studio 2008
- Both Object- and ValueAnimator jumps when Duration is set above API LvL 24
- How to avoid default initialization of objects in std::vector?
- second argument of the command line arguments in a format other than char** argv or char* argv[]
- How to improve efficiency of algorithm which generates next lexicographic permutation?
- Navigating to the another actvity app getting crash in android
- How to read the particular message format in android and store in sqlite database?
- Resetting inventory status after order is cancelled
- Efficiently compute powers of X in SSE/AVX
- Insert into an external database using ajax and php : POST 500 (Internal Server Error)
As a Rule of Thumb, Deep Neural Networks usually benefit from more data.
If you have a well described model and properly engineered your inputs, you will lose if you chose a smaller subset of your dataset.
However, you could always evaluate this by using metrics. Check how your loss decreases at every sample size, starting from your 8000 pairs.
For big problems, you always have to keep in mind that computation time is usually also big.