I have a set of classified data with three labels, 'd', 'e', and 'k'. I want to train a classifier to identify 'd' and remove them from the dataset, then identify 'e'. Currently, I'm splitting the data into thirds, which I'll call X1, X2, X3. I train a learner L1 on X1, use that learner to remove 'd' labels on X2, which I then use to train a second learner L2, which I test on X3. Is this a reasonable approach, and is there an accepted standard in this kind of scenario?
Cross-validating with two different algorithms on one data set
77 views Asked by johnklawlor At
1
There are 1 answers
Related Questions in VALIDATION
- Terraform valdiate that one of N variables is set to "true"
- How to validate if Confirm Password is same or not with the Password in React Native using ValidateJS?
- How to create yup schema for dynamic array of different objects
- Quintic Number Number Counting Hash Function
- DropdownButtonFormField doesn't apply custom InputDecoration style
- Is there a way to set a cells value based on the value this cell held at a certain time, even when the cell value changes over time?
- Multiple regex expressions to check mobile number in javascript
- Java Pojos - Setter-Call (Field Touched) Detection
- Input Field Required
- Angular restore ngModel input field to it's previous value
- Bean Validaton : org.springframework.web.bind.MethodArgumentNotValidException
- javax validation not working on spring boot
- How to show warning message for unmatched confirm password
- Flutter TextFormField validation with Firestore
- eval_set in CatBoostRegressor
Related Questions in MACHINE-LEARNING
- Trained ML model with the camera module is not giving predictions
- Keras similarity calculation. Enumerating distance between two tensors, which indicates as lists
- How to get content of BLOCK types LAYOUT_TITLE, LAYOUT_SECTION_HEADER and LAYOUT_xx in Textract
- How to predict input parameters from target parameter in a machine learning model?
- The training accuracy and the validation accuracy curves are almost parallel to each other. Is the model overfitting?
- ImportError: cannot import name 'HuggingFaceInferenceAPI' from 'llama_index.llms' (unknown location)
- Which library can replace causal_conv1d in machine learning programming?
- Fine-Tuning Large Language Model on PDFs containing Text and Images
- Sketch Guided Text to Image Generation
- My ICNN doesn't seem to work for any n_hidden
- Optuna Hyperband Algorithm Not Following Expected Model Training Scheme
- How can I resolve this error and work smoothly in deep learning?
- ModuleNotFoundError: No module named 'llama_index.node_parser'
- Difference between model.evaluate and metrics.accuracy_score
- Give Bert an input and ask him to predict. In this input, can Bert apply the first word prediction result to all subsequent predictions?
Related Questions in TRAINING-DATA
- higher coefficient of determination values in the testing phase compared to the training phase
- Loading the pre-trained model from the .h5 file (Works on Colab but does not work on Local)
- How to finetune the LLM to output the text with SSML tags?
- How to solve this problem in performing grid search?
- How can I fine tune the any generative model? Autotrain
- How many images should I label from the training set?
- Should I use training or validation set for parameter otimization?
- Generate TRAIN_DATA for spacy from xml
- Does scikit-learn train_test_split copy data?
- YOLOv8 custom model not making predictions
- python - How can I retrain an ONNX model?
- Why Val loss is not showing ? how to display it then plot it with training loss
- ValueError: Expected input data to be non-empty
- Problem with creating dataset for visual object tracker
- tesseract combine_tessdata eng. Combining tessdata files Error: traineddata file must contain at least (a unicharset file
Popular Questions
- How do I undo the most recent local commits in Git?
- How can I remove a specific item from an array in JavaScript?
- How do I delete a Git branch locally and remotely?
- Find all files containing a specific text (string) on Linux?
- How do I revert a Git repository to a previous commit?
- How do I create an HTML button that acts like a link?
- How do I check out a remote Git branch?
- How do I force "git pull" to overwrite local files?
- How do I list all files of a directory?
- How to check whether a string contains a substring in JavaScript?
- How do I redirect to another webpage?
- How can I iterate over rows in a Pandas DataFrame?
- How do I convert a String to an int in Java?
- Does Python have a string 'contains' substring method?
- How do I check if a string contains a specific word?
Trending Questions
- UIImageView Frame Doesn't Reflect Constraints
- Is it possible to use adb commands to click on a view by finding its ID?
- How to create a new web character symbol recognizable by html/javascript?
- Why isn't my CSS3 animation smooth in Google Chrome (but very smooth on other browsers)?
- Heap Gives Page Fault
- Connect ffmpeg to Visual Studio 2008
- Both Object- and ValueAnimator jumps when Duration is set above API LvL 24
- How to avoid default initialization of objects in std::vector?
- second argument of the command line arguments in a format other than char** argv or char* argv[]
- How to improve efficiency of algorithm which generates next lexicographic permutation?
- Navigating to the another actvity app getting crash in android
- How to read the particular message format in android and store in sqlite database?
- Resetting inventory status after order is cancelled
- Efficiently compute powers of X in SSE/AVX
- Insert into an external database using ajax and php : POST 500 (Internal Server Error)
Generally there are two popular techniques for evaluating your classifier's performance: cross-validation, which uses the entire data-set (using multiple "folds" of the data), and hold-out set, which excludes some of the data from training for evaluation. Typically, the hold-out set is much smaller than the data used for training (e.g. 80/20 or 70/30).
In this case one option would be to keep a holdout set; do whatever learning and changes on the learning set, i.e. train a classifier, remove 'd' elements, train another classifier, identify 'e' elements; Then test the entire process against your holdout set.