I have a set of classified data with three labels, 'd', 'e', and 'k'. I want to train a classifier to identify 'd' and remove them from the dataset, then identify 'e'. Currently, I'm splitting the data into thirds, which I'll call X1, X2, X3. I train a learner L1 on X1, use that learner to remove 'd' labels on X2, which I then use to train a second learner L2, which I test on X3. Is this a reasonable approach, and is there an accepted standard in this kind of scenario?
Cross-validating with two different algorithms on one data set
76 Views Asked by johnklawlor At
1
There are 1 best solutions below
Related Questions in VALIDATION
- Escape dot in jquery validate plugin
- PHP form validation: Where to plop the code
- i want to create a service that does the login functionality?
- Stray start tag head, Element style not allowed as child of element body in this context. (Suppressing further errors from this subtree.)
- Html File Input on Chrome for Android missing extension and mime type
- javascript check input fields are not blank and check input field length?
- Symfony 2 form - date widget and validator
- Bean Validation message interpolation with array constraint parameter used as variable in message
- Bash regular expression execution hangs on long expressions
- Accessing the main object in a javax.validation.ConstraintValidator
- RAILS: date_select validation
- How can I define items of an array in a form in AngularJS
- Validation DataGridView Windows Forms
- How to handle multiple if statements PHP
- Restrict comma in asp.net textbox
Related Questions in MACHINE-LEARNING
- How to cluster a set of strings?
- Enforcing that inputs sum to 1 and are contained in the unit interval in scikit-learn
- scikit-learn preperation
- Spark MLLib How to ignore features when training a classifier
- Increasing the efficiency of equipment using Amazon Machine Learning
- How to interpret scikit's learn confusion matrix and classification report?
- Amazon Machine Learning for sentiment analysis
- What Machine Learning algorithm would be appropriate?
- LDA generated topics
- Spectral clustering with Similarity matrix constructed by jaccard coefficient
- Speeding up Viterbi execution
- Memory Error with Classifier fit and partial_fit
- How to find algo type(regression,classification) in Caret in R for all algos at once?
- Difference between weka tool's correlation coefficient and scikit learn's coefficient of determination score
- What are the approaches to the Big-Data problems?
Related Questions in TRAINING-DATA
- Predictionio evaluation fails with empty.maxBy exception and training with java.lang.OutOfMemoryError
- Torch Lua: Why is my gradient descent not optimizing the error?
- Training a new LatentSVMDetector Models
- PredictionIO train error tokens must not be empty
- Leave one out crossvalind in Matlab
- How to copy training data n times during classification?
- How to build OCR data for Android app with known layout and font
- Combining tensorflow inception models/graphs
- Tensorflow AlexNet accuracy does not increase during training for ILSVRC2012 data set
- CNN model why the data is too large?
- Cross-validating with two different algorithms on one data set
- Huge datasets in machine learning sklearn
- How will the Imputers work if all the values in a column is missing in input vector in sklearn
- wit.ai is not training new examples / training status stays "clean"
- How can I generate training data on the fly in TensorFlow?
Trending Questions
- UIImageView Frame Doesn't Reflect Constraints
- Is it possible to use adb commands to click on a view by finding its ID?
- How to create a new web character symbol recognizable by html/javascript?
- Why isn't my CSS3 animation smooth in Google Chrome (but very smooth on other browsers)?
- Heap Gives Page Fault
- Connect ffmpeg to Visual Studio 2008
- Both Object- and ValueAnimator jumps when Duration is set above API LvL 24
- How to avoid default initialization of objects in std::vector?
- second argument of the command line arguments in a format other than char** argv or char* argv[]
- How to improve efficiency of algorithm which generates next lexicographic permutation?
- Navigating to the another actvity app getting crash in android
- How to read the particular message format in android and store in sqlite database?
- Resetting inventory status after order is cancelled
- Efficiently compute powers of X in SSE/AVX
- Insert into an external database using ajax and php : POST 500 (Internal Server Error)
Popular Questions
- How do I undo the most recent local commits in Git?
- How can I remove a specific item from an array in JavaScript?
- How do I delete a Git branch locally and remotely?
- Find all files containing a specific text (string) on Linux?
- How do I revert a Git repository to a previous commit?
- How do I create an HTML button that acts like a link?
- How do I check out a remote Git branch?
- How do I force "git pull" to overwrite local files?
- How do I list all files of a directory?
- How to check whether a string contains a substring in JavaScript?
- How do I redirect to another webpage?
- How can I iterate over rows in a Pandas DataFrame?
- How do I convert a String to an int in Java?
- Does Python have a string 'contains' substring method?
- How do I check if a string contains a specific word?
Generally there are two popular techniques for evaluating your classifier's performance: cross-validation, which uses the entire data-set (using multiple "folds" of the data), and hold-out set, which excludes some of the data from training for evaluation. Typically, the hold-out set is much smaller than the data used for training (e.g. 80/20 or 70/30).
In this case one option would be to keep a holdout set; do whatever learning and changes on the learning set, i.e. train a classifier, remove 'd' elements, train another classifier, identify 'e' elements; Then test the entire process against your holdout set.