Suppose there are 8 features in the dataset. I use PCA to and find out that 99% of the information is in the first 3 features using the cumulative sum of the explained variance ratio. Then why do I need to fit and transform these 3 features using PCA in order to use them for training my neural network ? Why cant I just use the three features as is ?
Using features without applying PCA
88 Views Asked by Tanmay Bhatnagar At
1
There are 1 best solutions below
Related Questions in MACHINE-LEARNING
- How to cluster a set of strings?
- Enforcing that inputs sum to 1 and are contained in the unit interval in scikit-learn
- scikit-learn preperation
- Spark MLLib How to ignore features when training a classifier
- Increasing the efficiency of equipment using Amazon Machine Learning
- How to interpret scikit's learn confusion matrix and classification report?
- Amazon Machine Learning for sentiment analysis
- What Machine Learning algorithm would be appropriate?
- LDA generated topics
- Spectral clustering with Similarity matrix constructed by jaccard coefficient
- Speeding up Viterbi execution
- Memory Error with Classifier fit and partial_fit
- How to find algo type(regression,classification) in Caret in R for all algos at once?
- Difference between weka tool's correlation coefficient and scikit learn's coefficient of determination score
- What are the approaches to the Big-Data problems?
Related Questions in NEURAL-NETWORK
- How to choose good SURF feature keypoints?
- How to avoid overfitting (Encog3 C#)?
- Run out of VRAM using Theano on Amazon cluster
- Calculating equation from image in Java
- Print output of a Theano network
- Torch Lua: Why is my gradient descent not optimizing the error?
- How can I train a neural (pattern recognition) network multiple times in matlab?
- Using Convolution Neural Net with Lasagne in Python error
- Random number of hidden units improves accuracy/F-score on test set
- Matlab example code for deep belief network for classification
- Pybrain Reinforcement Learning Example
- How to speed up caffe classifer in python
- Opencv mlp Same Data Different Results
- Word2Vec Data Setup
- How can I construct a Neural Network in Matlab with matrix of features extracted from images?
Related Questions in PCA
- How to choose good SURF feature keypoints?
- Spectral clustering with Similarity matrix constructed by jaccard coefficient
- Export PCA components in r
- How to export PCA from Weka
- Bad Orientation of Principal Axis of a Point Cloud
- Eigenfaces in OpenCV using C++
- How to do distributed Principal Components Analysis + Kmeans using Apache Spark?
- Significance of 99% of variance covered by the first component in PCA
- How to get the number of components needed in PCA with all extreme variance?
- Insufficient memory opencv
- Extract relevant attributes from postal addresses data in order to do PCA on those Data (using R)
- Unable to plot PCA data in R. Are scores defined by a given object/name to plot them specifically?
- Obtain unstandardized factor scores from factor analysis in R
- Why does classifier accuracy drop after PCA, even though 99% of the total variance is covered?
- R Biplot with clusters as colors
Related Questions in CROSS-VALIDATION
- computed initial MA coefficients are not invertible [Python] [TSA] [ARIMAX] [CrossValidation]
- Big accuracy difference between cross-validation and testing with a test set in weka? is it normal?
- How to predict labels for new data (test set) by the PartitionedEnsemble model in Matlab?
- h2o.runif() always returns the same vector
- Access indices of each CV fold for custom metric function in caret
- python sklearn cross_validation /number of labels does not match number of samples
- LDA cross validation and variable selection
- How to use cross validation in MATLAB
- Parameter selection of SVM
- Using Cross-Validation on a Scikit-Learn Classifer
- Can I extract y-values (data labels) from inside a cross-validation pipline in scikit-learn?
- Why does GridSearchCV give different optimums on repeated runs?
- How is cross validation implemented?
- Make a cross validation on a dataframe for an OLS regression model
- K fold cross validation using keras
Related Questions in DIMENSION-REDUCTION
- Dimension reduction for logical arrays
- In natural language processing (NLP), how do you make an efficient dimension reduction?
- Optimal perplexity for t-SNE with using larger datasets (>300k data points)
- Read Columnwise Matrix in R
- Global operator along a single dimension in Keras?
- Why is it ok to remove variables with low variance from a dataset
- Dimension reduction using PCA
- Looking for a function in R to sum rows and cols for matrix reduction
- I want to input 3d array(custom data) to sklearn-PCA function
- The lstm autoencoder does not use the full dimensions of the latent space for dimension reduction
- How to solve out of memory error?
- How to use QR-Decomposition to reduce the dimension of a dataset?
- Mapping a numerical function with two inputs onto one with one input
- Mapping N-dimensional value to a point on Hilbert curve
- Matlab: one-dim integral for a function @(x,y,z)
Trending Questions
- UIImageView Frame Doesn't Reflect Constraints
- Is it possible to use adb commands to click on a view by finding its ID?
- How to create a new web character symbol recognizable by html/javascript?
- Why isn't my CSS3 animation smooth in Google Chrome (but very smooth on other browsers)?
- Heap Gives Page Fault
- Connect ffmpeg to Visual Studio 2008
- Both Object- and ValueAnimator jumps when Duration is set above API LvL 24
- How to avoid default initialization of objects in std::vector?
- second argument of the command line arguments in a format other than char** argv or char* argv[]
- How to improve efficiency of algorithm which generates next lexicographic permutation?
- Navigating to the another actvity app getting crash in android
- How to read the particular message format in android and store in sqlite database?
- Resetting inventory status after order is cancelled
- Efficiently compute powers of X in SSE/AVX
- Insert into an external database using ajax and php : POST 500 (Internal Server Error)
Popular Questions
- How do I undo the most recent local commits in Git?
- How can I remove a specific item from an array in JavaScript?
- How do I delete a Git branch locally and remotely?
- Find all files containing a specific text (string) on Linux?
- How do I revert a Git repository to a previous commit?
- How do I create an HTML button that acts like a link?
- How do I check out a remote Git branch?
- How do I force "git pull" to overwrite local files?
- How do I list all files of a directory?
- How to check whether a string contains a substring in JavaScript?
- How do I redirect to another webpage?
- How can I iterate over rows in a Pandas DataFrame?
- How do I convert a String to an int in Java?
- Does Python have a string 'contains' substring method?
- How do I check if a string contains a specific word?
The reason is that when PCA tells you that 99% of the variance is explained by the first three components, it doesn't mean that it is explained by the first three features. PCA components are linear combinations of the features, but they are usually not the features themselves. For example, PCA components must be orthogonal to each other, while the features don't have to be.