For questions related to principal component analysis (PCA), which is commonly used in machine learning for dimensionality reduction.
Questions tagged [principal-component-analysis]
20 questions
5
votes
2 answers
What are examples of approaches to dimensionality reduction of feature vectors?
Given a pre-trained CNN model, I extract feature vector of images in reference and query dataset with several thousands of elements.
I would like to apply some augmentation techniques to reduce the feature vector dimension to speed up cosine…

Farid Alijani
- 299
- 3
- 10
3
votes
1 answer
When using PCA for dimensionality reduction of the feature vectors to speed up learning, how do I know that I'm not letting the model overfit?
I'm following Andrew Ng's course for Machine Learning and I just don't quite understand the following.
Using PCA to speed up learning
Using PCA to reduce the number of features, thus lowering the chances for overfitting
Looking at these two…

AfiJaabb
- 131
- 3
3
votes
0 answers
Looking for the proper algorithm to compress many lowres images of nearby locations
I have an optimization problem that I'm looking for the right algorithm to solve.
What I have: A large set of low-res 360 images that were taken on a regular grid within a certain area. each of these images is quite sparsely sampled and each of…

matthias_buehlmann
- 259
- 1
- 6
3
votes
1 answer
Do the eigenvectors represent the original features?
I've got a test dataset with 4 features and the PCA produces a set of 4 eigenvectors, e.g.,
EigenVectors: [0.7549043055910286, 0.24177972266822534, -0.6095588015369825, -0.01000612689310429]
EigenVectors: [0.0363767549959317, -0.9435613299702559,…

Crizly
- 131
- 1
2
votes
1 answer
Why does PCA of the vertices of a hexagon result in principal components of equal length?
I do PCA on the data points placed in the corners of a hexagon, and get the following principal components:
The PCA variance is $0.6$ and is the same for each component. Why is that? Shouldn't it be greater in the horizontal direction than in the…

Vladislav Gladkikh
- 472
- 5
- 14
2
votes
0 answers
Estimating dimensions to reduce input image size to in CNNs
Considering input images to a CNN that have a large dimension (e.g. 256X256), what are some possible methods to estimate the exact dimensions (e.g. 16X16 or 32X32) to which it can be condensed in the final pooling layer within the CNN network such…

Prishita Ray
- 51
- 4
2
votes
2 answers
How does PCA work when we reduce the original space to 2 or higher-dimensional space?
How does PCA work when we reduce the original space to a 2 or higher-dimensional space? I understand the case when we reduce the dimensionality to $1$, but not this case.
$$\begin{array}{ll} \text{maximize} & \mathrm{Tr}\left(…

VN Pikachu
- 21
- 3
1
vote
1 answer
Why does PCA work well while the total variance retained is small?
I'm learning machine learning by looking through other people's kernel on Kaggle, specifically this Mushroom Classification kernel.
The author first applied PCA to the transformed indicator matrix. He only used 2 principal components for…

Bicheng
- 111
- 1
1
vote
1 answer
How to perform PCA in the validation/test set?
I was using PCA on my whole dataset (and, after that, I would split it into training, validation, and test datasets). However, after a little bit of research, I found out that this is the wrong way to do it.
I have few questions:
Are there some…

LVoltz
- 121
- 1
- 5
1
vote
0 answers
How do I select the number of neurons for each layer in an auto-encoder for dimensionality reduction?
I am trying to apply an auto-encoder for dimensionality reduction. I wonder how it will be applied on a large dataset.
I have tried this code below. I have total of 8 features in my data and I want to reduce it to 3.
from keras.models import…

user12
- 111
- 3
1
vote
0 answers
Under what circumstances is a fully connected layer similar to PCA?
I am reading this paper on image retrieval where the goal is to train a network that produces highly discriminative descriptors (aka embeddings) for input images. If you are familiar with facial recognition architectures, it is similar in that the…

Alexander Soare
- 1,319
- 2
- 11
- 26
1
vote
0 answers
Is the 3d convolution associative given that it can be represented as matrix multiplication?
I'm trying to understand if a 3D convolution of the sort performed in a convolutional layer of a CNN is associative. Specifically, is the following true:
$$
X \otimes(W \cdot Q)=(X \otimes W) \cdot Q,
$$
where
$\otimes$ is a convolution,
$X$ is a…

HereItIs
- 11
- 2
1
vote
0 answers
Multiple-dimension scaling (MDS) objective for MDS and PCA
The following is the MDS Objective.
Let's think of a senario where I apply MDS with/from the solution I obtained from PCA. Then I calculate the objective function on the initial PCA solution and MDS solution (after applying MDS on the former PCA…

underdisplayname
- 111
- 2
1
vote
0 answers
What is the difference between principal component analysis and singular value decomposition in image processing?
What is the difference between principal component analysis and singular value decomposition in image processing? Which one performs better, and why?

DRV
- 1,573
- 2
- 11
- 18
0
votes
0 answers
Are the dimensions in embedding vectors ordered (similar to PCA)?
I am getting started with the vector embeddings. I have a general question about the embedding vectors generated by popular algorithms.
In PCA, usually, there is an implicit order of importance in the dimensions, with the most informative (by…

JackDaniels
- 101
- 1