Skip to main content

Questions tagged [principal-component-analysis]

Principal component analysis (PCA) is a linear dimensionality reduction technique. It reduces a multivariate dataset to a smaller set of constructed variables preserving as much information (as much variance) as possible. These variables, called principal components, are linear combinations of the input variables.

Filter by
Sorted by
Tagged with
0 votes
3 answers
112 views

I looked at the rules, I think it's not wrong to ask people to just explain something to you. (I hope the post won't get flagged). My knowledge level is as much as high-schooler. But I've been ...
Magical Briefcase's user avatar
0 votes
1 answer
71 views

Let $P \in \mathbb{R}^{d \times d}$ be an rank $p < d$ orthogonal projection matrix given by $P = VV^T$, $V \in \mathbb{R}^{d \times p}$. Do we have that $$ \underset{P^2 = P = P^T,\; \text{rank}(P)...
1809's user avatar
  • 71
0 votes
1 answer
153 views

I'm trying to get a deeper understanding of Principal Component Analysis (PCA), and I keep coming across the point that we must center the data around zero before determining the principal components. ...
Anil's user avatar
  • 11
0 votes
0 answers
46 views

I am currently working on the following problem: Assume that we observe n random vectors $X_1, . . . , X_n$ that are independent copies of a random vector $X \sim N (0, \Sigma)$ where $\Sigma$ is a $d\...
Hugo01's user avatar
  • 57
3 votes
1 answer
124 views

In these notes, the author formulates the PCA problem as follows. Given a matrix of data $X$ the PCA problem is: $$\text{min}_Y \|Y - X\|_F \qquad \text{rank}(Y) = k$$ This reads to me as obviously ...
Your neighbor Todorovich's user avatar
0 votes
1 answer
65 views

I have two data matrices $ X_1 $ and $ X_2 $ with sizes $n \times d $, and I apply PCA to them resulting in: $$ Z_1 = W_1 X_1 $$ $$ Z_2 = W_2 X_2 $$ Then, I apply a linear transformation in the PCA ...
Arya J's user avatar
  • 351
0 votes
0 answers
47 views

Given a data matrix $X\in\mathbb{R}^{n \times p}$ The PCA algorithm is computing the covariance matrix $C$ of the centered data matrix $\bar{X}$ and then diagonalizing it $$C=VLV^{T}$$ Are the "...
James's user avatar
  • 33
1 vote
1 answer
46 views

Suppose a probability density function over $\mathbb R^d$. Let its expectation be $\boldsymbol{\mu}$, and some erronous estimate of it $\tilde{\boldsymbol{\mu}} = \boldsymbol{\mu} + \boldsymbol{\...
Alonzo's user avatar
  • 11
1 vote
0 answers
33 views

Given a $d\times n$ real matrix $M$ and a “weights” vector $w\in \mathbb{R}^n$ such that $\sum w_i =1$ and $0\leq w_i\leq 1$ for all $i$, I'd like to find, for some fixed $m<n$, some $d\times m$ ...
R. Pullbacki's user avatar
0 votes
1 answer
52 views

Let's consider the following formulation of PCA. We have a set of points $\{x_1, ..., x_m\}$ in $\mathbb{R}^n$, and a function $L_{\mathbf{V}}: \mathbb{R}^n \to \mathbb{R}^k$, with $k < n$, where $...
dherrera's user avatar
  • 340
1 vote
1 answer
120 views

Assume a data matrix $X \in \mathbb{R}^{N \times p_X}$. Let it have some exact lower dimensional factor representation $X = A F$, where $F \in \mathbb{R}^{N \times p_F}$ and $p_F < p_X$. Let the ...
fes's user avatar
  • 1,678
1 vote
0 answers
61 views

Suppose we have a data distribution on $x ∈ R^d$. Suppose $E_x[x] = 0$, and let $Σ = E_x[xx^T]$ be the covariance of x. Let $Σ = USU^T$ be the spectral decomposition of Σ, with U orthonormal and S ...
MedMan27's user avatar
2 votes
1 answer
100 views

For a given covariance matrix $A$ (so $A$ is symmetrical and positive semidefinite), I want to find a vector $x$ such that $| x x^T - A |^2$ is minimized. Here $|M|^2$ just means the sum of squares ...
CuriousMind's user avatar
  • 1,916
1 vote
1 answer
118 views

Let $X \in \mathbb{R}^{n \times d}$ be our data matrix where $n$ is the number of examples and $d$ is the feature dimension. Applying PCA to $X$, we get a low-dimensional representation $A \in \mathbb{...
FountainTree's user avatar
1 vote
1 answer
72 views

So in this example here the Principal Components (PC) are: $$ \begin{align} PC_1 (Size) &= 0.707 * Height + 0.707 * Diameter \quad (1)\\ PC_2 (Shape) &= 0.707 * Height - 0.707 * Diameter \...
yyyLLL's user avatar
  • 33
1 vote
0 answers
47 views

As is shown in this question kernel pca eigenproblem and many other refernce materials about kernel PCA. They all point out that the solution of $K^2a_j=\lambda_jnKa_j$ and $Ka_j=\lambda_jna_j$ only ...
Wang Pengxin's user avatar
0 votes
0 answers
43 views

Let $\Sigma$ be the covariance matrix. The first principal axis of PCA, $u_1$, can be found by solving the optimization problem: $$\max_{u_{1}}=u^T_1\Sigma u_1$$ subject to the constraint $u^T_1u_1=1$....
s1m0n3's user avatar
  • 63
1 vote
1 answer
294 views

Here, the best-fitting plane is the plane that minimizes the sum of squared perpendicular distance from the data points to the plane. In other words, the best-fitting plane is the solution to the ...
iklnot's user avatar
  • 13
1 vote
0 answers
68 views

I am trying to visualize in MATLAB the relationship between the singular value decomposition (SVD) of a matrix of points. To simplify the problem, I am working in 2D and I am considering an ellipse ...
gwn's user avatar
  • 121
0 votes
1 answer
54 views

The expression comes from The Elements of Statistical Learningp66(3.49) Related part about this is shown in below figure. Where Using SVD, $X$ has the form: $X=UDV^T$. And $D$ is a diagonal matrix, ...
chenqile's user avatar
2 votes
0 answers
560 views

I am trying to understand a proof of the Eckart-Young Theorem (source in title). Let me add the definition and proof they provided, afterward I'll say what I do not understand. Proof: Some points I ...
Aaron's user avatar
  • 69
1 vote
1 answer
484 views

In spectral clustering, one computes sample-sample similarities, then from this computes a graph Laplacian matrix. (Typically, one uses the symmetrically normalized Laplacian matrix, but the pattern I'...
calmcc's user avatar
  • 271
0 votes
1 answer
181 views

Suppose that the variance-covariance matrix of a $p$ -dimensional random vector $X$ is $\Sigma=\sigma_{ij}$ for all $i,j=1, 2, ...,p$. Show that the coefficients of the first principal component have ...
RV math's user avatar
  • 67
0 votes
0 answers
83 views

For a given matrix $A$, the Principal Component Analysis (PCA) is done by finding the eigenvalues/eigenvectors of the covariance matrix associated with $A$. However, the entries of the covariance ...
user2167741's user avatar
0 votes
0 answers
58 views

There's a statement in the paper Phatak (1997) In most spectroscopic problems, the number of non-zero singular values is $n-1$ for $X\in \mathbb{R}^{n \times m}$ when $n\ll m$. Consequently, at most $...
WeiShan Ng's user avatar
0 votes
0 answers
273 views

As I was watching a video explaining how MDS works, the narrator mentioned that PCA is equivalent to MDS when Euclidean distances are used. I got confused as to how that's the case. My guess is that ...
baoiba's user avatar
  • 121
1 vote
0 answers
29 views

With a matrix $X_{n\times p}$ ($n>p$), we perform a principal component analysis: $T_{n\times p}=X_{n\times p}W_{p\times p}$ where $W$ is the loadings matrix while $T$ is the scores matrix for $X$. ...
bluepole's user avatar
  • 153
1 vote
1 answer
123 views

Let $X$ and $Y$ two linearly separable finite subsets of a $K$-dimensional real vector space $V$ with orthonormal basis $A = \{a_1,\ldots, a_K\}$. The covariance matrix $\Sigma_A$ of the set $X \cup Y$...
Alberto Carraro's user avatar
0 votes
0 answers
125 views

I was conducting the Karhunen-Loeve (K-L) Expansion for a random vector. Based on the KL expansion, I transformed the original random vector into a standardized random vector $\boldsymbol{X}=[X_1,X_2,\...
Alexander's user avatar
1 vote
1 answer
1k views

Please, I'm trying to understand why singular values (SV) are always hierarchical/descending. At the beginning of my studies, I thought that the hierarchy of sigmas ($ \sigma_1 \geq \sigma_2 \geq ... \...
Caio Rímoli's user avatar
0 votes
1 answer
62 views

Consider a function $F(x,y,z): \mathbb{R}^3 \mapsto \mathbb{R}^+$ (i.e., $F(x,y,z) > 0 ~~\forall~ x,y,z$) and consider a set of points in the (3D) space, $\{p_1, p_2, \cdots , p_N\}$. The problem ...
Cal-linux's user avatar
  • 144
0 votes
0 answers
199 views

First time asker here. First off, I know I should be doing this in R or Python. I will. For now I'm reading a textbook, using simple examples and Excel to try to learn the concepts of linear algebra. ...
jlewis's user avatar
  • 13
2 votes
0 answers
52 views

If we have a matrix $M$ and we want to find its least squares approximation as the product of two smaller (as in less rows or columns) matrices $M_1M_2$ of a given size, we can simply run SVD and pick ...
BluerFrog's user avatar
1 vote
1 answer
158 views

In studying principal components analysis, I am confused by one point. For a set of $N$ (zero-centered) data points of dimension $m$, projected to a dimension $k < m$, we want a set of vectors of ...
IntegrateThis's user avatar
2 votes
0 answers
192 views

I'd like to preface this post by saying that this is my first post on stack exchange, so if there is anything to improve, be it redaction or just the structuring of posts, I'm more than willing to ...
dizzyAfterMatrixRotations's user avatar
1 vote
1 answer
53 views

I am trying to fully understand/demystify EVD, SVD and PCA. Am I right to assume all of these are tools/methods to solve least squares (not only but for CG)? If I am not wrong, even there are multiple ...
Joan Venge's user avatar
2 votes
0 answers
34 views

I have a system of odes where the state vector has 6 elements. The system is a population biology model, where I am tracking the evolution of some competing species over time. Now I was trying to ...
krishnab's user avatar
  • 2,705
1 vote
0 answers
615 views

In my computational problem, I have a Riemann submanifold $S^{1000}$ embedded in $\mathbb{R}^{300000}$. I can numerically compute the induced metric tensor and the Jacobian. I have no analytical ...
Luke's user avatar
  • 11
1 vote
1 answer
494 views

In Bishop's Pattern Recognition And Machine Learning book, Chapter 12, Suppose $X$ is an uncentered data matrix and $\bar{x}=\frac{1}{m}\sum_ix_i$ is the sample mean of the columns of $X$. For the ...
Jayden Rice's user avatar
0 votes
1 answer
323 views

I tried comparing sklearn.decomposition.KernelPCA with a linear kernel to sklearn.decomposition.PCA on the same data set and got different eigenvectors. My understanding is that these should be ...
Galen's user avatar
  • 1,946
0 votes
1 answer
176 views

I'm reading "Introduction to linear algebra" of Gillbert Strang. PCA by SVD section. Text says that the sum of squared distances from the points to the line is a minimum and author is trying ...
Vanconts's user avatar
3 votes
1 answer
119 views

let's say I have a "data" matrix $X$ of $N$ rows and $p$ cols with $N \gg p$. Now PCA with $L$ components can be formulated as $$X_L = argmin_{Y:rank(Y) = L} ||X- Y||^2_F $$, where Y is ...
Vladimir Kirilin's user avatar
3 votes
1 answer
380 views

I have a certain XYZ set of points that make up an object. I chose a random point and make the nearest radius analysis and find the neighbors. From these neighbors, I get the green pointcloud curve ...
Hamzalihi's user avatar
0 votes
1 answer
2k views

I'm a final year maths undergrad doing a course in multivariate data analysis, but I'm really struggling with the linear algebra. In particular the “projection of the data along the 1st k principal ...
d89J_kl's user avatar
  • 39
1 vote
0 answers
39 views

I've seen multiple PCA derivations where the $\frac{1}{n}$ (for variance) or $\frac{1}{n-1}$ (for sample variance) is just omitted, e.g. here. I see that they are proportional to each other but is ...
jonithani123's user avatar
2 votes
0 answers
35 views

Suppose $x_i \in\mathbb{R}^2$ for $i=1,2,...9$ are unknown. I'm given the pair-wise distance matrix between these points $D$ which is a $9*9$ symmetric matrix. I want to learn these data points by ...
chesslad's user avatar
  • 2,633
1 vote
2 answers
2k views

I am reading the paper "Understanding dimensional collapse in contrastive self-supervised learning." The authors identified a dimensional collapse phenomenon: i.e. some dimension of ...
Noel's user avatar
  • 11
1 vote
1 answer
105 views

I've a question regarding pca variant. Let $X ∈ \Bbb R^{D×n}$ be a data matrix, $\{u_i\}^{d}i=1$ be the $d$ principal components of $X$, and where $μ ∈ \Bbb R^d$ is the sample mean vector and $1_n ∈ \...
joen joe's user avatar
0 votes
0 answers
57 views

I'm trying to find a suitable pairwise distance metric where the addition of correlated vectors results in (essentially) no change in the distance. Specifically, consider a set of $k$ vectors each of ...
user3037237's user avatar
2 votes
1 answer
293 views

Question Elaboration: When I say 'linear embeddings' I mean a lower-dimensional representation of variables resulting from an arbitrary linear transformation. And when I say 'explained variance' I ...
profPlum's user avatar
  • 387