similar to: Confidence intervals for PCA scores/eigenvalues

Displaying 20 results from an estimated 5000 matches similar to: "Confidence intervals for PCA scores/eigenvalues"

2010 Feb 04
0
pca in R: Problem Fixed
Good day all. This is to thank all those who have helped in fixing this problem. Starting with a text book was indeed a problem, however, that gave me a clue of what I was looking for. This, with your contributions added to other materials I got on the net, put me on the right track. Thank you so much. Warmest regards Ogbos On 31 January 2010 14:07, S Ellison <S.Ellison@lgc.co.uk> wrote:
2008 Sep 09
4
PCA and % variance explained
After doing a PCA using princomp, how do you view how much each component contributes to variance in the dataset. I'm still quite new to the theory of PCA - I have a little idea about eigenvectors and eigenvalues (these determine the variance explained?). Are the eigenvalues related to loadings in R? Thanks, Paul -- View this message in context:
2011 Nov 14
0
Fwd: How to compute eigenvectors and eigenvalues?
Inicio del mensaje reenviado: > De: Arnau Mir <arnau.mir@uib.es> > Fecha: 14 de noviembre de 2011 13:24:31 GMT+01:00 > Para: Martin Maechler <maechler@stat.math.ethz.ch> > Asunto: Re: [R] How to compute eigenvectors and eigenvalues? > > Sorry, but I can't explain very well. > > > The matrix 4*mp is: > > 4*mp > [,1] [,2] [,3] > [1,]
2013 Mar 14
2
Same eigenvalues but different eigenvectors using 'prcomp' and 'principal' commands
Dear all, I've used the 'prcomp' command to calculate the eigenvalues and eigenvectors of a matrix(gg). Using the command 'principal' from the 'psych' package  I've performed the same exercise. I got the same eigenvalues but different eigenvectors. Is there any reason for that difference? Below are the steps I've followed: 1. PRCOMP #defining the matrix
2011 Nov 14
2
How to compute eigenvectors and eigenvalues?
Hello. Consider the following matrix: mp <- matrix(c(0,1/4,1/4,3/4,0,1/4,1/4,3/4,1/2),3,3,byrow=T) > mp [,1] [,2] [,3] [1,] 0.00 0.25 0.25 [2,] 0.75 0.00 0.25 [3,] 0.25 0.75 0.50 The eigenvectors of the previous matrix are 1, 0.25 and 0.25 and it is not a diagonalizable matrix. When you try to find the eigenvalues and eigenvectors with R, R responses: > eigen(mp) $values [1]
2010 May 21
2
Data reconstruction following PCA using Eigen function
Hi all, As a molecular biologist by training, I'm fairly new to R (and statistics!), and was hoping for some advice. First of all, I'd like to apologise if my question is more methodological rather than relating to a specific R function. I've done my best to search both in the forum and elsewhere but can't seem to find an answer which works in practice. I am carrying out
2004 Jun 28
3
How to determine the number of dominant eigenvalues in PCA
Dear All, I want to know if there is some easy and reliable way to estimate the number of dominant eigenvalues when applying PCA on sample covariance matrix. Assume x-axis is the number of eigenvalues (1, 2, ....,n), and y-axis is the corresponding eigenvalues (a1,a2,..., an) arranged in desceding order. So this x-y plot will be a decreasing curve. Someone mentioned using the elbow (knee)
2008 Jun 18
2
highest eigenvalues of a matrix
DeaR list, I happily use eigen() to compute the eigenvalues and eigenvectors of a fairly large matrix (200x200, say), but it seems over-killed as its rank is limited to typically 2 or 3. I sort of remember being taught that numerical techniques can find iteratively decreasing eigenvalues and corresponding orthogonal eigenvectors, which would provide a nice alternative (once I have the
2004 Oct 19
3
matrix of eigenvalues
I thought that the function eigen(A) will return a matrix with eigenvectors that are independent of each other (thus forming a base and the matrix being invertible). This seems not to be the case in the following example A=matrix(c(1,2,0,1),nrow=2,byrow=T) eigen(A) ->ev solve(ev$vectors) note that I try to get the upper triangular form with eigenvalues on the diagonal and (possibly) 1 just
2010 May 05
3
Symbolic eigenvalues and eigenvectors
Let's say I had a matrix like this: library(Ryacas) x<-Sym("x") m<-matrix(c(cos (x), sin(x), -sin(x), cos(x)), ncol=2) How can I use R to obtain the eigenvalues and eigenvectors? Thanks, John [[alternative HTML version deleted]]
2011 May 16
1
help: Using hotelling for a confidence region for PCA scores
Hello everyone. In my last post I did not explained my problem quite well. I made a principal component analysis and took the 2 first principal components. I made ​​a chart of my points based on the score of the 2 PC. I would like to add on this graph a 95% confidence region. To do this I used the ellipse function as follows: pcsref=PC$score[data[,1]==ref,1:2] #matrix containing the scores
2012 Apr 25
1
pca biplot.princomp has a bug?
x=rmvnorm(2000, rep(0, 6), diag(c(5, rep(1,5)))) x=scale(x, center=T, scale=F) pc <- princomp(x) biplot(pc) There are a bunch of red arrows plotted, what do they mean? I knew that the first arrow labelled with "Var1" should be pointing the most varying direction of the data-set (if we think them as 2000 data points, each being a vector of size 6). I also read from
2004 Apr 07
1
eigenvalues for a sparse matrix
Hi, I have the following problem. It has two parts. 1. I need to calculate the stationary probabilities of a Markov chain, eg if the transition matrix is P, I need x such that xP = x in other words, the left eigenvectors of P which have an eigenvalue of one. Currently I am using eigen(t(P)) and then pick out the vectors I need. However, this seems to be an overkill (I only need a single
2010 Jun 28
2
Note on PCA (not directly with R)
Dear all, I am looking for some interactive study materials on Principal component analysis. Basically I would like to know what we are actually doing with PCA? What is happening within the dataset at the time of doing PCA. Probably a 3-dimensional interactive explanation would be best for me. I have gone through some online materials specially Wikipedia etc, however what I need a "movable
2003 Jan 30
3
Principal comp. scores in R
Hello, I am trying to run a PCA in R and I cannot get the PC scores for each of the values. Using pcX <- princomp(X) then loadings(pcX) I can get a listing of the eigenvectors but not the actual PC scores for each value in the dataset. I greatly appreciate any help anyone can offer Thanks Ken
2004 Apr 02
0
picking out eigenvalues of 1
After making E <- eigen( something ) I would like to extract those eigenvectors which have an eigenvalue of 1. If I had an isone() function, I would simply say E$vectors[,which(isone(E))] but the problem is that I have no such thing. I found all.equal, so I could test for all.equal(x, 1), but for complex numbers, I need to use something like all.equal(x, 1+0i), don't I? I tried
2010 Nov 10
2
prcomp function
Hello, I have a short question about the prcomp function. First I cite the associated help page (help(prcomp)): "Value: ... SDEV the standard deviations of the principal components (i.e., the square roots of the eigenvalues of the covariance/correlation matrix, though the calculation is actually done with the singular values of the data matrix). ROTATION the matrix of variable loadings
2011 May 28
1
prcomp & eigenvectors ... ??
Hi ... Please could you help with probably a very simple problem I have. I'm completely new to R and am trying to follow a tutorial using R for Force Distribution Analysis that I got from ... http://projects.eml.org/mbm/website/fda_gromacs.htm. Basically, the MDS I preform outputs a force matrix (.fm) from the force simulation I perform. Then, this matrix is read into R and prcomp is
2010 Nov 30
3
pca analysis: extract rotated scores?
Dear all I'm unable to find an example of extracting the rotated scores of a principal components analysis. I can do this easily for the un-rotated version. data(mtcars) .PC <- princomp(~am+carb+cyl+disp+drat+gear+hp+mpg, cor=TRUE, data=mtcars) unclass(loadings(.PC)) # component loadings summary(.PC) # proportions of variance mtcars$PC1 <- .PC$scores[,1] # extract un-rotated scores of
2011 Jun 18
1
"Justify" PCA? -- was: Bartlett's Test of Sphericity
Apologies for the obvious, but just to clarify: there is no reason to "justify" a PCA -- it's just an eigen decomposition of a matrix and is therefore "justified" by linear algebra. If one wants to determine whether some subset of the eigenvectors = principal components suffice to "represent" the data in some sense, then that is where distributional