similar to: Goodness of fit for hclust?

Displaying 20 results from an estimated 6000 matches similar to: "Goodness of fit for hclust?"

2011 Sep 13
2
help with hclust
Hello, how can I get the similarity value (i.e., the inner cluster similarity) that was used to cut a hierarchical tree at a specific height? I would appreciate your help! Best regards, Madeleine
2009 Nov 10
2
All possible combinations of functions within a function
Dear All, I wrote a function for cluster analysis to compute cophenetic correlations between dissimilarity matrices (using the VEGAN library) and cluster analyses of every possible clustering algorithm (SEE ATTACHED) http://old.nabble.com/file/p26288610/cor.coef.R cor.coef.R . As it is now, it is extremely long, and for the future I was hoping to find a more efficient way of doing this sort of
2004 May 11
1
stability measures for heirarchical clustering
Dear R users, I'm interested in measuring the stability of a heirarchical clustering, of the overall clustering and finding sub clusters (from cutting the heirarchical clustering at different levels) which demonstrate stability. I saw some postings on the R help from a while back about bootstrapping for clustering (using sample and generating a consesus tree with a web based tool CONSENSE)
2016 Apr 21
2
"cophenetic" function for objects of class "dendrogram"
Hello, I have been using the "cophenetic" function for objects of class "dendrogram" and I have realised that it gives different results when it is used with objects of class "hclust". For instance, running the first example in the help file of the "cophenetic" function, d1 <- dist(USArrests) hc <- hclust(d1, "ave") d2 <-
2016 Apr 21
1
"cophenetic" function for objects of class "dendrogram"
Note that cophenetic.default (which works on the output of hclust(dist(X))) uses the row names of X as labels. as.dendrogram.hclust does not retain those row names so cophenetic.dendrogram cannot use them (so it orders them based on the topology of the dendrogram). Bill Dunlap TIBCO Software wdunlap tibco.com On Thu, Apr 21, 2016 at 7:59 AM, William Dunlap <wdunlap at tibco.com> wrote:
2001 Jun 12
1
cophenetic matrix
Hello, I analyse some free-sorting data so I use hierarchical clustering. I want to compare my proximity matrix with the tree representation to evalute the fitting. (stress, cophenetic correlation (pearson's correlation)...) "The cophenetic similarity of two objects a and b is defined as the similarity level at wich objects a and b become members of the same cluster during the course of
2010 Feb 21
1
How to: Compare Two dendrograms (Hierarchical Clusterings) ?
Hello all, I wish to compare two dendrograms (representing Hierarchical Clusterings). My problems are several: 1) how do I manually create a dendrogram object ? That is, how can I reconstruct it as an "hclust" object that creates such a dendrogram, when all I have is the dendrogram image (but don't have the underlaying distance matrix that produced it) ? I see that there is a
2010 May 25
1
Hierarchical clustering using own distance matrices
Hey Everyone! I wanted to carry out Hierarchical clustering using distance matrices i have calculated ( instead of euclidean distance etc.) I understand as.dist is the function for this, but the distances in the dendrogram i got by using the following script(1) were not the distances defined in my distance matrices. script: var<-read.table("the distance matrix i calculated",
2001 Feb 23
4
hclust question
Dear all, I have a question with regard to the use of hclust. I would like to be able to specify my own distance matrix instead of asking R to compute the distance matrix for me. It is computationally easier for me this way. My question is: How can I get hclust to accept this? Thanks, Ranjan -- *************************************************************************** Ranjan
2006 Jan 11
1
hypothesis testing for rank-deficient linear models
Take the following example: a <- rnorm(100) b <- trunc(3*runif(100)) g <- factor(trunc(4*runif(100)),labels=c('A','B','C','D')) y <- rnorm(100) + a + (b+1) * (unclass(g)+2) m <- lm(y~a+b*g) summary(m) Here b is discrete but not treated as a factor. I am interested in computing the effect of b within groups defined by the
2004 Oct 19
1
plot.dendrogram and plot.hclust ZOOM into the height?
Hi, I clustered a distance matrix and would like to draw it using plot.hclust or plot.dendrogram. The dendrogram is not informative because I have a few extremely small dissimilarities in the distance matrix (e.g. 0), but most of the other distances are in the range 1e10+-5000. I would like to show the tree only for the height of 1e10+-5000 but unfortunately their are no parameter like
2003 Aug 04
1
hclust() and agnes() method="average" divergence (PR#3648)
This message is in MIME format. Since your mail reader does not understand this format, some or all of this message may not be legible. ------_=_NextPart_000_01C35A53.75780090 Content-Type: text/plain; charset="iso-8859-1" Anyone have a clue why hclust() and agnes() produce different results in the example below when both use method="average"?? I'm not able to reproduce
2012 Mar 29
2
hclust and plot functions work, cutree does not
Hi, I have the distance matrix computed and I feed it to hclust function. The plot function produces a dense dendrogram as well. But, the cutree function applied does not produce the desired list. Here is the code x=data.frame(similarity_matrix) colnames(x) = c(source_tags_vec) rownames(x) = c(source_tags_vec) clust_tree=hclust(as.dist(x),method="complete") plot(clust_tree)
2011 Dec 12
1
Is there a way to print branch distances for hclust function?
The R function hclust is used to do cluster analysis, but based on R help I see no way to print the actual fusion distances (that is, the vertical distances for each connected branch pairs seen in the cluster dendrogram). Any ideas? I'd like to use them test for significant differences from the mean fusion distance (i.e. The Best Cut Test). To perform a cluster analysis I'm using: x
2005 Jun 03
1
Hclust question
Hey, I am running hclust on several different distance matrices and I have a question thats more about labeling. I've been looking for a way to label the edge values on the graph with their distances between them. I've been looking through the documentation and I haven't found anything yet. Anyone know if there is a way to plot 'hclust' graphs with such edge values? Or
2009 Nov 17
1
hclust too slow?
Hi, I am new to clustering in R and I have a dataset with approximately 17,000 rows and 8 columns with each data point a numerical character with three decimal places. I would like to cluster the 8 columns so that I get a dendrogram as an output. So, I am simply creating a distance matrix of my data, using the 'hclust' function, and then plotting the results (see below, my data is
2012 Jul 04
1
Error in hclust?
Dear R users, I have noted a difference in the merge distances given by hclust using centroid method. For the following data: x<-c(1009.9,1012.5,1011.1,1011.8,1009.3,1010.6) and using Euclidean distance, hclust using centroid method gives the following results: > x.dist<-dist(x) > x.aah<-hclust(x.dist,method="centroid") > x.aah$merge [,1] [,2] [1,] -3 -6
2008 Feb 20
1
clustering problem
First I just want to say thanks for all the help I've had from the list so far..) I now have what I think is a clustering problem. I have lots of objects which I have measured a dissimilarity between. Now, this list only has one entry per pair, so it is not symmetrical. Example input: NameA NameB Dist 189_1C2 189_1C1 0 189_1C3 189_1C1 0.017 189_1C3 189_1C2 0.017 189_1C4 189_1C1 0
2010 Nov 15
2
hclust, does order of data matter?
Hello, I am using the hclust function to cluster some data. I have two separate files with the same data. The only difference is the order of the data in the file. For some reason, when I run the two files through the hclust function, I get two completely different results. Does anyone know why this is happening? Does the order of the data matter? Thanks, RC -- View this message in
2014 Jul 25
0
clustering with hclust
Hi everybody, I have a problem with a cluster analysis. I am trying to use hclust, method=ward. The Ward method works with SQUARED Euclidean distances. Hclust demands "a dissimilarity structure as produced by dist". Yet, dist does not seem to produce a table of squared euclidean distances, starting from cosines. In fact, computing manually the squared euclidean distances from cosines