Displaying 20 results from an estimated 6000 matches similar to: "Goodness of fit for hclust?"
2011 Sep 13
2
help with hclust
Hello,
how can I get the similarity value (i.e., the inner cluster similarity) that was used to cut a hierarchical tree at a specific height?
I would appreciate your help!
Best regards,
Madeleine
2009 Nov 10
2
All possible combinations of functions within a function
Dear All,
I wrote a function for cluster analysis to compute cophenetic correlations
between dissimilarity matrices (using the VEGAN library) and cluster
analyses of every possible clustering algorithm (SEE ATTACHED)
http://old.nabble.com/file/p26288610/cor.coef.R cor.coef.R . As it is now,
it is extremely long, and for the future I was hoping to find a more
efficient way of doing this sort of
2004 May 11
1
stability measures for heirarchical clustering
Dear R users,
I'm interested in measuring the stability of a heirarchical clustering, of
the overall clustering and finding sub clusters (from cutting the
heirarchical clustering at different levels) which demonstrate stability.
I saw some postings on the R help from a while back about bootstrapping for
clustering (using sample and generating a consesus tree with a web based
tool CONSENSE)
2016 Apr 21
2
"cophenetic" function for objects of class "dendrogram"
Hello,
I have been using the "cophenetic" function for objects of class "dendrogram" and I have realised that it gives different results when it is used with objects of class "hclust". For instance, running the first example in the help file of the "cophenetic" function,
d1 <- dist(USArrests)
hc <- hclust(d1, "ave")
d2 <-
2016 Apr 21
1
"cophenetic" function for objects of class "dendrogram"
Note that cophenetic.default (which works on the output of hclust(dist(X)))
uses the
row names of X as labels. as.dendrogram.hclust does not retain those row
names
so cophenetic.dendrogram cannot use them (so it orders them based on the
topology of the dendrogram).
Bill Dunlap
TIBCO Software
wdunlap tibco.com
On Thu, Apr 21, 2016 at 7:59 AM, William Dunlap <wdunlap at tibco.com> wrote:
2001 Jun 12
1
cophenetic matrix
Hello,
I analyse some free-sorting data so I use hierarchical
clustering.
I want to compare my proximity matrix with the tree
representation to evalute the fitting. (stress, cophenetic correlation
(pearson's correlation)...)
"The cophenetic similarity of two objects a and b is defined as the
similarity level at wich objects a and b become members of the same
cluster during the course of
2010 Feb 21
1
How to: Compare Two dendrograms (Hierarchical Clusterings) ?
Hello all,
I wish to compare two dendrograms (representing Hierarchical Clusterings).
My problems are several:
1) how do I manually create a dendrogram object ?
That is, how can I reconstruct it as an "hclust" object that creates such a
dendrogram, when all I have is the dendrogram image (but don't have the
underlaying distance matrix that produced it) ?
I see that there is a
2010 May 25
1
Hierarchical clustering using own distance matrices
Hey Everyone!
I wanted to carry out Hierarchical clustering using distance matrices i have
calculated ( instead of euclidean distance etc.)
I understand as.dist is the function for this, but the distances in the
dendrogram i got by using the following script(1) were not the distances
defined in my distance matrices.
script:
var<-read.table("the distance matrix i calculated",
2001 Feb 23
4
hclust question
Dear all,
I have a question with regard to the use of hclust. I would like to be
able to specify my own distance matrix instead of asking R to compute
the distance matrix for me. It is computationally easier for me this
way. My question is: How can I get hclust to accept this?
Thanks,
Ranjan
--
***************************************************************************
Ranjan
2006 Jan 11
1
hypothesis testing for rank-deficient linear models
Take the following example:
a <- rnorm(100)
b <- trunc(3*runif(100))
g <- factor(trunc(4*runif(100)),labels=c('A','B','C','D'))
y <- rnorm(100) + a + (b+1) * (unclass(g)+2)
m <- lm(y~a+b*g)
summary(m)
Here b is discrete but not treated as a factor. I am interested in
computing the effect of b within groups defined by the
2004 Oct 19
1
plot.dendrogram and plot.hclust ZOOM into the height?
Hi,
I clustered a distance matrix and would like to draw it using
plot.hclust or plot.dendrogram.
The dendrogram is not informative because I have a few extremely small
dissimilarities in the distance matrix (e.g. 0), but most of the other
distances are in the range 1e10+-5000.
I would like to show the tree only for the height of 1e10+-5000 but
unfortunately their are no parameter like
2003 Aug 04
1
hclust() and agnes() method="average" divergence (PR#3648)
This message is in MIME format. Since your mail reader does not understand
this format, some or all of this message may not be legible.
------_=_NextPart_000_01C35A53.75780090
Content-Type: text/plain;
charset="iso-8859-1"
Anyone have a clue why hclust() and agnes() produce different results in the
example below when both use method="average"?? I'm not able to reproduce
2012 Mar 29
2
hclust and plot functions work, cutree does not
Hi,
I have the distance matrix computed and I feed it to hclust function. The
plot function produces a dense dendrogram as well. But, the cutree function
applied does not produce the desired list.
Here is the code
x=data.frame(similarity_matrix)
colnames(x) = c(source_tags_vec)
rownames(x) = c(source_tags_vec)
clust_tree=hclust(as.dist(x),method="complete")
plot(clust_tree)
2011 Dec 12
1
Is there a way to print branch distances for hclust function?
The R function hclust is used to do cluster analysis, but based on R
help I see no way to print the actual fusion distances (that is, the
vertical distances for each connected branch pairs seen in the cluster
dendrogram).
Any ideas? I'd like to use them test for significant differences from
the mean fusion distance (i.e. The Best Cut Test).
To perform a cluster analysis I'm using:
x
2005 Jun 03
1
Hclust question
Hey,
I am running hclust on several different distance matrices and I have a
question thats more about labeling. I've been looking for a way to label
the edge values on the graph with their distances between them. I've
been looking through the documentation and I haven't found anything yet.
Anyone know if there is a way to plot 'hclust' graphs with such edge
values? Or
2009 Nov 17
1
hclust too slow?
Hi,
I am new to clustering in R and I have a dataset with approximately 17,000
rows and 8 columns with each data point a numerical character with three
decimal places. I would like to cluster the 8 columns so that I get a
dendrogram as an output. So, I am simply creating a distance matrix of my
data, using the 'hclust' function, and then plotting the results (see below,
my data is
2012 Jul 04
1
Error in hclust?
Dear R users,
I have noted a difference in the merge distances given by hclust using
centroid method.
For the following data:
x<-c(1009.9,1012.5,1011.1,1011.8,1009.3,1010.6)
and using Euclidean distance, hclust using centroid method gives the
following results:
> x.dist<-dist(x)
> x.aah<-hclust(x.dist,method="centroid")
> x.aah$merge
[,1] [,2]
[1,] -3 -6
2008 Feb 20
1
clustering problem
First I just want to say thanks for all the help I've had from the
list so far..)
I now have what I think is a clustering problem. I have lots of
objects which I have measured a dissimilarity between. Now, this list
only has one entry per pair, so it is not symmetrical.
Example input:
NameA NameB Dist
189_1C2 189_1C1 0
189_1C3 189_1C1 0.017
189_1C3 189_1C2 0.017
189_1C4 189_1C1 0
2010 Nov 15
2
hclust, does order of data matter?
Hello,
I am using the hclust function to cluster some data. I have two separate
files with the same data. The only difference is the order of the data in
the file. For some reason, when I run the two files through the hclust
function, I get two completely different results.
Does anyone know why this is happening? Does the order of the data matter?
Thanks,
RC
--
View this message in
2014 Jul 25
0
clustering with hclust
Hi everybody, I have a problem with a cluster analysis.
I am trying to use hclust, method=ward.
The Ward method works with SQUARED Euclidean distances.
Hclust demands "a dissimilarity structure as produced by dist".
Yet, dist does not seem to produce a table of squared euclidean distances,
starting from cosines.
In fact, computing manually the squared euclidean distances from cosines