similar to: to search the tree using the branch and bound mathod!!!

Displaying 20 results from an estimated 10000 matches similar to: "to search the tree using the branch and bound mathod!!!"

2011 Jul 30
0
search the clusterring tree using the branch and bound mathod!!!
hi, R users here is my problem, i want to make a clusterring tree, than do the searching through the tree using branch and bound method. the code to make the tree, but i dont know how to do the searching part. thanks for any helping... d <- dist(data, method = "euclidean") h1 <- hclust(d, method="ward") str(h1) plot(h1) mark -- View this message in context:
2013 Jan 18
1
Hclust tree to Figtree w/ branch lengths
Hi, I'm doing hierarchical clustering, and want to export my dendrogram to a tree-viewing/editing software. I can do this by converting the data to Newick format (hc2Newick in ctc package), but I can't get branch lengths to show in the resulting phylogram. I figured it might help to convert my hclust object into a phylo object (as.phylo in ape package), but the following lines give me
2010 Apr 26
2
Cluster analysis: dissimilar results between R and SPSS
Hello everyone! My data is composed of 277 individuals measured on 8 binary variables (1=yes, 2=no). I did two similar cluster analyses, one on SPSS 18.0 and one on R 2.9.2. The objective is to have the means for each variable per retained cluster. 1) the R analysis ran as followed: > call data > dist=dist(data,method="euclidean") >
2014 Jul 25
0
clustering with hclust
Hi everybody, I have a problem with a cluster analysis. I am trying to use hclust, method=ward. The Ward method works with SQUARED Euclidean distances. Hclust demands "a dissimilarity structure as produced by dist". Yet, dist does not seem to produce a table of squared euclidean distances, starting from cosines. In fact, computing manually the squared euclidean distances from cosines
2011 Dec 12
1
Is there a way to print branch distances for hclust function?
The R function hclust is used to do cluster analysis, but based on R help I see no way to print the actual fusion distances (that is, the vertical distances for each connected branch pairs seen in the cluster dendrogram). Any ideas? I'd like to use them test for significant differences from the mean fusion distance (i.e. The Best Cut Test). To perform a cluster analysis I'm using: x
2001 Apr 25
1
problems with a large data set
Hello, I have trouble with a data set that comprises 2136 lines of 20 columns. I would like to do a hierarchical clustering and I tried the following: ages.hclust <- hclust(dist(ages, method="euclidean"), "ward") but I get the following error message: Error: cannot allocate vector of size 17797 Kb When I try to do the dist() alone first without the hclust(), I get the
2001 Apr 27
0
weithed clustering (was: Re: problems with a large data set)
kmeans and clara work great. Thank you for the tip. I have another question: Is it possible to weight the observations in a cluster analysis ? I haven't found any mention of this in the kmeans of clara help texts. Moritz Lennert Charg? de recherche IGEAT - ULB t?l: 32-2-650.65.16 fax: 32-2-650.50.92 email: mlennert at ulb.ac.be > On Wed, 25 Apr 2001, Moritz Lennert wrote: >
2012 Nov 07
0
on Rclusterpp package usage
Hello I have a large dataset to make a cluster analysis and I'm trying to use Rclusterpp to do so. Nevertheless in a sample matrix analysis (see http://dl.dropbox.com/u/755659/gillnet.txt) I'm getting different results from Rclusterpp and hclust. Following the example from "An Introduction to Rclusterpp" (
2011 Nov 15
0
Bootstrap values for hierarchical tree based on distaance matrix
I would like to get an hierarchical clustering tree with bootstrap values indicated on the nodes, as in pvclust. The problem is that I have only distance matrix instead of the raw data, required for pvclust. Is there a way to get it? fit1 <- hclust(dist) # an object of class '"dist" plot(fit1) # dendogram without p values library(pvclust) fit2 <- pvclust(raw.data,
2007 Dec 07
1
pvclust warning message
Hi all I am trying to perform the follwing: fit<-pvclust(wq, method.hclust="ward", method.dist="euclidean") but get a strange error message that I just cant figure out. Has anyone come across this? Any help would be most appricieated Error in hclust(distance, method = method.hclust) : NA/NaN/Inf in foreign function call (arg 11) In addition: Warning message: NAs
2009 Feb 20
2
cluster analysis: mean values for each variable and cluster
Hi all! I'm new to R and don't know many about it. Because it is free, I managed to learn it a little bit. Here is my problem: I did a cluster analysis on 30 observations and 16 variables (monde, figaro, liberation, etc.). Here is the .txt data file:
2004 Oct 11
2
hclust title and paste - messed up
I use the following code to scan a (limited) parameter space of clustering strategies ... data <- read.table(... dataTranspose <- t(data) distMeth <- c("euclidean", "maximum", "manhattan", "canberra", "binary" ) clustMeth <- c("ward",
2009 Oct 21
2
squared euclidean distance
Dear R-Help-Team, I would like to cluster my data using the ward-method. In several papers I read (e.g. Bahrenberg) that it is neccesary to use the "squared euclidean distance" with the ward-method. Unfortunatelly I cannot find this term in r as a method for measuring the distance. Does anybody have an idea? Thanks in advance, Carolin [[alternative HTML version deleted]]
2001 Jun 12
1
cophenetic matrix
Hello, I analyse some free-sorting data so I use hierarchical clustering. I want to compare my proximity matrix with the tree representation to evalute the fitting. (stress, cophenetic correlation (pearson's correlation)...) "The cophenetic similarity of two objects a and b is defined as the similarity level at wich objects a and b become members of the same cluster during the course of
2013 Mar 28
2
hierarchical clustering with pearson's coefficient
Hello, I want to use pearson's correlation as distance between observations and then use any centroid based linkage distance (ex. Ward's distance) When linkage distances are formed as the Lance-Williams recursive formulation, they just require the initial distance between observations. See here: http://en.wikipedia.org/wiki/Ward%27s_method It is said that you have to use euclidean
2000 Dec 05
0
calculation of inertial difference with huygens theorem in ward clustering ?
Hello to the R people, within ward clustering the distance calculated to decide the clustering of 2 subsets (h1 and h2) is the variation of inertia : d(h1,h2)=I(h1Uh2)-I(h1)-I(h2); i've been said that a way to calculate faster this d(h1,h2) is using the huygens theorem decomposing the inertia into "the inertia to the centroid + the distance to an axe" (that's my version ...). My
2010 Jul 20
1
p-values pvclust maximum distance measure
Hi, I am new to clustering and was wondering why pvclust using "maximum" as distance measure nearly always results in p-values above 95%. I wrote an example programme which demonstrates this effect. I uploaded a PDF showing the results Here is the code which produces the PDF file: ------------------------------------------------------------------------------------- s <-
2011 Jul 27
0
Inversions in hierarchical clustering were they shouldn't be
Hi, I''m using heatmap.2 to cluster my data, using the centroid method for clustering and the maximum method for calculating the distance matrix: library("gplots") library("RColorBrewer") test <- matrix(c(0.96, 0.07, 0.97, 0.98, 0.50, 0.28, 0.29, 0.77, 0.08, 0.96, 0.51, 0.51, 0.14, 0.19, 0.41, 0.51), ncol=4, byrow=TRUE)
2011 Dec 02
0
what is used as height in hclust for ward linkage?
Dear R community, I am trying to understand how the ward linkage works from a quantitative point of view. To test it I have devised a simple 3-members set: G = c(0,2,10) The distances between all couples are: d(0,2) = 2 d(0,10) = 10 d(2,10) = 8 The smallest distance corresponds to merging 0 and 2. The corresponding ESS are: ESS(0,2) = 2*var(c(0,2)) = 4 ESS(0,10)
2012 Mar 15
1
Get Details About Clusters
Hi everybody! Anybody knows how can I get detalied information about clusters after using hclust? The issue is that if I have some items in different clusters, I would like to get the cluster where each item is placed. Taking into account that my data set is too large, it is not useful to have the dendogram or a graphic, and really I need something like a simple table with item label and cluster