similar to: PAM Clustering

Displaying 20 results from an estimated 4000 matches similar to: "PAM Clustering"

2017 Aug 17
2
PAM Clustering
Dear Germano, Thank you for your fast reply, In the above code, *MYData *is the actual data set. Do not we need to convert *MYData to *the dissimilarity matrix using *pam(as.dist(**MYData**), k = 10, diss = TRUE*)* code line?* *Regards.* On Thu, Aug 17, 2017 at 2:58 PM, Germano Rossi <germano.rossi at gmail.com> wrote: > try this > > MYdata <-
2017 Aug 17
0
PAM Clustering
Sorry, I never use pam. In the help, you can see that pam require a dataframe OR a dissimilarity matrix. If diss=FALSE then "euclidean" was use.So, I interpret that a matrix of dissimilarity is generated automatically. Problems may be in your data. Indeed pam(ruspini, 4)$diss write a dissimilaty matrix while pam(MYdata,10)$diss wite NULL 2017-08-17 16:03 GMT+02:00 Sema Atasever
2017 Aug 17
0
PAM Clustering
try this MYdata <- read.csv2("data.txt",dec='.') library(cluster) cluster.pam = pam(MYdata,10) table(cluster.pam$clustering) filenameclu = paste("clusters", ".txt") write.table(cluster.pam$clustering, file=filenameclu,sep=",") 2017-08-17 10:28 GMT+02:00 Sema Atasever <s.atasever at gmail.com>: > Dear Authorized Sir / Madam, > >
2017 Aug 17
2
PAM Clustering
[This email is either empty or too large to be displayed at this time]
2007 Nov 27
2
exporting clustering results to table
Hello list, the following approach did not work: clustersA <- pam(distances, nkA, diss=TRUE); gc(); filenameclu = paste("filenameclu", ".txt"); write.table(clustersA , file=filenameclu,sep=","); although it worked with clustersA <- hclust(distances, method="ward"); and a consecutive kclassA <- cutree(clustersA, k=nkA); filename =
2015 Nov 13
2
Fresh installation using usb
? 20151113_123827.mp4 <https://drive.google.com/file/d/0BwbqyaG4rXrCUXNfTWI3ZEk4N1k/view?usp=drive_web> ?We are using R630. Do you think it is better to install from DVD? Few times I waited for a long time. Bar was not moving in the screen. Please refer to the video. On Friday, November 13, 2015, Eero Volotinen <eero.volotinen at iki.fi> wrote: > what is model of your poweredge
2014 Sep 10
4
[LLVMdev] Leaks in PBQPBuilderWithCoalescing::build ?
Oooh. Neat. Thanks Dave. Please go ahead and commit that. Arnaud - I have no idea whether Dave's patch will help with this bug, but it's certainly worth testing. - Lang. On Wed, Sep 10, 2014 at 4:10 PM, David Blaikie <dblaikie at gmail.com> wrote: > While I'm not sure where the leak is, using some pre-canned memory > management might help... > > Attached is a
2011 Oct 22
1
cycling through a long list of files and names
R2.13.2, W7x64 Dear list, Excuse my ignorance, but I have gone through the R help (?parse, ?eval, etc.) and still really don't know how to do the following. I have the general following structure that I would like to automate [edited to make it shorter]: >>> city1997 <- dataCleaning(read.csv2("C:\\city\\year1997.txt")) city1997 <- wasteCalculations(city1997, year
2015 Dec 16
2
Mystifying SEGFAULT on 3.2.3 but not 3.2.2
Dear R developers, (Apologies about previous incomplete e-mail, I fat-fingered a keyboard shortcut)?? I recently upgraded to 3.2.3 and am getting mystifying errors with my in-development code (available at https://github.com/gertvv/gemtc/tree/feature/regression). Every few hundred LPs that I solve using the rcdd package will result in a SEGFAULT in the garbage collector (address 0x10, cause
2002 May 17
2
read.table
Hi, I have a data file with columns separated by ";" I read this file without any problem using read.csv2( ) but I had problems trying to read it with read.table( ... sep=";"). So it is not a problem for me, but I wonder if there is a bug here. drt <- read.csv2("t.txt", header=TRUE) # ok dcs <- read.table("t.txt", header=TRUE,
2009 Jun 18
3
Questíon regarding the use of write.csv2, write.table ...
Hi all, I use "write.csv" and "write.table" to write a data frame in a file like following: write.csv2(allRandomTestCase_XDroped, "allRandomTestCase.csv") But in the created file "allRandomTestCase.csv" an additional column with consecutive numbers is automatically added to the column of the data frame "allRandomTestCase_XDroped". That is why my
2012 Oct 12
2
Problem to read non-standard csv file
Hi all, I have a problem to read csv file with comma as decimal. The numbers were readed as strings. I used the following string in R, but I do not understand why it does not work. tab <- read.csv2("Specimen_RawData_1.csv", header = TRUE, sep = ";", dec = ",", nrows = -1) In addition, I copy/past into the post the link to the csv file generated by my instrument.
2009 Aug 10
2
(sans objet)
i have written this in R, > data1 <- read.csv2("c:/Total1.csv",sep=",") > data2 <- read.csv2("c:/GWPMax1.csv",sep=",") > M <- merge(data1, data2, by.x = "Policy.Number", by.y = "Policy.Number") > nrow(data1) [1] 20 > nrow(M) [1] 12 > NbOfPolicyWithoutGWPMax <- nrow(data1)-nrow(M) >
2009 Aug 10
3
how use cat() function?
i  want to print in the console and to have an excel file like this no_GWP                NbOfPolicyClass1[0-1000]     NbOfPolicyClass2[1000-3000]        NbOfPolicyClass3[> 3000] No_GWPMax=8    NbpolicyClass1=5                   NbpolicyClass2=4                            NbpolicyClass3 =3              i have do it like this:!!! data1 <-
2010 Apr 01
3
reading excel into R
Dear all, I am new R user and I am sure that this question has been asked quite often and I have also googled it and read about it! I understood that in order to read excel sheet into R you need to open it and saved it as csv or text, is this true? or you can use read.delim2 and read.csv2 to do this without the following error > dat <- read.csv2(file="C:\\Dokumente und
2001 Apr 06
2
automatic levels
Hello, I've imported a csv, semi-colon spearated file with read.csv2, containing one column of rownames and one column of floating point numbers. When I look at the column of data with framename$columnname, I get the values of the column plus level values. Are these level values created automatically ? The problem is when I try to calcluate the correlation coefficient between this set of data
2009 Nov 02
1
two small wishes (with code sugegstions) for R-core
Dear R developers, It would be great if you could implement the two minor code changes suggested below, which would help processing large objects in R. Jens Oehlschl?gel # Wish no. 1: let [.AsIs return the class AFTER subsetting, not the class of the original object # Wish no. 2: adjust write.csv and write.csv2 for multiple calls in chunked writing # Rationale no. 1: a couple of packages
2005 Nov 17
1
Help with read.csv2
Hello, I am importing the following file ;aa;bb;cc 1988;12;12;12 1989;78;78;12 1990;78;78;12 1991;78;78;12 1992;78;78;12 1993;78;78;12 1994;78;78;12 ------------------------------------------------ data<-read.csv2("test.csv",header=T) ------------------------------------------ it gives X aa bb cc 1 1988 12 12 12 2 1989 78 78 12 3 1990 78 78 12 4 1991 78 78 12 5 1992 78 78 12 6
2014 Jul 09
3
error com un archivo
Hola a todos, Me gustaria pedir vuestra ayuda a encontrar el error que no consigo encontrar en este archivo. He revisado todo mil veces y probado y no doy con ello.Adjunto el archivo con Google drive porque es muy grande. ? monicap_50.csv <https://docs.google.com/file/d/0B8o2KrPEgG7ATlBMc19lTVk1d3M/edit?usp=drive_web> ? Este es el script, y lo que no entiendo que pasa es que tengo 592044
2009 Aug 06
2
problem with r import data
Good moorning, yesterday, i asked for how we can import data into R i found that it is necessary to make the file in csv format then use L<- read.csv2("path of file") now i see that the dimension of my table importing into R is 1 colonne (i have realy 20) and (655555) the number of all my lines in the excel file (i want to have only the 2600 lines in my table) Also, when i say to do