Displaying 20 results from an estimated 4000 matches similar to: "PAM Clustering"
2017 Aug 17
2
PAM Clustering
Dear Germano,
Thank you for your fast reply,
In the above code, *MYData *is the actual data set.
Do not we need to convert *MYData to *the dissimilarity matrix using
*pam(as.dist(**MYData**), k = 10, diss = TRUE*)* code line?*
*Regards.*
On Thu, Aug 17, 2017 at 2:58 PM, Germano Rossi <germano.rossi at gmail.com>
wrote:
> try this
>
> MYdata <-
2017 Aug 17
0
PAM Clustering
Sorry, I never use pam. In the help, you can see that pam require a
dataframe OR a dissimilarity matrix. If diss=FALSE then "euclidean" was use.So,
I interpret that a matrix of dissimilarity is generated automatically.
Problems may be in your data. Indeed
pam(ruspini, 4)$diss
write a dissimilaty matrix
while
pam(MYdata,10)$diss
wite NULL
2017-08-17 16:03 GMT+02:00 Sema Atasever
2017 Aug 17
0
PAM Clustering
try this
MYdata <- read.csv2("data.txt",dec='.')
library(cluster)
cluster.pam = pam(MYdata,10)
table(cluster.pam$clustering)
filenameclu = paste("clusters", ".txt")
write.table(cluster.pam$clustering, file=filenameclu,sep=",")
2017-08-17 10:28 GMT+02:00 Sema Atasever <s.atasever at gmail.com>:
> Dear Authorized Sir / Madam,
>
>
2017 Aug 17
2
PAM Clustering
[This email is either empty or too large to be displayed at this time]
2007 Nov 27
2
exporting clustering results to table
Hello list,
the following approach did not work:
clustersA <- pam(distances, nkA, diss=TRUE);
gc();
filenameclu = paste("filenameclu", ".txt");
write.table(clustersA , file=filenameclu,sep=",");
although it worked with
clustersA <- hclust(distances, method="ward");
and a consecutive
kclassA <- cutree(clustersA, k=nkA);
filename =
2015 Nov 13
2
Fresh installation using usb
?
20151113_123827.mp4
<https://drive.google.com/file/d/0BwbqyaG4rXrCUXNfTWI3ZEk4N1k/view?usp=drive_web>
?We are using R630. Do you think it is better to install from DVD?
Few times I waited for a long time. Bar was not moving in the screen.
Please refer to the video.
On Friday, November 13, 2015, Eero Volotinen <eero.volotinen at iki.fi> wrote:
> what is model of your poweredge
2014 Sep 10
4
[LLVMdev] Leaks in PBQPBuilderWithCoalescing::build ?
Oooh. Neat. Thanks Dave. Please go ahead and commit that.
Arnaud - I have no idea whether Dave's patch will help with this bug, but
it's certainly worth testing.
- Lang.
On Wed, Sep 10, 2014 at 4:10 PM, David Blaikie <dblaikie at gmail.com> wrote:
> While I'm not sure where the leak is, using some pre-canned memory
> management might help...
>
> Attached is a
2011 Oct 22
1
cycling through a long list of files and names
R2.13.2, W7x64
Dear list,
Excuse my ignorance, but I have gone through the R help (?parse, ?eval,
etc.) and still really don't know how to do the following.
I have the general following structure that I would like to automate
[edited to make it shorter]:
>>>
city1997 <- dataCleaning(read.csv2("C:\\city\\year1997.txt"))
city1997 <- wasteCalculations(city1997, year
2015 Dec 16
2
Mystifying SEGFAULT on 3.2.3 but not 3.2.2
Dear R developers,
(Apologies about previous incomplete e-mail, I fat-fingered a keyboard
shortcut)??
I recently upgraded to 3.2.3 and am getting mystifying errors with my
in-development code (available at
https://github.com/gertvv/gemtc/tree/feature/regression). Every few hundred
LPs that I solve using the rcdd package will result in a SEGFAULT in the
garbage collector (address 0x10, cause
2002 May 17
2
read.table
Hi,
I have a data file with columns separated by ";" I read this file
without any problem using read.csv2( ) but I had problems trying to read
it with read.table( ... sep=";"). So it is not a problem for me, but I
wonder if there is a bug here.
drt <- read.csv2("t.txt", header=TRUE) # ok
dcs <- read.table("t.txt", header=TRUE,
2009 Jun 18
3
Questíon regarding the use of write.csv2, write.table ...
Hi all,
I use "write.csv" and "write.table" to write a data frame in a file like
following:
write.csv2(allRandomTestCase_XDroped, "allRandomTestCase.csv")
But in the created file "allRandomTestCase.csv" an additional column with
consecutive numbers is automatically added to the column of the data frame
"allRandomTestCase_XDroped".
That is why my
2012 Oct 12
2
Problem to read non-standard csv file
Hi all,
I have a problem to read csv file with comma as decimal. The numbers were
readed as strings.
I used the following string in R, but I do not understand why it does not
work.
tab <- read.csv2("Specimen_RawData_1.csv", header = TRUE, sep = ";", dec =
",", nrows = -1)
In addition, I copy/past into the post the link to the csv file generated by
my instrument.
2009 Aug 10
2
(sans objet)
i have written this in R,
> data1 <- read.csv2("c:/Total1.csv",sep=",")
> data2 <- read.csv2("c:/GWPMax1.csv",sep=",")
> M <- merge(data1, data2, by.x = "Policy.Number", by.y = "Policy.Number")
> nrow(data1)
[1] 20
> nrow(M)
[1] 12
> NbOfPolicyWithoutGWPMax <- nrow(data1)-nrow(M)
>
2009 Aug 10
3
how use cat() function?
i want to print in the console and to have an excel file like this
no_GWP NbOfPolicyClass1[0-1000] NbOfPolicyClass2[1000-3000] NbOfPolicyClass3[> 3000]
No_GWPMax=8 NbpolicyClass1=5 NbpolicyClass2=4 NbpolicyClass3 =3
i have do it like this:!!!
data1 <-
2010 Apr 01
3
reading excel into R
Dear all,
I am new R user and I am sure that this question has been asked quite often
and I have also googled it and read about it! I understood that in order to
read excel sheet into R you need to open it and saved it as csv or text, is
this true? or you can use read.delim2 and read.csv2 to do this without the
following error
> dat <- read.csv2(file="C:\\Dokumente und
2001 Apr 06
2
automatic levels
Hello,
I've imported a csv, semi-colon spearated file with read.csv2,
containing one column of rownames and one column of floating point
numbers. When I look at the column of data with framename$columnname, I
get the values of the column plus level values. Are these level values
created automatically ?
The problem is when I try to calcluate the correlation coefficient
between this set of data
2009 Nov 02
1
two small wishes (with code sugegstions) for R-core
Dear R developers,
It would be great if you could implement the two minor code changes suggested below, which would help processing large objects in R.
Jens Oehlschl?gel
# Wish no. 1: let [.AsIs return the class AFTER subsetting, not the class of the original object
# Wish no. 2: adjust write.csv and write.csv2 for multiple calls in chunked writing
# Rationale no. 1: a couple of packages
2005 Nov 17
1
Help with read.csv2
Hello,
I am importing the following file
;aa;bb;cc
1988;12;12;12
1989;78;78;12
1990;78;78;12
1991;78;78;12
1992;78;78;12
1993;78;78;12
1994;78;78;12
------------------------------------------------
data<-read.csv2("test.csv",header=T)
------------------------------------------
it gives
X aa bb cc
1 1988 12 12 12
2 1989 78 78 12
3 1990 78 78 12
4 1991 78 78 12
5 1992 78 78 12
6
2014 Jul 09
3
error com un archivo
Hola a todos,
Me gustaria pedir vuestra ayuda a encontrar el error que no consigo
encontrar en este archivo. He revisado todo mil veces y probado y no doy
con ello.Adjunto el archivo con Google drive porque es muy grande.
?
monicap_50.csv
<https://docs.google.com/file/d/0B8o2KrPEgG7ATlBMc19lTVk1d3M/edit?usp=drive_web>
?
Este es el script, y lo que no entiendo que pasa es que tengo 592044
2009 Aug 06
2
problem with r import data
Good moorning,
yesterday, i asked for how we can import data into R i found that it is necessary to make the file in csv format then use L<- read.csv2("path of file")
now i see that the dimension of my table importing into R is 1 colonne (i have realy 20) and (655555) the number of all my lines in the excel file (i want to have only the 2600 lines in my table)
Also, when i say to do