search for: sorrowli

Displaying 6 results from an estimated 6 matches for "sorrowli".

Did you mean: sorrowly
2007 Mar 30
2
Replacing slot of S4 class in method of S4 class?
Dear all, Assume that I have an S4 class "MyClass" with a slot "myname", which is initialized to: myname="" in method("initialize"): myclass <- new("MyClass", myname="") Assume that class "MyClass" has a method "mymethod": "mymethod.MyClass" <- function(object, myname=character(0), ...) {
2000 Feb 01
1
hdf5
Dear R-user group I am currently running R-0.90.1-1 for LinuxPPC on my PowerBook. Few days ago I have seen at CRAN that a new R-package exists, namely R-base-with-hdf5-0.90.1-2. Sorrowly, I could not find any explanation, what the difference to the other package is. I have downloaded the "Introduction to HDF5 Release 1.2" from hdf.ncsa.uiuc.edu/HDF5. It seems that HDF5 is a data
2003 Nov 19
11
Windows R 1.8.0 hangs when Mem Usage >1.8GB
I have a loop that increases the size of an object after each iteration. When the Windows Task Manager shows "Mem Usage" about 1.8GB, the Rgui.exe process no longer responds. I use: "C:\Program Files\R\rw1080\bin\Rgui.exe" --max-mem-size=4000M --min-vsize=10M --max-vsize=3000M --min-nsize=500k --max-nsize=1000M I have a dual Xeon 2.8GHz processor box with 4GB of memory and
2004 Jul 22
0
RE: Comparison of correlation coefficients - Details
Dear Ioannis Thank you very much for pointing me to meta-analysis. Although it may not solve my problem with the normalization, it gives me some other options to display the different correlation coefficients. One possibility is the use of Funnel plots, which are even available in library(rmeta). Another possibility is the use of forest-plots, as implemented in rmeta as metaplot. Sorrowly,
2001 Nov 29
2
memory issue trying to solve too large a problem using hclust
Hi, all. I'm trying to cluster 12,500 objects using hclust from package mva. The distance matrix takes up nearly 600 MB. The distance matrix also needs to be copied when being passed to the fortran routine that actually does the clustering (it's modified during the clustering), so that's 1200 MB. I'm actually on a machine with 2.5 GB of memory (and nothing else running), so I
2003 Dec 04
2
RE: R performance questions
Hi-- While I agree that we cannot agree on the ideal algorithms, we should be taking practical steps to implement microarrays in the clinic. I think we can all agree that our algorithms have some degree of efficacy over and above conventional diagnostic techniques. If patients are dying from lack of diagnostic accuracy, I think we have to work hard to use this technology to help them, if we