Displaying 6 results from an estimated 6 matches for "sorrowli".
Did you mean:
sorrowly
2007 Mar 30
2
Replacing slot of S4 class in method of S4 class?
Dear all,
Assume that I have an S4 class "MyClass" with a slot "myname", which
is initialized to: myname="" in method("initialize"):
myclass <- new("MyClass", myname="")
Assume that class "MyClass" has a method "mymethod":
"mymethod.MyClass" <-
function(object, myname=character(0), ...) {
2000 Feb 01
1
hdf5
Dear R-user group
I am currently running R-0.90.1-1 for LinuxPPC on my PowerBook. Few days
ago
I have seen at CRAN that a new R-package exists, namely
R-base-with-hdf5-0.90.1-2.
Sorrowly, I could not find any explanation, what the difference to the
other package is.
I have downloaded the "Introduction to HDF5 Release 1.2" from
hdf.ncsa.uiuc.edu/HDF5.
It seems that HDF5 is a data
2003 Nov 19
11
Windows R 1.8.0 hangs when Mem Usage >1.8GB
I have a loop that increases the size of an object after each iteration. When the Windows Task Manager shows "Mem Usage" about 1.8GB, the Rgui.exe process no longer responds.
I use:
"C:\Program Files\R\rw1080\bin\Rgui.exe" --max-mem-size=4000M --min-vsize=10M --max-vsize=3000M --min-nsize=500k --max-nsize=1000M
I have a dual Xeon 2.8GHz processor box with 4GB of memory and
2004 Jul 22
0
RE: Comparison of correlation coefficients - Details
Dear Ioannis
Thank you very much for pointing me to meta-analysis. Although it
may not solve my problem with the normalization, it gives me some
other options to display the different correlation coefficients.
One possibility is the use of Funnel plots, which are even available
in library(rmeta). Another possibility is the use of forest-plots,
as implemented in rmeta as metaplot. Sorrowly,
2001 Nov 29
2
memory issue trying to solve too large a problem using hclust
Hi, all.
I'm trying to cluster 12,500 objects using hclust from package mva. The
distance matrix takes up nearly 600 MB. The distance matrix also needs to
be copied when being passed to the fortran routine that actually does the
clustering (it's modified during the clustering), so that's 1200 MB. I'm
actually on a machine with 2.5 GB of memory (and nothing else running), so I
2003 Dec 04
2
RE: R performance questions
Hi--
While I agree that we cannot agree on the ideal algorithms, we should be
taking practical steps to implement microarrays in the clinic. I think
we can all agree that our algorithms have some degree of efficacy over
and above conventional diagnostic techniques. If patients are dying
from lack of diagnostic accuracy, I think we have to work hard to use
this technology to help them, if we