Displaying 20 results from an estimated 900 matches similar to: "memory trouble"
2001 Aug 23
2
Missing Values from Spss-Data
Hello,
i'm a newbie of the really fantastic "R" , but have some
problems to define Missing Values from Spss-Files which i import
with the read.spss(data) command !
What should i do to recode i.e. the values 9,99 & 999 to NA.
Something like this
x <- c(9,99,999,NA)
x==NA
...do not what i want, because the data.frame show
the values 9,99 & 999 and not NA in this fields ?
2001 Jan 19
2
bringToTop() fails (PR#818)
windows()
plot(1,1)
windows()
plot(1:2,1:2)
dev.list() # we have devices 2 and 3
# device 3 is active and in front
bringToTop(2) # does not work
bringToTop(3) # does not work also
Regards
Jens Oehlschlaegel
--please do not edit the information below--
Version:
platform = i386-pc-mingw32
arch = x86
os = Win32
system = x86, Win32
status =
major = 1
minor = 2.1
year = 2001
month = 01
2001 Aug 29
2
Matching Data & Results (Fuzzy-Cluster-Analysis)
Hello,
i use i.e. the cluster-package and the
fanny-object to construct some cluster's and get
the membership'S for every "Person" & Cluster !
My finally problem is to match efficient the data
and the results's togehter in one table!
Until now i use the ODBC-Package and
save both in a access-file and copy them together,
but i can't be sure that the sorting is the
2010 Nov 04
1
Memory Management under Linux
Dear all,
I am using ubuntu linux 32 with 4 Gb. I am running a very small script and I always got the same error message: CAN NOT ALLOCATE A VECTOR OF SIZE 231.8 Mb.
I have reading carefully the instruction in ?Memory. Using the function gc() I got very low numbers of memory (please sea below). I know that it has been posted several times at r-help
2010 Nov 05
1
improve R memory under linux
Dear all,
I am using ubuntu linux 32 with 4 Gb. I am running a very small script and I always got the same error message: CAN NOT ALLOCATE A VECTOR OF SIZE 231.8 Mb.
I have reading carefully the instruction in ?Memory. Using the function gc() I got very low numbers of memory (please sea below). I know that it has been posted several times at r-help
2010 Nov 08
0
R memory allocation ubuntu
Dear all,
I am using ubuntu linux 32 with 4 Gb. I am running a
very small script and I always got the same error message: CAN NOT
ALLOCATE A VECTOR OF SIZE 231.8 Mb.
I have reading carefully the
instruction in ?Memory. Using the function gc() I got very low numbers
of memory (please sea below). I know that it has been posted several
times at r-help
2010 Nov 05
1
R memory allocation in Linux
Dear all,
I am using ubuntu linux 32 with 4 Gb. I am running a very small script and I always got the same error message: CAN NOT ALLOCATE A VECTOR OF SIZE 231.8 Mb.
I have reading carefully the instruction in ?Memory. Using the function gc() I got very low numbers of memory (please sea below). I know that it has been posted several times at r-help
2013 Aug 21
2
[PATCH 1/3] Rationalise whitespace to 4 space indentation with no trailing spaces
RHSrvAny.c was using a mixture of 4 space indentation, and tabs with a width of
4. This commit rationalises the whitespace to use only 4 space indentation, and
removes trailing whitespace.
---
RHSrvAny/RHSrvAny.c | 537 ++++++++++++++++++++++++++--------------------------
RHSrvAny/RHSrvAny.h | 1 -
RHSrvAny/resource.h | 2 +-
3 files changed, 269 insertions(+), 271 deletions(-)
diff --git
2007 Oct 28
1
tree problem
I am trying to use tree to partition a data set. The data set has 3924
observations. Partitioning seems to work for small subsets of the data,
but when I use the entire data set, no partitioning occurs. The
variables are:
RESP respondent to a survey (0 = not a respondent, 1 =
respondent)
AGE_P Age (continuous)
ORIGIN_I Hispanic Ethnicity (1 = Hispanic, 2 = non-Hispanic)
RACRECI2 Race
2005 Jun 29
3
Memory Management under Linux: Problems to allocate large amounts of data
Dear Group
I'm still trying to bring many data into R (see older postings). After solving some troubles with the database I do most of the work in MySQL. But still I could be nice to work on some data using R. Therefore I can use a dedicated Server with Gentoo Linux as OS hosting only R. This Server is a nice machine with two CPU and 4GB RAM which should do the job:
Dual Intel XEON 3.06 GHz
1999 Apr 27
2
Memory management
Dear all,
I don't get it:
First of all, the help doesn't say what are the memory limits of
R. Say, what's the max heap size for instance ????
Secondly, I invoke R with the following commands each time:
rgui --vsize 30M --nsize 1000K
rgui --vsize 30M --nsize 2000K
rgui --vsize 30M --nsize 3000K
rgui --vsize 30M --nsize 4000K
I try to open a matrix 8000x8000 by issuing
2005 Jul 07
2
r: LOOPING
hi all
i know that one should try and limit the amount of looping in R
programs. i have supplied some code below. i am interested in seeing how
the code cold be rewritten if we dont use the loops.
a brief overview of what is done in the code.
==============================================
==============================================
==============================================
1. the input
2013 Aug 29
5
[PATCH 1/6] Rationalise whitespace to 4 space indentation with no trailing spaces
RHSrvAny.c was using a mixture of 4 space indentation, and tabs with a width of
4. This commit rationalises the whitespace to use only 4 space indentation, and
removes trailing whitespace.
---
RHSrvAny/RHSrvAny.c | 537 ++++++++++++++++++++++++++--------------------------
RHSrvAny/RHSrvAny.h | 1 -
RHSrvAny/resource.h | 2 +-
3 files changed, 269 insertions(+), 271 deletions(-)
diff --git
2010 Jan 19
2
Server hanging despite efforts to correct memory limits
My group is working with datasets between 100 Mb and 1 GB in size, using
multiple log ins. From the documentation, it appears that vsize is limited
to 2^30-1, which tends to prove too restrictive for our use. When we drop
that restriction (set vsize = NA) we end up hanging the server, which
requires a restart. Is there any way to increase the memory limits on R
while keeping our jobs from
2001 Aug 22
1
Huge workspace cannot be opened
Hi everyone,
I have a problem that some people may have already encountered but i did not
find the solution yet.
As I use R to simulate several arrays of data, my workspace is now 35Mb big and
I cannot launch R with it.
An "xdr real data read error occured" and R tells me to delete .RData or
increase memory. I WON'T delete this file and changing the max-nsize to 40600k
did not
2013 Aug 07
2
Override master service settigs with spaces
I'd like to override one setting for a master service in
conf.d/10-master.conf. Unfortunately, said setting contains spaces, and I
do not know how to escape them properly.
Here's what I've tried so far. (Note: This is just the easiest/silliest
test case I could come up with; not the actual setting or service I want to
overwrite.)
conf.d/10-master.conf:
service quota-status {
2006 Mar 03
0
Memory problem
Hi list,
I am analysing a large dataset using random coefficient (using nlme) and
fixed effects (using lm function) models. I have problem with my R version
2. 2. 1 due to memory allocation difficulties. When I try to expand the
memory I get the following error message.
> R --min-vsize=10000000 --max-vsize=1000000000 --min-nsize=500
--max-nsize=10000000
Error: target of assignment expands
2010 Jan 18
0
R jobs keep hanging linux server despite mem.limits modifcations
My group is working with datasets between 100 Mb and 1 GB in size, using
multiple log ins. From the documentation, it appears that vsize is limited
to 2^30-1, which tends to prove too restrictive for our use. When we drop
that restriction (set vsize = NA) we end up hanging the server, which
requires a restart. Is there any way to increase the memory limits on R
while keeping our jobs from
2005 Dec 20
1
Problems in batch mode
Dear R-users,
I am trying to run some simulations in batch mode. In an older version
of the program, I used
rterm --vsize=100M --nsize=5000K --restore --save <input file> output file,
however, in the new version R 2.2.0 , the parameters vsize and nsize are
ignored.
I can use the command memory.limit to increase memory, but I am not sure if
this corresponds to vsize and nsize.
2017 Nov 22
2
function pointers?
We have a project that calls for the creation of a list of many
distribution objects. Distributions can be of various types, with
various parameters, but we ran into some problems. I started testing
on a simple list of rnorm-based objects.
I was a little surprised at the RAM storage requirements, here's an example:
N <- 10000
closureList <- vector("list", N)
nsize = sample(x