similar to: .Rdata questions

Displaying 20 results from an estimated 2000 matches similar to: ".Rdata questions"

1999 Jan 08
1
avoiding (another) error
Dear all, Many thanks to all who always so quickly respond with helpful hints! Currently I am working with R 0.63.1 on SuSE 5.2 which I installed this morning within about half an hour: my deepest compliments to the developpers! (I had still 0.62.2 before.) This time I am trying the following: > xy <- list (a= c (1, 2, NA), b= c (2, 3)) > list.name <- "xy" > get
1999 Jan 08
1
character vectors in data.frames
Dear all, well, today I just seem to run into these problems ... (I am still working on R 0.63.1 on a SuSE 5.2). I would like to have a data.frame that also contains character (and NOT factor) vectors. Two things that I tried (the second worked in R 0.62.2): > test.df <- data.frame (X= character (4)) > test.df$X [1] <- "a" Warning: invalid factor level, NAs generated or:
1999 Jan 25
2
graphical paramters to boxplot
Dear all, is this a bug or do a miss something (this concerns version 0.63.1 on Linux)? > boxplot (c (1, 2, 3, 4, 5), ylim= c (-5, 5)) Warning: parameter "ylim" couldn't be set in high-level plot() function > boxplot (c (1, 2, 3, 4, 5), pars= list (ylim= c (-5, 5))) Warning: parameter "ylim" couldn't be set in high-level plot() function > and actually in
1999 Jan 04
1
Error messages
Dear all, The following works with a simple result of lm but gives the error with a more complicated one (could this be due to singularities that occurr in the model?). > cooks.distance (X) Error: NAs in foreign function call (arg 6) > class (X) [1] "lm" But, actually, I would like to ask a more general question: Is there a place (I searched in the CRAN site) where I find
1998 Jul 16
1
R-beta: which Linux - again
Dear all, I think this question has been raised very recently but I think I didn't see repsonses on the list nor did I find the thread in the archives. (If it is there I'd appreciate directions ...) I am about to install Linux on a Compaq notebook. One important thing I want to do is being able to run R. I have fiddled with different distributions of Linux (LST, SuSE, Debian) and can
1998 Dec 03
1
image ()
Dear all, I am working with R Version 0.62.2 (July 10, 1998) on a SuSE Linux 5.2. I have troubles in understanding the function image (). Can someone enlighten me? I'd like to draw a square image-plot (with every little part of the image a square as well) and add some (straight) lines to it. Also, I create an empty z matrix which I need to fill step by step. Thus I tried to find out how
1999 Jan 22
0
lm with rank-deficient X matrix
Dear all, I would like to fit an lm in which a subset of the explanatory variables are linearly dependent. Thus I would like to include the restriction that all betas of these variables sum to 1. Is there a way to this in R? Or is this what happens automatically if I set singular.ok= T (which is the default, I believe)? Thanks for your help. Lorenz --
1999 Mar 30
0
idea for (further) debugging ace ()?
Dear all, I am trying to use ace on a 446 x 14 Matrix, with two categorical variables in the first two columns. If I try to run it I get: > aceiH <- ace (data.mat, resp, cat= 1:length(x.cat)) Error: NAs in foreign function call (arg 3) There is only one .Fortran call in ace () with the third argument being x = t(x). Thus I tried to debug ace, i.e. I went through ace until just before
1999 Apr 27
2
Memory management
Dear all, I don't get it: First of all, the help doesn't say what are the memory limits of R. Say, what's the max heap size for instance ???? Secondly, I invoke R with the following commands each time: rgui --vsize 30M --nsize 1000K rgui --vsize 30M --nsize 2000K rgui --vsize 30M --nsize 3000K rgui --vsize 30M --nsize 4000K I try to open a matrix 8000x8000 by issuing
2010 Jan 19
2
Server hanging despite efforts to correct memory limits
My group is working with datasets between 100 Mb and 1 GB in size, using multiple log ins. From the documentation, it appears that vsize is limited to 2^30-1, which tends to prove too restrictive for our use. When we drop that restriction (set vsize = NA) we end up hanging the server, which requires a restart. Is there any way to increase the memory limits on R while keeping our jobs from
2001 Aug 22
1
Huge workspace cannot be opened
Hi everyone, I have a problem that some people may have already encountered but i did not find the solution yet. As I use R to simulate several arrays of data, my workspace is now 35Mb big and I cannot launch R with it. An "xdr real data read error occured" and R tells me to delete .RData or increase memory. I WON'T delete this file and changing the max-nsize to 40600k did not
2000 Nov 09
3
maximum of nsize=20000k ??
Dear R-ers, somehow it is not possible to increase nsize to more than 20000k. When I specify e.g. > R --vsize=10M --nsize=21000K the result is: free total (Mb) Ncells 99658 350000 6.7 Vcells 1219173 1310720 10.0 Maybe I have overlooked s.th.... Marcus -- +------------------------------------------------------- | Marcus Eger | E-Mail: eger.m at gmx.de (NEW) |
1999 Jan 27
1
cant restore .Rdata
Hi Folks, I loaded a couple of quite large data sets into an R session and then quit (after saving the image). Now I get: Error: a read error occured Fatal error: unable to restore saved data (remove .RData or increase memory) after trying to start my R session using something like: R --vsize XXX --nsize 1000000 For any value of XXX (I went up to 300 or 400, which is as high as I could go.
2005 Dec 20
1
Problems in batch mode
Dear R-users, I am trying to run some simulations in batch mode. In an older version of the program, I used rterm --vsize=100M --nsize=5000K --restore --save <input file> output file, however, in the new version R 2.2.0 , the parameters vsize and nsize are ignored. I can use the command memory.limit to increase memory, but I am not sure if this corresponds to vsize and nsize.
1999 May 15
2
vsize and nsize
I am running R version ??? under Redhat 5.2. It seems as though the --nsize object has no effct on the size of the allocated Ncells as determined using gc(). Yes, I have that much data.... That is if I envoke R with R --vsize 100 --nsize 5000000 then type gc() I get free total Ncells 92202 200000 Vcells 12928414 13107200 Thanks Tony Long Ecology and Evolutionary Biology Steinhaus
2009 May 07
1
increasing memory for R bg job
Hi, Is the following command used to increase the memory or any other command when a background R job is run? R --min-vsize=vl --max-vsize=vu --min-nsize=nl --max-nsize=nu --max-ppsize=N source: http://stat.ethz.ch/R-manual/R-patched/library/base/html/Memory.html Thx Carol [[alternative HTML version deleted]]
2009 Nov 30
1
allocating vector memory > 1 GByte on Windows XP / Vista / 7
Let me begin stating that I read all help files and faq's on the subject matter (there aren't more than about a dozen) but either did not find solutions or found them not to work. Here is the issue. I am trying to run a spatial regression on a medium-sized dataset. Part of the functions in the spdep package I use require me to allocate a vector of 1.1 Gb (mine is not a spatial SIG
2001 Mar 01
3
How do you expand memory capability (Was: R crashes in Windows ME)
Hello- Since my data bank in SPSS has > 40 variables, I think that R crashes because of the memory limit. In Maindonald?s UsingR text, on pg 3, there?s a footnote that reads: "If you want larger memory space than the default you may want a target akin to <path to binary>\rw091\bin\rgui.exe --visize 30M --nsize 1000K [The default is --vsize 6M --nsize 250K
1999 Oct 06
2
R --nsize 2M runs havoc (under linux)
Dear All, I am running R version 0.65.0 under a) Suse-Linux 6.1, and Suse-Linux 6.2, compiler gcc-2.95, CPUs pentium pro 200, 128MB, and pentium II 450, 128MB b) Solaris 5.7, compiler gcc-2.95, cpu SUN sparc, 4000MB When I set --nsize to more than 1M, R's internal storage management runs havoc. gc() indicates the requested sizes, but the overall process size is much too big: Running R with
2004 Mar 08
2
memory problem
I am trying to upload into R 143 Affymetrix chips onto using R on the NIH Nimbus server. I can load 10 chips without a problem, however, when I try to load 143 I receive a error message: cannot create a vector of 523263 KB. I have expanded the memory of R as follows: R --min-vsize=10M --max-vsize=2500M --min-nsize=10M -max-nsize=50M (as specified in help in R). After running this command the