Displaying 20 results from an estimated 1000 matches similar to: "Windows Memory"
2001 Jan 14
2
Help
Dear sir,
I am using R in windows. I want to extend R Memory
size.
I use the following command, but unfortunately it
doesn't work.
-- vsize=15M --nsize=1000K
Your help is appreciated.
Thanks,
Esmail Amiri.
__________________________________________________
Do You Yahoo!?
Get email at your own domain with Yahoo! Mail.
http://personal.mail.yahoo.com/
2002 Nov 11
3
Wind roses
Are there any R function or packages that generate wind roses? These are
``star''-like plots that summarize the frequency, direction and magnigude
of windspeeds.
Thanks,
Matt
--
Matt Pocernich
University of Colorado - Denver
Department of Mathematics
-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-
r-help mailing list -- Read
2000 Nov 09
3
maximum of nsize=20000k ??
Dear R-ers,
somehow it is not possible to increase nsize to more than
20000k. When I specify e.g.
> R --vsize=10M --nsize=21000K
the result is:
free total (Mb)
Ncells 99658 350000 6.7
Vcells 1219173 1310720 10.0
Maybe I have overlooked s.th....
Marcus
--
+-------------------------------------------------------
| Marcus Eger
| E-Mail: eger.m at gmx.de (NEW)
|
2001 Mar 01
3
How do you expand memory capability (Was: R crashes in Windows ME)
Hello-
Since my data bank in SPSS has > 40 variables, I think that R crashes because of the memory limit.
In Maindonald?s UsingR text, on pg 3, there?s a footnote that reads:
"If you want larger memory space than the default you may want a target akin to
<path to binary>\rw091\bin\rgui.exe --visize 30M --nsize 1000K
[The default is --vsize 6M --nsize 250K
2001 Aug 22
1
Huge workspace cannot be opened
Hi everyone,
I have a problem that some people may have already encountered but i did not
find the solution yet.
As I use R to simulate several arrays of data, my workspace is now 35Mb big and
I cannot launch R with it.
An "xdr real data read error occured" and R tells me to delete .RData or
increase memory. I WON'T delete this file and changing the max-nsize to 40600k
did not
2000 Aug 25
3
unexpected R crash - again
Sorry, but I lost this thread, so I sending this as a new message.
This is really a follow-up to a post from a couple days ago saying that
fisher.test from the ctest library crashed on the following data set:
> T
[,1] [,2]
[1,] 2 1
[2,] 2 1
[3,] 4 0
[4,] 8 0
[5,] 6 0
[6,] 0 0
[7,] 1 0
[8,] 1 1
[9,] 7 1
[10,] 8 2
[11,]
1999 Apr 27
2
Memory management
Dear all,
I don't get it:
First of all, the help doesn't say what are the memory limits of
R. Say, what's the max heap size for instance ????
Secondly, I invoke R with the following commands each time:
rgui --vsize 30M --nsize 1000K
rgui --vsize 30M --nsize 2000K
rgui --vsize 30M --nsize 3000K
rgui --vsize 30M --nsize 4000K
I try to open a matrix 8000x8000 by issuing
2001 Mar 21
3
memory allocation error
Hi,
I have recently installed R-1.2.2 for windows (16MB RAM, P-166) and I
getting the following message after processing my data (6 variables and
1200 observations):
>Error: cannot allocate vector of size 4 Kb
>In addition: Warning message:
>Reached total allocation of 15Mb: see help(memory.size)
Then, the program close.
With the last version, 1.1.1 (I think) I didn't have this
1999 Apr 12
3
--nsize and --vsize
Martin M has suggested I widen this discussion to R-devel, and
> I agree that we should increase them,
> but I'm not sure at all about the amount.
>
> The default could even depend on the architecture (via "./configure")..
Views, please.
------------- Begin Forwarded Message -------------
Is is not time we increased the defaults a bit? As the base gets bigger
I hit
2010 Nov 04
1
Memory Management under Linux
Dear all,
I am using ubuntu linux 32 with 4 Gb. I am running a very small script and I always got the same error message: CAN NOT ALLOCATE A VECTOR OF SIZE 231.8 Mb.
I have reading carefully the instruction in ?Memory. Using the function gc() I got very low numbers of memory (please sea below). I know that it has been posted several times at r-help
2010 Nov 05
1
improve R memory under linux
Dear all,
I am using ubuntu linux 32 with 4 Gb. I am running a very small script and I always got the same error message: CAN NOT ALLOCATE A VECTOR OF SIZE 231.8 Mb.
I have reading carefully the instruction in ?Memory. Using the function gc() I got very low numbers of memory (please sea below). I know that it has been posted several times at r-help
2010 Nov 05
1
R memory allocation in Linux
Dear all,
I am using ubuntu linux 32 with 4 Gb. I am running a very small script and I always got the same error message: CAN NOT ALLOCATE A VECTOR OF SIZE 231.8 Mb.
I have reading carefully the instruction in ?Memory. Using the function gc() I got very low numbers of memory (please sea below). I know that it has been posted several times at r-help
2001 Feb 10
3
Scatterplot - symbols coded by factor
Hi - I would like to create a scatterplot with the symbols coded by a factor in each record. Is their a better way than what I have used below? I attempted to use the plot.factor, but cannot force it to make a scatterplot instead of boxplots. I have considerd making a function with a 'for' statement for each level of the factor. Is there a better way for when I am dealing with more
2003 Dec 19
2
generic/method consistency
Hi,
I realize the answer is very likely in the section Generic functions and
methods (or Adding new generics), but I'm not clear what to do with the
following. Running R CMD check, I get the following warnings for my
generic functions. Does this mean I need the argument
* checking generic/method consistency ... WARNING
leps:
function(x, ...)
leps.default:
function(x, pred, titl,
2001 Apr 02
1
Run out of memory
I am trying to use R to cluster 7129 samples, my data set is a 7129 x 38 matrix, when I trying to get the distance matrix using function dist( ), the memory exhausted, and I tried to set the memory when I run R by
R --vsize=250M --nsize=1000k
no matter what I set for vsize the result is the same, it says:
Error: heap memory (256000kb) exhausted [need 198498Kb more]
2001 Nov 07
2
Filtering data
Hello,
I am having difficulty filtering data. I am working with flow data
collected at a stream gage. For each record, I have a date and flow
value. I have filtered this data to only include days when flow values
exceed a given threshold.
Here is my problem. Within this subset of data, I often have several
consecutive days above the threshold. From this group of days, I wish to
select the
2004 Jul 27
1
lattice.device in loop
Hi,
I am having problems creating a pdf file of a lattice graph. Things work
fine for a single image, but I am having trouble using the commands in a
loop. To illustrate by example
This works with both ps and pdf files.
dat = list(x= 1:10, y = 1:10)
trellis.device(postscript, file = "/d1/pocernic/test.ps")
xyplot(y~x, data = dat)
dev.off()
This does not. It produces a very
2004 Jul 28
2
lattice variable by page
Hi,
Using lattice's xyplot, is it possible to specify a variable to group
plots by page. For example, if I have
xyplot(y~x|A*B*C)
could I get a page created for each unique value in
variable C ? I am hoping to avoid having pages with the same strip above
each plot on a page.
Thanks,
M
Matt Pocernich
NCAR - Research Applications Program
303-497-8312
1999 Nov 12
1
R-0.65.1 Startup
Dear R users,
I have noticed that my R startup is extremely slow. It takes almost 3
minutes from "double-click" to R prompt. I have been running R-0.64.1 till
recently and it took about 30 sec. I still have access to R-0.64.1. When I
started it up, it took about 25 sec. Can anyone tell me if this is a bug in
R or a problem with my machine?
Note: This is after bootup with R being the
2004 Mar 04
2
lattice, different plotting symbols
Hi,
I am trying to plot different variables from a data.frame using lattice's
xyplot using code like that below. How do I specify a symbol and color
for the variable 'prob' and different one's for 'll.prob'?
Thanks, Matt
xyplot( prob + ll.prob ~ time.eff |stat.id + time.out ,data = OUT,
allow.multiple = TRUE,
layout = c(6,3), as.table = TRUE ,