Displaying 20 results from an estimated 3000 matches similar to: "Memory management"
1999 Apr 16
1
Heap size
Hi,
I am trying to create a matrix 8000x8000 but cannot
make R work properly with:
rgui --vsize 30 --nsize 1000000
I can't pass an nsize value more than 1000000.
What are the limits uder windows 98 ?
Best,
Costas
--
-------------------------------------------------------------------------
Konstantinos E. Vorloou | Tel: +44 (0)191 374 1821
Department of Economics &
1999 Feb 17
3
Variable Names
Dear R users,
That's probably a silly question even for a newbie but :
Is it possible to assign variable prices as variable names ?
For instance, would something like the following work :
for (i in 1:4)
{
error.i <- some_calculation
}
so "error.i" variables would be created for i=1,2,3,4
i.e. error.1 error.2 error.3 etc. etc.
Costas
--
1999 Apr 07
1
Console Output while running scripts
Hi,
I am using 0.63.2 on Windows. I have some functions
programmed to do repetitive tasks and have some
print(variables) statements to check how my program is progressing
but get no output on screen (console) while the program is running.
I tried CTRL-W as the help says but I get no screen output
before each run terminates. Probably some kind of setting needs to be
predefined correctly.
Any
1999 Sep 06
1
Printer button
Hi,
I am not able to print from within windows 98 (graphics and text) in
the latest R. Menus and buttons don't function...
platform Windows
arch x86
os Win32
system x86, Win32
status
status.rev 0
major 0
minor 65.0
year 1999
month August
day 27
language R
Anybody encountering the same problem ?
Also windows manipulation by the environment
1999 May 20
1
Wavethresh + accessC > Error: wd.structure has no class
Hi,
Whenever I issue this commands (as in the help for the threshold
function),
#
# Do a 1D decomposition
#
tdecomp <- wd(test.data)
#
# Threshold it
#
tdecomp.thresh <- threshold(tdecomp)
#
# Reconstruct from the thresholded coefficients
#
trecons <- wr(tdecomp.thresh)
#
#
1999 Mar 23
1
rts, cts, tsplot
Is plot(ts.object) equivalent to the S-Plus ts.plot ?
--
-------------------------------------------------------------------------
Konstantinos E. Vorloou | Tel: +44 (0)191 374 1821
Department of Economics & Finance | Fax: +44 (0)191 374 7289
University of Durham, | email: K.E.Vorloou at durham.ac.uk
23/26 Old Elvet, | or : vorlow at rocketmail.com
1999 May 12
1
Beran's Stats for Long memory processes book code
Hello,
Has anybody transferred Bearn's S+ code to R ?
Best,
Costas
--
-------------------------------------------------------------------------
Konstantinos E. Vorloou | Tel: +44 (0)191 374 1821
Department of Economics & Finance | Fax: +44 (0)191 374 7289
University of Durham, | email: K.E.Vorloou at durham.ac.uk
23/26 Old Elvet, | or : vorlow at
2001 Mar 01
3
How do you expand memory capability (Was: R crashes in Windows ME)
Hello-
Since my data bank in SPSS has > 40 variables, I think that R crashes because of the memory limit.
In Maindonald?s UsingR text, on pg 3, there?s a footnote that reads:
"If you want larger memory space than the default you may want a target akin to
<path to binary>\rw091\bin\rgui.exe --visize 30M --nsize 1000K
[The default is --vsize 6M --nsize 250K
2000 Jan 23
1
size limits
Hi,
I have a few questions about how to handle large data sets in R.
What is the size of the largest matrix that R can comfortably deal with?
Is this size limit imposed by R's software, or is it a question
of the machine that one runs on?
How does one go about choosing reasonable values of vsize
and nsize?
I have a data set with about 1,000,000 rows, and 30
2010 Jan 19
2
Server hanging despite efforts to correct memory limits
My group is working with datasets between 100 Mb and 1 GB in size, using
multiple log ins. From the documentation, it appears that vsize is limited
to 2^30-1, which tends to prove too restrictive for our use. When we drop
that restriction (set vsize = NA) we end up hanging the server, which
requires a restart. Is there any way to increase the memory limits on R
while keeping our jobs from
2000 Aug 25
3
unexpected R crash - again
Sorry, but I lost this thread, so I sending this as a new message.
This is really a follow-up to a post from a couple days ago saying that
fisher.test from the ctest library crashed on the following data set:
> T
[,1] [,2]
[1,] 2 1
[2,] 2 1
[3,] 4 0
[4,] 8 0
[5,] 6 0
[6,] 0 0
[7,] 1 0
[8,] 1 1
[9,] 7 1
[10,] 8 2
[11,]
2001 Aug 22
1
Huge workspace cannot be opened
Hi everyone,
I have a problem that some people may have already encountered but i did not
find the solution yet.
As I use R to simulate several arrays of data, my workspace is now 35Mb big and
I cannot launch R with it.
An "xdr real data read error occured" and R tells me to delete .RData or
increase memory. I WON'T delete this file and changing the max-nsize to 40600k
did not
2000 Nov 09
3
maximum of nsize=20000k ??
Dear R-ers,
somehow it is not possible to increase nsize to more than
20000k. When I specify e.g.
> R --vsize=10M --nsize=21000K
the result is:
free total (Mb)
Ncells 99658 350000 6.7
Vcells 1219173 1310720 10.0
Maybe I have overlooked s.th....
Marcus
--
+-------------------------------------------------------
| Marcus Eger
| E-Mail: eger.m at gmx.de (NEW)
|
1999 May 15
2
vsize and nsize
I am running R version ??? under Redhat 5.2. It seems as though the
--nsize object has no effct on the size of the allocated Ncells as
determined using gc(). Yes, I have that much data....
That is if I envoke R with
R --vsize 100 --nsize 5000000
then type
gc()
I get
free total
Ncells 92202 200000
Vcells 12928414 13107200
Thanks
Tony Long
Ecology and Evolutionary Biology
Steinhaus
2009 Nov 30
1
allocating vector memory > 1 GByte on Windows XP / Vista / 7
Let me begin stating that I read all help files and faq's on the subject
matter (there aren't more than about a dozen) but either did not find
solutions or found them not to work.
Here is the issue. I am trying to run a spatial regression on a
medium-sized dataset. Part of the functions in the spdep package I use
require me to allocate a vector of 1.1 Gb (mine is not a spatial SIG
2009 May 07
1
increasing memory for R bg job
Hi,
Is the following command used to increase the memory or any other command when a background R job is run?
R --min-vsize=vl --max-vsize=vu --min-nsize=nl --max-nsize=nu --max-ppsize=N
source:
http://stat.ethz.ch/R-manual/R-patched/library/base/html/Memory.html
Thx
Carol
[[alternative HTML version deleted]]
2005 Dec 20
1
Problems in batch mode
Dear R-users,
I am trying to run some simulations in batch mode. In an older version
of the program, I used
rterm --vsize=100M --nsize=5000K --restore --save <input file> output file,
however, in the new version R 2.2.0 , the parameters vsize and nsize are
ignored.
I can use the command memory.limit to increase memory, but I am not sure if
this corresponds to vsize and nsize.
1999 Nov 12
1
R-0.65.1 Startup
Dear R users,
I have noticed that my R startup is extremely slow. It takes almost 3
minutes from "double-click" to R prompt. I have been running R-0.64.1 till
recently and it took about 30 sec. I still have access to R-0.64.1. When I
started it up, it took about 25 sec. Can anyone tell me if this is a bug in
R or a problem with my machine?
Note: This is after bootup with R being the
2004 Mar 08
2
memory problem
I am trying to upload into R 143 Affymetrix chips onto using R on the NIH
Nimbus server. I can load 10 chips without a problem, however, when I try
to load 143 I receive a error message: cannot create a vector of 523263 KB.
I have expanded the memory of R as follows: R --min-vsize=10M
--max-vsize=2500M --min-nsize=10M -max-nsize=50M (as specified in help in
R). After running this command the
2005 Jun 29
3
Memory Management under Linux: Problems to allocate large amounts of data
Dear Group
I'm still trying to bring many data into R (see older postings). After solving some troubles with the database I do most of the work in MySQL. But still I could be nice to work on some data using R. Therefore I can use a dedicated Server with Gentoo Linux as OS hosting only R. This Server is a nice machine with two CPU and 4GB RAM which should do the job:
Dual Intel XEON 3.06 GHz