Displaying 20 results from an estimated 3000 matches similar to: "--nsize and --vsize"
2015 Jan 15
2
default min-v/nsize parameters
Just wanted to start a discussion on whether R could ship with more
appropriate GC parameters. Right now, loading the recommended package
Matrix leads to:
> library(Matrix)
> gc()
used (Mb) gc trigger (Mb) max used (Mb)
Ncells 1076796 57.6 1368491 73.1 1198505 64.1
Vcells 1671329 12.8 2685683 20.5 1932418 14.8
Results may vary, but here R needed 64MB of N cells and 15MB
2015 Jan 18
2
default min-v/nsize parameters
On Thu, Jan 15, 2015 at 3:55 PM, Michael Lawrence
<lawrence.michael at gene.com> wrote:
> Just wanted to start a discussion on whether R could ship with more
> appropriate GC parameters.
I've been doing a number of similar measurements, and have come to the
same conclusion. R is currently very conservative about memory usage,
and this leads to unnecessarily poor performance on
2015 Jan 20
1
default min-v/nsize parameters
>>>>> Peter Haverty <haverty.peter at gene.com>
>>>>> on Mon, 19 Jan 2015 08:50:08 -0800 writes:
> Hi All, This is a very important issue. It would be very
> sad to leave most users unaware of a free speedup of this
> size. These options don't appear in the R --help
> output. They really should be added there.
Indeed,
1999 Dec 17
1
R CMD check --help
This example from the INSTALL help seems to be broken in R 0.90.1 (on Solaris):
gilp/dse : R CMD check --help
Usage: R CMD check [options] [-l lib] pkg_1 ... pkg_n
I'm trying to figure out how to request more nsize and vsize when using R CMD
check.
Paul Gilbert
-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-
r-devel mailing list -- Read
1999 Sep 24
2
R's startup : .Rprofile & .Renviron -- info and RFC
[RFC = Request for Comments]
{ Yes, the documentation for .Renviron is really not there (but the FAQ...);
the rest is in ?Startup }
In R's Startup (on Unix only??) {unless --no-environ is specified}
~/.Renviron (if there) is read as an 'sh' script before R is called,
then R looks ((for the site-wide Rprofile and then))
for .Rprofile in the current directory and then for
2000 Mar 13
1
check does not accept --vsize option (PR#481)
Full_Name: Markus Neteler
Version: 1.0.0
OS: Linux 2.2.10/i686
Submission from: (NULL) (130.75.72.37)
Hi,
I wanted to "check" the R.GRASS GIS interface from Roger Bivand:
http://www.geog.uni-hannover.de/grass/statsgrasslist.html
using
R CMD check --vsize=10M GRASS
but:
[error message shortened]
> G <- gmeta()
Error: heap memory (6144 Kb) exhausted [needed 1024 Kb more]
1999 May 15
2
vsize and nsize
I am running R version ??? under Redhat 5.2. It seems as though the
--nsize object has no effct on the size of the allocated Ncells as
determined using gc(). Yes, I have that much data....
That is if I envoke R with
R --vsize 100 --nsize 5000000
then type
gc()
I get
free total
Ncells 92202 200000
Vcells 12928414 13107200
Thanks
Tony Long
Ecology and Evolutionary Biology
Steinhaus
2004 Jul 20
1
--max-vsize and --max-nsize linux?
Hi,
somtimes i have trivial recodings like this:
> dim(tt)
[1] 252382 98
system.time(for(i in 2:length(tt)){
tt[,i][is.na(tt[,i])] <- 0
})
...and a win2000(XP2000+,1GB) machine makes it in several minutes, but
my linux notebook (XP2.6GHZ,512MB) don't get success after some hours.
I recognize that the cpu load is most time relative small, but the hardisk
1999 Jul 08
1
Gnome interface status report
Hi,
The Gnome version now compiles, and it should also be working (at least as much as it
ever has). I've changed Makefile.in to the new system, which is very cool. What I
want to work on now is:
- Graphics. I want to move to the Gnome canvas for this, which should be reasonably
easy. This will give us rotated text (which I never got going properly before) and
the option for
2015 Jan 19
0
default min-v/nsize parameters
Hi All,
This is a very important issue. It would be very sad to leave most users
unaware of a free speedup of this size. These options don't appear in the
R --help output. They really should be added there. Additionally, if the
garbage collector is working very hard, might it emit a note about better
setting for these variables?
It's not really my place to comment on design philosophy,
1999 Aug 30
1
interface w/ emacs (PR#261)
Full_Name: Laurent Gautier
Version: 0.65.0
OS: mips SGI-Irix 6.5
Submission from: (NULL) (195.110.4.98)
Using R through emacs with ess5.1.8, I cannot set R workspace (--vsize and
--nsize).
So far I was using R0.64.2 without such a problem. I am aware my bug report is a
bit light,
but just let me know if anything I could do with my R and emacs would be of any
help for\
specifying better what is
2000 Nov 09
3
maximum of nsize=20000k ??
Dear R-ers,
somehow it is not possible to increase nsize to more than
20000k. When I specify e.g.
> R --vsize=10M --nsize=21000K
the result is:
free total (Mb)
Ncells 99658 350000 6.7
Vcells 1219173 1310720 10.0
Maybe I have overlooked s.th....
Marcus
--
+-------------------------------------------------------
| Marcus Eger
| E-Mail: eger.m at gmx.de (NEW)
|
1999 Jan 27
1
cant restore .Rdata
Hi Folks,
I loaded a couple of quite large data sets into an R session and then
quit (after saving the image). Now I get:
Error: a read error occured
Fatal error: unable to restore saved data
(remove .RData or increase memory)
after trying to start my R session using something like:
R --vsize XXX --nsize 1000000
For any value of XXX (I went up to 300 or 400, which is as high as I could
go.
2001 Jan 07
2
"Invalid character 32" problem with Pager??
Hi all,
I'm having a very minor problem. Most every time I use the pager (I think) I
get the following error about character 32. It works fine, but the error is
a bit annoying.
> ?version
sh: invalid character 32 in exportstr for export R_NSIZE
sh: invalid character 32 in exportstr for export R_VSIZE
R.Version package:base R Documentation
(and then the
1999 Oct 06
2
R --nsize 2M runs havoc (under linux)
Dear All,
I am running R version 0.65.0 under
a) Suse-Linux 6.1, and Suse-Linux 6.2, compiler gcc-2.95, CPUs pentium pro
200, 128MB, and pentium II 450, 128MB
b) Solaris 5.7, compiler gcc-2.95, cpu SUN sparc, 4000MB
When I set --nsize to more than 1M, R's internal storage management runs
havoc. gc() indicates the requested sizes, but the overall process size is
much too big: Running R with
2000 May 30
6
heap size trouble
Hi ,
I ''ve got a trouble with using R.
When I want to load a file that contains 93 thousand raws and 22 colums
of data (essentially float)
R shows me this error message
"heap size trouble"
Does anyone could tell me what parameter shall I precise before
launching R in order to load my big file.
Thanks a lot
-------------- next part --------------
A non-text attachment was
2001 Aug 22
1
Huge workspace cannot be opened
Hi everyone,
I have a problem that some people may have already encountered but i did not
find the solution yet.
As I use R to simulate several arrays of data, my workspace is now 35Mb big and
I cannot launch R with it.
An "xdr real data read error occured" and R tells me to delete .RData or
increase memory. I WON'T delete this file and changing the max-nsize to 40600k
did not
2001 Jan 14
2
Help
Dear sir,
I am using R in windows. I want to extend R Memory
size.
I use the following command, but unfortunately it
doesn't work.
-- vsize=15M --nsize=1000K
Your help is appreciated.
Thanks,
Esmail Amiri.
__________________________________________________
Do You Yahoo!?
Get email at your own domain with Yahoo! Mail.
http://personal.mail.yahoo.com/
1998 Aug 22
1
R-beta: re -n -v wr0613b - windows dynload
When I use the -v I can modify the size of the heap, as assessed by
gc(), but the -n key seems to be without effect ?
On a machine with 48 mB ram I can load the libraries without problem,
but on my own 36 mB ram machine, I get dynload problems with the larger,
eg survival4.
Any suggestions ?
Troels
-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-
r-help mailing list
1998 Aug 22
1
R-beta: re -n -v wr0613b - windows dynload
When I use the -v I can modify the size of the heap, as assessed by
gc(), but the -n key seems to be without effect ?
On a machine with 48 mB ram I can load the libraries without problem,
but on my own 36 mB ram machine, I get dynload problems with the larger,
eg survival4.
Any suggestions ?
Troels
-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-
r-help mailing list