similar to: R memory usage and size limits

Displaying 20 results from an estimated 30000 matches similar to: "R memory usage and size limits"

2016 Jan 22
3
Return options used to configure R
So in English what you are saying is that there's a file called Makeconf in the etc directory under the R_HOME that contains the information. That certainly seems to be true for Linux. Thanks. -----Original Message----- From: G?bor Cs?rdi [mailto:csardi.gabor at gmail.com] Sent: 22 January 2016 11:37 To: Tom Quarendon <tom.quarendon at teamwpc.co.uk> Cc: r-devel at r-project.org
2016 Jan 22
2
Return options used to configure R
I did have a search, but searching for "configure" just brought up loads of results for how to configure R. I'm sure it is in the manual, once you know where to look for it and if you know all the manuals inside out. Looking at the help for R CMD config, it's not actually clear that these relate to the options specified on configure, and it doesn't tell you things like
2016 Jan 21
2
Return options used to configure R
Is there any way to print out the options that where passed to configure when R was built? I want to build a later version to the one I have on my machine, and I want to build it in the same way, with the same configure options. Thanks.
2007 Jan 04
2
memory limits in R loading a dataset and using the package tree
? stato filtrato un testo allegato il cui set di caratteri non era indicato... Nome: non disponibile Url: https://stat.ethz.ch/pipermail/r-help/attachments/20070104/6e94ce08/attachment.pl
2023 Apr 04
6
on lexical scoping....
Dear Members, I have the following code typed at the console prompt: y <- x*10 X has not been defined and the above code throws an object not found error. That is, the global environment does not contain x. Why doesn't it look further in the environment stack, like that of packages? There are thousands of packages that contain the variable named x. Of
2008 Jun 05
1
memory.size() for large memory usage (PR#11596)
This amusing behaviour is from R2.6.1 on WinXP Pro SP2 running with boot.in= i /3GB flag, exploring memory limits. =20 When actual (object, not total) memory usage hits 2048 MB, memory.size star= ts counting down again, but reports a negative amount, e.g. -2046.333. Is t= his an intended (or unavoidable) feature?) =20 Thanks, =20 Rory Bowden bowden at stats.ox.ac.uk =20 [[alternative
2007 Jan 05
0
memory limits in R loading a dataset and using the packagetree
I agree about sampling, but.. You can go a little further with your hardware. The defaults in R is to play nice and limit your allocation to half the available RAM. Make sure you have a lot of disk swap space (at least 1G with 2G of RAM) and you can set your memory limit to 2G for R. See help(memory.size) and use the memory.limit function Hugues P.s. Someone let me use their 16Gig of RAM
2006 May 11
0
Echo Canceller Memory Usage, Frame Size
(from thread Re: [Speex-dev] Speex echo canceller on TI C55 DSP, but this is a more general topic) >> With these changes, the link was successful, using testecho.c with some >> modifications for the C55 environment. The code and data memory >> requirements were a lot more than I had hoped (>20kbytes of dynamic data >> memory for block size=128, tail length = 1024),
2023 Apr 04
2
on lexical scoping....
The following *might* be of use to you. If you can predict what the various function invocations will do, I think you have a reasonable grasp of how lexical scoping works in R (contrary or supplementary opinions welcome). It is the sort of thing you will find in the references also. If this is all obvious, sorry for wasting your time. ####################### search() ls() dat <- list(x =2)
2013 Feb 28
1
rsync memory usage with block-size option
I have been using rsync to sync a SQL database dump(127GB in size) on a server (cygwin to linux) for many years now w/o a problem, until they put a maintence plan to reindex the database once a week. This is all done over a 2Mbit link normally took 3-4 hrs now I noticed it was running the next day and took days so I looked into it more. After I discovered the --only-write-batch it makes a 52 GB
2003 Mar 31
2
point-biserial correlation
Dear list, has anyone written a package/function in R for computing a point- biserial resp. biserial correlation? Thanks in advance Bernd
2010 Jan 19
2
Server hanging despite efforts to correct memory limits
My group is working with datasets between 100 Mb and 1 GB in size, using multiple log ins. From the documentation, it appears that vsize is limited to 2^30-1, which tends to prove too restrictive for our use. When we drop that restriction (set vsize = NA) we end up hanging the server, which requires a restart. Is there any way to increase the memory limits on R while keeping our jobs from
2006 Mar 08
1
memory limits in Windows
Hello, I apologize, since I know that variations of this question come up often, but I have searched both the archives and the rw-FAQ, and haven't had any luck solving my problem. I am trying to run some generalized additive mixed models (using the mgcv library) with some large datasets, and get error messages regarding memory allocation. An example dataset has around 45,000 points
2010 May 07
2
ZFS root ARC memory usage on VxFS system...
Hi Folks.. We have started to convert our Veritas clustered systems over to ZFS root to take advantage of the extreme simplification of using Live Upgrade. Moving the data of these systems off VxVM and VxFS is not in scope for reasons to numerous to go into.. One thing my customers noticed immediately was a reduction in "free" memory as reported by ''top''. By way
2013 Aug 15
0
Quota to use limits from LDAP and usage from FS
Hi, I'm new to Dovecot (switching from Cyrus). I'm using Dovecot 2.1.7 from Ubuntu 13.04. Maildir data is stored on a Linux ext4 filesystem with quotas. I would like to use mail quota in a way, so that the usage values are taken from the filesystem's quota system by the "fs" backend, while the per-user limits get read from LDAP. Due to debugging messages, it's my
2007 Mar 14
5
aaf batch_size limits indexing on mssql to 1000 records
hi! after wondering why i can''t find alot of records i eventually found the problem in the sqlserver_adapters implementation of "add_limit_offset!". the problem is when using MSSQL with the sqlserver_adapter paging will only work when at least one column is defined in ":order". for example i was indexing a table with 2912 records, the generated sql for the batches
2023 Apr 04
2
on lexical scoping....
obviously, everyone has different opinions on what's useful but I always found this document quite helpful. I think, in the past, someone said that there are some incorrect statements in but I'm not sure what they are. https://askming.github.io/study_notes/Stats_Comp/Note-How%20R%20searches%20and%20finds%20stuff.html On Tue, Apr 4, 2023 at 7:06?PM Bert Gunter <bgunter.4567 at
2008 Aug 12
1
[LLVMdev] Using JIT to construct an invocation of a non-JIT function, or how do I interface to GCC exception handling
I'm just starting to have a look at LLVM, so forgive me if this is a simple question. What I'd like to do is use the JIT api to construct a "program" that calls a series of functions. In other words I'm wanting to translate a scripting language into some executable assembler where each primitive of my scripting language is implemented with a "normal" function.
2023 Apr 05
1
on lexical scoping....
R *does* search the environment stack. > search() [1] ".GlobalEnv" "package:stats" "package:graphics" [4] "package:grDevices" "package:utils" "package:datasets" [7] "package:methods" "Autoloads" "package:base What you seem to be missing is that a package may contain bindings that it
2012 Mar 23
1
Memory limits for MDSplot in randomForest package
Hello, I am struggling to produce an MDS plot using the randomForest package with a moderately large data set. My data set has one categorical response variables, 7 predictor variables and just under 19000 observations. That means my proximity matrix is approximately 133000 by 133000 which is quite large. To train a random forest on this large a dataset I have to use my institutions high