Displaying 20 results from an estimated 4000 matches similar to: "Memory limit for Windows 64bit build of R"
2006 Mar 31
2
Small local dir refuses to sync, others AOK.
Hi guys,
I'm having a problem rsyncing between two directories on local disk.
I wish to sync the local directory /local/CA/eTrustDirectory/dxserver/ldif
to the other local directory /local/CA/archive.
'rsync -vvvr --progress /local/CA/eTrustDirectory/dxserver/ldif
/local/CA/archive' pauses at:
make_file(4,ldif/MELLDRBIB11-ldsau004wm1/ldbau004v1_20060330.0204.raw)
2005 Mar 18
4
Sweave/margin
Hi!
I am currently using Sweave for writing my bachelor thesis - and I have a
problem:
I am using a LaTeX style (report) with quite big margin spaces. The Sweave
generated LaTeX code "floats" into the margin - and it looks ugly. The text
is blocked and fine... then there comes some flattering code running over
the margin... and blocked text again.
Considering the LaTeX output, I guess
2011 Aug 27
3
Exception while using NeweyWest function with doMC
Dear R users,
I am using R right now for a simulation of a model that needs a lot of
memory. Therefore I use the *bigmemory* package and - to make it faster -
the *doMC* package. See my code posted on http://pastebin.com/dFRGdNrG
Now, if I use the foreach loop with the addon %do% (for sequential run) I
have no problems at all - only here and there some singularities in
regressor matrices which
2011 Nov 30
1
Error adding Bigmemory package
I am trying to add the bigmemory packages but I get the following error
message:
In file included from bigmemory.cpp:14:0:
../inst/include/bigmemory/isna.hpp: In function 'bool neginf(double)':
../inst/include/bigmemory/isna.hpp:22:57: error: 'isinf' was not declared in
this scope
make: *** [bigmemory.o] Error 1
ERROR: compilation failed for package 'bigmemory'
* removing
2011 Jun 24
1
Installation of bigmemory fails
Hello All,
I tried to intall the bigmemory package from a CRAN mirror site and
received the following output while installing. Any idea what's going
on and how to fix it? The system details are provided below.
--------------------- begin error messages -----------------------
* installing *source* package 'bigmemory' ...
checking for Sun Studio compiler...no
checking for
2013 Apr 29
2
bigmemory and R 3.0
Dear helpers,
Does anyone have information on the status of bigmemory and R3.0? Will it
just take time for the devs to re-code for the new environment? Or is there
an alternative for this new version?
Thanks
Ben Caldwell
[[alternative HTML version deleted]]
2011 Sep 02
2
Advice on large data structures
Friends
I am starting on a (section of the) project where I need to build a matrix
with on the order of 5 million rows and 200 columns
I am wondering if I can stay in R.
I need to do rollapply type operations on the columns, including some that
will be functions of (windows of) two columns.
I have been looking at the ff and bigmemory packages but am not sure that
they will do.
Before I get
2010 Aug 11
1
Bigmemory: Error Running Example
Hi,
I am trying to run the bigmemory example provided on the
http://www.bigmemory.org/
The example runs on the "airline data" and generates summary of the csv
files:-
library(bigmemory)
library(biganalytics)
x <- read.big.matrix("2005.csv", type="integer", header=TRUE,
backingfile="airline.bin",
descriptorfile="airline.desc",
2009 Jul 20
2
kmeans.big.matrix
Hi,
I'm playing around with the 'bigmemory' package, and I have finally
managed to create some really big matrices. However, only now I
realize that there may not be functions made for what I want to do
with the matrices...
I would like to perform a cluster analysis based on a big.matrix.
Googling around I have found indications that a certain
kmeans.big.matrix() function should
2010 Mar 11
2
ANNOUNCE--Rdsm package, a threads-like environment for R
My long-promised Rdsm package is now on CRAN. Some of you may recall
that I made a prototype available on my own Web page last July. This is
the official version, much evolved since I released the prototype.
The CRAN description states:
Provides a threads-like programming environment for R, usable both on
a multicore machine and across a network of multiple machines. The
package
2010 Feb 22
1
big panel: filehash, bigmemory or other
Dear R-list
I'm on my way to start a new project on a rather big panel, consisting
of approximately 8 million observations in 30 waves of data and about
15 variables. I have a similar data set that is approximately 7
gigabytes in size.
Until now I have done my data management in SAS, and Stata, mostly
identifying spells, counting events in intervals, and a like, but I
would like to
2012 May 05
2
looking for adice on bigmemory framework with C++ and java interoperability
I work with problems that have rather large data requirements -- typically
a bunch of multigig arrays. Given how generous R is with using memory, the
only way for me to work with R has been to use bigmatrices from bigmemory
package. One thing that is missing a bit is interoperability of bigmatrices
with C++ and possibly java. What i mean by that is API that would allow
read and write filebacked
2010 Sep 08
1
bigmemory doubt
Hi,
Is it possible for me to read data from shared memory created by a vc++ program into R using bigmemory?
[[alternative HTML version deleted]]
2010 Apr 23
2
bigmemory package woes
I have pretty big data sizes, like matrices of .5 to 1.5GB so once i need to
juggle several of them i am in need of disk cache. I am trying to use
bigmemory package but getting problems that are hard to understand. I am
getting seg faults and machine just hanging. I work by the way on Red Hat
Linux, 64 bit R version 10.
Simplest problem is just saving matrices. When i do something like
2012 May 11
1
bigmemory
To answer your first question about read.big.matrix(), we don't know what
your acc3.dat file is, but it doesn't appear to have been detected as a
standard file (like a CSV file) or -- perhaps -- doesn't even exist (or
doesn't exist in your current directory)?
Next:
> In addition, I am planning to do a multiple imputation with MICE package
> using the data read by bigmemory
2019 Jul 19
1
difficulty with sanitizer using bigmemory
Dear all,
bigKRLS, which has been on CRAN for a couple of years, had to be pulled
recently due to what seems to be a sanitizer issue stemming from its use of
bigmemory. bigKRLS works fine (we?ve used it ourselves on many different
platforms and have had over 15,000 downloads without an end user reporting
difficulties because of this issue). Unfortunately, we have been unable to
reproduce the
2009 Sep 11
3
Working with large matrix
Dear All,
I have large matrix (46000 x 11250). I would like to do the linear regression for each row. I wrote a simple function that has lm() and used apply(mat,1,func). The issue is that it takes ages to load the file and also to finish the lm. I am using LINUX 64 bit with 32G mem. Is there an elegant and fast way of completing this task?
Thanks in advance.
Kind regards,
Ezhil
2015 May 13
4
example fails during R CMD CHECK but works interactively?
Thank you Dan but it isn't my tests that are failing (all of them pass
without problem) but one of the examples from the inst/examples directory.
I did try, however, to start R with the environmental variables as you
suggest but it had no effect on my tests.
Charles
On Wed, May 13, 2015 at 1:51 PM, Dan Tenenbaum <dtenenba at fredhutch.org>
wrote:
>
>
> ----- Original Message
2010 Dec 17
1
[Fwd: adding more columns in big.matrix object of bigmemory package]
Hi,
With reference to the mail below, I have large datasets, coming from various
different sources, which I can read into filebacked big.matrix using library
bigmemory. I want to merge them all into one 'big.matrix' object. (Later, I
want to run regression using library 'biglm').
I am unsuccessfully trying to do this from quite some time now. Can you
please
2012 Oct 18
3
bigmemory for dataframes?
Hi Folks,
I've been bumping my head against the 4GB limit for 32-bit R. I can't
go to 64-bit R due to package compatibility issues (ROBDC - possible but
painful, xlsReadWrite - not possible, and others). I have a number of
big dataframes whose columns all sorts of data types - factor,
character, integer, etc. I run and save models that keep copies of the
modeled data inside the model