Displaying 20 results from an estimated 20 matches for "filebackings".
Did you mean:
filebacking
2009 Jun 02
2
bigmemory - extracting submatrix from big.matrix object
I am using the library(bigmemory) to handle large datasets, say 1 GB,
and facing following problems. Any hints from anybody can be helpful.
_Problem-1:
_
I am using "read.big.matrix" function to create a filebacked big matrix
of my data and get the following warning:
> x =
read.big.matrix("/home/utkarsh.s/data.csv",header=T,type="double",shared=T,backingfile
2010 Dec 17
1
[Fwd: adding more columns in big.matrix object of bigmemory package]
Hi,
With reference to the mail below, I have large datasets, coming from various
different sources, which I can read into filebacked big.matrix using library
bigmemory. I want to merge them all into one 'big.matrix' object. (Later, I
want to run regression using library 'biglm').
I am unsuccessfully trying to do this from quite some time now. Can you
please
2012 Jun 19
1
how to manipulate dput output format
I am reading into Java dput output for a matrix, more specifically for a
file backed big-matrix. I basically need to lift dimnames for a matrix from
dput output. It's no big deal, but the code is very 'hackish' due to the
need to get rid of quotes, endlines, parenthesis, etc. I was wondering if i
could manipulate to an extent dput output with some options that define it,
for example,
2010 Aug 11
1
Bigmemory: Error Running Example
Hi,
I am trying to run the bigmemory example provided on the
http://www.bigmemory.org/
The example runs on the "airline data" and generates summary of the csv
files:-
library(bigmemory)
library(biganalytics)
x <- read.big.matrix("2005.csv", type="integer", header=TRUE,
backingfile="airline.bin",
descriptorfile="airline.desc",
2012 May 05
2
looking for adice on bigmemory framework with C++ and java interoperability
I work with problems that have rather large data requirements -- typically
a bunch of multigig arrays. Given how generous R is with using memory, the
only way for me to work with R has been to use bigmatrices from bigmemory
package. One thing that is missing a bit is interoperability of bigmatrices
with C++ and possibly java. What i mean by that is API that would allow
read and write filebacked
2011 Sep 29
1
efficient coding with foreach and bigmemory
I recently learned about the bigmemory and foreach packages and am trying
to use them to help me create a very large matrix. Without those
packages, I can create the type of matrix that I want with 10 columns and
5e6 rows. I would like to be able to scale up to 5e9 rows, or more, if
possible.
I have created a simplified example of what I'm trying to do, below. The
first part of the
2010 Oct 12
0
merging and working with big data sets
...big.matrix of the proper size.
- Fill it in chunks (typically a column at a time if you can afford
the RAM overhead, or a portion of a column at a time). Column
operations are more efficient than row operations (again, because of
the internals of the data structure).
- Because you'll be using filebackings, RAM limitations won't matter
other than the overhead of copying each chunk.
I should note: if you used separated=TRUE, each column would have a
separate binary file, and a "smart" cbind() would be possible simply
by manipulating the descriptor file. Again, not something we advise
o...
2012 Jan 02
0
Reading mcmc/coda into a big.matrix efficiently
I'm trying to read CODA/mcmc files (see the coda package), as
generated by jags/WinBUGS/OpenBUGS, into a big.matrix. I can't load
the whole mcmc object produced by read.coda() into memory since I'm
using a laptop for this analysis (currently I'm unfunded).
Right now I'm doing it by creating the filebacked.big.matrix, reading
a chunk of data at a time from the chain
1999 Oct 29
4
SMBMount and backup application
I would like some advice. I am considering the best way to restructure my
backup scheme on my NT network. I would like to see if it would be
possible/feasible to use cron, smbmount, some SCSI tape drives, and some type
of archive tool to backup my NT shares to a RedHat box using Samba. What do
all of you do out there? I would envision a situation where each night a
script is run by cron that
2010 Apr 23
2
bigmemory package woes
I have pretty big data sizes, like matrices of .5 to 1.5GB so once i need to
juggle several of them i am in need of disk cache. I am trying to use
bigmemory package but getting problems that are hard to understand. I am
getting seg faults and machine just hanging. I work by the way on Red Hat
Linux, 64 bit R version 10.
Simplest problem is just saving matrices. When i do something like
2009 Feb 25
3
Using very large matrix
Dear friends,
I have to use a very large matrix. Something of the sort of
matrix(80000,80000,n) .... where n is something numeric of the sort 0.xxxxxx
I have not found a way of doing it. I keep getting the error
Error in matrix(nrow = 80000, ncol = 80000, 0.2) : too many elements specified
Any suggestions? I have searched the mailing list, but to no avail.
Best,
--
Corrado Topi
Global
2019 Jul 19
1
difficulty with sanitizer using bigmemory
Dear all,
bigKRLS, which has been on CRAN for a couple of years, had to be pulled
recently due to what seems to be a sanitizer issue stemming from its use of
bigmemory. bigKRLS works fine (we?ve used it ourselves on many different
platforms and have had over 15,000 downloads without an end user reporting
difficulties because of this issue). Unfortunately, we have been unable to
reproduce the
2009 Apr 16
0
Major bigmemory revision released.
The re-engineered bigmemory package is now available (Version 3.5
and above) on CRAN. We strongly recommend you cease using
the older versions at this point.
bigmemory now offers completely platform-independent support for
the big.matrix class in shared memory and, optionally, as filebacked
matrices for larger-than-RAM applications. We're working on updating
the package vignette, and a
2009 Apr 16
0
Major bigmemory revision released.
The re-engineered bigmemory package is now available (Version 3.5
and above) on CRAN. We strongly recommend you cease using
the older versions at this point.
bigmemory now offers completely platform-independent support for
the big.matrix class in shared memory and, optionally, as filebacked
matrices for larger-than-RAM applications. We're working on updating
the package vignette, and a
2009 Oct 30
1
Multicore package: sharing/modifying variable accross processes
Hi,
I want to parallelize some computations when it's possible on multicore
machines.
Each computation produces a big objects that I don't want to store if
not necessary: in the end only the object that best fits my data have to
be returned. In non-parallel mode, a single gloabl object is updated if
the current computation gets a better result than the best previously found.
My plan
2010 Dec 16
0
adding more columns in big.matrix object of bigmemory package
Hi all,
Is there any way I can add more columns to an existing filebacked big.matrix
object.
In general, I want a way to modify an existing big.matrix object, i.e., add
rows/columns, rename colnames, etc.
I tried the following:
> library(bigmemory)
> x =
read.big.matrix("test.csv",header=T,type="double",shared=T,backingfile="test
2009 Jul 20
2
kmeans.big.matrix
Hi,
I'm playing around with the 'bigmemory' package, and I have finally
managed to create some really big matrices. However, only now I
realize that there may not be functions made for what I want to do
with the matrices...
I would like to perform a cluster analysis based on a big.matrix.
Googling around I have found indications that a certain
kmeans.big.matrix() function should
2012 May 09
2
ergm model, nodematch with diff=T
Dear all,
I am new to network analysis, but since I have good data I started to read
about it and learned how to use the ergm and related packages. I generally
get interesting results, but when I run a model including sociality and
selective mixing effects for different groups, the model runs (and
converges) but I get a warning as follows:
mod <- ergm(network ~ edges + gwesp(0, fixed=T) +
2009 Sep 04
3
asking for suggestions: interface for a C++ class
Dear All,
I would like to have an advice for designing an R library, and thought
that R-devel may be the best place to ask given so many people who are
highly expert in R are around.
We are at an early stage of designing an R library, which is
effectively an interface to a C++ library providing fast access to
large matrices stored on HDD as binary files. The core of the C++
library is
2011 Aug 27
3
Exception while using NeweyWest function with doMC
Dear R users,
I am using R right now for a simulation of a model that needs a lot of
memory. Therefore I use the *bigmemory* package and - to make it faster -
the *doMC* package. See my code posted on http://pastebin.com/dFRGdNrG
Now, if I use the foreach loop with the addon %do% (for sequential run) I
have no problems at all - only here and there some singularities in
regressor matrices which