Displaying 20 results from an estimated 10000 matches similar to: "extendable arrays in R"
2012 Jul 17
3
complexity of operations in R
Hello!
I am optimizing my code in R and for this I need to know a bit more about
the internals. It would help tremendously if someone could link me to a
page with O()-complexities of all the operations.
In this particular case, I need something like a linked list with O(1)
insertLast/First ability. I can't preallocate a vector since I do not know
the final size of the list ahead of time.
The
2019 Jan 15
2
preallocate working incorrectly in 3.1.3
I believe that the changes to support --preallocate and --sparse together
have broken --preallocate by itself (commit
f3873b3d88b61167b106e7b9227a20147f8f6197)
The previous behavior of --preallocate was to do just that: reserve blocks
in the filesystem WITHOUT setting the size of the file to the final
length. The reported filesize would change as the preallocated blocks were
actually written.
2008 Apr 09
1
preallocating matrices and rda read-back objects
I've read in Phil Spector's new book that it's a good idea to
preallocate a big matrix, like
u <- matrix(0,nrow,ncol) # (1)
Now, I read contents of a huge matrix from a Fortran binary dump.
u <- readBin(con,what="double",n=nrow*ncol) # (2)
If I do (1) and then (2), u is a vector, obviously it's either
reallocated or its matrix nature is lost -- overridden?
2006 Aug 06
2
File fragmentation
I've been running some tests on files created by rsync and noticing
fragmentation issues. I started the testing because our 5TB array started
performing very slowly and it appears fragmentation was the culprit. The
test I conducted was straighforward:
1. Copy over a 49GB file. Analyzed with contig (from sysinternals), no
fragments.
2. Ran rsync and the file was recreated normally (rsync
2008 Apr 03
1
combining --preallocate and --fuzzy
Greetings,
I would like to write a patch for rsync but need some help getting
started. Here is my situation. I am using cwrsync to copy files from
one Windows server to another Windows server. One file that I need
to backup is 130 GB. The daily changes occur all throughout the file,
not just at the end of the file. File names look like this:
Db_20080402_0003_DB.BAK
Db_20080403_0003_DB.BAK
2018 Mar 04
2
file contents cause rsync to fail (with certains args and dir structure)
script to reproduce:
#!/bin/bash
#tested to fail as below: ArchLinux's rsync-3.1.3-1-x86_64.pkg.tar.xz
#tested to fail as below: ArchLinux's rsync-3.1.3pre1-1-x86_64.pkg.tar.xz
#tested to work ok : ArchLinux's rsync-3.1.2-8-x86_64.pkg.tar.xz
if test "$1" == "clean"; then
rm -vrf destdir sourcedir sourcedir2 sourcedir3
exit 0
fi
echo '!! test 1:'
2007 Nov 05
13
preallocate CPU usage - pre4
When I use the preallocate patch and create a 77GB file using the function I
get a CPU spike on the server-side. The spike lasts about 20 minutes and
uses about 20%-25% of the cpu associated with the rsync instance creating
the file. The spike is directly linked to the time it takes to create the
file.
I compiled rsync using cygwin CVS. I initially suspected the implementation
of
2012 Feb 23
1
[Bug 8775] New: Preallocate option doesn't work with recursive?
https://bugzilla.samba.org/show_bug.cgi?id=8775
Summary: Preallocate option doesn't work with recursive?
Product: rsync
Version: 3.1.0
Platform: x64
OS/Version: Linux
Status: NEW
Severity: normal
Priority: P5
Component: core
AssignedTo: wayned at samba.org
ReportedBy: wgordonharris at
2010 Feb 12
1
[RFC] add support for fallocate()
fallocate() is linux specific and will preallocate the space on disk for
the entire file. FALLOC_FL_KEEP_SIZE does not change the filesize as
reported by stat(). An aborted transfer will have preallocated disk space
which is not "visible" via stat(). This shouldn't matter unless the user
does complet his transfer.
An alternative would be to use ftruncate() and shorten the file to the
2003 Jul 30
6
reverse array indexing
Hi,
Suppose I have a multidimensional array:
tmp <- array(1:8, c(2,2,2))
is there a function out there that, given a one-dimensional array index,
will
return the separate indices for each array dimension?
for instance, tmp[8] is equivalent to tmp[2,2,2]. I'd like to derive the
vector (2,2,2)
from the index 8.
thanks,
Brad Buchsbaum
2020 Feb 14
3
RFC Storing BB order in llvm::Instruction for faster local dominance
Hello again. :)
There has been renewed interest in having instructions track their own
order in basic blocks to help make dominance queries fast. I have a very
simple naive implementation of this here:
https://reviews.llvm.org/D51664
Essentially, every instruction will carry an integer order number, and
inserting new instructions invalidates the ordering. I know there are
better algorithms for
2010 Oct 20
1
minimize mbox mdbox fragmentation
Hi Timo,
Any chance the mbox/mdbox writer code could be modified to do physical
preallocation on files to help avoid file(system) fragmentation?
Constantly appending a file is the prime recipe for causing
fragmentation, and mbox is notorious for this--not a fault of Dovecot
but the nature of the mbox beast. Obviously maildir doesn't have such a
problem, but some (many?) of us still prefer
2018 Mar 05
5
[Bug 13320] New: file contents cause rsync to fail (with certains args and dir structure)
https://bugzilla.samba.org/show_bug.cgi?id=13320
Bug ID: 13320
Summary: file contents cause rsync to fail (with certains args
and dir structure)
Product: rsync
Version: 3.1.3
Hardware: All
OS: All
Status: NEW
Severity: normal
Priority: P5
Component: core
2018 Mar 05
1
file contents cause rsync to fail (with certains args and dir structure)
Problem was introduced with this commit:
commit f3873b3d88b61167b106e7b9227a20147f8f6197
Author: Wayne Davison <wayned at samba.org>
Date: Mon Oct 10 11:49:50 2016 -0700
Support --sparse combined with --preallocate or --inplace.
The new code tries to punch holes in the destination file using newer
Linux fallocate features. It also supports a --whole-file
2015 Nov 03
26
[Bug 11588] New: missing option: preallocate for all files except for sparse
https://bugzilla.samba.org/show_bug.cgi?id=11588
Bug ID: 11588
Summary: missing option: preallocate for all files except for
sparse
Product: rsync
Version: 3.1.2
Hardware: x64
OS: Linux
Status: NEW
Severity: enhancement
Priority: P5
Component: core
2010 Jan 05
2
Fast nested List->data.frame
I have very large data sets given in a format similar to d below. Converting
these to a data frame is a bottleneck in my application. My fastest version
is given below, but it look clumsy to me.
Any ideas?
Dieter
# -----------------------
len = 100000
d = replicate(len, list(pH = 3,marker = TRUE,position = "A"),FALSE)
# Data are given as d
# preallocate vectors
pH =rep(0,len)
marker
2009 Feb 12
3
Looping multiple output values to dataframe
Dear R users,
I have various vector geometry operations to perform on 3-D coordinate data
located on multiple (500+) csv files. The code I have written for the
calculations works just fine. I have written a 'for' loop to automate the
task of extracting the coordinates from the files and perform the analyses.
The loop works reasonable well, but if the number of csv files is greater
than
2023 Oct 14
1
Create new data frame with conditional sums
That's very helpful and instructive, thank you!
Jason Stout, MD, MHS
Box 102359-DUMC
Durham, NC 27710
FAX 919-681-7494
________________________________
From: John Fox <jfox at mcmaster.ca>
Sent: Saturday, October 14, 2023 10:13 AM
To: Jason Stout, M.D. <jason.stout at duke.edu>
Cc: r-help at r-project.org <r-help at r-project.org>
Subject: Re: [R] Create new data frame with
2023 Oct 15
2
Create new data frame with conditional sums
Under the hood, sapply() is also a loop (at the interpreted level). As
is lapply(), etc.
-- Bert
On Sun, Oct 15, 2023 at 2:34?AM Jason Stout, M.D. <jason.stout at duke.edu> wrote:
>
> That's very helpful and instructive, thank you!
>
> Jason Stout, MD, MHS
> Box 102359-DUMC
> Durham, NC 27710
> FAX 919-681-7494
> ________________________________
> From: John
2009 Jun 12
2
Automate a data load and merge
Hi R list,
I would like to automate, or speed up the process from which I take
several separate datasets, stored in .csv formate, import and merge
them by a common variable. So far I have greatly sped up the loading
process but cannot think of a way to automate the merger of all
datasets into a common data.frame.
My apologies if this has been covered, any R search suggestions are