similar to: R Bug: write.table for matrix of more than 2, 147, 483, 648 elements

Displaying 20 results from an estimated 500 matches similar to: "R Bug: write.table for matrix of more than 2, 147, 483, 648 elements"

2018 Apr 19
3
R Bug: write.table for matrix of more than 2, 147, 483, 648 elements
Le 19/04/2018 ? 09:30, Tomas Kalibera a ?crit?: > On 04/19/2018 02:06 AM, Duncan Murdoch wrote: >> On 18/04/2018 5:08 PM, Tousey, Colton wrote: >>> Hello, >>> >>> I want to report a bug in R that is limiting my capabilities to >>> export a matrix with write.csv or write.table with over >>> 2,147,483,648 elements (C's int limit). I found
2018 Apr 18
1
R Bug: write.table for matrix of more than 2, 147, 483, 648 elements
Hi Colton, You could divide your write task into chunks that do not violate the 2^31-1 limit. write.table has an append argument (default FALSE). Figure out a row chunk size nri < nr such that nri * nc is under 2^31-1 and use write.table() to write that out. Then use write.table( append = TRUE, ) for the next chunk of rows, looping over chunks until done. Two chunks will get your 2.8
2018 Apr 19
0
R Bug: write.table for matrix of more than 2, 147, 483, 648 elements
On 04/19/2018 02:06 AM, Duncan Murdoch wrote: > On 18/04/2018 5:08 PM, Tousey, Colton wrote: >> Hello, >> >> I want to report a bug in R that is limiting my capabilities to >> export a matrix with write.csv or write.table with over 2,147,483,648 >> elements (C's int limit). I found this bug already reported about >> before:
2018 Apr 19
0
R Bug: write.table for matrix of more than 2, 147, 483, 648 elements
On 04/19/2018 11:47 AM, Serguei Sokol wrote: > Le 19/04/2018 ? 09:30, Tomas Kalibera a ?crit?: >> On 04/19/2018 02:06 AM, Duncan Murdoch wrote: >>> On 18/04/2018 5:08 PM, Tousey, Colton wrote: >>>> Hello, >>>> >>>> I want to report a bug in R that is limiting my capabilities to >>>> export a matrix with write.csv or write.table
2018 Apr 18
0
R Bug: write.table for matrix of more than 2, 147, 483, 648 elements
Hello, I want to report a bug in R that is limiting my capabilities to export a matrix with write.csv or write.table with over 2,147,483,648 elements (C's int limit). I found this bug already reported about before: https://bugs.r-project.org/bugzilla/show_bug.cgi?id=17182. However, there appears to be no solution or fixes in upcoming R version releases. The error message is coming from the
2018 Apr 19
2
R Bug: write.table for matrix of more than 2, 147, 483, 648 elements
Le 19/04/2018 ? 12:15, Tomas Kalibera a ?crit?: > On 04/19/2018 11:47 AM, Serguei Sokol wrote: >> >> replace >> ??? tmp = EncodeElement2(x, i + j*nr, quote_col[j], qmethod, >> ??? ??? ??? ??? ??? &strBuf, sdec); >> by >> ??? tmp = EncodeElement2(VECTOR_ELT(x, (R_xlen_t)i + j*nr), 0, >> quote_col[j], qmethod, >> ??? ??? ??? ??? ???
2010 Jun 28
2
Not-In operator
I would like to use grep to return all the lines of a data frame that do not contain the letters HD. I have tried the ^ inside brackets as well as !. The data frame is one column consisting of spaces,numbers, and letters with several thousand rows. Thank you! Colton [[alternative HTML version deleted]]
2010 Jun 22
3
Project
I am an intern in Virginia Beach and I am currently working on a project that involves converting a script of Pearl to R. The code takes one input text file and outputs six text files. If you think you can help, I will be able to provide more detailed information. Thank you for your time! Colton Hunt
2016 Nov 15
1
creating a long list puts R in a state where many things stop working
Hi, After I create a long list e.g. with x <- vector(mode="list", length=3e9) many bad things start to happen e.g. some things stop working with a spurious error message: gc() # Error in gc() : # long vectors not supported yet: /home/hpages/src/R-3.3.2/src/main/memory.c:1137 x_lens <- lengths(x) # Error in lengths(x) : # long vectors not supported yet:
2007 May 24
1
video quality problem, encodeing with ffmpeg2theora -p pro
Hello all and thanks, I have encoded several .vob with " ffmpeg2theoar -p preveiw " with no problems at all. The quality is exellent. These .vob were ripped with vobcopy. The problem I am having is when I encode a .vob with " ffmpeg2theora -p pro ". The resulting .ogg will not play correctly with vlc or kaffeine. The vedio is jerky and the sound go out of sycro just
2015 Sep 20
2
Long vectors: Missing values and R_xlen_t?
Is there a missing value constant defined for R_xlen_t, cf. NA_INTEGER (== R_NaInt == INT_MIN) for int(eger)? If not, is it correct to assume that missing values should be taken care/tested for before coercing from int or double? Thank you, Henrik
2012 Mar 15
4
replicating C example from the Extensions Manual problem
Dear R People: Here is something that I am sure is very simple. I'm just trying to re-create the C convolution example in the Extensions manual. Here is the subroutine: void convolve(double *a, int *na, double *b, int *nb, double *ab) { R_len_t i, j, nab = *na + *nb - 1; for(i = 0; i < nab; i++) ab[i] = 0.0; for(i = 0; i < *na; i++) for(j = 0; j < *nb; j++) ab[i + j] += a[i] *
2013 Jul 12
1
robustbase compilation problem: probably boneheaded? maybe 32-bit?
With a recent SVN build (R Under development (unstable) (2013-07-10 r63264) -- "Unsuffered Consequences"), I'm having trouble installing the robustbase package. The bottom line is that I *think* it's a 32-bit-system problem, but I could easily be mistaken. robustbase is passing its package checks: http://cran.r-project.org/web/checks/check_results_robustbase.html ... but from
2004 Jul 19
2
SSLeay
Could someone tell me where i can download the latest version of SSLeay. I have tried the following sites: * <ftp://ftp.psy.uq.oz.au/pub/Crypto/SSL/> * <ftp://ftp.uni-mainz.de/pub/internet/security/ssl> * <ftp://ftp.cert.dfn.de/pub/tools/crypt/sslapps> * <ftp://ftp.funet.fi/pub/crypt/mirrors/ftp.psy.uq.oz.au> * <ftp://ftp.sunet.se/ftp/pub/security/tools/crypt/ssleay>
2019 Oct 23
2
Unexpected behavior when using macro to loop over vector
Hi all, I found an unexpected behavior when I was trying to use the macro defined in "R_ext/Itermacros.h" to loop over an atomic vector. Here is a minimum example: C++ code ``` #include "R_ext/Itermacros.h" #define GET_REGION_BUFSIZE 2 //Redefine the macro since C++ is not happy with the implicit type conversion #define ITERATE_BY_REGION_PARTIAL(sx, px, idx, nb, etype,
2008 Jan 04
1
Evaluating R expressions from C
I am currently puzzled by a passage in the R Extensions manual, section 5.10: SEXP lapply(SEXP list, SEXP expr, SEXP rho) { R_len_t i, n = length(list); SEXP ans; if(!isNewList(list)) error("`list' must be a list"); if(!isEnvironment(rho)) error("`rho' should be an environment"); PROTECT(ans = allocVector(VECSXP, n));
2015 Jun 01
2
sum(..., na.rm=FALSE): Summing over NA_real_ values much more expensive than non-NAs for na.rm=FALSE? Hmm...
I'm observing that base::sum(x, na.rm=FALSE) for typeof(x) == "double" is much more time consuming when there are missing values versus when there are not. I'm observing this on both Window and Linux, but it's quite surprising to me. Currently, my main suspect is settings in on how R was built. The second suspect is my brain. I hope that someone can clarify the below
2019 Oct 25
2
Unexpected behavior when using macro to loop over vector
On 10/25/19 11:01 AM, Tomas Kalibera wrote: > On 10/23/19 6:45 AM, Wang Jiefei wrote: >> Hi all, >> >> I found an unexpected behavior when I was trying to use the macro >> defined >> in "R_ext/Itermacros.h"? to loop over an atomic vector. Here is a >> minimum >> example: >> >> C++ code >> ``` >> #include
2013 Apr 09
2
Behaviors of diag() with character vector in R 3.0.0
Dear all, According to CHANGES IN R 3.0.0: o diag() as used to generate a diagonal matrix has been re-written in C for speed and less memory usage. It now forces the result to be numeric in the case diag(x) since it is said to have 'zero off-diagonal entries'. diag(x) does not work for character vector in R 3.0.0 any more. For example, v <- c("a",
2023 Feb 11
1
scan(..., skip=1e11): infinite loop; cannot interrupt
On Fri, 10 Feb 2023 23:38:55 -0600 Spencer Graves <spencer.graves at prodsyse.com> wrote: > I have a 4.54 GB file that I'm trying to read in chunks using > "scan(..., skip=__)". It works as expected for small values of > "skip" but goes into an infinite loop for "skip=1e11" and similar > large values of skip: I cannot even interrupt it; I