search for: chunksiz

Displaying 20 results from an estimated 52 matches for "chunksiz".

Did you mean: chunksize
2010 Nov 03
2
[PATCH 1/2] Ocfs2: Add a new code 'OCFS2_INFO_FREEINODE' for o2info ioctl.
The new code is dedicated to calculate free inodes number of all inode_allocs, then return the info to userpace in terms of an array. Specially, flag 'OCFS2_INFO_FL_NON_COHERENT', manipulated by '--cluster-coherent' from userspace, is now going to be involved. setting the flag on means no cluster coherency considered, usually, userspace tools choose none-coherency strategy by
2005 Aug 13
1
retrieving large columns using RODBC
...t;index" which is used to order rows. Unfortunately, sqlQuery can't return all the values from a column at once (RODBC complains about lack of memory). So I am using the following code: getcolumns <- function(channel, tablename, colnames, totalrows, ordered=TRUE,chunksize=1e5) { r <- matrix(double(0),totalrows,length(colnames)) for (i in 1:ceiling(totalrows/chunksize)) { cat(".") r[((i-1)*chunksize+1):(i*chunksize)] <- as.matrix( sqlQuery(channel, paste("SELECT", paste(colnames,collapse=", "),...
2013 Sep 23
28
[PATCH 0/2] add LZ4 kernel decompression support
Linux 3.11 added respective support, so I think we should follow suit. 1: xen: add LZ4 decompression support 2: libxc: add LZ4 decompression support Signed-off-by: Jan Beulich <jbeulich@suse.com>
2001 Aug 30
1
Problems with vorbisfile and files smaller than CHUNKSIZE
...imple application to decompress a set of Ogg files to raw PCM format, using vorbisfile, because that is the simplest way of doing it. These Ogg files are sounds for a game, and thus sometimes very short (about .25 s). However, vorbisfile has severe problems with Ogg files that are smaller than the CHUNKSIZE which is #defined at the top of vorbisfile.c (8150 bytes at the moment). In that case, the ov_open() call takes an unusually long amount of time (10-30 s on a P3-450), and the first ov_read() call after that will _always_ return 0, indicating EOF. The result is that decoding with vorbisfile takes...
2008 May 16
2
Fetching Binary data from SQL Server
...", colsize); //PRINTING COLSIZE ON CLI if (fd > -1) { char tmp[1]=""; lseek(fd, fdlen - 1, SEEK_SET); if (write(fd, tmp, 1) != 1) { close(fd); res = -1; goto free_res; } } if (fd > -1){ //Trying to fetch data in chunks for (offset = 0; offset < colsize; offset += CHUNKSIZE) { if ((fdm = mmap(NULL, CHUNKSIZE, PROT_READ | PROT_WRITE, MAP_SHARED, fd, offset)) == MAP_FAILED) { ast_log(LOG_WARNING, "Could not mmap the output file: %s (%d)\n", strerror(errno), errno); goto free_res; } else { res = SQLGetData(stmt, x + 1, SQL_BINARY, fdm, CHUNKSIZE, NULL);...
2007 Jun 29
1
Comparison: glm() vs. bigglm()
...obably a naive assumption? Anyways, I've been using bigglm() on some datasets I have available. One of the sets has >15M observations. I have 3 continuous predictors (A, B, C) and a binary outcome (Y). And tried the following: m1 <- bigglm(Y~A+B+C, family=binomial(), data=dataset1, chunksize=10e6) m2 <- bigglm(Y~A*B+C, family=binomial(), data=dataset1, chunksize=10e6) imp <- m1$deviance-m2$deviance For my surprise "imp" was negative. I then tried the same models, using glm() instead... and as I expected, "imp" was positive. I also noticed differences on...
2009 Apr 03
1
bigglm "update" with ff
Hi, since bigglm doesn't have update, I was wondering how to achieve something like (similar to the example in ff package manual using biglm): first <- TRUE ffrowapply ({ if (first) { first <- FALSE fit <- bigglm(eqn, as.data.frame(bigdata[i1:i2,,drop=FALSE]), chunksize = 10000, family = binomial()) } else { fit <- update(fit, as.data.frame(bigdata[i1:i2,,drop=FALSE]), chunksize = 10000, family = binomial()) } }, X=bigdata, VERBOSE = TRUE, BATCHSIZE = nmax) Many thanks. Yuesheng [[alternative HTML version deleted]]
2010 Mar 02
1
bigglm Memory Issues
...lp with an issue I'm having. I'm trying to model a glm using bigglm, but in spite of my best efforts, I cannot get it to work! Here is the particular line of code that is giving me trouble: >mod = bigglm(Pres/wt ~ Xdes, data=dat, family=poisson(), weights = ~wt, maxit=100, tol=1.e-10, chunksize=(dim(Xdes)[1] + 1)) When I attempt to run this, it gives me the following error: >Error: cannot allocate vector of size 255.9 Mb However, I had set the maximum memory to 4095 Mb and had used only 926 Mb to that point. If I try to use a smaller chunksize, I get this error: >Error in mod...
2009 Feb 08
1
[libvorbis] Bugfix: seek broken for large files and on pure 32 bit architectures
...n int ov_pcm_seek_page(OggVorbis_File *vf,ogg_int64_t pos) the calculation of the approximate seek point causes an overflow for very large files and on pure 32 bit architectures. /* take a (pretty decent) guess. */ bisect=begin + (target-begintime)*(end-begin)/(endtime-begintime) - CHUNKSIZE; The temporary expression (target-begintime)*(end-begin) grows with the square of the file size. This effectively limits the file size to the square root of the maximum value of ogg_int64_t. In consequence the file is scanned from the very beginning which is pretty slow over a network. FIX:...
2010 Jul 02
2
unable to get bigglm working, ATTN: Thomas Lumley
...999.000 -0.100 8 0 988887 2007-03-14 2007-06-01 90 3 5.450 205500.00 999.00 999.000 -0.100 9 0 988887 2007-03-15 2007-06-01 90 3 5.450 205500.00 999.00 999.000 -0.100 10 0 988887 2007-03-16 2007-06-01 90 3 5.450 205500.00 999.00 999.000 -0.100 11 0 the code is: make.data <- function (filename, chunksize, ...) { conn<-NULL; function (reset=FALSE) { if (reset) { if (!is.null(conn)) { close(conn); }; conn <<- file (description=filename, open="r"); } else { rval <- read.table (conn, nrows=chunksize,sep=' ', sk...
2002 Nov 12
2
Vorbisfile: Small Files broken in 1.0?
...les. I have a file called "silence" that's just a silent wav file. It compresses down to barely more than the header, so it's _really_ small. When I try to play it, it goes into an infinite loop in vorbisfile.c, line 159. In RC3, that code was: while(offset==-1){ begin-=CHUNKSIZE; if(begin<0) begin=0; _seek_helper(vf,begin); while(vf->offset<begin+CHUNKSIZE){ ret=_get_next_page(vf,og,begin+CHUNKSIZE-vf->offset); if(ret==OV_EREAD)return(OV_EREAD); if(ret<0){ break; }else{ offset=ret; } } }...
2009 Jul 03
2
bigglm() results different from glm()
...making package available to us. I am facing few problems if you can give some hints: Problem-1: The model summary and residual deviance matched (in the mail below) but I didn't understand why AIC is still different. > AIC(m1) [1] 532965 > AIC(m1big_longer) [1] 101442.9 Problem-2: chunksize argument is there in bigglm but not in biglm, consequently, udate.biglm is there, but not update.bigglm Is my observation correct? If yes, why is this difference? Regards Utkarsh / / From: Thomas Lumley <tlumley_at_u.washington.edu <mailto:tlumley_at_u.washington.edu?Subject=Re:%20%5BR...
2012 May 31
2
bigglm binomial negative fitted value
Hi, there Since glm cannot handle factors very well. I try to use bigglm like this: logit_model <- bigglm(responser~var1+var2+var3, data, chunksize=1000, family=binomial(), weights=~trial, sandwich=FALSE) fitted <- predict(logit_model, data) only var2 is factor, var1 and var3 are numeric. I expect fitted should be a vector of value falls in (0,1) However, I get something like this: str(fitted) num [1:260617, 1] -0.0564 -0.0564 -0.1817...
2007 Jan 22
1
Example function for bigglm (biglm) data input from file
...ented example function for use in the data argument to the bigglm(biglm) function, when you want to read the data from a file (instead of a URL), or rescale or modify the data before fitting the model. In the hope that this may be of help to someone out there. make.data <- function (filename, chunksize, ...) { conn<-NULL; function (reset=FALSE) { if (reset) { if (!is.null(conn)) { close(conn); }; # This is for a file. # For other methods, see: help("connections") # and replace the following definition of conn # (and possibly the r...
2009 Oct 08
3
foreach loop - rejection method
...and2[l-1],phi,lambda[l-1]) p2<-cand2[l-1]^-1.5*exp(-y[l]^2/(2*cand2[l-1]))*exp(-((log(cand2[l-1])-mu_t_cand[l-1])^2)/2*sigmasq) ratio<-p2/(c[l-1]*q2) if (runif(1)< ratio) {break} } f2[l-1]<-min(p2,c[l-1]*q2) }} foreach(l=2:(n-1),.combine=c,.packages='MCMCpack',.options.nws=li st(chunkSize=250)) %dopar% test() even though i've created the array earlier called cand2 when i run this code i get the error object "cand2" not found. What I'm trying to do is the rejection method for my cand2. i need cand2 and f2 as vectors as output both of length l-2 in the correct ord...
2013 May 23
11
raid6: rmw writes all the time?
Hi all, we got a new test system here and I just also tested btrfs raid6 on that. Write performance is slightly lower than hw-raid (LSI megasas) and md-raid6, but it probably would be much better than any of these two, if it wouldn''t read all the during the writes. Is this a known issue? This is with linux-3.9.2. Thanks, Bernd -- To unsubscribe from this list: send the line
2006 Apr 04
0
Need help debugging (long)
...ar_ptr = arena_for_chunk(p); (gdb) 3445 (void)mutex_lock(&ar_ptr->mutex); (gdb) 3447 _int_free(ar_ptr, mem); (gdb) Couldn't get registers: No such process. (gdb) [Switching to Thread -1216079680 (LWP 4132)] 4291 p = mem2chunk(mem); (gdb) 4278 { (gdb) 4292 size = chunksize(p); (gdb) 4298 if (__builtin_expect ((uintptr_t) p > (uintptr_t) -size, 0) (gdb) 4307 if (__builtin_expect (size < MINSIZE, 0)) (gdb) 4320 if ((unsigned long)(size) <= (unsigned long)(get_max_fast ()) (gdb) 4360 else if (!chunk_is_mmapped(p)) { (gdb) 4365 if (__...
2012 Jul 02
0
S4 class for chunked data that inherits from numeric- numeric methods won't work
...numeric"), contains="numeric" #so that it inherits from ‘vector’ without assuming logical data ) setGeneric("set.chunk", function(x,...) standardGeneric("set.chunk")) setMethod("set.chunk", signature(x = "range.vec"), function (x, chunksize=100, chunkpos=1) { #This function extracts a chunk of data from the range.vec object. begin <- x@start + (chunkpos - 1)*x@step end <- x@start + (chunkpos + chunksize - 2)*x@step data <- seq(begin, end, x@step) #calculate values in data chunk #get r...
2004 Aug 09
5
How to import specific column(s) using "read.table"?
Dear R people, I have a very big tab-delim txt file with header and I only want to import several columns into R. I checked the options for "read.table" and only found "nrows" which lets you specify the maximum number of rows to read in. Although I can use some text editors (e.g., wordpad) to edit the txt file first before running R, I feel it?s not very convenient. The
2010 Oct 01
2
Format details for a raid partition....
So I have been playing with a RAID 10 f2 ( 2 disks far layout) setup...thanks for all of the advice..Now I am playing with the format and want to make sure I have it setup the best that I can, my raid was built using the raid 10 option with 2 disks with the layout=far, chunk size 512....now I read all of the docs I could find about format and stride and stripe size and this is what i came up