Displaying 20 results from an estimated 120 matches similar to: "[PATCH 1/2] Ocfs2: Add a new code 'OCFS2_INFO_FREEINODE' for o2info ioctl."
2008 Nov 20
2
Removing rows with rowsums==0 (I can't figure this out)
##I want to remove the rows where the row sums are zero and this is as
far as I have gotten
ffg <- (structure(list(CD = c(0, 0, 0, 0, 3.125, 0, 0, 0, 0, 1.6, 3.125,
0, 0, 6.25, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 3.125, 0, 0, 0, 0,
0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 1.6, 0, 0, 0, 0, 0,
0, 0, 0, 0, 0, 0, 1.6, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,
0, 0, 0, 0, 0, 3.125, 0, 0, 0, 0, 0, 0, 0, 0,
2002 May 20
3
Samba & Symantec Ghost
Hello,
I saw this posting by Christopher Robert Woods. I am having the identical
problem with Samba on Suse Linux 7.2 (OS/390). Is there any known
workaround?
----------------------------------------------------------------------------
--------------------
To: Multiple recipients of list SAMBA <samba@samba.org>
Subject: Symantec Ghost and Samba
From: Christopher Robert Woods
2011 Mar 14
1
Javascript & Rails future
Hi,
David H.H. announced recently that jquery is going to be the default in Rails 3.1, and that Prototype helpers / RJS are going into a gem.
What does that mean for the future? Should we progressively forget about things like javascript helpers, RJS, and all these fun parts of rails? And start coding with jquery/json in mind?
I''m not against that at all, but I have to admit I find RJS
2013 Sep 23
28
[PATCH 0/2] add LZ4 kernel decompression support
Linux 3.11 added respective support, so I think we should follow
suit.
1: xen: add LZ4 decompression support
2: libxc: add LZ4 decompression support
Signed-off-by: Jan Beulich <jbeulich@suse.com>
2005 Aug 13
1
retrieving large columns using RODBC
Hi,
I have a large table in Postgresql (result of an MCMC simulation, with 1
million rows) and I would like to retrive colums (correspond to variables)
using RODBC. I have a column called "index" which is used to order rows.
Unfortunately, sqlQuery can't return all the values from a column at once
(RODBC complains about lack of memory). So I am using the following code:
2007 Jun 29
1
Comparison: glm() vs. bigglm()
Hi,
Until now, I thought that the results of glm() and bigglm() would
coincide. Probably a naive assumption?
Anyways, I've been using bigglm() on some datasets I have available.
One of the sets has >15M observations.
I have 3 continuous predictors (A, B, C) and a binary outcome (Y).
And tried the following:
m1 <- bigglm(Y~A+B+C, family=binomial(), data=dataset1, chunksize=10e6)
2013 May 23
11
raid6: rmw writes all the time?
Hi all,
we got a new test system here and I just also tested btrfs raid6 on
that. Write performance is slightly lower than hw-raid (LSI megasas) and
md-raid6, but it probably would be much better than any of these two, if
it wouldn''t read all the during the writes. Is this a known issue? This
is with linux-3.9.2.
Thanks,
Bernd
--
To unsubscribe from this list: send the line
2009 Oct 08
3
foreach loop - rejection method
Hi Everybody,
Thanks in advance for your help.
This is my first time using the foreach statement and I cant get it to work
properly so here is what i have
test<-function(){
repeat {
cand2[l-1]<-rinvgamma(1,phi,lambda[l-1])
q2<-dinvgamma(cand2[l-1],phi,lambda[l-1])
p2<-cand2[l-1]^-1.5*exp(-y[l]^2/(2*cand2[l-1]))*exp(-((log(cand2[l-1])-mu_t_cand[l-1])^2)/2*sigmasq)
2010 Mar 02
1
bigglm Memory Issues
Hi all,
I'm somewhat of a novice in terms of programming, so I thought I'd come here to seek some help with an issue I'm having.
I'm trying to model a glm using bigglm, but in spite of my best efforts, I cannot get it to work!
Here is the particular line of code that is giving me trouble:
>mod = bigglm(Pres/wt ~ Xdes, data=dat, family=poisson(), weights = ~wt, maxit=100,
2012 May 31
2
bigglm binomial negative fitted value
Hi, there
Since glm cannot handle factors very well. I try to use bigglm like this:
logit_model <- bigglm(responser~var1+var2+var3, data, chunksize=1000,
family=binomial(), weights=~trial, sandwich=FALSE)
fitted <- predict(logit_model, data)
only var2 is factor, var1 and var3 are numeric.
I expect fitted should be a vector of value falls in (0,1)
However, I get something like this:
2009 Feb 08
1
[libvorbis] Bugfix: seek broken for large files and on pure 32 bit architectures
In file vorbisfile.c, function
int ov_pcm_seek_page(OggVorbis_File *vf,ogg_int64_t pos)
the calculation of the approximate seek point causes an overflow for
very large files and on pure 32 bit architectures.
/* take a (pretty decent) guess. */
bisect=begin +
(target-begintime)*(end-begin)/(endtime-begintime) - CHUNKSIZE;
The temporary expression (target-begintime)*(end-begin)
2009 Apr 03
1
bigglm "update" with ff
Hi, since bigglm doesn't have update, I was wondering how to achieve
something like (similar to the example in ff package manual using biglm):
first <- TRUE
ffrowapply ({
if (first) {
first <- FALSE
fit <- bigglm(eqn, as.data.frame(bigdata[i1:i2,,drop=FALSE]), chunksize =
10000, family = binomial())
} else {
fit <- update(fit,
2004 Aug 09
5
How to import specific column(s) using "read.table"?
Dear R people,
I have a very big tab-delim txt file with header and I only want to import
several columns into R. I checked the options for "read.table" and only
found "nrows" which lets you specify the maximum number of rows to read in.
Although I can use some text editors (e.g., wordpad) to edit the txt file first
before running R, I feel it?s not very convenient. The
2010 Jul 02
2
unable to get bigglm working, ATTN: Thomas Lumley
I am using an example posted in this help forum to work with a file. the head
of the file looks like:
988887 2007-03-05 2007-06-01 90 3 5.450 205500.00 999.00 999.000 0.000 0 0
988887 2007-03-06 2007-06-01 90 3 5.450 205500.00 999.00 999.000 0.000 1 0
988887 2007-03-07 2007-06-01 90 3 5.450 205500.00 999.00 999.000 -0.100 2 0
988887 2007-03-08 2007-06-01 90 3 5.450 205500.00 999.00 999.000 -0.100
2010 Oct 01
2
Format details for a raid partition....
So I have been playing with a RAID 10 f2 ( 2 disks far layout)
setup...thanks for all of the advice..Now I am playing with the format and
want to make sure I have it setup the best that I can, my raid was built
using the raid 10 option with 2 disks with the layout=far, chunk size
512....now I read all of the docs I could find about format and stride and
stripe size and this is what i came up
2009 Jul 03
2
bigglm() results different from glm()
Hi Sir,
Thanks for making package available to us. I am facing few problems if
you can give some hints:
Problem-1:
The model summary and residual deviance matched (in the mail below) but
I didn't understand why AIC is still different.
> AIC(m1)
[1] 532965
> AIC(m1big_longer)
[1] 101442.9
Problem-2:
chunksize argument is there in bigglm but not in biglm, consequently,
2004 Aug 06
2
Darkice memory leak
Unfortunately I can't reproduce your memory leak reports. For example,
here is the line about my production darkice instance from top:
PID USER PRI NI SIZE RSS SHARE STAT %CPU %MEM TIME COMMAND
27516 root 20 0 3564 3564 1340 S 95.6 1.3 47:06 darkice
<p>this as after approx 50 minutes from start, but the memory load doesn't
change for the whole 4 hours it
2008 May 16
2
Fetching Binary data from SQL Server
I am trying to write a customized app using C that would fetch voice file from SQL Server 2000 using ODBC and FREETDS.
Currently I am only able to fetch first 63 KB chunk from the DB, and not able to fetch the rest of the file, below is the code that i am using to do so,
fd = open(fullpath, O_RDWR | O_CREAT | O_TRUNC, 0770);
if (fd < 0) {
ast_log(LOG_WARNING, "Failed to write
2009 Dec 18
1
Could Asterisk be crashing under high context switches?
Hello!
I have been struggling with Asterisk 1.6 and DAHDI for the past few weeks. We are an outgoing call center with 30 internal analog phones hooked up to 2 Rhino CB24 channel banks. The banks are connected to a Rhino R4T1 card in a Dell 2950 server with 8 gigs of RAM. The 2 other ports on the R4T1 go to our 2 PRIs.
In this configuration, we have trouble maintaining stability. It may be fine
2009 Mar 17
2
bigglm() results different from glm()
Dear all,
I am using the bigglm package to fit a few GLM's to a large dataset (3
million rows, 6 columns). While trying to fit a Poisson GLM I noticed
that the coefficient estimates were very different from what I obtained
when estimating the model on a smaller dataset using glm(), I wrote a
very basic toy example to compare the results of bigglm() against a
glm() call. Consider the