Displaying 20 results from an estimated 2000 matches similar to: "Format details for a raid partition...."
2013 May 23
11
raid6: rmw writes all the time?
Hi all,
we got a new test system here and I just also tested btrfs raid6 on
that. Write performance is slightly lower than hw-raid (LSI megasas) and
md-raid6, but it probably would be much better than any of these two, if
it wouldn''t read all the during the writes. Is this a known issue? This
is with linux-3.9.2.
Thanks,
Bernd
--
To unsubscribe from this list: send the line
2010 Nov 03
2
[PATCH 1/2] Ocfs2: Add a new code 'OCFS2_INFO_FREEINODE' for o2info ioctl.
The new code is dedicated to calculate free inodes number of all inode_allocs,
then return the info to userpace in terms of an array.
Specially, flag 'OCFS2_INFO_FL_NON_COHERENT', manipulated by '--cluster-coherent'
from userspace, is now going to be involved. setting the flag on means no cluster
coherency considered, usually, userspace tools choose none-coherency strategy by
2013 Sep 23
28
[PATCH 0/2] add LZ4 kernel decompression support
Linux 3.11 added respective support, so I think we should follow
suit.
1: xen: add LZ4 decompression support
2: libxc: add LZ4 decompression support
Signed-off-by: Jan Beulich <jbeulich@suse.com>
2005 Aug 13
1
retrieving large columns using RODBC
Hi,
I have a large table in Postgresql (result of an MCMC simulation, with 1
million rows) and I would like to retrive colums (correspond to variables)
using RODBC. I have a column called "index" which is used to order rows.
Unfortunately, sqlQuery can't return all the values from a column at once
(RODBC complains about lack of memory). So I am using the following code:
2007 Jun 29
1
Comparison: glm() vs. bigglm()
Hi,
Until now, I thought that the results of glm() and bigglm() would
coincide. Probably a naive assumption?
Anyways, I've been using bigglm() on some datasets I have available.
One of the sets has >15M observations.
I have 3 continuous predictors (A, B, C) and a binary outcome (Y).
And tried the following:
m1 <- bigglm(Y~A+B+C, family=binomial(), data=dataset1, chunksize=10e6)
2010 Mar 02
1
bigglm Memory Issues
Hi all,
I'm somewhat of a novice in terms of programming, so I thought I'd come here to seek some help with an issue I'm having.
I'm trying to model a glm using bigglm, but in spite of my best efforts, I cannot get it to work!
Here is the particular line of code that is giving me trouble:
>mod = bigglm(Pres/wt ~ Xdes, data=dat, family=poisson(), weights = ~wt, maxit=100,
2009 Feb 08
1
[libvorbis] Bugfix: seek broken for large files and on pure 32 bit architectures
In file vorbisfile.c, function
int ov_pcm_seek_page(OggVorbis_File *vf,ogg_int64_t pos)
the calculation of the approximate seek point causes an overflow for
very large files and on pure 32 bit architectures.
/* take a (pretty decent) guess. */
bisect=begin +
(target-begintime)*(end-begin)/(endtime-begintime) - CHUNKSIZE;
The temporary expression (target-begintime)*(end-begin)
2009 Apr 03
1
bigglm "update" with ff
Hi, since bigglm doesn't have update, I was wondering how to achieve
something like (similar to the example in ff package manual using biglm):
first <- TRUE
ffrowapply ({
if (first) {
first <- FALSE
fit <- bigglm(eqn, as.data.frame(bigdata[i1:i2,,drop=FALSE]), chunksize =
10000, family = binomial())
} else {
fit <- update(fit,
2010 Jul 02
2
unable to get bigglm working, ATTN: Thomas Lumley
I am using an example posted in this help forum to work with a file. the head
of the file looks like:
988887 2007-03-05 2007-06-01 90 3 5.450 205500.00 999.00 999.000 0.000 0 0
988887 2007-03-06 2007-06-01 90 3 5.450 205500.00 999.00 999.000 0.000 1 0
988887 2007-03-07 2007-06-01 90 3 5.450 205500.00 999.00 999.000 -0.100 2 0
988887 2007-03-08 2007-06-01 90 3 5.450 205500.00 999.00 999.000 -0.100
2009 Oct 08
3
foreach loop - rejection method
Hi Everybody,
Thanks in advance for your help.
This is my first time using the foreach statement and I cant get it to work
properly so here is what i have
test<-function(){
repeat {
cand2[l-1]<-rinvgamma(1,phi,lambda[l-1])
q2<-dinvgamma(cand2[l-1],phi,lambda[l-1])
p2<-cand2[l-1]^-1.5*exp(-y[l]^2/(2*cand2[l-1]))*exp(-((log(cand2[l-1])-mu_t_cand[l-1])^2)/2*sigmasq)
2008 May 16
2
Fetching Binary data from SQL Server
I am trying to write a customized app using C that would fetch voice file from SQL Server 2000 using ODBC and FREETDS.
Currently I am only able to fetch first 63 KB chunk from the DB, and not able to fetch the rest of the file, below is the code that i am using to do so,
fd = open(fullpath, O_RDWR | O_CREAT | O_TRUNC, 0770);
if (fd < 0) {
ast_log(LOG_WARNING, "Failed to write
2012 May 31
2
bigglm binomial negative fitted value
Hi, there
Since glm cannot handle factors very well. I try to use bigglm like this:
logit_model <- bigglm(responser~var1+var2+var3, data, chunksize=1000,
family=binomial(), weights=~trial, sandwich=FALSE)
fitted <- predict(logit_model, data)
only var2 is factor, var1 and var3 are numeric.
I expect fitted should be a vector of value falls in (0,1)
However, I get something like this:
2009 Jul 03
2
bigglm() results different from glm()
Hi Sir,
Thanks for making package available to us. I am facing few problems if
you can give some hints:
Problem-1:
The model summary and residual deviance matched (in the mail below) but
I didn't understand why AIC is still different.
> AIC(m1)
[1] 532965
> AIC(m1big_longer)
[1] 101442.9
Problem-2:
chunksize argument is there in bigglm but not in biglm, consequently,
2004 Aug 09
5
How to import specific column(s) using "read.table"?
Dear R people,
I have a very big tab-delim txt file with header and I only want to import
several columns into R. I checked the options for "read.table" and only
found "nrows" which lets you specify the maximum number of rows to read in.
Although I can use some text editors (e.g., wordpad) to edit the txt file first
before running R, I feel it?s not very convenient. The
2004 Aug 06
2
Darkice memory leak
Unfortunately I can't reproduce your memory leak reports. For example,
here is the line about my production darkice instance from top:
PID USER PRI NI SIZE RSS SHARE STAT %CPU %MEM TIME COMMAND
27516 root 20 0 3564 3564 1340 S 95.6 1.3 47:06 darkice
<p>this as after approx 50 minutes from start, but the memory load doesn't
change for the whole 4 hours it
2010 Jul 28
6
Read ahead / prefetching
Hi,
I am trying to educate myself on prefetching/readahead algorithm for
Lustre''s read. For a starter I only have two simple questions.
1 - Does Lustre detect linear or random I/O pattern or it always triggers
readahead?
2 - If readahead is triggered, how many pages are read in addition to what
is necessary?
Thanks,
Arifa.
2014 Mar 08
2
Re: questions regarding file-system optimization for sortware-RAID array
Andreas,
why is it relevant only in case of RAID5 or RAID6?
regards,
Martin
On Fri, Mar 7, 2014 at 5:57 PM, Andreas Dilger <adilger@dilger.ca> wrote:
> Note that stride and stripe width only make sense for RAI-5/6 arrays.
> For RAID-1 it doesn't really matter.
>
> Cheers, Andreas
>
>> On Mar 6, 2014, at 13:46, Martin T <m4rtntns@gmail.com> wrote:
>>
2014 Mar 06
2
questions regarding file-system optimization for sortware-RAID array
Hi,
I created a RAID1 array of two physical HDD's with chunk size of 64KiB
under Debian "wheezy" using mdadm. As a next step, I would like to create
an ext3(or ext4) file-system to this RAID1 array using mke2fs utility.
According to RAID-related tutorials, I should create the file-system like
this:
# mkfs.ext3 -v -L myarray -m 0.5 -b 4096 -E stride=16,stripe-width=32
/dev/md0
2007 Jun 05
1
Calculating stride values?
All,
I have a question about calculating the value for the -E stride option
to mke2fs.
The mke2fs man page says
stride=stripe-size
Configure the filesystem for a RAID array with stripe-size filesystem blocks per stripe.
So stride = size of stripe/blocksize.
The size of a stripe is the RAID chunk size * the number of drives in the RAID.
My question: are parity disks
2009 Dec 18
1
Could Asterisk be crashing under high context switches?
Hello!
I have been struggling with Asterisk 1.6 and DAHDI for the past few weeks. We are an outgoing call center with 30 internal analog phones hooked up to 2 Rhino CB24 channel banks. The banks are connected to a Rhino R4T1 card in a Dell 2950 server with 8 gigs of RAM. The 2 other ports on the R4T1 go to our 2 PRIs.
In this configuration, we have trouble maintaining stability. It may be fine