similar to: handling time units--hours, minutes, days--from file times

Displaying 20 results from an estimated 10000 matches similar to: "handling time units--hours, minutes, days--from file times"

2006 Nov 17
0
difftime for handling time units--hours, minutes, days--from file times
Thanks, Tim To get the time difference in the proper units, it's as simple as this: difftime(myfileinfo$mtime[2], myfileinfo$mtime[1], units="hours") Cheers, Warren On 11/13/06, Tim Calkins <tcalkins@gmail.com> wrote: > hardly the most efficient way to go, but consider using a substring > function to extract the time bits from your data, then reading them as POSIX
2020 Oct 27
2
Expert to work on load issue
Jon, We are only using FastAgi. On the second system (running Asterisk 16) there are no agi's running (just some bash scripts on call hangup). I did add some hackey code (netstat -nua | grep -v 'udp 0 0' | grep -v udp6 | grep -v ' 0 0.0.0.0' | grep udp) to my bash script to check out the packet queue (with the help of
2007 Sep 12
1
reshape help
Hi, I'm trying to use reshape but I cannot quite understand how it works. Could somebody help me on this? Example, my data is something like: mydat <- data.frame(tree= 1:10, serra=rep(1:2, c(5,5)), bt01= 101:110, bt02= 201:210, bt03= 301:310, mm01= 9101:9110, mm02= 9201:9210, mm03= 9301:9310) > mydat tree serra bt01 bt02 bt03 mm01 mm02 mm03 1 1 1 101 201 301 9101 9201
2017 Nov 09
2
GlusterFS healing questions
Hi, We ran a test on GlusterFS 3.12.1 with erasurecoded volumes 8+2 with 10 bricks (default config,tested with 100gb, 200gb, 400gb bricksizes,10gbit nics) 1. Tests show that healing takes about double the time on healing 200gb vs 100, and abit under the double on 400gb vs 200gb bricksizes. Is this expected behaviour? In light of this would make 6,4 tb bricksizes use ~ 377 hours to heal. 100gb
2017 Nov 09
0
GlusterFS healing questions
Hi Rolf, answers follow inline... On Thu, Nov 9, 2017 at 3:20 PM, Rolf Larsen <rolf at jotta.no> wrote: > Hi, > > We ran a test on GlusterFS 3.12.1 with erasurecoded volumes 8+2 with 10 > bricks (default config,tested with 100gb, 200gb, 400gb bricksizes,10gbit > nics) > > 1. > Tests show that healing takes about double the time on healing 200gb vs > 100, and
2017 Nov 09
2
GlusterFS healing questions
Hi, You can set disperse.shd-max-threads to 2 or 4 in order to make heal faster. This makes my heal times 2-3x faster. Also you can play with disperse.self-heal-window-size to read more bytes at one time, but i did not test it. On Thu, Nov 9, 2017 at 4:47 PM, Xavi Hernandez <jahernan at redhat.com> wrote: > Hi Rolf, > > answers follow inline... > > On Thu, Nov 9, 2017 at
2017 Nov 09
0
GlusterFS healing questions
Someone on the #gluster-users irc channel said the following : "Decreasing features.locks-revocation-max-blocked to an absurdly low number is letting our distributed-disperse set heal again." Is this something to concider? Does anyone else have experience with tweaking this to speed up healing? Sent from my iPhone > On 9 Nov 2017, at 18:00, Serkan ?oban <cobanserkan at
2012 Nov 12
3
arrange data
Dear r-users,   I have daily rainfall data from 1971 to 2000.  I would like to extract november and december data only.  I would also like to do column bind for november and december, therefore I would like to delete 31 December from december data so that the length of november and december are the same.  Hope somebody can help me.  I tried this below:   > kuantan.dt.1 <-
2001 Mar 26
1
duplicated lines in serverloop.c? (openssh252p2)
Hi, I was looking through the source, and I noticed that the following code appears twice in the file serverloop.c. Is it supposed to, and if not, would there be any ill effect? +289 +290 /* Read and buffer any available stdout data from the program. */ +291 if (!fdout_eof && FD_ISSET(fdout, readset)) { +292 len = read(fdout, buf,
2010 Jul 16
1
Troubles with DBI's dbWriteTable in RMySQL
I am feeling rather dumb right now. I created what I thought was a data.frame as follows: aaa <- lapply(split(moreinfo,list(moreinfo$m_id),drop = TRUE), fun_m_id) m_id_default_res <- do.call(rbind, aaa) print("==========================================") m_id_default_res print("==========================================") ndf <- m_id_default_res[, c('mid',
2013 Nov 26
7
[PATCH RESEND 0/1] libxl: introduce an option for disabling the non-O_DIRECT
I think I posted this patch before, but it looks like it was in December 2012 (!). 1/1 libxl: introduce an option for disabling the non-O_DIRECT workaround Ideally it would go into 4.4, at least. Provided the corresponding qemu part has gone into qemu-xen, which I think it has. Can anyone confirm ?
2020 Oct 29
0
Expert to work on load issue
Anyone have any other ideas? On Tue, Oct 27, 2020 at 1:27 PM Dovid Bender <dovid at telecurve.com> wrote: > Jon, > > We are only using FastAgi. On the second system (running Asterisk 16) > there are no agi's running (just some bash scripts on call hangup). I did > add some hackey code (netstat -nua | grep -v 'udp 0 0' | grep > -v udp6 | grep -v
2008 May 09
1
FLAC__metadata_get_picture()
Hi all! I have a problem to get _all_ images from flac file. I have flac file with 3 images inside ( FLAC__METADATA_TYPE_PICTURE ) How to get _all_ images with bool FLAC::Metadata::get_picture() func? This function is wrapper FLAC__bool FLAC__metadata_get_picture(). flac-1.2.1/src/libFLAC/metadata_iterators.c: ... 285 FLAC_API FLAC__bool FLAC__metadata_get_picture(const char *filename,
2020 Feb 26
1
Re: *** buffer overflow detected *** accessing invalid FD in libguestfs
On Wednesday, 26 February 2020 10:43:27 CET Richard W.M. Jones wrote: > On Wed, Feb 26, 2020 at 11:21:18AM +0200, Veselin Kozhuharski wrote: > > Hallo Rich, > > > > Here is the fd list and total number just before collectd application > > crashes. Before that the number of used fd's is constantly increasing. It > > looks like a fd leak inside libguestfs to me.
2010 Jul 29
1
R Equivalent of SAS Datastep Line-Hold (@@) Specifier?
Hello Everyone,   Below is some SAS code that uses a "line hold specifier" to read multiple observations from each of several input lines of data. There are 3 patients per line in the in-stream data.   Is there a way in R to read this kind of data? I've looked in my books and online but haven't found anything   Thanks,   Paul     DATA EXAMP.TRIAL; INPUT TRT $ CENTER PAT SEX $ AGE
2020 Oct 27
2
Expert to work on load issue
Hi, Sorry in advance that I am emailing the users list and not the biz list I think I will find my target audience here. We are looking to hire a consultant to help us figure out an issue. We are having what seems are "random load" issues with bare metal boxes that are dedicated to Asterisk and a few Perl AGI's. We went after all the usual suspects (CPU IDLE, memory usage etc.). I
2010 May 13
1
results of pnorm as either NaN or Inf
I stumbled across this and I am wondering if this is unexpected behavior or if I am missing something. > pnorm(-1.0e+307, log.p=TRUE) [1] -Inf > pnorm(-1.0e+308, log.p=TRUE) [1] NaN Warning message: In pnorm(q, mean, sd, lower.tail, log.p) : NaNs produced > pnorm(-1.0e+309, log.p=TRUE) [1] -Inf I don't know C and am not that skilled with R, so it would be hard for me to look into
2011 Dec 27
1
differences between 1.7 and 1.7.1 glmnet versions
Dear All, ? I have found differences between glmnet versions 1.7 and 1.7.1 which, in my opinion, are not cosmetic and do not appear in the ChangeLog. If I am not mistaken, glmnet appears to return different number of selected input variables, i.e. nonzeroCoef(fit$beta[[1]]) differes between versions. The code below is the same for 1.7.1 and 1.7, but you can see that outputs differ. I would
2015 Jan 14
1
Questions regarding imaptest
Hi, The measurements were created under the following conditions : ? operating system : Red Hat Enterprise Linux Server release 6.6 (Santiago) kernel in version 2.6.32-504.el6.x86_64 ? virtual server (VMware) with an Intel(R) Xeon(R) 4vCPU E5649 x 2,526 GHz (2 cores per virtual socket) and 4 GB RAM ? 7,200 RPM SATA 1TB (FC SAN IBM System Storage N3400) ? all file systems had been formated in
2012 Oct 29
4
replace repeated id in a pedigree list
Hello, I have a pedigree file such this: FAMID ID FA ID MO ID SEX STATUS 1 1 0 0 2 0 1 2 3 1 2 2 1 4 3 1 1 2 1 5 3 1 1 0 1 6 3 1 1 0 1 7 3 1 2 2 1 8 3 1 1 0 1 9 3 1 1 0 1 10 3 1 2 0 1 11 3 1 1 0 1 3 0 0 1 0 2 12 13