similar to: Problem With Model.Tables Function

Displaying 20 results from an estimated 150 matches similar to: "Problem With Model.Tables Function"

2008 May 29
1
plotting zoo using datetime as xlim
is there a way to use the actual index value for plotting zoo objects this is the way that the index is set up and a sample range of what I would like to plot 01/01/06 00:00:00 - 01/01/06 23:45:00 { library(zoo) # chron library(chron) fmt.chron <- function(x) { chron(sub(" .*", "", x), gsub(".* (.*)", "\\1:00", x)) }} x <- structure(c(15.57, 15.5,
2009 Jul 17
19
I need help saving table data from a rake task
I need to find out how I can create and save a large dataset to a table based on multiple returned arrays from a Rake task. Here is my example using just two arrays (there are 14 in this particular rake task): update_tsos_offense = TsosOffense.new to_team_id, to_ppcs = update_tsos_offense.calculate_tsos(TotalOffense, "ydspgm", "desc") ro_team_id, ro_ppcs =
2012 Aug 15
4
boxplot help
Hi, im a newbie with very wobbly coding abilities. Tearing my hair out over getting the boxplot i want... I have a dataset called 'eagle' which consists of year (2011 or 2012), month (jan - dec), roof (TT6, TT13 or BARE) and temp (the continuous variable that i want to plot). So i want boxplots of the three roof treatments in every month organised in chronical order along x axis 2011 -
2008 Jan 09
1
reading table
Hi, I am encountering a problem in reading a file, the file looks like that: T 3 0 -- -- -- T -- -- -- 18.98 3 1 6.75 4.39 39 -- -- -- 18.58 3 2 6.90 4.90 43 -- -- -- 18.63 3 3 7.07 5.39 48 -- -- -- 18.78 G 4 0 7.41 5.54 47 G
2011 Feb 15
4
string parsing
I am trying to get stock metadata from Yahoo finance (or maybe there is a better source?) here is what I did so far: yahoo.url <- "http://finance.yahoo.com/d/quotes.csv?f=j1jka2&s="; stocks <- c("IBM","NOIZ","MSFT","LNN","C","BODY","F"); # just some samples socket <-
2008 Dec 04
2
How to optimize this codes ?
How to optimize the for-loop to be reasonably fast for sample.size=100000000 ? You may want to change sample.size=1000 to have an idea what I am achieving. set.seed(143) A <- matrix(sample(0:1, sample.size, TRUE), ncol=10, dimnames=list(NULL, LETTERS[1:10])) B <- list() for(i in 1:10) { B[[i]] <- apply(combn(LETTERS[1:10], i), 2, function(x) { sum(apply(data.frame(A[,x]), 1,
2010 Sep 21
1
Abysmal sparse file performance!
I have a 5.5GB file, mostly sparse. Tar performs far[!] better than rsync. I have no ideas yet, so just an FYI as to current state. FreeBSD 8.1 i386 zfs Yes, I know the blocks used differs but don't know why yet, could be just how zfs does things or related to the large amount of sparseness. There are no media errors, CPU/IO load or anything like that and the source and dest paths are on the
2012 Dec 22
1
Character Variable in X axis scatter plot
I am very new to R statistics. Have installed R-2.15.2 ; Rcmdr 1.9-2 ; RStudio 0.97.237 on Debian Squeeze and also windows7 I can Import from Excel File OK .Workbook <- loadWorkbook("/media/4C90-B739/Oct13-Dec21Bsl.xls") JJData <- readWorksheet(.Workbook, "Oct13-Dec21Bsl") remove(.Workbook) have a data frame with following. DATEEVENT
2014 Sep 01
1
Correlation Matrix with a Covariate
R Help - I'm trying to run a correlation matrix with a covariate of "age" and will at some point will also want to covary other variables concurrently. I'm using the "psych" package and have tried other methods such as writing a loop to extract semi-partial correlations, but it does not seem to be working. How can I accomplish this? library(psych) > set.cor(y =
2016 May 27
2
Slow RAID Check/high %iowait during check after updgrade from CentOS 6.5 -> CentOS 7.2
All of our Kafka clusters are fairly write-heavy. The cluster in question is our second-heaviest ? we haven?t yet upgraded the heaviest, due to the issues we?ve been experiencing in this one. Here is an iostat example from a host within the same cluster, but without the RAID check running: [root at r2k1 ~] # iostat -xdmc 1 10 Linux 3.10.0-327.13.1.el7.x86_64 (r2k1) 05/27/16 _x86_64_ (32 CPU)
2015 Apr 20
4
Samba 4 slow write
Hi Jones, many thanks for your reply. Yepp, I'm using ACL's (Posix ACL's) on filesystem to regulate the access of users for files and directories. Do you mean that this is the "normal" way? These geteuid and getegid system calls are so slow? On the previous system (Debian Wheezy, Samba 3) I've also used ACL's, but there were much faster than this... What should
2005 Jun 24
5
Memory limits using read.table on Windows XP Pro
Hello, When I try: geno <-read.table("2500.geno.tab",header=TRUE,sep="\t",na.strings=".",quote=" ",comment.char="",colClasses=c("factor"),nrows=2501) I get, after hour(s) of work: Error: cannot allocate vector of size 9 Kb I have: Rgui.exe --max-mem-size=3Gb and multi(0)disk(0)rdisk(0)partition(1)\WINDOWS="Microsoft
2016 May 25
6
Slow RAID Check/high %iowait during check after updgrade from CentOS 6.5 -> CentOS 7.2
I?ve posted this on the forums at https://www.centos.org/forums/viewtopic.php?f=47&t=57926&p=244614#p244614 - posting to the list in the hopes of getting more eyeballs on it. We have a cluster of 23 HP DL380p Gen8 hosts running Kafka. Basic specs: 2x E5-2650 128 GB RAM 12 x 4 TB 7200 RPM SATA drives connected to an HP H220 HBA Dual port 10 GB NIC The drives are configured as one large
2004 Feb 13
3
Calculate Closest 5 Cases?
I've only begun investigating R as a substitute for SPSS. I have a need to identify for each CASE the closest (or most similar) 5 other CASES (not including itself as it is automatically the closest). I have a fairly large matrix (50000 cases by 50 vars). In SPSS, I can use Correlate > Distances to generate a matrix of similarity, but only on a small sample. The entire matrix can not
2012 Jun 05
2
[LLVMdev] [PATCH] add x32 psABI support
If you are interesting to play around X32, you may refer to http://sourceware.org/glibc/wiki/x32 to bootstrap a local environment on Linux. Yours - Michael -----Original Message----- From: cfe-commits-bounces at cs.uiuc.edu [mailto:cfe-commits-bounces at cs.uiuc.edu] On Behalf Of Liao, Michael Sent: Monday, June 04, 2012 5:09 PM To: llvm-commits at cs.uiuc.edu; cfe-commits at cs.uiuc.edu
2004 Dec 02
3
Tbench benchmark numbers seem to be limiting samba performance in the 2.4 and 2.6 kernel.
Hi, I'm getting horrible performance on my samba server, and I am unsure of the cause after reading, benchmarking, and tuning. My server is a K6-500 with 43MB of RAM, standard x86 hardware. The OS is Slackware 10.0 w/ 2.6.7 kernel I've had similar problems with the 2.4.26 kernel. I use samba version 3.0.5. I've listed my partitions below, as well as the drive models. I have a
2011 May 16
2
wireframe advice - with reproducible code
Dear List, i am trying to produce a 3d plot using wireframe using the code: wireframe(Residuals_FD ~ Elevation * Temperature, data = data2, scales = list(arrows = FALSE), drape = TRUE, colorkey = TRUE) As you can see when the code (using the data below) is run the plot area is set-up correctly but the actual surface is missing? Any help would be greatly appreciated. Chris #data Elevation
2013 Sep 13
2
[LLVMdev] [Polly] Compile-time and Execution-time analysis for the SCEV canonicalization
At 2013-09-09 13:07:07,"Tobias Grosser" <tobias at grosser.es> wrote: >On 09/09/2013 05:18 AM, Star Tan wrote: >> >> At 2013-09-09 05:52:35,"Tobias Grosser" <tobias at grosser.es> wrote: >> >>> On 09/08/2013 08:03 PM, Star Tan wrote: >>> Also, I wonder if your runs include the dependence analysis. If this is >>> the
2005 Sep 05
1
convergence for proportional odds model
Hey, everyone, I am using proportional odds model for ordinal responses in dose-response experiments. For some samll data, SAS can successfully provide estimators of the parameters, but the built-in function polr() in R fails. Would you like to tell me how to make some change so I can use polr() to obtain the estimators? Or anyone can give me a hint about the conditions for the existance of MLE
2009 Jun 01
3
Within Subject ANOVA question
Dear R users, I have copied for following table from an article on "Using confidence intervals in within-subject designs": Subject 1sec 2sec 5sec 1 10 13 13 12.00 2 6 8 8 7.33 3 11 14 14 13.00 4 22 23 25 23.33 5 16 18 20 18.00 6 15 17 17 16.33 7 1 1 4 2.00 8 12 15 17 14.67 9 9 12 12 11.00 10 8 9 12 9.67 I rearranged the data this way: