search for: 0.0367

Displaying 10 results from an estimated 10 matches for "0.0367".

Did you mean: 0.0360
2009 Feb 08
0
Initial values of the parameters of a garch-Model
Dear all, I'm using R 2.8.1 under Windows Vista on a dual core 2,4 GhZ with 4 GB of RAM. I'm trying to reproduce a result out of "Analysis of Financial Time Series" by Ruey Tsay. In R I'm using the fGarch library. After fitting a ar(3)-garch(1,1)-model > model<-garchFit(~arma(3,0)+garch(1,1), analyse) I'm saving the results via > result<-model
2009 Sep 09
1
Forecast - How to create variables with summary() results parameters
Hi, I would like to create variables in R containing parameters of summary(*Forecast Results*). Using the following code: library(forecast) data <- AirPassengers xets <- ets(data, model="ZZZ", damped=NULL) xfor <- forecast(xets,h=12, level=c(80,95)) summary(xfor) the output is: Forecast method: ETS(M,A,M) Model Information: ETS(M,A,M) Call: ets(y = data, model =
2017 Dec 20
2
outlining (highlighting) pixels in ggplot2
Using the small reproducible example below, I'd like to know if one can somehow use the matrix "sig" (defined below) to add a black outline (with lwd=2) to all pixels with a corresponding value of 1 in the matrix 'sig'? So for example, in the ggplot2 plot below, the pixel located at [1,3] would be outlined by a black square since the value at sig[1,3] == 1. This is my first
2002 Sep 12
1
dropterm, binomial.glm, F-test
Hi there - I am using R1.5.1 on WinNT and the latest MASS (Venables and Ripley) library. Running the following code: >minimod<-glm(miniSF~gtbt*f.batch+log(mxjd),data=gtbt,family="binomial") >summary(minimod,cor=F) Coefficients: Estimate Std. Error z value Pr(>|z|) (Intercept) 0.91561 0.32655 2.804 0.005049 ** gtbtgt 0.47171
2017 Dec 20
0
outlining (highlighting) pixels in ggplot2
Hi Eric, you can use an annotate-layer, eg ind<-which(sig>0,arr.ind = T) ggplot(m1.melted, aes(x = Month, y = Site, fill = Concentration), autoscale = FALSE, zmin = -1 * zmax1, zmax = zmax1) + geom_tile() + coord_equal() + scale_fill_gradient2(low = "darkred", mid = "white", high = "darkblue",
2011 Jul 24
2
[LLVMdev] [llvm-testresults] bwilson__llvm-gcc_PROD__i386 nightly tester results
A big compile time regression. Any ideas? Ciao, Duncan. On 22/07/11 19:13, llvm-testresults at cs.uiuc.edu wrote: > > bwilson__llvm-gcc_PROD__i386 nightly tester results > > URL http://llvm.org/perf/db_default/simple/nts/253/ > Nickname bwilson__llvm-gcc_PROD__i386:4 > Name curlew.apple.com > > Run ID Order Start Time End Time > Current 253 0 2011-07-22 16:22:04
2006 Nov 23
2
random effect question and glm
consider p as random effect with 5 levels, what is difference between these two models? > p5.random.p <- lmer(Y ~p+(1|p),data=p5,family=binomial,control=list(usePQL=FALSE,msV=1)) > p5.random.p1 <- lmer(Y ~1+(1|p),data=p5,family=binomial,control=list(usePQL=FALSE,msV=1)) in addtion, I try these two models, it seems they are same. what is the difference between these two model. Is
2011 Jul 24
0
[LLVMdev] [llvm-testresults] bwilson__llvm-gcc_PROD__i386 nightly tester results
On Jul 24, 2011, at 3:02 AM, Duncan Sands wrote: > A big compile time regression. Any ideas? > > Ciao, Duncan. False alarm. For some reason that I have not yet been able to figure out, these tests run significantly more slowly when I run them during the daytime, which I did for that run. I checked a few of the worst regressions reported here and they all recovered in subsequent
2015 Feb 26
5
[LLVMdev] [RFC] AArch64: Should we disable GlobalMerge?
Hi all, I've started looking at the GlobalMerge pass, enabled by default on ARM and AArch64. I think we should reconsider that, at least for AArch64. As is, the pass just merges all globals together, in groups of 4KB (AArch64, 128B on ARM). At the time it was enabled, the general thinking was "it's almost free, it doesn't affect performance much, we might as well use it".
2011 May 15
5
Question on approximations of full logistic regression model
Hi, I am trying to construct a logistic regression model from my data (104 patients and 25 events). I build a full model consisting of five predictors with the use of penalization by rms package (lrm, pentrace etc) because of events per variable issue. Then, I tried to approximate the full model by step-down technique predicting L from all of the componet variables using ordinary least squares