Displaying 20 results from an estimated 700 matches similar to: "biglm() and NeweyWest()"
2011 Sep 28
1
Robust covariance matrix with NeweyWest()
Dear R-users,
I would like to compute a robust covariance matrix of two series of realizations of random variables:
###Begin Example###
data <- cbind(rnorm(100), rnorm(100))
model <- lm(data ~ 1)
vcov(model)
library(sandwich)
NeweyWest(model) #produces an error
###End Example###
NeweyWest() produces an error but sandwich(), vcovHAC(), kernHAC, weave(),... do not produce any errors. It
2010 Jun 27
1
NeweyWest
I want to calculate Newey West robust standard error using NeweyWest. Comparing the results to what I get in STATA, in order to get the same results in I need to specify "prewhite=0". Can someone explain what this prewhite command means?
Thanks
[[alternative HTML version deleted]]
2009 Dec 10
2
Problem with coeftest using Newey West estimator
Hi,
I want to calculate the t- and p-values for a linear model using the Newey West estimator.
I tried this Code and it usually worked just fine:
> oberlm <- lm(DYH ~ BIP + Infl + EOil, data=HU_H)
> coeftest(oberlm, NeweyWest(oberlm, lag=2))
t test of coefficients:
Estimate Std. Error t value Pr(>|t|)
(Intercept) 0.1509950 0.0743832 2.0300 0.179486
BIP
2010 May 02
1
question about 2SLS
Hi All,
I am using R 2.11.0 on a Ubuntu machine. I estimated a model using "tsls"
from the package "sem". Is there a way to get Newey West standard errors for
the parameter estimates?
When estimating the model by OLS, I used "NeweyWest" from the package
"sandwich" to get HAC standard errors. But, I am not able to use the same
method with the results of the
2010 Sep 23
1
Newey West and Singular Matrix + library(sandwich)
thank you, achim. I will try chol2inv.
sandwich is a very nice package, but let me make some short
suggestions. I am not a good econometrician, so I do not know what
prewhitening is, and the vignette did not explain it. "?coeftest" did
not work after I loaded the library. automatic bandwidth selection
can be a good thing, but is not always.
as to my own little function, I like the
2011 Sep 09
1
Exception in NeweyWest - Pre-Whitening necessary?
Hi guyz,
I have run my algorithm in R (see http://pastebin.com/q84Tujfg) and got the following error:
Error in ar.ols(x, aic = aic, order.max = order.max, na.action = na.action, :
'order.max' must be < 'n.used'
I am pretty sure, that the error comes from the NeweyWest function in line 45, as the NeweyWest function uses the ar.ols() function for pre whitening. Does anyone
2010 Sep 22
1
Newey West and Singular Matrix
dear R experts: ?I am writing my own little newey-west standard error
function, with heteroskedasticity and arbitrary x period
autocorrelation corrections. ?including my function in this post here
may help others searching for something similar. it is working quite
well, except on occasion, it complains that
Error in solve.default(crossprod(x.na.omitted, x.na.omitted)) :
system is
2011 Aug 27
3
Exception while using NeweyWest function with doMC
Dear R users,
I am using R right now for a simulation of a model that needs a lot of
memory. Therefore I use the *bigmemory* package and - to make it faster -
the *doMC* package. See my code posted on http://pastebin.com/dFRGdNrG
Now, if I use the foreach loop with the addon %do% (for sequential run) I
have no problems at all - only here and there some singularities in
regressor matrices which
2008 May 22
1
How to account for autoregressive terms?
Hi,
how to estimate a the following model in R:
y(t)=beta0+beta1*x1(t)+beta2*x2(t)+...+beta5*x5(t)+beta6*y(t-1)+beta7*y(t-2)+beta8*y(t-3)
1) using "lm" :
dates <- as.Date(data.df[,1])
selection<-which(dates>=as.Date("1986-1-1") & dates<=as.Date("2007-12-31"))
dep <- ts(data.df[selection,c("dep")])
indep.ret1
2010 Jun 15
1
help biglm.big.matrix; problem with weights
Hello colleagues,
I have tried to use the package biglm. I want to specify a
multivariate regression with a weight.
I have imported a large dataset with the library(bigmemory). I load
the library (biglm) and specified a regression with a weight. But I
get everytime a error message like ?object not found? or ?`weights'
must be a formula? or "error in eval(expr, envir, enclos)". I
2007 Feb 12
0
predict on biglm class
Hi Everyone,
I often use the 'safe prediction' feature available through glm().
Now, I'm at a situation where I must use biglm:::bigglm.
## begin example
library(splines)
library(biglm)
ff <- log(Volume)~ns(log(Girth), df=5)
fit.glm <- glm(ff, data=trees)
fit.biglm <- bigglm(ff, data=trees)
predict(fit.glm, newdata=data.frame(Girth=2:5))
## -1.3161465 -0.2975659
2010 Oct 31
1
biglm: how it handles large data set?
I am trying to figure out why 'biglm' can handle large data set...
According to the R document - "biglm creates a linear model object that uses
only p^2 memory for p variables. It can be updated with more data using
update. This allows linear regression on data sets larger than memory."
After reading the source code below? I still could not figure out how
'update'
2009 Feb 19
1
Questions about biglm
Hello folks,
I am very excited to have discovered R and have been exploring its
capabilities. R's regression models are of great interest to me as my
company is in the business of running thousands of linear regressions
on large datasets.
I am using biglm to run linear regressions on datasets that are as
large as several GB's. I have been pleasantly surprised that biglm
runs the
2010 Jun 16
0
biglm.big.matrix: Problem with weighting
Hello colleagues,
I have tried to use the package bigmemory, biganalytics and biglm. I
want to specify a multivariate regression with a weight.
I have imported a large dataset with the library(bigmemory). I load the
library (biglm) and specified a regression with a weight. But I get
everytime an error message like "object not found" or "`weights' must be
a
2012 Jan 03
0
Biglm source code alternatives (E.g. Call to Fortran)
Hi everyone,
I have been looking at the Bigglm (Basically does Generalised Linear Models
for big data under the Biglm package) command and I have done some profiling
on this code and found that to do a GLM on a 100mb file (9 million rows by 5
columns matrix(most of the numbers were either a 0,1 or 2 randomly
generated)) it took about 2 minutes on a linux machine with 8gb of RAM and 4
cores.
2009 Apr 27
0
VIF's in R using BIGLM
Dear R-help
This is a follow-up to my previous post here:
http://groups.google.com/group/r-help-archive/browse_thread/thread/d9b6f87ce06a9fb7/e9be30a4688f239c?lnk=gst&q=dobomode#e9be30a4688f239c
I am working on developing an open-source automated system for running
batch-regressions on very large datasets. In my previous post, I posed
the question of obtaining VIF's from the output of
2011 Nov 03
0
anova or liklihood ratio test from biglm output
(Sorry if this is a repost, I got a bounce reply from the r-help server)
Hi,
I’m using the biglm() function to create some linear models for a very
large data set than lm() can’t fit due to memory issues (the problem is
with the number of interactions, I can fit the main effects model)
I need to determine if the 2-way interactions are necessary or not. Ideally
I’d like to use anova() to
2009 Feb 25
0
leaps and biglm
New versions of leaps and biglm are percolating through CRAN.
The new version of biglm fixes a bug in sandwich standard errors with weights, and adds predict(), deviance() and AIC() methods [based on code from Christophe Dutang].
The new version of leaps adds a regsubsets() method for biglm objects, so that the subset selection algorithms can be run efficiently on large data sets.
-thomas
2009 Feb 25
0
leaps and biglm
New versions of leaps and biglm are percolating through CRAN.
The new version of biglm fixes a bug in sandwich standard errors with weights, and adds predict(), deviance() and AIC() methods [based on code from Christophe Dutang].
The new version of leaps adds a regsubsets() method for biglm objects, so that the subset selection algorithms can be run efficiently on large data sets.
-thomas
2007 Oct 23
0
Residuals from biglm package
Hi all,
first of all, I'm not an expert on R, I'm still learning, so sorry if this is a stupid question...
I have a large dataset that is to big for my computer memory, and I found quite useful the package biglm. Now everything is working perfectly. But if I want the residuals, how I can do it?
Let's say that we are running the example:
> data(trees)>