Displaying 20 results from an estimated 1000 matches similar to: "Use of bread() function"
2010 May 10
2
Robust SE & Heteroskedasticity-consistent estimation
Hi,
I'm using maxlik with functions specified (L, his gradient & hessian).
Now I would like determine some robust standard errors of my estimators.
So I 'm try to use vcovHC, or hccm or robcov for example
but in use one of them with my result of maxlik, I've a the following
error message :
Erreur dans terms.default(object) : no terms component
Is there some attributes
2006 Dec 07
1
Heteroscedasticity consistent standard errors for Spatial error models
Hello,
Could anyone please tell me how to estimate Heteroscedasticity
Consistent standard errors for a Spatial error model? All the functions
I have looked at only works for lm objects.
Thank you very much!
- Oshadhi
2007 Oct 26
1
Newey-West and SUR regression models
Is anyone aware of a procedure to apply Newey-West corrections for
autocorrelation to a SUR regression model? The SANDWICH package seems to be
applicable only to LM or GLM models.
Thanks,
Richard Saba
Department of Economics
Auburn University
Email: sabaric at auburn.edu
2007 Oct 30
1
Some matrix and sandwich questions
Dear R-help,
I have a four-part question about regression, matrices, and sandwich package.
1) In the sandwich package, I would like to better understand the
meat() function.
>From the bread() documentation, for a simple OLS regression, bread() returns
(1/n * X'X)^(-1)
That is, for a simple regression (per the documentation on bread()):
MyLM <- lm(y ~ x)
bread(MyLM)
2010 May 14
1
Creating an S3 method when the generic function is defined in another (imported) package
Hi,
In one of my packages (maxLik), I would like to add an S3 method,
where the generic function (estfun) is defined in another package
(sandwich). Everything works fine if my package "Depends" on the other
package and I import the generic function "estfun" from the "sandwich"
package and define the new method in the NAMESPACE file. However, I
prefer not to load the
2011 Sep 19
1
"could not find function" after import
I am trying to build a package (GWASTools, submitted to Bioconductor)
that uses the "sandwich" package. I have references to "sandwich" in
DESCRIPTION:
Imports: methods, DBI, RSQLite, sandwich, survival, DNAcopy
and NAMESPACE:
import(sandwich)
In the code itself is a call to vcovHC:
Vhat <- vcovHC(mod, type="HC0")
I have sandwich version 2.2-7 installed.
2010 Jun 08
2
how to ignore rows missing arguments of a function when creating a function?
Hi,
I am relatively new to R; when creating functions, I run into problems with
missing values. I would like my functions to ignore rows with missing values
for arguments of my function) in the analysis (as for example is the case in
STATA). Note that I don't want my function to drop rows if there are missing
arguments elsewhere in a row, ie for variables that are not arguments of my
2007 Oct 08
2
estfun & df
Hello EVERYONE,
I need an URGENT help from you please!
How can I see the "estfun" (empirical estimating function) and "df" (degree of freedom) from the following mixed-model please?
(fm1 <- lmer2(Reaction ~ Days + (Days|Subject), sleepstudy))
Many thanks in advance for your kind help.
Sattar
2008 Oct 08
1
Error in spdep: system is computationally singular
Hi all,
I am trying to run an autologistic model using the function errorsarlm from spdep package.
**I built an XY matrix extracting the two colums from matriz**
coords1<-matriz[matriz$casos1==1, c(4,5)]
coords1<-as.matrix(coords1)
**I identify neighbours of region points**
nb20<-dnearneigh(coords1,0,20,longlat=TRUE)
** I build a neighbours list with spatial weights**
2009 Jun 06
0
SMACOF joint configuration plot with bread data?
Dear R-helpers,
I have dist class objects for 10 individuals rating the
dissimilarities (on a 100-point scale) of the same 10 faces (analogous
to the bread data). I would like to get an individual differences
scaling jointly for the individual judges and the faces, and plot them
on the same axes.
This is the example:
library(smacof)
data(breakfast)
res.rect<-smacofRect(breakfast,
2004 May 04
1
spdep question
Dear list,
(also sent to Roger Bivand, but perhaps somebody of you can help me also)
I am trying to use package spdep for fitting an SAR model with errorsarlm.
However, I am not sure how to make a valid nb object out of my
neighborhood. As far as I have seen, there is no documentation for
nb.object.
I have done the following:
class(pschmid$nb) <- "nb"
# pschmid is a prab object as
2010 Jun 26
1
All a column to a data frame with a specific condition
Hi, folks,
Please first look at the codes:
plan_a=c('apple','orange','apple','apple','pear','bread')
plan_b=c('bread','bread','orange','bread','bread','yogurt')
value=1:6
data=data.frame(plan_a,plan_b,value)
library(plyr)
library(reshape)
mm=melt(data, id=c('plan_a','plan_b'))
2012 Mar 18
2
word frequency count
Hi:
I have a dataframe containing comma seperated group of words such as
milk,bread
bread,butter
beer,diaper
beer,diaper
milk,bread
beer,diaper
I want to output the frequency of occurrence of comma separated words
for each row and collapse duplicate rows, to make the output as shown
in the following dataframe:
milk,bread 2
bread,butter 1
beer,diaper 3
milk,bread 2
Thanks for help!
deb
2010 Sep 01
1
transaction object - how to coerce this data
Hi,
I am wanting to look at frequent item sets using the arules package. I need
to transform my data into a "transactions" object. The data I read in from a
file has 2 columns, an ID and an item. How do I convert data like this into
a transactions object?
I've tried
class? transactions
but it only confuses me.
My data is like this....
basketID item
1 bread
1 cheese
1 milk
2
2012 Sep 23
1
ruby koans don't understand the principle sandwhich code
hello,
Im still working on ruby koans.
Now I have to do some sandwhich code.
The exercise looks like this :
require File.expand_path(File.dirname(__FILE__) + ''/edgecase'')
class AboutSandwichCode < EdgeCase::Koan
def count_lines(file_name)
file = open(file_name)
count = 0
while line = file.gets
count += 1
end
count
ensure
file.close if
2006 Nov 01
4
extract values from a vector
Hello,
I'm looking for a solution for the following problem:
I have two vectors
V1 <- c("apple","honey","milk","bread","butter")
V2 <- c("bread","milk")
now, I would like to know for each element in V1 if it's equal to one of
the elements in V2
I could do:
which(V1 == V2[1] | V1 == V2[2])
but what if I
2010 May 02
1
question about 2SLS
Hi All,
I am using R 2.11.0 on a Ubuntu machine. I estimated a model using "tsls"
from the package "sem". Is there a way to get Newey West standard errors for
the parameter estimates?
When estimating the model by OLS, I used "NeweyWest" from the package
"sandwich" to get HAC standard errors. But, I am not able to use the same
method with the results of the
2013 Mar 30
1
vcovHC and arima() output
Dear all,
how can I use vcovHC() to get robust/corrected standard errors from an
arima() output?
I ran an arima model with AR(1) and got the estimate, se, zvalue and
p-value using coeftest(arima.output).
However, I cannot use vcovHC(arima.output) to get corrected standard
errors. It seems vcovHC works only with lm and plm objects?
Is there another way I can get robust/corrected
2009 Jun 08
0
SMACOF joint configuration plot with bread data? (Michael Kubovy)
Hi Michael,
with res.uc$conf you'll get the single configurations for each rater.
You can use these to produce the plot you want to have.
Best,
Patrick
r-help-request at r-project.org wrote:
> Send R-help mailing list submissions to
> r-help at r-project.org
>
> To subscribe or unsubscribe via the World Wide Web, visit
> https://stat.ethz.ch/mailman/listinfo/r-help
>
2005 Apr 25
1
multiple autocorrelation coefficients in spdep?
Hello,
Has anyone modified the errorsarlm in the R package spdep to
allow for more than a single spatial autocorrelation coefficient (i.e.
'lambda')?
Or, if not, any initial suggestions on how to make that
modification? I have looked at the source code for the function and realize
that any attempt to do it on my own would require much dedication, so would
like to