Displaying 20 results from an estimated 10000 matches similar to: "binomial glm for relevant feature selection?"
2005 Jul 02
2
Is it possible to use glm() with 30 observations?
I have a very simple problem. When using glm to fit
binary logistic regression model, sometimes I receive
the following warning:
Warning messages:
1: fitted probabilities numerically 0 or 1 occurred
in: glm.fit(x = X, y = Y, weights = weights, start =
start, etastart = etastart,
2: fitted probabilities numerically 0 or 1 occurred
in: glm.fit(x = X, y = Y, weights = weights, start =
start,
2001 Oct 10
1
What kind of test in summary(glm)?
Hello R Users,
when I use summary(glm) for a logistic regression model with logit as link
function I get one column "z value". What kind of test does R use? (I would
have expected a t-test).
Thanks, Anne
--
GMX - Die Kommunikationsplattform im Internet.
http://www.gmx.net
-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-
r-help mailing list -- Read
2006 Dec 31
2
zero random effect sizes with binomial lmer [sorry, ignore previous]
I am fitting models to the responses to a questionnaire that has
seven yes/no questions (Item). For each combination of Subject and
Item, the variable Response is coded as 0 or 1.
I want to include random effects for both Subject and Item. While I
understand that the datasets are fairly small, and there are a lot of
invariant subjects, I do not understand something that is happening
here, and in
2010 Dec 29
1
logistic regression with response 0,1
Dear Masters,
first I'd like to wish u all a great 2011 and happy holydays by now,
second (here it come the boring stuff) I have a question to which I hope u
would answer:
I run a logistic regression by glm(), on the following data type
(y1=1,x1=x1); (y2=0,x2=x2);......(yn=0,xn=xn), where the response (y) is
abinary outcome on 0,1 amd x is any explanatory variable (continuous or not)
2006 Oct 21
1
logistic regression with a sample missing subjects with a value of an independent variable
Dear R-help,
I am trying to make logistic regression analysis using the R function
"glm", with the parameter family set to binomial, in order to use a
logistic regression model.
I have 70 samples. The dependent variables has two levels (0 and 1) and
one of the independent variables has too two levels (0 and 1).
The variables associate in the way shown in the table:
2006 Mar 05
1
glm gives t test sometimes, z test others. Why?
I just ran example(glm) and happened to notice that models based on
the Gamma distribution gives a t test, while the Poisson models give a
z test. Why?
Both are b/s.e., aren't they?
I can't find documentation supporting the claim that the distribution
is more like t in one case than another, except in the Gaussian case
(where it really is t).
Aren't all of the others approximations
2000 Aug 14
2
conf. int. for lm() and Up-arrow
Dear all,
Is there any function for calculating confidence limits
for coefficients in an lm() object? I know of the
confint() function in the MASS library working very
well on my binomial GLMs and I have tried it (using glm
() , family=gaussian) but it gives NAs according to
below. Does the confint() function not accept gaussian
GLMs? Could there be convergence problems in the GLM?
Note the
2003 May 09
1
Tolerances in glm.control
I have tightened the tolerances in glm.control in R-devel (aka 1.8.0 Under
Development) from epsilon = 1e-4 to 1e-8, and increases maxit from 10 to
25.
Normally the effect is to do one more iteration and get more accurate
results. However, in cases of partial separation several more iterations
will be done and it will be clearer from the results which parameters are
converging and which are
2003 May 08
1
A problem in a glm model
Hallo all,
I have the following glm model:
f1 <- as.formula(paste("factor(y.fondi)~",
"flgsess + segmeta2 + udm + zona.geo + ultimo.prod.",
"+flg.a2 + flg.d.na2 + flg.v2 + flg.cc2",
" +(flg.a1 + flg.d.na1 + flg.v1 + flg.cc1)^2",
" + flg.a2:flg.d.na2 + flg.a2:flg.v2 +
2002 May 16
1
glm(y ~ -1 + c, "binomial") question
This is a question about removing the intercept in a binomial
glm() model with categorical predictors. V&R (3rd Ed. Ch7) and
Chambers & Hastie (1993) were very helpful but I wasn't sure I
got all the answers.
In a simplistic example suppose I want to explore how disability
(3 levels, profound, severe, and mild) affects the dichotomized
outcome. The glm1 model (see below) is
2000 May 09
4
Dispersion in summary.glm() with binomial & poisson link
Following p.206 of "Statistical Models in S", I wish to change
the code for summary.glm() so that it estimates the dispersion
for binomial & poisson models when the parameter dispersion is
set to zero. The following changes [insertion of ||dispersion==0
at one point; and !is.null(dispersion) at another] will do the trick:
"summary.glm" <-
function(object, dispersion =
2013 Dec 17
1
ggplot2: stat_smooth for family=binomial with cbind(Y, N) formula
With ggplot2, I can plot the glm stat_smooth for binomial data when the
response is binary or
a two-level factor as follows:
data("Donner", package="vcdExtra")
ggplot(Donner, aes(age, survived)) +
geom_point(position = position_jitter(height = 0.02, width = 0)) +
stat_smooth(method = "glm", family = binomial, formula = y ~ x,
alpha = 0.2, size=2)
But how can I
2010 Mar 31
2
interpretation of p values for highly correlated logistic analysis
Dear list,
I want to perform a logistic regression analysis with multiple
categorical predictors (i.e., a logit) on some data where there is a
very definite relationship between one predicator and the
response/independent variable. The problem I have is that in such a
case the p value goes very high (while I as a naive newbie would
expect it to crash towards 0).
I'll illustrate my problem
2006 Sep 13
3
unexpected result in glm (family=poisson) for data with an only zero response in one factor
Dear members,
here is my trouble: My data consists of counts of trapped insects in different attractive traps. I usually use GLMs with a poisson error distribution to find out the differences between my traitments (and to look at other factor effects). But for some dataset where one traitment contains only zeros, GLM with poisson family fail to find any difference between this particular traitment
2007 Aug 15
0
Negative Binomial: glm.nb
Hi Folks,
I'm playing with glm.nb() in MASS.
Reference: the negative binomial distribution
P(y) = (Gamma(theta+y)/(Gamma(theta)*y!))*(p^theta)*(1-p)^y
y = 0,1,2,...
in the notation of the MASS book (section 7.4), where
p = theta/(mu + theta) so (1-p) = mu/(mu + theta)
where mu is the expected value of Y. It seems from ?glm.nb that
an initial value of theta is either supplied, or
2009 Jun 22
1
How to make try to catch warnings in logistic glm
Dear list,
>From an earlier post I got the impression that one could promote
warnings from a glm to errors (presumably by putting
options(warn=1)?), then try() would flag them as errors. I?ve spent
half the day trying to do this, but no luck. Do you have an explicit
solution?
My problems is that I am trying to figure out during what conditions
one may find 5 significant parameters in a
2018 Apr 02
2
What is the universal (world wide) understanding behind degaussing harddisks?
Good evening from Singapore!
The foremost question which I want to ask is, what is the universal
(world wide) understanding behind degaussing hard drives?
I work for No Secrets Agency (NSA) Pte Ltd (fictitious company name
used). My sales manager Edward Joseph Snowden (fictitious individual
name used) had *promised* our customer Leave Me in the Lurch (S) Pte
Ltd (fictitious company name used)
2018 Apr 02
2
What is the universal (world wide) understanding behind degaussing harddisks?
Hello,
On Mon, 2 Apr 2018 10:01:56 -0400 m.roth at 5-cent.us wrote:
> Turritopsis Dohrnii Teo En Ming wrote:
> > Good evening from Singapore!
> >
> > The foremost question which I want to ask is, what is the universal
> > (world wide) understanding behind degaussing hard drives?
> >
> > I work for No Secrets Agency (NSA) Pte Ltd (fictitious company name
2008 Mar 13
1
strange results from binomial lmer?
I'm running lmer repeatedly on artificial data with two fixed factors (called
'gender' and 'stress') and one random factor ('speaker'). Gender is a
between-speaker variable, stress is a within-speaker variable, if that matters.
Each dataset has 100 rows from each of 20 speakers, 2000 rows in all.
About 5% of the time I get a strange result, where the lmer() model with
2005 Feb 03
1
If this is should be posted elsewhere, please advise
Hi,
I am puzzled by the relationship between the p-values asociated with the
coefficients of a univariate logistic regression involving categorical
variables and the p-value I get from Fisher's exact test of the
associated 2 x 2 contingency table.
(1) The 2-sided p-value for the table is ~ 0.0015, whereas the p-value
for the independent is 0.101 and the p-value for the intercept is