similar to: Incorrect p value for binom.test?

Displaying 20 results from an estimated 190 matches similar to: "Incorrect p value for binom.test?"

2006 Oct 11
2
expression as a parameter of binom.test (PR#9288)
Full_Name: Petr Savicky Version: 2.4.0 OS: Fedora Core release 2 Submission from: (NULL) (62.24.91.47) the error is > binom.test(0.56*10000,10000) Error in binom.test(0.56 * 10000, 10000) : 'x' must be nonnegative and integer while > binom.test(5600,10000) yields correct result. The same error occurrs for > binom.test(0.57*10000,10000)
2000 Oct 02
2
binom.test bug?
R. 1.1.0 The example below is self explanatory. ## 1 ## # works fine > binom.test((50*.64),50,.5,alt='g') ... Exact binomial test ... ## 2 ## # WHAT ! ? > binom.test((50*.65),50,.5,alt='g') Error in binom.test((50 * 0.65), 50, 0.5, alt = "g") : x must be an
1999 Jan 28
1
bug in the ctest package: binom.test
R 0630 for windows > library(ctest) > binom.test(7,10,p=0.3, alternative="two.sided") returns a p-value of =< 2.2e-016 and a warning In Splus 3.4 > binom.test(7,10,p=0.3, alternative="two.sided") returns a p-value of 0.0106 I think it is the max(v[v<=(1+eps)*PVAL]) causing the problem... max() of an empty vector....... Mai Z
2002 Mar 22
1
binom.test and small N
running R 1.4.1 on MAC and 1.2.2 on Linux When I use run binom.test with small N the results are a little perplexing to me >binom.test(9,20,p=0.5) gives the below plus other stuff 95 percent confidence interval: 0.2305779 0.6847219 Now: >pbiom(9,20,0.6847219) [1] 0.02499998 # i.e., lower 2.5% of distribution >pbinom(9,20,0.2305779) [1] 0.9923132 >pbinom(8,20,0.2305779)
2012 Aug 20
1
The difference between chisq.test binom.test and pbinom
Hello all, I am trying to understand the different results I am getting from the following 3 commands: chisq.test(c(62,50), p = c(0.512,1-0.512), correct = F) # p-value = 0.3788 binom.test(x=62,n=112, p= 0.512) # p-value = 0.3961 2*(1-pbinom(62,112, .512)) # p-value = 0.329 Well, the binom.test was supposed to be "exact" and give the same results as the pbinom, while the chisq.test
2008 May 29
1
Accessing Value of binom.test
With this line: > binom.test(x=12, n=50, p=12/50, conf.level = 0.90) I get this output: > Exact binomial test > > data: 12 and 50 > number of successes = 12, number of trials = 50, p-value = 1 > alternative hypothesis: true probability of success is not equal to 0.24 > 90 percent confidence interval: > 0.1447182 0.3596557 > sample estimates: > probability
2001 Jun 09
1
AW: binom.test appropriate?
No, since I'd like to test null: p <= p0 alternative: p > p0. and my understanding is that binom.test tests null: p = p0 (can only be a "simple" null hypothesis according to help(binom.test)) alternative: p > p0 (or p < p0 or p != p0). Thanks, Mirko. > -----Urspr?ngliche Nachricht----- > Von: Douglas Bates [mailto:bates at stat.wisc.edu] >
2002 Sep 22
3
binom.test()
Hello everybody. Does anyone else find the last test in the following sequence odd? Can anyone else reproduce it or is it just me? > binom.test(100,200,0.13)$p.value [1] 2.357325e-36 > binom.test(100,200,0.013)$p.value [1] 6.146546e-131 > binom.test(100,200,0.0013)$p.value [1] 1.973702e-230 > binom.test(100,200,0.00013)$p.value [1] 0.9743334 (R 1.5.1, Linux RedHat 7.1) --
2001 Jun 08
1
binom.test appropriate?
Hi there, as part of a 2 x 2 contingency table analysis I would like to estimate conditional probabilities (success rates) in a Bernoulli experiment. In particular I want to test a null hypothesis p <= p0 versus the alternative hypothesis p > p0. As far as I understand the subject, there are UMPU tests for these types of hypotheses. Now I know about R's "binom.test" but the
2006 Oct 19
5
binom.test
R-experts: A quick question, please. >From a lab exp, I got 12 positives out of 50. To get 90% CI for this , I think binom.test might be the one to be used. Is there a better way or function to calculate this? > binom.test(x=12, n=50, p=12/50, conf.level = 0.90) Exact binomial test data: 12 and 50 number of successes = 12, number of trials = 50, p-value = 1 alternative
2003 Jan 22
2
small bug in binom.test?
Hi all, I am wondering whether there is a small bug in the binom.test function of the ctest library (I'm using R 1.6.0 on windows 2000, but Splus 2000 seems to have the same behaviour). Or perhaps I've misunderstood something. the command binom.test(11,100,p=0.1) and binom.test(9,100,p=0.1) give different p-values (see below). As 9 and 11 are equidistant from 10, the mean of the
2007 Apr 05
1
binom.test() query
Hi Folks, The recent correspondence about "strange fisher.test result", and especially Peter Dalgaard's reply on Tue 03 April 2007 (which I want to investigate further) led me to take a close look at the code for binom.test(). I now have a query! The code for the two-sided case computes the p-value as follows: if (p == 0) (x == 0) else if (p == 1) (x == n)
2010 Feb 11
1
Zero-inflated Negat. Binom. model
Dear R crew: I am sorry this question has been posted before, but I can't seem to solve this problem yet. I have a simple dataset consisting of two variables: cestode intensity and chick size (defined as CAPI). Intensity is a count and clearly overdispersed, with way too many zeroes. I'm interested in looking at the association between these two variables, i.e. how well does chick
2007 Feb 24
1
Woolf's test, Odds ratio, stratification
Just a general question concerning the woolf test (package vcd), when we have stratified data (2x2 tables) and when the p.value of the woolf-test is below 0.05 then we assume that there is a heterogeneity and a common odds ratio cannot be computed? Does this mean that we have to try to add more stratification variables (stratify more) to make the woolf-test p.value insignificant? Also in the
2004 Apr 19
0
One inflated Poisson or Negative Binomal regression
Dr. Flom, I was searching the web for any examples of one-inflated negative binomial regression, and ran across your post. Fittingly, I am working on the analysis of data from the NIDA Cooperative Agreement where I had the pleasure of working with Sherry Deren and other folks at NDRI. NBR does a poor job of modeling number of sex partners. (I am using Stata.) Did you have any luck modeling a
2009 Oct 19
1
lmer family=binomal p-values
Hi, If any one has time I need some help understanding the P-values given in the lmer output. Using AIC for model selection I find my minimal model is FOLLOW~MOVERSTATUS+DISTANCE however it appears DISTANCE is not significant at 95% confidence, see output quoted below. However, removing DISTANCE gives a higher AIC=433.5, therefore I will keep it in, but am confused as to what is adds to the
2003 Oct 29
1
One inflated Poisson or Negative Binomal regression
Hello I am interested in Poisson or (ideally) Negative Binomial regression with an inflated number of 1 responses I have seen JK Lindsey's fmr function in the gnlm library, which fits zero inflated Poisson (ZIP) or zero inflated negative binomial regression, but the help file states that for ' Poisson or related distributions the mixture involves the zero category'. I had thought
2012 May 26
2
Assessing interaction effects in GLMMs
Dear R gurus I am running a GLMM that looks at whether chimpanzees spend time in shade more than sun (response variable 'y': used cbind() on counts in the sun and shade) based on the time of day (Time) and the availability of shade (Tertile). I've included some random factors too which are the chimpanzee in question (Individual) and where they are in a given area (Zone). There are
2013 Apr 16
1
assistant
Dear Sir/Ma, I Adelabu.A.A, one of the R-users from Nigeria. When am running a coxph command the below error was generated, and have try some idea but not going through. kindly please assist: > cox1 <- coxph(Surv(tmonth,status) ~ sex + age + marital + sumassure, X) Warning message: In fitter(X, Y, strats, offset, init, control, weights = weights, : Ran out of iterations and did not
2006 Jul 04
1
problem getting R 2.3.1 svn r38481 to pass make check-all
Hi, I noticed this problem on my home desktop running FC4 and again on my laptop running FC5. Both have previously compiled and passed make check-all on 2.3.1 svn revisions from 10 days ago or so. On both these machines, make check-all is consistently failing (4 out of 4 attempts on the FC 4 desktop and 3 out of 3 on the FC 5 laptop) in the p-r-random-tests tests. This is with both default