Displaying 20 results from an estimated 200 matches similar to: "Fligner-Policello robust rank test"
2004 Apr 05
1
fligner.test (ctest) (PR#6739)
Full_Name: Karel Zvara
Version: 1.8.1
OS: MS Winows 2000
Submission from: (NULL) (195.113.30.163)
The test statistics of the fligner.test (ctest package) depends on the order of
cases:
> fligner.test(count~spray,data=InsectSprays)
Fligner-Killeen test for homogeneity of variances
data: count by spray
Fligner-Killeen:med chi-squared = 14.4828, df = 5, p-value =
0.01282
>
2012 Jan 10
1
different results from fligner.test
I've made fligner test with the same data, changing the orders of the
variables, and this what i get
> fligner.test(rojos~edadysexo*zona*ano*estacion)
Fligner-Killeen test of homogeneity of variances
data: rojos by edadysexo by zona by ano by estacion
Fligner-Killeen:med chi-squared = 15.7651, df = 2, p-value = 0.0003773
> fligner.test(rojos~ano*edadysexo*zona*estacion)
2016 Apr 04
4
Fligner-Killeen test on binary data
Hello,
I investigate survival until the following year (0,1) and I wish to test if
the variance in survival for two or more groups are significantly different
from each other.
I read that the Fligner-Killeen test is a non-parametric test which is very
robust against departures from normality but is it correct (valuable
technique for publication) to use it on binary data?
In other
2016 Apr 04
0
Fligner-Killeen test on binary data
That's not an R question but a stats question, but I wouldn't do it. For one thing: The variance of binary data is a function of the mean, so the research question is dubious in the first place. Secondly, the test is based on ranking and comparing absolute differences from the group median, which for binary data is generally 0 or 1, so all absolute differences will be 1.... Put
2008 Sep 17
2
Unexpected behaviour when testing for independence with multiple factors
Hi, I'm a new user of R. My background is Electrical Engineering, so
please bear with me if this is a silly question.
I'm trying to assess whether the results of an experiment satisfy the
hypothesis of homoscedasticity (my ultimate goal is to use ANOVA).
The result of the experiment is mean delay (dT), which depends on
three factors, topology, drift, and lambda. The first two factors are
2008 Aug 22
1
Test of Homogeneity of Variances
I am testing the homogeneity of variances via bartlett.test and fligner.test. Using the following example, how should I interpret the p-value in order to accept or reject the null hypothesis ?
set.seed(5)
x <- rnorm(20)
bartlett.test(x, rep(1:5, each=4))
Bartlett test of homogeneity of variances
data: x and rep(1:5, each = 4)
Bartlett's K-squared = 1.7709, df = 4, p-value =
2000 Sep 01
1
Levene's test
> From: Peter Dalgaard BSA <p.dalgaard at biostat.ku.dk>
> Date: 01 Sep 2000 09:54:59 +0200
>
> Prof Brian D Ripley <ripley at stats.ox.ac.uk> writes:
Important omission: specification from Murray Jorgensen
The test that I was thinking of basically does an anova on a modified
response variable that is the absolute value of the difference between an
observation
2019 Jun 21
1
[Suggested patch] to fligner.test - constant values can produce significant results
In specific cases fligner.test() can produce a small p-value even when both
groups have constant variance.
Here is an illustration:
fligner.test(c(1,1,2,2), c("a","a","b","b"))
# p-value = NA
But:
fligner.test(c(1,1,1,2,2,2), c("a","a","a","b","b","b"))
# p-value < 2.2e-16
2010 Feb 15
1
Difference in Levene's test between R and SPSS
Hello,
I notice that when I do Levene's test to test equality of variances across
levels of a factor, I get different answers in R and SPSS 16.
e.g.: For the chickwts data, in R, levene.test(weight, feed) gives
F=0.7493, p=0.5896.
SPSS 16 gives F=0.987, p=0.432
Why this difference? Which one should I believe? (I would like to believe
R :)
Ravi
--
View this message in context:
2003 Feb 15
2
How to code a bootstrap version of the Wilcoxon-Mann-Whitney test (and variants)?
Hello,
can someone please help me with coding a function for a bootstrap WMW test (package boot, R under Windows, version 1.6.2)?
2006 Mar 28
2
Welch test for equality of variance
Hello
Using R 2.2.1 on a Windows machine.
Has anyone programmed the Welch test for equality of variances?
I tried RSiteSearch, but this gave references to t test and
oneway.test, which are not quite what I need.....I need the Welch test
itself, for use in a meta-analysis (to determine if variances are
equal).
TIA
Peter
Peter L. Flom, PhD
Assistant Director, Statistics and Data Analysis
2007 Jul 05
4
Levene Test with R
Hi All,
is there Levene' test in R ? If not ,Could you give me some
advice about Levene test with R?
Thanks a lot! I am waiting for yours.
2011 Sep 23
1
Significance test
I have a bunch of benchmark measurements that look something like this:
sample.1 0.0000066660 0.0000062500 0.0000058330 0.0000058330
0.0000058330
sample.2 0.0000058330 0.0000058330 0.0000058330 0.0000058330
0.0000058330
sample.3 0.0000062500 0.0000062500 0.0000070830 0.0000062500
0.0000066660
i.e each measurement take on one of a set of values. The set values isn't
fixed, but
2019 Jun 18
0
Small bug in fligner.test - constant values can produce significant results (patch attached)
In specific cases fligner.test() can produce a small p-value even when both groups have constant variance.
Here is an illustration:
fligner.test(c(1,1,2,2), c("a","a","b","b"))
# p-value = NA
But:
fligner.test(c(1,1,1,2,2,2), c("a","a","a","b","b","b"))
# p-value < 2.2e-16
2008 Jan 23
2
from a normal bivariate distribution to the marginal one
Hello,
I'm quite new with R and so I would like to know if there is a command
to calculate an integral.
In particular I simulated a bivariate normal distribution using these
simple lines:
rbivnorm <- function(n, # sample size
mux, # expected value of x
muy, # expected value of Y
sigmax, # standard deviation of
2003 May 04
1
port of Pan to R
I'm looking for a port of Schafer's PAN module for multiple imputation of
nested data.
It is written in S-Plus, and I would like to use it in R.
Any pointers most appreciated.
Best wishes,
Paul von Hippel
2008 Mar 10
2
question for aov and kruskal
Hi R users!
I have the following problem: how appropriate is my aov model under the violation of anova assumptions?
Example:
a<-c(1,1,1,1,1,1,1,1,1,1,2,2,2,3,3,3,3,3,3,3)
b<-c(101,1010,200,300,400, 202, 121, 234, 55,555,66,76,88,34,239, 30, 40, 50,50,60)
z<-data.frame(a, b)
fligner.test(z$b, factor(z$a))
aov(z$b~factor(z$a))->ll
TukeyHSD(ll)
Now from the aov i found that my model
2011 Jan 12
2
Don´t know what test i have to use
Hello,
I?m starting with my PhD and I have to stop because i got a little knowledge
in R and statistics.
I?ve got a model of this kind:
binary response variable: prevalence of infection (0/1)
3 categorical independent variables: sex, month and name of the area
I was trying with a full model like this, before the simplification
model<-aov(prevalencia~sex*month*area)
but the Fligner test
2008 Sep 21
0
Unexpected behaviour when testing for independence, with multiple factors
>Ben Bolker <bolker <at> ufl.edu> writes:
>
>I would try
>
>fligner.test(dT ~ Topology:Drift:lambda)
>
>in response to:
>
>Javier Acuna <javier.acuna.o <at> gmail.com> writes:
>
> Hi, I'm a new user of R. My background is Electrical Engineering, so
> please bear with me if this is a silly question.
>
> I'm trying to assess
2009 Feb 04
1
holidays effect
how can I eliminate the influence of the festivities in a time series with
daily data?I tried to remove them and replace their value with a value of
interpolation using na.approx (). There is an alternative method?
--
View this message in context: http://www.nabble.com/holidays-effect-tp21830785p21830785.html
Sent from the R help mailing list archive at Nabble.com.