Displaying 20 results from an estimated 10000 matches similar to: "Kolmogorov Smirnov Test"
2007 Feb 23
4
using "integrate" in a function definition
Dear list members,
I'm quite new to R, and though I tried to find the answer to my probably
very basic question through the available resources (website, mailing
list archives, docs, google), I've not found it.
If I try to use the "integrate" function from within my own functions,
my functions seem to misbehave in some contexts. The following example
is a bit silly, but
2011 Feb 19
3
Kolmogorov-smirnov test
Is the kolmogorov-smirnov test valid on both continuous and discrete data?
I don't think so, and the example below helped me understand why.
A suggestion on testing the discrete data would be appreciated.
Thanks,
a <- rnorm(1000, 10, 1);a # normal distribution a
b <- rnorm(1000, 12, 1.5);b # normal distribution b
c <- rnorm(1000, 8, 1);c # normal distribution c
d <- rnorm(1000,
2009 Apr 29
2
Kolmogorov-Smirnov test
I got a distribution function and a empirical distribution function. How do I
make to Kolmogorov-Smirnov test in R.
Lets call the empirical distribution function >Fn on [0,1]
and the distribution function >F on [0,1]
ks.test( )
thanks for the help
--
View this message in context: http://www.nabble.com/Kolmogorov-Smirnov-test-tp23296096p23296096.html
Sent
2011 Apr 27
3
Kolmogorov-Smirnov test
Hi,
I have a problem with Kolmogorov-Smirnov test fit. I try fit distribution to
my data. Actualy I create two test:
- # First Kolmogorov-Smirnov Tests fit
- # Second Kolmogorov-Smirnov Tests fit
see below. This two test return difrent result and i don't know which is
properly. Which result is properly? The first test return lower D = 0.0234
and lower p-value = 0.00304. The lower 'D'
2011 Jul 29
1
How to interpret Kolmogorov-Smirnov stats
Hi,
Interpretation problem ! so what i did is by using the:
>fit1 <- fitdist(vectNorm,"beta")
Warning messages:
1: In dbeta(x, shape1, shape2, log) : NaNs produced
2: In dbeta(x, shape1, shape2, log) : NaNs produced
3: In dbeta(x, shape1, shape2, log) : NaNs produced
4: In dbeta(x, shape1, shape2, log) : NaNs produced
5: In dbeta(x, shape1, shape2, log) : NaNs produced
6: In
2003 May 15
2
kolmogorov-smirnov
Hello,
I got a rather simple question: Can I find somewhere in R the
significance values for a Kolmogorov distribution (I know the degrees
of freedom and I have already the maximum deviation). ks.test is not
really doing what I want. All I need is the values, like one can get
the values for a chi-squared distribution by 'qchisq(0.05, 375)'.
tnx,
Kurt.
2009 Oct 12
1
Kolmogorov smirnov test
Hi r-users,
I would like to use Kolmogorov smirnov test but in my observed data(xobs) there are ties. I got the warning message. My question is can I do something about it?
ks.test(xobs, xsyn)
Two-sample Kolmogorov-Smirnov test
data: xobs and xsyn
D = 0.0502, p-value = 0.924
alternative hypothesis: two-sided
Warning message:
In ks.test(xobs, xsyn) : cannot compute correct
2011 Jun 10
3
Test if data uniformly distributed (newbie)
Hello,
I have a bunch of files containing 300 data points each with values from 0
to 1 which also sum to 1 (I don't think the last element is relevant
though). In addition, each data point is annotated as an "a" or a "b".
I would like to know in which files (if any) the data is uniformly
distributed.
I used Google and found out that a Kolmogorov-Smirnov or a Chi-square
2010 Aug 05
1
Kolmogorov-Smirnov test, which one to use?
Hi,
I have two sets of data, an observed data and generated data.
The generated data is obtained from the model where the parameters is estimated
from the observed data.
So I'm not sure which to use either
one-sample test
ks.test(x+2, "pgamma", 3, 2) # two-sided, exact
or
two-sample test
ks.test(x, x2, alternative="l")
If I use the one-sample test I need to
2007 May 27
1
Parametric bootstrapped Kolmogorov-Smirnov GoF: what's wrong
Dear R-users,
I want to perform a One-Sample parametric bootstrapped Kolmogorov-Smirnov
GoF test (note package "Matching" provides "ks.boot" which is a 2-sample
non-parametric bootstrapped K-S version).
So I wrote this code:
---[R Code] ---
ks.test.bootnp <- function( x, dist, ..., alternative=c("two.sided", "less",
"greater"), B = 1000 )
{
2005 Oct 07
1
permutational Kolmogorov-Smirnov p-value for paired data
Dear List,
I am new to R and find it very powerful. I would like to compute the
permutational p-value for paired data using Kolmogorov-Smirnov, but
the built-in ks.test does not have this option, unlike the t.test
which has a paired=TRUE flag. Has someone written a library or a
routine that does this? Alternatively, if someone could show me how to
do pair-wise permutations in R, then I can
2002 Jul 01
1
modified kolmogorov-smirnov
I'm trying to use modified Kolmogorov-Smirnov test with a Normal which I
don't know it's parameters. Somebody told me about the lilifor function in
R, but just can't find it.
Does anybody know how I can test with the modified Kolmogorov-Smirnov
test?
Porqu? usar una base de datos relacional cualquiera,
si pod?s usar PostgreSQL?
2002 Jun 23
1
Kolmogorov-Smirnov tests: overflow
Dear All,
I've got a problem with ks.test. I've two realy large vectors, that I'd
like to test, but I get an overflow, and the p-value cannot be
calculated:
> length(genomesv)
[1] 390025
> length(scopv)
[1] 140002
> ks.test(genomesv, scopv)
Two-sample Kolmogorov-Smirnov test
data: genomesv and scopv
D = 0.2081, p-value = NA
alternative hypothesis: two.sided
2006 Apr 28
1
Checking Goodness of Fit With Kolmogorov-Smirnov
Hi,
I'm using the power.law.fit function from the igraph package to fit a
power law distribution to some data. This function returns the power
law exponent as it's only result. I would like to have some sort of
goodness-of-fit and/or error estimate of the exponent returned. This
paper:
http://www.edpsciences.org/articles/epjb/pdf/2004/18/b04111.pdf
suggests using the
2009 Sep 08
1
Unexpected behavior in friedman.test and ks.test
I have to start by saying that I am new to R, so I might miss something crucial here. It seems to me that the results of friedman.test and ks.test are "wrong". Now, obviously, the first thing which crossed my mind was "it can't be, this is a package used by so many, someone should have observed", but I can't figure out what it might be.
Problem: let's start with
2011 Oct 06
2
KS test and theoretical distribution
> x <- runif(100)
> y <- runif(100)
> ks.test(x,y)
Two-sample Kolmogorov-Smirnov test
data: x and y
D = 0.11, p-value = 0.5806
alternative hypothesis: two-sided
ok I expected that, but:
> ks.test(runif(100), "runif")
One-sample Kolmogorov-Smirnov test
data: runif(100)
D = 0.9106, p-value < 2.2e-16
alternative hypothesis: two-sided
How
2001 Jul 02
2
Shapiro-Wilk test
Hi,
does the shapiro wilk test in R-1.3.0 work correctly? Maybe it does, but can
anybody tell me why the following sample doesn't give "W = 1" and
"p-value = 1":
R> x<-1:9/10;x
[1] 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9
R> shapiro.test(qnorm(x))
Shapiro-Wilk normality test
data: qnorm(x)
W = 0.9925, p-value = 0.9986
I can't imagine a sample being
2012 May 26
1
Kolmogorov-Smirnov test and the plot of max distance between two ecdf curves
Hi all,
given this example
#start
a<-c(0,70,50,100,70,650,1300,6900,1780,4930,1120,700,190,940,
760,100,300,36270,5610,249680,1760,4040,164890,17230,75140,1870,22380,5890,2430)
length(a)
b<-c(0,0,10,30,50,440,1000,140,70,90,60,60,20,90,180,30,90,
3220,490,20790,290,740,5350,940,3910,0,640,850,260)
length(b)
out<-ks.test(log10(a+1),log10(b+1))
# max distance D
2007 Nov 16
2
ks.test
Hello,
I want to do normality test on my data
I write this but I don't understand the display of the results
ks.test(data,"pnorm")
In fact I want to know if my data is a normal distribution. I have to check the p-value or D?
Thanks.
_____________________________________________________________________________
l
[[alternative HTML version deleted]]
2005 Nov 22
1
Kolmogorov-Smirnov test help
Hi
I am conducting 2-sample Kolmogorov Smirnov tests for my Masters project to
determine if two independant tree populations have the same size-class
distribution or not. The trees have been placed into size-class categories
based on their basal diameters. Once I started running the stats on my data,
I got confused with the results. Just to show an example of what I was
testing I ran stats