similar to: Weighted Kappa for m Raters

Displaying 20 results from an estimated 200 matches similar to: "Weighted Kappa for m Raters"

2009 Jul 13
3
Help With Fleiss Kappa
Hi All, I am using fleiss kappa for inter rater agreement. Are there any know issues with Fleiss kappa calculation in R? Even when I supply mock data with total agreement among the raters I do not get a kappa value of 1. instead I am getting negative values. I am using the irr package version 0.70 Any help is much appreciated. Thanks and Regards M [[alternative HTML version deleted]]
2006 May 16
1
Cannot load irr package
The irr package seems to install correctly: > install.packages("irr") trying URL 'http://cran.us.r-project.org/src/contrib/irr_0.61.tar.gz' Content type 'application/x-tar' length 13848 bytes opened URL ================================================== downloaded 13Kb * Installing *source* package 'irr' ... ** R ** data ** help >>>
2008 Aug 22
3
simple generation of artificial data with defined features
Dear R-colleagues, I am quite a newbie to R fighting my stupidity to solve a probably quite simple problem of generating artificial data with defined features. I am conducting a study of inter-observer-agreement in child-bronchoscopy. One of the most important measures is Kappa according to Fleiss, which is very comfortable available in R through the irr-package. Unfortunately medical doctors
2010 May 25
1
Cohen's Kappa for beginners
Hi, I've got two vectors with ratings from two coders, like this: x<-c("red", "yellow", "blue", "red") #coder number 1 y<-c("red", "blue", "blue", "red") #coder number 2 I want to find Cohen's Kappa using the wkappa function in the psych package. The only example in the docs is using a matrix, which
2009 Jul 13
0
Fleiss Kappa
Hello Everyone I am calculating Fleiss Kappa, I have 28 raters, 5 Subjects and 5 ratings. The problem is that there are 2 missing values in the data. Would it better to replace those with "0" or should those be omitted? By omission I will be left wit only 3 subjects. and my second problem is that overall agreement comes to zero, whereas the data is not showing agreement to be close
2008 Oct 31
1
stratified kappa (measure agreement or interrater reliability)?
Hi All: Could anyone point me to a package that can calculate stratified kappa? My design is like this, 4 raters, 30 types of diagnosis scores, 20 patients. Each rater will rate each patient for each type of diagnosis score. The rater's value is nominal. I know I can measure the agreement between raters for each type of diagnosis score, e.g., calculate out 30 kappa values. My problem is I
2012 Feb 01
1
Function to compute multi-response, multi-rater kappa?
I'm looking for a function in R that extends kappa to multiple raters when there is more than one response per subject. For example, say a group of doctors have to assign diseases to patients. Each patient will be assigned one to many diseases, and the number of doctors assigning diseases to any one patient will be two to many. Here's an extremely simple example of the type of data I
2012 Feb 01
0
Multi-response, multi-rater kappa?
I'm looking for an extension of kappa to measure agreement among multiple raters when there can be more than one response per subject. For example, say a group of doctors assign diseases to patients. Each patient will be assigned one to many diseases, and the number of doctors assigning diseases to any one patient will be two to many. Here's an extremely simple example of the type of
2018 Apr 26
2
kappa
Buenas, ¿alguien sabe qué puede ser este error? data es una df y las columnas 23 y 24 son las que quiero comparar; están bien. testK<-kappa2(data[,c(23,24)], "equal") Error in sort.list(y) : 'x' must be atomic for 'sort.list' Have you called 'sort' on a list? -- Dr Manuel Mendoza Department of Biogeography and Global Change National Museum of Natural
2012 Apr 25
1
Using apply() with a function involving ode()
Hello, I am trying to get the output from the numerical simulation of a system of ordinary differential equations for a range of values for three parameters. I am using the ode() function (deSolve package) to run the numerical simulation and apply() to run the simulation function for each set of parameter values. I am having trouble getting the apply() function to work. Here is an
2008 Sep 01
3
convenient way to calculate specificity, sensitivity and accuracy from raw data
Dear R-colleagues, this is a question from a R-newbie medical doctor: I am evaluating data on inter-observer-reliability in endoscopy. 20 medical doctors judged 42 videos filling out a multiple choice survey for each video. The overall-data is organized in a classical way: observations (items from the multiple choice survey) as columns, each case (identified by the two columns "number of
2006 Apr 28
1
Where do I find Cohen´s kappa???
Hello, I?m looking for a way to measure the goodness of fit of my model with Cohen?s Kappa (scaling between 0 and 1). The kappa function does not give the results I?m looking for. Heres the code: z<-glm(x~y,binomial) kappa(z, exact = T) Does anyone know more? many thanks Christian -- _______________________________________________ Search for businesses by name, location, or phone number.
2006 Jun 18
1
bug with boot.sw98 function (PR#8999)
Full_Name: Nuno Monteiro Version: 2.3.1 OS: Windows XP HE Submission from: (NULL) (84.9.38.207) I'm using the FEAR library to perform Data Envelopment analysis with a 36,000 obs dataset. The function dea is working fine but then when I try to use the boot.sw98 to come up with some sensitivity analysis I get the following error:
2008 Jan 03
1
KernSmooth: bkde and dpik bandwidth questions
Hi, I have two separate questions relating to the KernSmooth package. I am using the dpik function from the KernSmooth package and receive the error Warning message: In kappam * Gcounts : longer object length is not a multiple of shorter object length I saw an earlier post , but the issue was using the bkde fxn and the person appeared to be using too small of a bandwidth.
2007 Jun 26
3
inter-rater agreement index kappa
Is there a function that calculates the inter-rater agreement index (kappa) in R? Thanks ../Murli [[alternative HTML version deleted]]
2010 Jun 16
1
shrout & fleiss ICC´s with varying numbers of judges
Win7, R2.11.0 I am working on a report together with several co-authors. The data concern several performance measures on a set of groups. These measures are scored by external judges. We report findings on several datasets, including several of the 6 ICC's discussed by Shrout & Fleiss (1979). We determine these using the icc function from the irr package. We have also used the ICC
2005 Apr 26
0
psy version 0.65 released
Dear R users, psy version 0.65 is on CRAN now. psy provides several methods used in psychometry (kappa, icc, cronbach, screeplot (with simulations), non linear mapping, etc.) A bug has been fixed in function wkappa (weighted kappa): in particular circumstances, the 2*2 table presented levels in an order different to what was suggested in the help file. The fpca function (PCA plot with
2005 Apr 26
0
psy version 0.65 released
Dear R users, psy version 0.65 is on CRAN now. psy provides several methods used in psychometry (kappa, icc, cronbach, screeplot (with simulations), non linear mapping, etc.) A bug has been fixed in function wkappa (weighted kappa): in particular circumstances, the 2*2 table presented levels in an order different to what was suggested in the help file. The fpca function (PCA plot with
2007 Jun 06
2
Multiple color schemes for barchart (lattice)
Hello R-help. I am trying to make a stacked barplot where the color of the sections of each bar depend on another variable. > myData[1:11,] score percent marker cellType Malignant 1 0 100.00000 ESR1 (ER) Bladder.M(5) TRUE 2 0 80.00000 PAX8 Bladder.M(5) TRUE 3 1 20.00000 PAX8 Bladder.M(5) TRUE 4 0 100.00000 ESR1 (ER) Brain.N(3) FALSE 5 0
2006 Oct 02
2
R function to compute kappa statistics for two vector
Dear R-user, Did anybody know how to calculate the kappa statistics of two vectors for the agreement? for example: a<-c(0,1,0,0,1,0) b<-c(0,1,1,0,0,1) I know the percent of agreement is 3/6, but how to get the kappa? [[alternative HTML version deleted]]