Hi all,
I am conducting research and need some help with doing the Fleiss Kappa analysis. First and foremost let me give a bit of a layout of the study. We have a sample of 75 students in the social sciences who were asked to categorize potential items for a new measure based on descriptions of 2 theoretical constructs. This was meant to be a preliminary check to help with construct validity and take it a step beyond just the research team validating items initially. The items are supposed to characterize coping behaviours following romantic loss. The constructs are loss-oriented coping and restoration-oriented coping, however, we also gave the option for participants to say potential items were both types of coping behaviour or something else altogether (basically an 'other' option). Therefore we have 75 participants categorizing items into one of 4 categories. We want to have some measure of the inter-rater reliability of all these participants though. My understanding is that Fleiss Kappa would be an appropriate tool for this purpose, however, if anyone is aware of other statistics which may be a better fit for our model then please let me know. I also need to the syntax for the Fleiss Kappa statistic if anyone has it available. Thanks in advance for any help people can provide. |
I am currently out of the office and have limited access to email. I will get back to you as soon as possible.
|
In reply to this post by BAM86
I can send you a macro for Fleiss' kappa, which will give you the overall as well as the category kappa's, based on Fleiss' 1971 Psych. Bulletin article. It uses, however, the updated 1979 formula for variance. How many items will the raters categorize? You have other options, e.g., Krippendorf's alpha and Gwet's AC1 statistic. Although Fleiss is the most used, there's some use of these as well in the literature. If you're interested in Cohen's kappa, there are some multiple rater models, most notably Light (averages all the two-rater kappas) and Hubert (averages the two-rater Pe's). Let me know which you want, and I'll get them off to you.
Brian -----Original Message----- From: SPSSX(r) Discussion [mailto:[hidden email]] On Behalf Of BAM86 Sent: Wednesday, January 09, 2013 3:25 PM To: [hidden email] Subject: Fleiss Kappa (Inter-rater reliability questions) Hi all, I am conducting research and need some help with doing the Fleiss Kappa analysis. First and foremost let me give a bit of a layout of the study. We have a sample of 75 students in the social sciences who were asked to categorize potential items for a new measure based on descriptions of 2 theoretical constructs. This was meant to be a preliminary check to help with construct validity and take it a step beyond just the research team validating items initially. The items are supposed to characterize coping behaviours following romantic loss. The constructs are loss-oriented coping and restoration-oriented coping, however, we also gave the option for participants to say potential items were both types of coping behaviour or something else altogether (basically an 'other' option). Therefore we have 75 participants categorizing items into one of 4 categories. We want to have some measure of the inter-rater reliability of all these participants though. My understanding is that Fleiss Kappa would be an appropriate tool for this purpose, however, if anyone is aware of other statistics which may be a better fit for our model then please let me know. I also need to the syntax for the Fleiss Kappa statistic if anyone has it available. Thanks in advance for any help people can provide. -- View this message in context: http://spssx-discussion.1045642.n5.nabble.com/Fleiss-Kappa-Inter-rater-reliability-questions-tp5717299.html Sent from the SPSSX Discussion mailing list archive at Nabble.com. ===================== To manage your subscription to SPSSX-L, send a message to [hidden email] (not to SPSSX-L), with no body text except the command. To leave the list, send the command SIGNOFF SPSSX-L For a list of commands to manage subscriptions, send the command INFO REFCARD ===================== To manage your subscription to SPSSX-L, send a message to [hidden email] (not to SPSSX-L), with no body text except the command. To leave the list, send the command SIGNOFF SPSSX-L For a list of commands to manage subscriptions, send the command INFO REFCARD |
Administrator
|
Brian, how does your macro compare to this one written by David Nichols?
http://imaging.mrc-cbu.cam.ac.uk/statswiki/FAQ/kappa/multiple Thanks, Bruce
--
Bruce Weaver bweaver@lakeheadu.ca http://sites.google.com/a/lakeheadu.ca/bweaver/ "When all else fails, RTFM." PLEASE NOTE THE FOLLOWING: 1. My Hotmail account is not monitored regularly. To send me an e-mail, please use the address shown above. 2. The SPSSX Discussion forum on Nabble is no longer linked to the SPSSX-L listserv administered by UGA (https://listserv.uga.edu/). |
In reply to this post by bdates
Brian,
I have a related kappa question. I have a case reading instrument with 120 items. Most of the items have three answer choices, but some have two and some have four. There is also one with five. We did an inter-rater reliability study with five readers reading the same case. I haven't looked at the data yet, but I'll need to soon. I was going to use Cohen's Kappa, which SPSS gives you. I was going to set up the data file so that similar items--related topic areas with the same number of answer choices--were treated as separate cases, analogous to a set of MRIs rated by different clinicians, as I've seen in a number of articles. I'm not happy about that because the content for the item may not be similar enough. Also, this approach doesn't allow me to identify items that are troublesome. Can you see a better way of approaching the task? Thanks, Henry -----Original Message----- From: SPSSX(r) Discussion [mailto:[hidden email]] On Behalf Of Dates, Brian Sent: Wednesday, January 09, 2013 4:07 PM To: [hidden email] Subject: Re: Fleiss Kappa (Inter-rater reliability questions) I can send you a macro for Fleiss' kappa, which will give you the overall as well as the category kappa's, based on Fleiss' 1971 Psych. Bulletin article. It uses, however, the updated 1979 formula for variance. How many items will the raters categorize? You have other options, e.g., Krippendorf's alpha and Gwet's AC1 statistic. Although Fleiss is the most used, there's some use of these as well in the literature. If you're interested in Cohen's kappa, there are some multiple rater models, most notably Light (averages all the two-rater kappas) and Hubert (averages the two-rater Pe's). Let me know which you want, and I'll get them off to you. Brian -----Original Message----- From: SPSSX(r) Discussion [mailto:[hidden email]] On Behalf Of BAM86 Sent: Wednesday, January 09, 2013 3:25 PM To: [hidden email] Subject: Fleiss Kappa (Inter-rater reliability questions) Hi all, I am conducting research and need some help with doing the Fleiss Kappa analysis. First and foremost let me give a bit of a layout of the study. We have a sample of 75 students in the social sciences who were asked to categorize potential items for a new measure based on descriptions of 2 theoretical constructs. This was meant to be a preliminary check to help with construct validity and take it a step beyond just the research team validating items initially. The items are supposed to characterize coping behaviours following romantic loss. The constructs are loss-oriented coping and restoration-oriented coping, however, we also gave the option for participants to say potential items were both types of coping behaviour or something else altogether (basically an 'other' option). Therefore we have 75 participants categorizing items into one of 4 categories. We want to have some measure of the inter-rater reliability of all these participants though. My understanding is that Fleiss Kappa would be an appropriate tool for this purpose, however, if anyone is aware of other statistics which may be a better fit for our model then please let me know. I also need to the syntax for the Fleiss Kappa statistic if anyone has it available. Thanks in advance for any help people can provide. -- View this message in context: http://spssx-discussion.1045642.n5.nabble.com/Fleiss-Kappa-Inter-rater-reliability-questions-tp5717299.html Sent from the SPSSX Discussion mailing list archive at Nabble.com. ===================== To manage your subscription to SPSSX-L, send a message to [hidden email] (not to SPSSX-L), with no body text except the command. To leave the list, send the command SIGNOFF SPSSX-L For a list of commands to manage subscriptions, send the command INFO REFCARD ===================== To manage your subscription to SPSSX-L, send a message to [hidden email] (not to SPSSX-L), with no body text except the command. To leave the list, send the command SIGNOFF SPSSX-L For a list of commands to manage subscriptions, send the command INFO REFCARD Confidentiality Notice: This e-mail communication, and any attachments, contains confidential and privileged information for the exclusive use of the recipient(s) named above. If you are not an intended recipient, or the employee or agent responsible to deliver it to an intended recipient, you are hereby notified that you have received this communication in error and that any review, disclosure, dissemination, distribution or copying of it or its contents is prohibited. If you have received this communication in error, please notify me immediately by replying to this message and delete this communication from your computer. Thank you. ===================== To manage your subscription to SPSSX-L, send a message to [hidden email] (not to SPSSX-L), with no body text except the command. To leave the list, send the command SIGNOFF SPSSX-L For a list of commands to manage subscriptions, send the command INFO REFCARD |
Administrator
|
In reply to this post by bdates
Hi Brian,
Would it be possible to post those here in the NG as an attachment? David --
Please reply to the list and not to my personal email.
Those desiring my consulting or training services please feel free to email me. --- "Nolite dare sanctum canibus neque mittatis margaritas vestras ante porcos ne forte conculcent eas pedibus suis." Cum es damnatorum possederunt porcos iens ut salire off sanguinum cliff in abyssum?" |
In reply to this post by BAM86
You might consider adding Krippendorff's alpha as a custom dialog to SPSS: Search for KALPHA on the website linked above and download the kalpha.spd file. The documentation for KALPHA is here: Ryan On Wed, Jan 9, 2013 at 3:24 PM, BAM86 <[hidden email]> wrote: Hi all, |
Right. I know I’ve been in touch with Andy, and all he asks is that he’s cited appropriately. I think the correct citation may even
be in his syntax. B From: SPSSX(r) Discussion [mailto:[hidden email]]
On Behalf Of R B You might consider adding Krippendorff's alpha as a custom dialog to SPSS: Search for KALPHA on the website linked above and download the kalpha.spd file. The documentation for KALPHA is here: Ryan On Wed, Jan 9, 2013 at 3:24 PM, BAM86 <[hidden email]> wrote: Hi all, |
In reply to this post by Bruce Weaver
I used Fleiss' 1979 variance, and I don't know which one ('71 or '79) David used. I added syntax for category kappa's, which Fleiss developed, and an upper and lower 95% confidence interval on all of them. I also report on the number of responses each rater has in each category, the number of agreements each rater has with the other rater(s) in each category, the number of subjects, number of categories, number of raters, and raw proportion of agreement. So there's more information, but the overall kappa's still the same.
B -----Original Message----- From: SPSSX(r) Discussion [mailto:[hidden email]] On Behalf Of Bruce Weaver Sent: Wednesday, January 09, 2013 4:42 PM To: [hidden email] Subject: Re: Fleiss Kappa (Inter-rater reliability questions) Brian, how does your macro compare to this one written by David Nichols? http://imaging.mrc-cbu.cam.ac.uk/statswiki/FAQ/kappa/multiple Thanks, Bruce Dates, Brian wrote > I can send you a macro for Fleiss' kappa, which will give you the overall > as well as the category kappa's, based on Fleiss' 1971 Psych. Bulletin > article. It uses, however, the updated 1979 formula for variance. How > many items will the raters categorize? You have other options, e.g., > Krippendorf's alpha and Gwet's AC1 statistic. Although Fleiss is the most > used, there's some use of these as well in the literature. If you're > interested in Cohen's kappa, there are some multiple rater models, most > notably Light (averages all the two-rater kappas) and Hubert (averages the > two-rater Pe's). Let me know which you want, and I'll get them off to > you. > > Brian > > -----Original Message----- > From: SPSSX(r) Discussion [mailto: > SPSSX-L@.UGA > ] On Behalf Of BAM86 > Sent: Wednesday, January 09, 2013 3:25 PM > To: > SPSSX-L@.UGA > Subject: Fleiss Kappa (Inter-rater reliability questions) > > Hi all, > > I am conducting research and need some help with doing the Fleiss Kappa > analysis. > > First and foremost let me give a bit of a layout of the study. We have a > sample of 75 students in the social sciences who were asked to categorize > potential items for a new measure based on descriptions of 2 theoretical > constructs. This was meant to be a preliminary check to help with > construct > validity and take it a step beyond just the research team validating items > initially. > > The items are supposed to characterize coping behaviours following > romantic > loss. The constructs are loss-oriented coping and restoration-oriented > coping, however, we also gave the option for participants to say potential > items were both types of coping behaviour or something else altogether > (basically an 'other' option). > > Therefore we have 75 participants categorizing items into one of 4 > categories. We want to have some measure of the inter-rater reliability of > all these participants though. My understanding is that Fleiss Kappa would > be an appropriate tool for this purpose, however, if anyone is aware of > other statistics which may be a better fit for our model then please let > me > know. > > I also need to the syntax for the Fleiss Kappa statistic if anyone has it > available. Thanks in advance for any help people can provide. > > > > -- > View this message in context: > http://spssx-discussion.1045642.n5.nabble.com/Fleiss-Kappa-Inter-rater-reliability-questions-tp5717299.html > Sent from the SPSSX Discussion mailing list archive at Nabble.com. > > ===================== > To manage your subscription to SPSSX-L, send a message to > LISTSERV@.UGA > (not to SPSSX-L), with no body text except the > command. To leave the list, send the command > SIGNOFF SPSSX-L > For a list of commands to manage subscriptions, send the command > INFO REFCARD > > ===================== > To manage your subscription to SPSSX-L, send a message to > LISTSERV@.UGA > (not to SPSSX-L), with no body text except the > command. To leave the list, send the command > SIGNOFF SPSSX-L > For a list of commands to manage subscriptions, send the command > INFO REFCARD ----- -- Bruce Weaver [hidden email] http://sites.google.com/a/lakeheadu.ca/bweaver/ "When all else fails, RTFM." NOTE: My Hotmail account is not monitored regularly. To send me an e-mail, please use the address shown above. -- View this message in context: http://spssx-discussion.1045642.n5.nabble.com/Fleiss-Kappa-Inter-rater-reliability-questions-tp5717299p5717302.html Sent from the SPSSX Discussion mailing list archive at Nabble.com. ===================== To manage your subscription to SPSSX-L, send a message to [hidden email] (not to SPSSX-L), with no body text except the command. To leave the list, send the command SIGNOFF SPSSX-L For a list of commands to manage subscriptions, send the command INFO REFCARD ===================== To manage your subscription to SPSSX-L, send a message to [hidden email] (not to SPSSX-L), with no body text except the command. To leave the list, send the command SIGNOFF SPSSX-L For a list of commands to manage subscriptions, send the command INFO REFCARD |
In reply to this post by bdates
In response to how to how to cite his macro, see the top of this document: Ryan On Wed, Jan 9, 2013 at 5:39 PM, Dates, Brian <[hidden email]> wrote:
|
Hello there!
First time poster and very glad I found this mailing list. I'm currently working on my graduation thesis for which I want to use Fleiss Kappa and maybe Krippendorph's alpha to calculate inter-rater-agreement. Would you send me those macros too, if you could be so kind? Thank you very much, Max |
Max,
I sent them out yesterday to the entire listserv, so Fleiss should be in the archives. If you have trouble, contact me offline, and I'll send it directly to you rather than load up the listserv. Andrew Hayes at the Ohio State University has a macro for Krippendorf's alpha. You can download it from http://afhayes.com/spss-sas-and-mplus-macros-and-code.html. The name of the macro is KALPHA. Brian -----Original Message----- From: SPSSX(r) Discussion [mailto:[hidden email]] On Behalf Of Maximilian Koenig Sent: Thursday, January 10, 2013 11:35 AM To: [hidden email] Subject: Re: Fleiss Kappa (Inter-rater reliability questions) Hello there! First time poster and very glad I found this mailing list. I'm currently working on my graduation thesis for which I want to use Fleiss Kappa and maybe Krippendorph's alpha to calculate inter-rater-agreement. Would you send me those macros too, if you could be so kind? Thank you very much, Max -- View this message in context: http://spssx-discussion.1045642.n5.nabble.com/Fleiss-Kappa-Inter-rater-reliability-questions-tp5717299p5717320.html Sent from the SPSSX Discussion mailing list archive at Nabble.com. ===================== To manage your subscription to SPSSX-L, send a message to [hidden email] (not to SPSSX-L), with no body text except the command. To leave the list, send the command SIGNOFF SPSSX-L For a list of commands to manage subscriptions, send the command INFO REFCARD ===================== To manage your subscription to SPSSX-L, send a message to [hidden email] (not to SPSSX-L), with no body text except the command. To leave the list, send the command SIGNOFF SPSSX-L For a list of commands to manage subscriptions, send the command INFO REFCARD |
Administrator
|
Hi Brian,
I did not see the post. Do you have a link to the posting? I didn't see anything with an attachment. David On Thu, Jan 10, 2013 at 12:20 PM, Dates, Brian [via SPSSX Discussion] <[hidden email]> wrote: > Max, > > I sent them out yesterday to the entire listserv, so Fleiss should be in the > archives. If you have trouble, contact me offline, and I'll send it > directly to you rather than load up the listserv. Andrew Hayes at the Ohio > State University has a macro for Krippendorf's alpha. You can download it > from http://afhayes.com/spss-sas-and-mplus-macros-and-code.html. The name > of the macro is KALPHA. > > Brian > > -----Original Message----- > From: SPSSX(r) Discussion [mailto:[hidden email]] On Behalf Of Maximilian > Koenig > Sent: Thursday, January 10, 2013 11:35 AM > To: [hidden email] > Subject: Re: Fleiss Kappa (Inter-rater reliability questions) > > Hello there! > > First time poster and very glad I found this mailing list. I'm currently > working on my graduation thesis for which I want to use Fleiss Kappa and > maybe Krippendorph's alpha to calculate inter-rater-agreement. Would you > send me those macros too, if you could be so kind? > > Thank you very much, > > Max > > > > -- > View this message in context: > http://spssx-discussion.1045642.n5.nabble.com/Fleiss-Kappa-Inter-rater-reliability-questions-tp5717299p5717320.html > > Sent from the SPSSX Discussion mailing list archive at Nabble.com. > > ===================== > To manage your subscription to SPSSX-L, send a message to > [hidden email] (not to SPSSX-L), with no body text except the > command. To leave the list, send the command > SIGNOFF SPSSX-L > For a list of commands to manage subscriptions, send the command > INFO REFCARD > > ===================== > To manage your subscription to SPSSX-L, send a message to > [hidden email] (not to SPSSX-L), with no body text except the > command. To leave the list, send the command > SIGNOFF SPSSX-L > For a list of commands to manage subscriptions, send the command > INFO REFCARD > > > ________________________________ > If you reply to this email, your message will be added to the discussion > below: > http://spssx-discussion.1045642.n5.nabble.com/Fleiss-Kappa-Inter-rater-reliability-questions-tp5717299p5717321.html > To unsubscribe from Fleiss Kappa (Inter-rater reliability questions), click > here. > NAML
Please reply to the list and not to my personal email.
Those desiring my consulting or training services please feel free to email me. --- "Nolite dare sanctum canibus neque mittatis margaritas vestras ante porcos ne forte conculcent eas pedibus suis." Cum es damnatorum possederunt porcos iens ut salire off sanguinum cliff in abyssum?" |
In reply to this post by BAM86
You have had a lot of response about computing kappas, but I have seen
nothing about, How appropriate is this? - I have some doubts about the appropriateness of what you described. First, I will mention that kappas are often problematic when they are not the simple ones on 2x2 tables. My experience is with Cohen's kappa, but it looks like the same problems should exist with Fleiss's. Second, I will say that your "four categories" seem to reduce to two dichotomies -- No/Yes for Loss-oriented, and No/Yes for Restoration- oriented. (The category Other seems to denote the No-No cell, if you draw up the two categories as another 2x2 table.) Therefore -- I think I would restructure this as two analyses done separately for Loss and Restoration. You can look for reliability of high versus low counts on one dimension at a time. In a minimal sense, this would support the reliability - and face validity? - of the "constructs" taken separately. This approach does not yet tackle the "discriminative" ability of the items, if you want to separate the two kinds of coping. It also does not look at the chance that there are additional latent factors in the items that exist for all 75 subjects or for subsets of them. -- Rich Ulrich > Date: Wed, 9 Jan 2013 12:24:32 -0800 > From: [hidden email] > Subject: Fleiss Kappa (Inter-rater reliability questions) > To: [hidden email] > > Hi all, > > I am conducting research and need some help with doing the Fleiss Kappa > analysis. > > First and foremost let me give a bit of a layout of the study. We have a > sample of 75 students in the social sciences who were asked to categorize > potential items for a new measure based on descriptions of 2 theoretical > constructs. This was meant to be a preliminary check to help with construct > validity and take it a step beyond just the research team validating items > initially. > > The items are supposed to characterize coping behaviours following romantic > loss. The constructs are loss-oriented coping and restoration-oriented > coping, however, we also gave the option for participants to say potential > items were both types of coping behaviour or something else altogether > (basically an 'other' option). > > Therefore we have 75 participants categorizing items into one of 4 > categories. We want to have some measure of the inter-rater reliability of > all these participants though. My understanding is that Fleiss Kappa would > be an appropriate tool for this purpose, however, if anyone is aware of > other statistics which may be a better fit for our model then please let me > know. > > I also need to the syntax for the Fleiss Kappa statistic if anyone has it > available. Thanks in advance for any help people can provide. > ... |
Free forum by Nabble | Edit this page |