Cargando…

Evaluation of Inter-Observer Reliability of Animal Welfare Indicators: Which Is the Best Index to Use?

SIMPLE SUMMARY: In order to be effective, on-farm welfare assessment protocols should always rely on reliable, as well as valid and feasible, indicators. Inter-observer reliability refers to the extent to which two or more observers are observing and recording data in the same way. The present study...

Descripción completa

Detalles Bibliográficos
Autores principales: Giammarino, Mauro, Mattiello, Silvana, Battini, Monica, Quatto, Piero, Battaglini, Luca Maria, Vieira, Ana C. L., Stilwell, George, Renna, Manuela
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2021
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8157558/
https://www.ncbi.nlm.nih.gov/pubmed/34069942
http://dx.doi.org/10.3390/ani11051445
_version_ 1783699708711010304
author Giammarino, Mauro
Mattiello, Silvana
Battini, Monica
Quatto, Piero
Battaglini, Luca Maria
Vieira, Ana C. L.
Stilwell, George
Renna, Manuela
author_facet Giammarino, Mauro
Mattiello, Silvana
Battini, Monica
Quatto, Piero
Battaglini, Luca Maria
Vieira, Ana C. L.
Stilwell, George
Renna, Manuela
author_sort Giammarino, Mauro
collection PubMed
description SIMPLE SUMMARY: In order to be effective, on-farm welfare assessment protocols should always rely on reliable, as well as valid and feasible, indicators. Inter-observer reliability refers to the extent to which two or more observers are observing and recording data in the same way. The present study focuses on the problem of assessing inter-observer reliability in the case of dichotomous (e.g., yes/no) welfare indicators and the presence of two observers, in order to decide about the inclusion of indicators in welfare assessment protocols. We compared the performance of the most popular currently available agreement indexes. Some widely used indexes showed their inappropriateness to evaluate the inter-observer reliability when the agreement between observers was high. Other less used indexes, such as Bangdiwala’s [Formula: see text] or Gwet’s [Formula: see text] , were found to perform better and are therefore suggested to assess the inter-observer reliability of dichotomous indicators. ABSTRACT: This study focuses on the problem of assessing inter-observer reliability (IOR) in the case of dichotomous categorical animal-based welfare indicators and the presence of two observers. Based on observations obtained from Animal Welfare Indicators (AWIN) project surveys conducted on nine dairy goat farms, and using udder asymmetry as an indicator, we compared the performance of the most popular agreement indexes available in the literature: Scott’s [Formula: see text] , Cohen’s [Formula: see text] , [Formula: see text] , Holsti’s [Formula: see text] , Krippendorff’s [Formula: see text] , Hubert’s [Formula: see text] , Janson and Vegelius’ [Formula: see text] , Bangdiwala’s [Formula: see text] , Andrés and Marzo’s [Formula: see text] , and Gwet’s [Formula: see text]. Confidence intervals were calculated using closed formulas of variance estimates for [Formula: see text] , [Formula: see text] , [Formula: see text]   [Formula: see text]   [Formula: see text] , [Formula: see text] , [Formula: see text] , and [Formula: see text] , while the bootstrap and exact bootstrap methods were used for all the indexes. All the indexes and closed formulas of variance estimates were calculated using Microsoft Excel. The bootstrap method was performed with R software, while the exact bootstrap method was performed with SAS software. [Formula: see text] , [Formula: see text] and [Formula: see text] exhibited a paradoxical behavior, showing unacceptably low values even in the presence of very high concordance rates. [Formula: see text] and [Formula: see text] showed values very close to the concordance rate, independently of its value. Both bootstrap and exact bootstrap methods turned out to be simpler compared to the implementation of closed variance formulas and provided effective confidence intervals for all the considered indexes. The best approach for measuring IOR in these cases is the use of [Formula: see text] or [Formula: see text] , with bootstrap or exact bootstrap methods for confidence interval calculation.
format Online
Article
Text
id pubmed-8157558
institution National Center for Biotechnology Information
language English
publishDate 2021
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-81575582021-05-28 Evaluation of Inter-Observer Reliability of Animal Welfare Indicators: Which Is the Best Index to Use? Giammarino, Mauro Mattiello, Silvana Battini, Monica Quatto, Piero Battaglini, Luca Maria Vieira, Ana C. L. Stilwell, George Renna, Manuela Animals (Basel) Article SIMPLE SUMMARY: In order to be effective, on-farm welfare assessment protocols should always rely on reliable, as well as valid and feasible, indicators. Inter-observer reliability refers to the extent to which two or more observers are observing and recording data in the same way. The present study focuses on the problem of assessing inter-observer reliability in the case of dichotomous (e.g., yes/no) welfare indicators and the presence of two observers, in order to decide about the inclusion of indicators in welfare assessment protocols. We compared the performance of the most popular currently available agreement indexes. Some widely used indexes showed their inappropriateness to evaluate the inter-observer reliability when the agreement between observers was high. Other less used indexes, such as Bangdiwala’s [Formula: see text] or Gwet’s [Formula: see text] , were found to perform better and are therefore suggested to assess the inter-observer reliability of dichotomous indicators. ABSTRACT: This study focuses on the problem of assessing inter-observer reliability (IOR) in the case of dichotomous categorical animal-based welfare indicators and the presence of two observers. Based on observations obtained from Animal Welfare Indicators (AWIN) project surveys conducted on nine dairy goat farms, and using udder asymmetry as an indicator, we compared the performance of the most popular agreement indexes available in the literature: Scott’s [Formula: see text] , Cohen’s [Formula: see text] , [Formula: see text] , Holsti’s [Formula: see text] , Krippendorff’s [Formula: see text] , Hubert’s [Formula: see text] , Janson and Vegelius’ [Formula: see text] , Bangdiwala’s [Formula: see text] , Andrés and Marzo’s [Formula: see text] , and Gwet’s [Formula: see text]. Confidence intervals were calculated using closed formulas of variance estimates for [Formula: see text] , [Formula: see text] , [Formula: see text]   [Formula: see text]   [Formula: see text] , [Formula: see text] , [Formula: see text] , and [Formula: see text] , while the bootstrap and exact bootstrap methods were used for all the indexes. All the indexes and closed formulas of variance estimates were calculated using Microsoft Excel. The bootstrap method was performed with R software, while the exact bootstrap method was performed with SAS software. [Formula: see text] , [Formula: see text] and [Formula: see text] exhibited a paradoxical behavior, showing unacceptably low values even in the presence of very high concordance rates. [Formula: see text] and [Formula: see text] showed values very close to the concordance rate, independently of its value. Both bootstrap and exact bootstrap methods turned out to be simpler compared to the implementation of closed variance formulas and provided effective confidence intervals for all the considered indexes. The best approach for measuring IOR in these cases is the use of [Formula: see text] or [Formula: see text] , with bootstrap or exact bootstrap methods for confidence interval calculation. MDPI 2021-05-18 /pmc/articles/PMC8157558/ /pubmed/34069942 http://dx.doi.org/10.3390/ani11051445 Text en © 2021 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Giammarino, Mauro
Mattiello, Silvana
Battini, Monica
Quatto, Piero
Battaglini, Luca Maria
Vieira, Ana C. L.
Stilwell, George
Renna, Manuela
Evaluation of Inter-Observer Reliability of Animal Welfare Indicators: Which Is the Best Index to Use?
title Evaluation of Inter-Observer Reliability of Animal Welfare Indicators: Which Is the Best Index to Use?
title_full Evaluation of Inter-Observer Reliability of Animal Welfare Indicators: Which Is the Best Index to Use?
title_fullStr Evaluation of Inter-Observer Reliability of Animal Welfare Indicators: Which Is the Best Index to Use?
title_full_unstemmed Evaluation of Inter-Observer Reliability of Animal Welfare Indicators: Which Is the Best Index to Use?
title_short Evaluation of Inter-Observer Reliability of Animal Welfare Indicators: Which Is the Best Index to Use?
title_sort evaluation of inter-observer reliability of animal welfare indicators: which is the best index to use?
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8157558/
https://www.ncbi.nlm.nih.gov/pubmed/34069942
http://dx.doi.org/10.3390/ani11051445
work_keys_str_mv AT giammarinomauro evaluationofinterobserverreliabilityofanimalwelfareindicatorswhichisthebestindextouse
AT mattiellosilvana evaluationofinterobserverreliabilityofanimalwelfareindicatorswhichisthebestindextouse
AT battinimonica evaluationofinterobserverreliabilityofanimalwelfareindicatorswhichisthebestindextouse
AT quattopiero evaluationofinterobserverreliabilityofanimalwelfareindicatorswhichisthebestindextouse
AT battaglinilucamaria evaluationofinterobserverreliabilityofanimalwelfareindicatorswhichisthebestindextouse
AT vieiraanacl evaluationofinterobserverreliabilityofanimalwelfareindicatorswhichisthebestindextouse
AT stilwellgeorge evaluationofinterobserverreliabilityofanimalwelfareindicatorswhichisthebestindextouse
AT rennamanuela evaluationofinterobserverreliabilityofanimalwelfareindicatorswhichisthebestindextouse