In: 0

A procedure to improve the credibility of the data by comparing the independent observations of two or more people from the same event. The IOA is calculated by calculating the number of agreements between independent observers and divided by the total number of agreements plus disagreements. The coefficient is then multiplied by 100 to calculate the percentage (%) Consent. J. Cohen: Cohen. A coefficient of agreement for nominal scales. Educational and psychological measure, 20, 37-46. Langenbucher, J., Labouvie, E., Morgenstern, J. (1996).

Methodological evolution: measurement of the diagnostic agreement. Journal of Consulting and Clinical Psychology, 64, 1285-1289. Berk, R. A. (1979). Generalization of behavioural observations: a clarification of the Interobserver agreement and the reliability of the inter-observer. American Journal of Mental Deficiency, 83, 460-472. Behaviouralists have developed a sophisticated method for assessing changes in behaviour that depend on accurate measurement of behaviour. Direct observation of behaviour is traditionally one of the carriers of behavioural measurement. Therefore, researchers need to address psychometric properties, such as the .B the Interobserver Agreement, of observational measures to ensure a reliable and valid measurement.

Of the many indices of the Interobserver agreement, the percentage of the agreement is the most popular. Its use persists despite repeated reminders and empirical evidence that suggests that it is not the most psychometric statistic that determines interobserver agreement because of its inability to take into account chance. Cohens Kappa (1960) has long been proposed as a more psychometric statistic for the evaluation of the Interobserver agreement. Kappa is described and calculation methods are presented. Shrout, P. E., Spitzer, R. L., Fleiss, J. L. (1987). Comment: Quantification of compliance in the resumed psychiatric diagnosis. Archives of General Psychiatry, 44, 172-178.

Suen, H. K., Lee, P.S. (1985). Impact of the use of a percentage agreement on behavioural observation: a reassessment. Journal of Psychopathology and Behavioral Assessment, 7, 221-234. Hartmann, D. P. (1977, Spring). Reflections in the choice of the reliability estimates of inter-observers. Journal of Applied Behavior Analysis, 10, 103-116. Landis, J.

R., Koch, G. G. (1977). The measure of the compliance agreement for categorical data. Biometrics, 33, 159-174. Fiss, J. L. (1971). Measure of the scale rated correspondence between many advisors. Psychological Bulletin, 76, 378-382. Dunn, G., Everitt, B. (1995).

Clinical Biostatistics: An Introduction to Evidence-Based Medicine. Edward Arnold in London. Your email address will not be published. The required fields are marked in Gresham, F.M. (1998). Designs to assess changes in behaviour. In T. S. Watson – F.M. Gresham (Eds.), Child Behaviour Manual.

NY: Plenum. This is an overview of the content of the subscription, log in to check access. The definition reproduced with the permission of Amanda N. Kelly at Behaviorbabe. Save my name, email address and website in this browser for the next time I will make a comment. Nelson, L.D., Cicchetti, D. V. (1995). Emotional function assessment in people with brain disease. Psychological assessment, 7, 404-413. Ciminero, A.

R., Calhoun, K. S., Adams, H. E. (Eds.). (1986). Behavioural Assessment Manual (2. New York: Wiley. Wasik, B. H., Loven, M.

D. (1980). Classroom observational data: sources of inaccuracy and proposed solutions. Behavioral Assessment, 2, 211-227. Cicchetti, D. V. (1994). Guidelines, criteria and ground rules for evaluating standardized and standardized assessment instruments in psychology.

Psychological assessment, 6, 284-290. Reliable data is data that gives the same results to each measurement.