Observer ratings of instructional quality: Do they fulfill what they promise?

Research output: Journal contributionsJournal articlesResearchpeer-review

Standard

Observer ratings of instructional quality: Do they fulfill what they promise? / Praetorius, Anna Katharina; Lenske, Gerlinde; Helmke, Andreas.
In: Learning and Instruction, Vol. 22, No. 6, 12.2012, p. 387-400.

Research output: Journal contributionsJournal articlesResearchpeer-review

Harvard

APA

Vancouver

Praetorius AK, Lenske G, Helmke A. Observer ratings of instructional quality: Do they fulfill what they promise? Learning and Instruction. 2012 Dec;22(6):387-400. doi: 10.1016/j.learninstruc.2012.03.002

Bibtex

@article{9d18b635daca4003ad112e8652ec348a,
title = "Observer ratings of instructional quality: Do they fulfill what they promise?",
abstract = "Despite considerable interest in the topic of instructional quality in research as well as practice, little is known about the quality of its assessment. Using generalizability analysis as well as content analysis, the present study investigates how reliably and validly instructional quality is measured by observer ratings. Twelve trained raters judged 57 videotaped lesson sequences with regard to aspects of domain-independent instructional quality. Additionally, 3 of these sequences were judged by 390 untrained raters (i.e., student teachers and teachers). Depending on scale level and dimension, 16-44% of the variance in ratings could be attributed to instructional quality, whereas rater bias accounted for 12-40% of the variance. Although the trained raters referred more often to aspects considered essential for instructional quality, this was not reflected in the reliability of their ratings. The results indicate that observer ratings should be treated in a more differentiated manner in the future.",
keywords = "Generalizability theory, Instructional quality, Observer ratings, Reliability, Validity, Educational science",
author = "Praetorius, {Anna Katharina} and Gerlinde Lenske and Andreas Helmke",
year = "2012",
month = dec,
doi = "10.1016/j.learninstruc.2012.03.002",
language = "English",
volume = "22",
pages = "387--400",
journal = "Learning and Instruction",
issn = "0959-4752",
publisher = "Netherlands : Elsevier Science",
number = "6",

}

RIS

TY - JOUR

T1 - Observer ratings of instructional quality

T2 - Do they fulfill what they promise?

AU - Praetorius, Anna Katharina

AU - Lenske, Gerlinde

AU - Helmke, Andreas

PY - 2012/12

Y1 - 2012/12

N2 - Despite considerable interest in the topic of instructional quality in research as well as practice, little is known about the quality of its assessment. Using generalizability analysis as well as content analysis, the present study investigates how reliably and validly instructional quality is measured by observer ratings. Twelve trained raters judged 57 videotaped lesson sequences with regard to aspects of domain-independent instructional quality. Additionally, 3 of these sequences were judged by 390 untrained raters (i.e., student teachers and teachers). Depending on scale level and dimension, 16-44% of the variance in ratings could be attributed to instructional quality, whereas rater bias accounted for 12-40% of the variance. Although the trained raters referred more often to aspects considered essential for instructional quality, this was not reflected in the reliability of their ratings. The results indicate that observer ratings should be treated in a more differentiated manner in the future.

AB - Despite considerable interest in the topic of instructional quality in research as well as practice, little is known about the quality of its assessment. Using generalizability analysis as well as content analysis, the present study investigates how reliably and validly instructional quality is measured by observer ratings. Twelve trained raters judged 57 videotaped lesson sequences with regard to aspects of domain-independent instructional quality. Additionally, 3 of these sequences were judged by 390 untrained raters (i.e., student teachers and teachers). Depending on scale level and dimension, 16-44% of the variance in ratings could be attributed to instructional quality, whereas rater bias accounted for 12-40% of the variance. Although the trained raters referred more often to aspects considered essential for instructional quality, this was not reflected in the reliability of their ratings. The results indicate that observer ratings should be treated in a more differentiated manner in the future.

KW - Generalizability theory

KW - Instructional quality

KW - Observer ratings

KW - Reliability

KW - Validity

KW - Educational science

UR - http://www.scopus.com/inward/record.url?scp=84865576473&partnerID=8YFLogxK

U2 - 10.1016/j.learninstruc.2012.03.002

DO - 10.1016/j.learninstruc.2012.03.002

M3 - Journal articles

AN - SCOPUS:84865576473

VL - 22

SP - 387

EP - 400

JO - Learning and Instruction

JF - Learning and Instruction

SN - 0959-4752

IS - 6

ER -