How to get really smart: Modeling retest and training effects in ability testing using computer-generated figural matrix items

Research output: Journal contributionsJournal articlesResearchpeer-review

Standard

How to get really smart: Modeling retest and training effects in ability testing using computer-generated figural matrix items. / Freund, Philipp Alexander; Holling, Heinz.
In: Intelligence, Vol. 39, No. 4, 07.2011, p. 233-243.

Research output: Journal contributionsJournal articlesResearchpeer-review

Harvard

APA

Vancouver

Bibtex

@article{3f131713a30c4416813a82a38b05ad28,
title = "How to get really smart: Modeling retest and training effects in ability testing using computer-generated figural matrix items",
abstract = "The interpretation of retest scores is problematic because they are potentially affected by measurement and predictive bias, which impact construct validity, and because their size differs as a function of various factors. This paper investigates the construct stability of scores on a figural matrices test and models retest effects at the level of the individual test taker as a function of covariates (simple retest vs. training, use of identical vs. parallel retest forms, and general mental ability). A total of N=189 subjects took two tests of matrix items that were automatically generated according to a strict construction rationale. Between test administrations, participants in the intervention groups received training, while controls did not. The Rasch model fit the data at both time points, but there was a lack of item difficulty parameter invariance across time. Training increased test performance beyond simple retesting, but there was no large difference between the identical and parallel retest forms at the individual level. Individuals varied greatly in how they profited from retest experience, training, and the use of identical vs. parallel retest forms. The results suggest that even with carefully designed tasks, it is problematic to directly compare scores from initial tests and retests. Test administrators should emphasize learning potential instead of state level assessment, and inter-individual differences with regard to test experience should be taken into account when interpreting test results.",
keywords = "Economics, empirical/statistics, Figural matrix items, Individual change, Rational item construction, Retest effects, Training effects",
author = "Freund, {Philipp Alexander} and Heinz Holling",
year = "2011",
month = jul,
doi = "10.1016/j.intell.2011.02.009",
language = "English",
volume = "39",
pages = "233--243",
journal = "Intelligence",
issn = "0160-2896",
publisher = "Elsevier Ltd",
number = "4",

}

RIS

TY - JOUR

T1 - How to get really smart: Modeling retest and training effects in ability testing using computer-generated figural matrix items

AU - Freund, Philipp Alexander

AU - Holling, Heinz

PY - 2011/7

Y1 - 2011/7

N2 - The interpretation of retest scores is problematic because they are potentially affected by measurement and predictive bias, which impact construct validity, and because their size differs as a function of various factors. This paper investigates the construct stability of scores on a figural matrices test and models retest effects at the level of the individual test taker as a function of covariates (simple retest vs. training, use of identical vs. parallel retest forms, and general mental ability). A total of N=189 subjects took two tests of matrix items that were automatically generated according to a strict construction rationale. Between test administrations, participants in the intervention groups received training, while controls did not. The Rasch model fit the data at both time points, but there was a lack of item difficulty parameter invariance across time. Training increased test performance beyond simple retesting, but there was no large difference between the identical and parallel retest forms at the individual level. Individuals varied greatly in how they profited from retest experience, training, and the use of identical vs. parallel retest forms. The results suggest that even with carefully designed tasks, it is problematic to directly compare scores from initial tests and retests. Test administrators should emphasize learning potential instead of state level assessment, and inter-individual differences with regard to test experience should be taken into account when interpreting test results.

AB - The interpretation of retest scores is problematic because they are potentially affected by measurement and predictive bias, which impact construct validity, and because their size differs as a function of various factors. This paper investigates the construct stability of scores on a figural matrices test and models retest effects at the level of the individual test taker as a function of covariates (simple retest vs. training, use of identical vs. parallel retest forms, and general mental ability). A total of N=189 subjects took two tests of matrix items that were automatically generated according to a strict construction rationale. Between test administrations, participants in the intervention groups received training, while controls did not. The Rasch model fit the data at both time points, but there was a lack of item difficulty parameter invariance across time. Training increased test performance beyond simple retesting, but there was no large difference between the identical and parallel retest forms at the individual level. Individuals varied greatly in how they profited from retest experience, training, and the use of identical vs. parallel retest forms. The results suggest that even with carefully designed tasks, it is problematic to directly compare scores from initial tests and retests. Test administrators should emphasize learning potential instead of state level assessment, and inter-individual differences with regard to test experience should be taken into account when interpreting test results.

KW - Economics, empirical/statistics

KW - Figural matrix items

KW - Individual change

KW - Rational item construction

KW - Retest effects

KW - Training effects

UR - http://www.scopus.com/inward/record.url?scp=79957661124&partnerID=8YFLogxK

U2 - 10.1016/j.intell.2011.02.009

DO - 10.1016/j.intell.2011.02.009

M3 - Journal articles

VL - 39

SP - 233

EP - 243

JO - Intelligence

JF - Intelligence

SN - 0160-2896

IS - 4

ER -

Recently viewed

Publications

  1. Textkohäsion und deren Bedeutung für das Textverständnis: Wie reagieren Lernende auf temporale Kohäsion am Beispiel eines Sachtextes?
  2. It Is Belief in Dualism, and Not Free Will, That Best Predicts Helping: A Conceptual Replication and Extension of Baumeister et al. (2009)
  3. Religionsunterricht weiterdenken. Innovative Ansätze für eine zukunftsfähige Religionsdidaktik, hg. v. Bernhard Grümme / Manfred L. Pirner
  4. Integriertes Flächenmanagement überschwemmungsgefährdeter Gebiete an der niedersächsischen Elbe unter Verwendung OGC-konfomer Services
  5. Wie unterstützt eine Schulinspektion die Entwicklung schulischer Qualität? Aufbereitung von Daten zur Ergebnisrückmeldung in der Praxis.
  6. The Influence of Personality on Students‘ Career Decisiveness – A Comparison between Chinese and German Economics and Management Student
  7. Nachhaltige Entwicklung durch gesellschaftliche Partizipation und Kooperation? – eine kritische Revision zentraler Theorien und Konzepte
  8. Die Notwendigkeit der Verhandlung widersprüchlicher Anforderungen an das Lehren von Chemie an einer inklusiven Schule – Eine Fallstudie
  9. Die UN-Dekade "Bildung für nachhaltige Entwicklung" als Plattform für das Thema biologische Vielfalt. Empirische Daten und Erfolgsfaktoren
  10. Stakeholder engagement in Water Framework Directive planning in the United Kingdom: Two case studies from Northern Ireland and Scotland
  11. Fingerspitzengefühl inklusiv(e) - Schiedsrichtertätigkeit im inklusiven Wettkampfsport am Beispiel der Handballinitiative Freiwurf Hamburg
  12. Die Ausstrahlung des Europäischen Gemeinschaftsrechts auf die Rechtsordnungen der Beitrittskandidaten am Beispiel des Wettbewerbsrechts