How to get really smart: Modeling retest and training effects in ability testing using computer-generated figural matrix items

Publikation: Beiträge in ZeitschriftenZeitschriftenaufsätzeForschungbegutachtet

Standard

How to get really smart: Modeling retest and training effects in ability testing using computer-generated figural matrix items. / Freund, Philipp Alexander; Holling, Heinz.
in: Intelligence, Jahrgang 39, Nr. 4, 07.2011, S. 233-243.

Publikation: Beiträge in ZeitschriftenZeitschriftenaufsätzeForschungbegutachtet

Harvard

APA

Vancouver

Bibtex

@article{3f131713a30c4416813a82a38b05ad28,
title = "How to get really smart: Modeling retest and training effects in ability testing using computer-generated figural matrix items",
abstract = "The interpretation of retest scores is problematic because they are potentially affected by measurement and predictive bias, which impact construct validity, and because their size differs as a function of various factors. This paper investigates the construct stability of scores on a figural matrices test and models retest effects at the level of the individual test taker as a function of covariates (simple retest vs. training, use of identical vs. parallel retest forms, and general mental ability). A total of N=189 subjects took two tests of matrix items that were automatically generated according to a strict construction rationale. Between test administrations, participants in the intervention groups received training, while controls did not. The Rasch model fit the data at both time points, but there was a lack of item difficulty parameter invariance across time. Training increased test performance beyond simple retesting, but there was no large difference between the identical and parallel retest forms at the individual level. Individuals varied greatly in how they profited from retest experience, training, and the use of identical vs. parallel retest forms. The results suggest that even with carefully designed tasks, it is problematic to directly compare scores from initial tests and retests. Test administrators should emphasize learning potential instead of state level assessment, and inter-individual differences with regard to test experience should be taken into account when interpreting test results.",
keywords = "Economics, empirical/statistics, Figural matrix items, Individual change, Rational item construction, Retest effects, Training effects",
author = "Freund, {Philipp Alexander} and Heinz Holling",
year = "2011",
month = jul,
doi = "10.1016/j.intell.2011.02.009",
language = "English",
volume = "39",
pages = "233--243",
journal = "Intelligence",
issn = "0160-2896",
publisher = "Elsevier Ltd",
number = "4",

}

RIS

TY - JOUR

T1 - How to get really smart: Modeling retest and training effects in ability testing using computer-generated figural matrix items

AU - Freund, Philipp Alexander

AU - Holling, Heinz

PY - 2011/7

Y1 - 2011/7

N2 - The interpretation of retest scores is problematic because they are potentially affected by measurement and predictive bias, which impact construct validity, and because their size differs as a function of various factors. This paper investigates the construct stability of scores on a figural matrices test and models retest effects at the level of the individual test taker as a function of covariates (simple retest vs. training, use of identical vs. parallel retest forms, and general mental ability). A total of N=189 subjects took two tests of matrix items that were automatically generated according to a strict construction rationale. Between test administrations, participants in the intervention groups received training, while controls did not. The Rasch model fit the data at both time points, but there was a lack of item difficulty parameter invariance across time. Training increased test performance beyond simple retesting, but there was no large difference between the identical and parallel retest forms at the individual level. Individuals varied greatly in how they profited from retest experience, training, and the use of identical vs. parallel retest forms. The results suggest that even with carefully designed tasks, it is problematic to directly compare scores from initial tests and retests. Test administrators should emphasize learning potential instead of state level assessment, and inter-individual differences with regard to test experience should be taken into account when interpreting test results.

AB - The interpretation of retest scores is problematic because they are potentially affected by measurement and predictive bias, which impact construct validity, and because their size differs as a function of various factors. This paper investigates the construct stability of scores on a figural matrices test and models retest effects at the level of the individual test taker as a function of covariates (simple retest vs. training, use of identical vs. parallel retest forms, and general mental ability). A total of N=189 subjects took two tests of matrix items that were automatically generated according to a strict construction rationale. Between test administrations, participants in the intervention groups received training, while controls did not. The Rasch model fit the data at both time points, but there was a lack of item difficulty parameter invariance across time. Training increased test performance beyond simple retesting, but there was no large difference between the identical and parallel retest forms at the individual level. Individuals varied greatly in how they profited from retest experience, training, and the use of identical vs. parallel retest forms. The results suggest that even with carefully designed tasks, it is problematic to directly compare scores from initial tests and retests. Test administrators should emphasize learning potential instead of state level assessment, and inter-individual differences with regard to test experience should be taken into account when interpreting test results.

KW - Economics, empirical/statistics

KW - Figural matrix items

KW - Individual change

KW - Rational item construction

KW - Retest effects

KW - Training effects

UR - http://www.scopus.com/inward/record.url?scp=79957661124&partnerID=8YFLogxK

U2 - 10.1016/j.intell.2011.02.009

DO - 10.1016/j.intell.2011.02.009

M3 - Journal articles

VL - 39

SP - 233

EP - 243

JO - Intelligence

JF - Intelligence

SN - 0160-2896

IS - 4

ER -

DOI

Zuletzt angesehen

Publikationen

  1. Problem solving in mathematics education
  2. Can measurement errors explain variance in the relationship between muscle- and tendon stiffness and range of motion?—a blinded reliability and objectivity study
  3. A Study on the Performance of Adaptive Neural Networks for Haze Reduction with a Focus on Precision
  4. Performance incentives in activity-based management
  5. Implementation of formative assessment
  6. Within-individual leaf trait variation increases with phenotypic integration in a subtropical tree diversity experiment
  7. The role of task meaning on output in groups
  8. The effect of structural complexity on large mammal occurrence in revegetation
  9. Analysis of a phase‐field finite element implementation for precipitation
  10. Walk counts, labyrinthicity, and complexity of acyclic and cyclic graphs and molecules.
  11. Putting adaptive planning into practice: A meta-analysis of current applications
  12. Baseflow recession and recharge as nonlinear storage processes
  13. Applying Quarter-Vehicle Model Simulation for Road Elevation Measurements Utilizing the Vehicle Level Sensor
  14. Tree mixtures mediate negative effects of introduced tree species on bird taxonomic and functional diversity
  15. What is normal?
  16. Multinomial choice models based on Archimedean copulas
  17. As cast microstructures on the mechanical and corrosion behaviour of ZK40 modified with Gd and Nd additions
  18. Learning through evaluation
  19. Coupling ordination techniques and GAM to spatially predict vegetation assemblages along a climatic gradient in an ENSO-affected region of extremely high climate variability
  20. Media Review: Extrapolations - A View from OS4F
  21. Integration of Sustainability into Universities - Good Practices and Benchmarking for Integration

Presse / Medien

  1. Rio+20