How to get really smart: Modeling retest and training effects in ability testing using computer-generated figural matrix items

Publikation: Beiträge in ZeitschriftenZeitschriftenaufsätzeForschungbegutachtet

Standard

How to get really smart: Modeling retest and training effects in ability testing using computer-generated figural matrix items. / Freund, Philipp Alexander; Holling, Heinz.
in: Intelligence, Jahrgang 39, Nr. 4, 07.2011, S. 233-243.

Publikation: Beiträge in ZeitschriftenZeitschriftenaufsätzeForschungbegutachtet

Harvard

APA

Vancouver

Bibtex

@article{3f131713a30c4416813a82a38b05ad28,
title = "How to get really smart: Modeling retest and training effects in ability testing using computer-generated figural matrix items",
abstract = "The interpretation of retest scores is problematic because they are potentially affected by measurement and predictive bias, which impact construct validity, and because their size differs as a function of various factors. This paper investigates the construct stability of scores on a figural matrices test and models retest effects at the level of the individual test taker as a function of covariates (simple retest vs. training, use of identical vs. parallel retest forms, and general mental ability). A total of N=189 subjects took two tests of matrix items that were automatically generated according to a strict construction rationale. Between test administrations, participants in the intervention groups received training, while controls did not. The Rasch model fit the data at both time points, but there was a lack of item difficulty parameter invariance across time. Training increased test performance beyond simple retesting, but there was no large difference between the identical and parallel retest forms at the individual level. Individuals varied greatly in how they profited from retest experience, training, and the use of identical vs. parallel retest forms. The results suggest that even with carefully designed tasks, it is problematic to directly compare scores from initial tests and retests. Test administrators should emphasize learning potential instead of state level assessment, and inter-individual differences with regard to test experience should be taken into account when interpreting test results.",
keywords = "Economics, empirical/statistics, Figural matrix items, Individual change, Rational item construction, Retest effects, Training effects",
author = "Freund, {Philipp Alexander} and Heinz Holling",
year = "2011",
month = jul,
doi = "10.1016/j.intell.2011.02.009",
language = "English",
volume = "39",
pages = "233--243",
journal = "Intelligence",
issn = "0160-2896",
publisher = "Elsevier Ltd",
number = "4",

}

RIS

TY - JOUR

T1 - How to get really smart: Modeling retest and training effects in ability testing using computer-generated figural matrix items

AU - Freund, Philipp Alexander

AU - Holling, Heinz

PY - 2011/7

Y1 - 2011/7

N2 - The interpretation of retest scores is problematic because they are potentially affected by measurement and predictive bias, which impact construct validity, and because their size differs as a function of various factors. This paper investigates the construct stability of scores on a figural matrices test and models retest effects at the level of the individual test taker as a function of covariates (simple retest vs. training, use of identical vs. parallel retest forms, and general mental ability). A total of N=189 subjects took two tests of matrix items that were automatically generated according to a strict construction rationale. Between test administrations, participants in the intervention groups received training, while controls did not. The Rasch model fit the data at both time points, but there was a lack of item difficulty parameter invariance across time. Training increased test performance beyond simple retesting, but there was no large difference between the identical and parallel retest forms at the individual level. Individuals varied greatly in how they profited from retest experience, training, and the use of identical vs. parallel retest forms. The results suggest that even with carefully designed tasks, it is problematic to directly compare scores from initial tests and retests. Test administrators should emphasize learning potential instead of state level assessment, and inter-individual differences with regard to test experience should be taken into account when interpreting test results.

AB - The interpretation of retest scores is problematic because they are potentially affected by measurement and predictive bias, which impact construct validity, and because their size differs as a function of various factors. This paper investigates the construct stability of scores on a figural matrices test and models retest effects at the level of the individual test taker as a function of covariates (simple retest vs. training, use of identical vs. parallel retest forms, and general mental ability). A total of N=189 subjects took two tests of matrix items that were automatically generated according to a strict construction rationale. Between test administrations, participants in the intervention groups received training, while controls did not. The Rasch model fit the data at both time points, but there was a lack of item difficulty parameter invariance across time. Training increased test performance beyond simple retesting, but there was no large difference between the identical and parallel retest forms at the individual level. Individuals varied greatly in how they profited from retest experience, training, and the use of identical vs. parallel retest forms. The results suggest that even with carefully designed tasks, it is problematic to directly compare scores from initial tests and retests. Test administrators should emphasize learning potential instead of state level assessment, and inter-individual differences with regard to test experience should be taken into account when interpreting test results.

KW - Economics, empirical/statistics

KW - Figural matrix items

KW - Individual change

KW - Rational item construction

KW - Retest effects

KW - Training effects

UR - http://www.scopus.com/inward/record.url?scp=79957661124&partnerID=8YFLogxK

U2 - 10.1016/j.intell.2011.02.009

DO - 10.1016/j.intell.2011.02.009

M3 - Journal articles

VL - 39

SP - 233

EP - 243

JO - Intelligence

JF - Intelligence

SN - 0160-2896

IS - 4

ER -

DOI

Zuletzt angesehen

Publikationen

  1. Evaluating OWL 2 reasoners in the context of checking entity-relationship diagrams during software development
  2. Supporting discourse in a synchronous learning environment
  3. Optimizing price levels in e-commerce applications with respect to customer lifetime values
  4. Topic Embeddings – A New Approach to Classify Very Short Documents Based on Predefined Topics
  5. Changing Data Collection Methods Means Different Kind of Data
  6. Constraints are the solution, not the problem
  7. Practical guide to SAP Netweaver PI-development
  8. Q-Adaptive Control of the nonlinear dynamics of the cantilever-sample system of an Atomic Force Microscope
  9. Representation for interactive exercises
  10. On the Difficulty of Forgetting
  11. Effects of an expressive writing intervention (EWI) with women treated for breast cancer explored with recurrence quantification analysis (RQA) of changes in text structure - a proof-of-concept study
  12. Modeling Efficient Grounding in Chat-based CSCL
  13. Legitimizing Digital Transformation: From System Integration to Platformization
  14. What can we learn from bargaining models about union power?
  15. Differences in the sophistication of Value-based Management
  16. Implementation of a Blended-Learning Course as Part of Faculty Development
  17. How people explain their own and others’ behavior:
  18. How numeric advice precision affects advice taking
  19. The Creation of the Concept through the Interaction of Philosophy with Science and Art
  20. Learning and Re-learning in Chat-based CSCL
  21. Leaf trait variation within individuals mediates the relationship between tree species richness and productivity
  22. Embedded, not plugged-in
  23. Learning in Real-World Laboratories: A Systematic Impulse for Discussion