Rahmenprogramm des BMBF zur Förderung der empirischen Bildungsforschung

Literaturdatenbank

Vollanzeige

    Pfeil auf den Link... Verfügbarkeit 
Autoren Funke, Linda; Kähler, Jana; Hahn, Inga  
Institution Leibniz-Institut für Bildungsverläufe  
Titel NEPS technical report for science: Scaling results of starting cohort 3 in 6th grade.  
URL http://nbn-resolving.de/urn:nbn:de:0111-pedocs-127138  
URN, persistent urn:nbn:de:0111-pedocs-127138  
Erscheinungsjahr 2016  
Seitenzahl 28 S.  
Verlag Bamberg: Leibniz Institute for Educational Trajectories  
Dokumenttyp Monographie; online  
Sprache englisch  
Forschungsschwerpunkt Bildungspanel (NEPS)  
Schlagwörter Erhebungsinstrument; Langzeituntersuchung; Längsschnittuntersuchung; Skalierung; Testreliabilität; Kompetenzmessung; Item-Response-Theorie; Testentwicklung; Schuljahr 06; Wissenschaftsverständnis; Wissenschaftliches Arbeiten; Wissenschaftliches Denken; Lebensspanne; Kompetenzentwicklung; Qualitätssicherung; Skalenkonstruktion; Deutschland  
Abstract The National Educational Panel Study (NEPS) aims at investigating the development of competences across the whole life span and designs tests for assessing these different competence domains. In order to evaluate the quality of the competence tests, a wide range of analyses have been performed based on item response theory (IRT). This paper describes the data on scientific literacy for starting cohort 3 in grade 6. Besides presenting descriptive statistics for the data, the scaling model applied to estimate competence scores and analyses performed to investigate the quality of the scale as well as the results of these analyses are also explained. The science test in grade 6 originally consisted of 27 multiple choice and complex multiple choice items and covered two knowledge domains as well as three different contexts. The test was administered to 4,871 students. A Partial Credit Model was used for scaling the data. Item fit statistics, differential item functioning, Rasch-homogeneity, and the tests' dimensionality were evaluated to ensure the quality of the test. Two items had to be eliminated due to insufficient item discrimination. The results of the remaining 25 items illustrate good item fit values and measurement invariance across various subgroups. Moreover, the test showed a high reliability. As the correlations between the two knowledge domains are very high in a multidimensional model, the assumption of unidimensionality seems adequate. Among the challenges of this test is the lack of very difficult items. But overall, the results emphasize the good psychometric properties of the science test, thus supporting the estimation of reliable scientific literacy scores. In this paper, the data available in the Scientific Use File are described and the ConQuest-Syntax for scaling the data is provided. (IPN/Orig.).  
Förderkennzeichen 01GJ0888