Conference Agenda

Overview and details of the sessions of this conference. Please select a date or location to show only sessions at that day or location. Please select a single session for detailed view (with abstracts and downloads if available).

Please note that all times are shown in the time zone of the conference. The current conference time is: 17th May 2024, 07:19:24am GMT

 
 
Session Overview
Session
12 SES 05.5 A: General Poster Session
Time:
Wednesday, 23/Aug/2023:
12:15pm - 1:15pm

Location: Gilbert Scott, Hunter Halls [Floor 2]


General Poster Session

Show help for 'Increase or decrease the abstract text size'
Presentations
12. Open Research in Education
Poster

Finding Wood in the Trees: Mapping and Connecting 25 Years of Large Scale Assessment Data

Maximilian Brinkmann1, Nora Huth-Stöckle2, Nakia El-Sayed1

1University of Hildesheim, Germany; 2University of Wuppertal, Germany

Presenting Author: Brinkmann, Maximilian

Large Scale Assessment Data (LSA) is a rich source of data on educational achievement, educational inequalities and other socio-psychological indicators surrounding schooling and learning across the world (e.g. Mullis & Martin, 2017; OECD, 2019). Exemplary are the six PISA cycles between 2000 and 2015 which resulted in over 650 peer-reviewed articles in the English language only (Hopfenbeck et al. 2018, p.337). LSA studies have also been increasingly pooled to create quasi-longitudinal data (i.e. repeated cross-sections of representative student populations) or to track a cohort from primary to secondary school (e.g. by pooling PIRLS and PISA studies). These approaches extend the boundaries of cross-sectional data and allow for closer approximations of causal inference. Productive examples of this research include studies on the effect of educational systems and institutions (e.g. Hanushek & Woessmann, 2006; Teltemann & Schunck, 2020). LSA studies like PISA, TIMSS and PIRLS therefore provide increasing opportunities for secondary and quasi-longitudinal data analysis, covering various research areas.

Yet, the richness and complexity of LSA data can also be a challenge for researchers, since over 16 000 items have been collected in 24 studies over the last 25 years. Our research project aims at helping researchers find their way around the dense jungle of observed items and constructs in LSA studies.

Working with LSA is generally not trivial, since it involves complex sampling structures and (often) so-called plausible values as measures of student achievement. However, another factor is the diversity and sheer number of items which have been collected since the first TIMSS study in 1995. LSA studies provide items from the perspective of students, parents, teachers and principals which are collected by different questionnaires that can vary considerably within (e.g. PISA 2012 to 2015) and between studies (e.g. TIMSS-8 2015 and PISA 2015). According to our own estimations, the items collected by TIMSS, PIRLS and PISA between 1995 and 2019 sum up to roughly 16 000.

The amount of items but also the differences in questionnaires over time and across studies pose obstacles for researchers, especially those who are interested in pooling several cycles and/or studies. In our own experience, pooling several LSA studies is quite challenging partly, because it is unclear from the beginning when and how certain constructs were collected and whether they can be compared across studies. As a consequence, the rich potential of LSA data is still not fully exploited. The aim of our project is to provide a comprehensive database of all items covered in regular cycles of PISA, TIMSS and PIRLS since 1995. The resulting database would include the technical name of the items, the question text, answer categories, official scales, countries covered and rates of missingness.

Most importantly, however, our research attempts to connect similar and comparable items within and across studies in order to facilitate harmonization. This step creates an underlying network structure of our database, in which ties between single items indicate their resemblance to each other. By making our data publicly available, we hope to facilitate further research with LSA data in general and research with pooled (e.g. pseudo-longitudinal) LSA studies in particular.

We believe that such a database is an important contribution to educational research. Whereas several valuable contributions exist that facilitate the use of LSA’s complex sampling structure (e.g. Breit & Schreiner, 2016), plausible values (e.g. Breit & Schreiner, 2016) or survey weights (Jerrim et al., 2017), there is so far very limited help to find a way around the dense jungle of 16 000 LSA survey items.


Methodology, Methods, Research Instruments or Sources Used
The challenge of mapping and connecting 16 000 items can be separated into two steps. The first step involves the compilation of the “data”, hence the items and their meta information. The second step involves matching items based on their similarity.

In the first step, we gathered all questionnaires and codebooks at the  student, parent, teacher and principal level for all 24 studies. Because questionnaires are only available in PDF format, retrieval of information from questionnaires is challenging, since PDF data structure does not correspond to any regular structure representing tables. We therefore employed the new Excel function of Windows 365, which allowed us to transfer questionnaires automatically from PDF to Excel. Retrieving information from the LSA codebooks was even more challenging, since PDF formats changed throughout the years. We therefore collected the data from the actual LSA datasets (i.e. stored variable and value labels) and matched them with our data on question texts.

In the second step we prepare for the matching of similar items. This step is complex because, for instance, questions from the principal’s questionnaires alone sum up to about 4000 items, resulting in a number of roughly 4000²/2 theoretically possible matching combinations. We tested several approaches to handle this task, including machine learning approaches and string matching based on item questions and labels. However, machine learning was ruled out due to the lack of training data and the comparatively short and interpretative data input. We considered string-matching algorithms based on item questions and labels, but initial results turned out to be unsatisfactory.

Therefore, Instead of sophisticated data-wrangling tools, we decided to use an inductive method and exploited the inner structure of LSA surveys. A lot of items or constructs have been designed to reflect specific dimensions of education, organization of education, school life, learning activities and so on. We therefore started to inductively code single items into larger, mutually-exclusive categories. Where feasible, these categories have been further distinguished to represent mutually-exclusive sub-categories. This allows us to reduce the number of possible combinations considerably. So far, we have employed this approach for the principal’s questionnaires and obtained satisfactory results.


Conclusions, Expected Outcomes or Findings
We are still in the process of retrieving the complete LSA data (i.e. teacher data) and preparing the matching of items with each. Initial results of the matching process with the principal's questions indicate that our approach of sorting items into mutually exclusive categories is both practical and effective to match similar items. We expect to have finished the matching process by August and are confident to present first preliminary results.

We are inspired by the goal of providing a public resource for educational researchers. Thus, we are particularly interested in receiving feedback and comments from researchers who worked or wish to work with this kind of data in order to create a public resource that is tailored towards the needs of practitioners.

This project is the base and first part of a larger research project that aims to harmonize items and concepts across existing LSA studies in order to facilitate research with pooled LSA data.

References
Breit, S., & Schreiner, C. (Eds.). (2016). Large-Scale Assessment mit R: methodische Grunglagen der österreichischen Bildungsstandard-Überprüfung. facultas.

Hanushek, E. A., & W ößmann, L. (2006). Does educational tracking affect performance and inequality? Differences‐in‐differences evidence across countries. The Economic Journal, 116(510), C63-C76.
Hopfenbeck, T. N., Lenkeit, J., El Masri, Y., Cantrell, K., Ryan, J., & Baird, J. A. (2018). Lessons learned from PISA: A systematic review of peer-reviewed articles on the programme for international student assessment. Scandinavian Journal of Educational Research, 62(3), 333-353.
Jerrim, J., Lopez-Agudo, L. A., Marcenaro-Gutierrez, O. D., & Shure, N. (2017, June). To weight or not to weight?: the case of PISA data. In Proceedings of the XXVI Meeting of the Economics of Education Association, Murcia, Spain (pp. 29-30).

Mullis, I. V. S., & Martin, M. O. (Eds.). (2017). TIMSS 2019 Assessment Frameworks. Retrieved from Boston College, TIMSS & PIRLS International Study Center website: http://timssandpirls.bc.edu/timss2019/frameworks/
OECD (2019), PISA 2018 Assessment and Analytical Framework, PISA, OECD Publishing, Paris,
https://doi.org/10.1787/b25efab8-en.
Teltemann, J., & Schunck, R. (2020). Standardized Testing, Use of Assessment Data, and Low Reading Performance of Immigrant and Non-immigrant Students in OECD Countries. Frontiers in sociology, 5.


 
Contact and Legal Notice · Contact Address:
Privacy Statement · Conference: ECER 2023
Conference Software: ConfTool Pro 2.6.149+TC
© 2001–2024 by Dr. H. Weinreich, Hamburg, Germany