Conference Agenda

Overview and details of the sessions of this conference. Please select a date or location to show only sessions at that day or location. Please select a single session for detailed view (with abstracts and downloads if available).

Please note that all times are shown in the time zone of the conference. The current conference time is: 17th May 2024, 05:43:47am GMT

 
 
Session Overview
Session
09 SES 09 B: Advancing Assessment Methods and Insights for Education Systems
Time:
Thursday, 24/Aug/2023:
9:00am - 10:30am

Session Chair: Stefan Johansson
Location: Gilbert Scott, 253 [Floor 2]

Capacity: 40 persons

Paper Session

Show help for 'Increase or decrease the abstract text size'
Presentations
09. Assessment, Evaluation, Testing and Measurement
Paper

Measuring and Misrepresenting the Missing Millions: the OECD’s Assessment of out-of-School Youth in PISA for Development

Xiaomin Li

Beijing Normal University, China, People's Republic of

Presenting Author: Li, Xiaomin

As the education agenda of global agencies changed after 2015 to emphasise minimum standards of quality for all countries to be delivered by 2030, the OECD has sought to expand its most successful comparative instrument, the Programme for International Student Assessment (PISA), to include low- and middle-income countries (LMICs). In 2014, it introduced PISA for Development (PISA-D) as the means to establish PISA as a universal measure of learning, and in 2020, it declared PISA-D a success. The most innovative feature of PISA-D was that the assessment included out-of-school youth (OOSY); that task was sub-contracted to Educational Testing Service (ETS). The OOSY population is a geographically dispersed group which present considerable challenges to any researchers seeking to access them. Given this, we ask: who did the OECD assess? More specifically, how did the OECD define the target population of PISA-D out-of-school assessment, what was the sampling frame, and were they accurately represented in the PISA-D OOSY sample?

Much of the existing literature on the OECD influence is based on different theoretical positions. These differences in perspective have real consequences, however, often determining which legitimation dynamics researchers see and which they overlook. In this paper, we seek to adopt a holistic approach by drawing on Suchman’s (1995) framework for analysing the multiple sources of organisational legitimacy and the means by which it is promoted and repaired.

In applying Suchman’s framework, we argue that PISA-D was a macro level exercise designed to legitimate the OECD’s extension of PISA into LMICs and to establish its role in a new arena. The incorporation of OOSY in the assessment was a key micro level endeavour which would allow the OECD to achieve that end and, if not done properly, would challenge aspects of its legitimacy. For example, not assessing sufficient OOSY would debase the quality of the OECD’s products and services; this would also damage the OECD’s moral claims with regard to monitoring the attainment of the SDGs and promoting an inclusive approach. In parallel, at the cognitive level, this would challenge the whole logic of the novelty and value of PISA-D. Overall, the successful identification and assessment of OOSY was vital to ensuring its legitimacy. This would require the OECD to either address the considerable difficulties of accessing OOSY or find a tactical solution which obscured the many challenges to its legitimacy.

Suchman (1995) also analyses how organisations respond to challenges to their legitimacy and identifies three broad approaches: (a) offer normalising accounts; (b) restructure, and (c) don’t panic. He suggested that although legitimacy crises may coalesce around performance issues, most challenges ultimately rest on failures of meaning, where ‘audiences begin to suspect that putatively desirable outputs are hazards, that putatively efficacious procedures are tricks, or that putatively genuine structures are facades’ (1995, 597). Consequently, the initial task in mending a breach of legitimacy usually will be ‘to formulate a normalising account’ that separates the threatening revelation from larger assessments of the organisation as a whole. He identified ‘justifications’ and ‘explanations’ as the two main types of normalising accounts. Suchman also noted that organisations may also re-establish legitimacy through micro-level strategic restructuring, in the sense that ‘narrowly tailored changes that mesh with equally focused normalising accounts can serve as effective damage-containment techniques’ (ibid., 598).


Methodology, Methods, Research Instruments or Sources Used
To understand what challenges the OECD encountered and how it managed to address them, we firstly draw on two categories of documents: the first are the UNICEF and UNESCO Institute for Statistics (UIS) publications on the Out-of-school Children Initiative (OOSCI) and Lewin’s (2011) work as part of the Consortium for Research on Educational Access, Transitions and Equity (CREATE) initiative which provide the standard approaches to identifying OOSY and describing their characteristics which PISA-D draws upon. The second are the OECD publications which explain the PISA-D out-of-school sample design and selection plans , and which present the final results . We also draw on three interviews with: a key member of the PISA-D team at the OECD, a technical expert who had undertaken OOSY surveys, and a lead analyst from one of the piloting nations.
Conclusions, Expected Outcomes or Findings
We argue that, as an organisation with no experience with assessing OOSY and working in poorer nations, the OECD was faced by a ‘disruptive event’ (Suchman 1995) as it was unable to effectively sample youth based on their initial definition. This event, if not addressed immediately, would interrupt its ongoing PISA-D legitimation activities and may severely deplete its long-term legitimacy. In line with Suchman’s (1995) analysis of how organisations respond to challenges to their legitimacy, we demonstrate that the OECD pursued a normalising strategy by tailoring and justifying how OOSY were defined and by minimising coverage of its tactical changes. Consequently, it avoided addressing the many problems which face researchers on OOSY by quietly imposing a sampling frame which matched its available sources of data and established methodologies.
The analysis builds on our earlier work which identified the broader strategies that the OECD employed to create the legitimacy to monitor SDG 4 (Li and Morris 2022) and extends it by focusing on legitimacy maintenance and repair work. It also contributes to the important work of others who have critiqued the validity and impact of various assessments undertaken by global agencies.

References
Addey, Camilla. 2017. ‘Golden Relics & Historical Standards: How the OECD Is Expanding Global Education Governance through PISA for Development’. Critical Studies in Education 0 (0): 1–15. https://doi.org/10.1080/17508487.2017.1352006.
Auld, Euan, and Paul Morris. 2021. ‘A NeverEnding Story: Tracing the OECD’s Evolving Narratives within a Global Development Complex’. Globalisation, Societies and Education 19 (2): 183–97. https://doi.org/10.1080/14767724.2021.1882959.
Berten, John, and Matthias Kranke. 2019. ‘Studying Anticipatory Practices of International Organisations: A Framework for Analysis’. Framing Paper for Workshop on Anticipatory Governance at 6th European Workshops in International Studies. Kraków.
Carr-Hill, Roy. 2013. ‘Missing Millions and Measuring Development Progress’. World Development 46 (June): 30–44. https://doi.org/10.1016/j.worlddev.2012.12.017.
Grek, Sotiria. 2009. ‘Governing by Numbers: The PISA “Effect” in Europe’. Journal of Education Policy 24 (1): 23–37. https://doi.org/10.1080/02680930802412669.
Grey, Sue, and Paul Morris. 2018. ‘PISA: Multiple “Truths” and Mediatised Global Governance’. Comparative Education 54 (2): 109–31. https://doi.org/10.1080/03050068.2018.1425243.
Lewin, Keith. 2011. ‘Making Rights Realities: Researching Educational Access, Transitions and Equity’. Project Report. Brighton: University of Sussex. http://www.create-rpc.org/pdf_documents/Making-Rights-Realities-Keith-Lewin-September-2011.pdf.
Li, Xiaomin, and Euan Auld. 2020. ‘A Historical Perspective on the OECD’s “Humanitarian Turn”: PISA for Development and the Learning Framework 2030’. Comparative Education 56 (4): 503–21. https://doi.org/10.1080/03050068.2020.1781397.
Hamilton, Mary. 2017. ‘How International Large-Scale Skills Assessments Engage with National Actors: Mobilising Networks through Policy, Media and Public Knowledge’. Critical Studies in Education 58 (3): 280–94. https://doi.org/10.1080/17508487.2017.1330761.
Martens, Kerstin. 2007. ‘How to Become an Influential Actor - The “comparative Turn” in OECD Education Policy’. In New Arenas of Education Governance: The Impact of International Organisations and Markets on Educational Policy Making, edited by Kerstin Martens, Alessandra Rusconi, and Kathrin Leuze. Basingstoke: Macmillan.
Li, Xiaomin, and Paul Morris. 2022. ‘Generating and Managing Legitimacy: How the OECD Established Its Role in Monitoring Sustainable Development Goal 4’. Compare: A Journal of Comparative and International Education 0 (0): 1–18. https://doi.org/10.1080/03057925.2022.2142038.
Robertson, Susan L. 2020. ‘Guardians of the Future: International Organisations, Anticipatory Governance and Education’. Presented at the International Webinar on UNESCO’s and OECD’s Ambition to Govern the Future of Education, Copenhagen, April 23.
Suchman, Mark C. 1995. ‘Managing Legitimacy: Strategic and Institutional Approaches’. The Academy of Management Review 20 (3): 571–610. https://doi.org/10.2307/258788.
Zapp, Mike. 2020. ‘The Authority of Science and the Legitimacy of International Organisations: OECD, UNESCO and World Bank in Global Education Governance’. Compare: A Journal of Comparative and International Education, 1–20. https://doi.org/10.1080/03057925.2019.1702503.


09. Assessment, Evaluation, Testing and Measurement
Paper

Tinkering towards an assessment of Global Competence

Harsha Chandir, Radhika Gorur, Jill Blackmore

Deakin University, Australia

Presenting Author: Blackmore, Jill

PISA has become the “world’s premier yardstick” against which the “quality, equity and efficiency” of national education systems are evaluated (Gurría in OECD, 2018a, p. 2). PISA claims to be able to compare, on a single scale, the performance of education systems around the globe. These comparative measures have “contributed to the constituting of a global commensurate space of educational performance” (Rizvi & Lingard, 2010, p. 135), regardless of the varying political, economic, social, and cultural contexts of participating nations. Data from PISA are being used by countries to identify “gaps” in their education systems and to develop policies to “move up” on the league tables (Meyer & Benavot, 2013; Wiseman, 2013). A recent development in this space has been the OECD’s development of an internationally comparable measure of Global Competence.

The use of PISA measures as benchmarks for shaping global policies and governing education makes it important to examine the process of how “PISA knowledge” is arrived at. Developing a set of measures requires normative decisions about what the concept encompasses. The assessment of global competence provides a useful example of examining the development of this particular form of global knowledge. Given the multifaceted definitions and understandings of this term, this paper empirically examines the challenges such efforts at stabilising the definition faced, and the ways in which these were negotiated. Locating our study in the interdisciplinary field of Science and Technology Studies (STS), and deploying the concept of tinkering (Knorr Cetina, 1981), we attend to the practices that stabilised the assessment of Global Competence in PISA 2018.

To make globally acceptable knowledge, various epistemic, cultural and political perspectives are brought together in relations of mutual learning and construction, and through iterative processes of expert consultation, country feedback, committee endorsement, etc. These encounters, where diverse perspectives are brought together, have the potential to be hijacked by more outspoken or forceful participants. Moreover, these processes typically take several months during which a range of unexpected events may occur or challenges posed to the successful completion of the endeavour. Tinkering is the way in which actors and events are managed – through cajoling, placating, compromising, modifying, etc. to ensure that the project does not collapse.

Drawing on empirical data relating to the development of the assessment of global competencies, we provide examples tinkering in the development of PISA’s tests of global competency. We highlight three key tinkering moves by the OECD during the process of developing the assessment. In the first move, the OECD replaced the initial Global Competence Expert Group with another group of experts to placate the PISA Governing Board, which objected to the heave economic slant of the first expert presentation. In the second tinkering move, the OECD retrospectively aligned the PISA assessments of global competency with the global competence framework with the UN SDGs. This enabled the OECD to gather more (of the right) allies to support its efforts, and provided “a moral legitimacy the OECD has not enjoyed with the traditional PISA initiative and its narrow economic focus” (Auld & Morris 2019b. p.11). A third tinkering move was the push by the OECD to administer the assessment even when only a minority of the countries decided to participate, arguing that more nations might join subsequent rounds.


Methodology, Methods, Research Instruments or Sources Used
This paper offers data from publicly available documents as well as semi-structured interviews with key OECD officials and members of PISA 2018’s Global Competence Expert Group, to highlight three tinkering moves. By tracing the practices of the assessment development, this study aims to understand how a particular ontology of global competence was stabilised in PISA 2018.
Conclusions, Expected Outcomes or Findings
By enrolling experts and considering feedback from countries, PISA can be said to be a collaborative and democratic global process. However, a closer examination reveals that in spaces when there is uncertainty, and when stalemates develop between different groups, decision making lies with the OECD which tinkers to steer actors in ways that primarily benefit the organisation’s pre-determined agenda.
Tracing this process of making global knowledge of global competence allows for an exploration of “which kind of society and which idea of humanity is pursued and enacted” (d'Agnese, 2018, p. 16) in the OECD’s assessments of global competence – and more generally the PISA project. As the OECD attempts to develop other “global” measures of literacies (OECD, 2018b), it is important to open up the politics of their production. By putting centre-stage the controversies and negotiations, the processes that stabilise these assessments can be opened up to critical scrutiny.

References
Auld, E., & Morris, P. (2019a). Science by streetlight and the OECD’s measure of global competence: A new yardstick for internationalisation? Policy Futures in Education, 17(6), 677-698. https://doi.org/10.1177/1478210318819246
Auld, E., & Morris, P. (2019b). The OECD’s Assessment of Global Competence: Measuring and making global elites. In L. C. Engel, C. Maxwell, & M. Yemini (Eds.), The Machinery of School Internationalisation in Action (pp. 17-35). Routledge
d'Agnese, V. (2018). Reclaiming education in the age of PISA: Challenging OECD’s educational order. Routledge.
Knorr Cetina, K. (1981). The manufacture of knowledge: An essay on the constructivist and contextual nature of science. Pergamon.
Meyer, H. D., & Benavot, A. (Eds.). (2013). PISA, power, and policy: The emergence of global educational governance. Symposium books
Organisation for Economic Co-operation and Development. (2018a). PISA 2015 results in focus. https://www.oecd.org/pisa/pisa-2015-results-in-focus.pdf
Organisation for Economic Co-operation and Development. (2018b). The future of education and skills 2030: The future we want. https://www.oecd.org/education/2030/E2030%20Position%20Paper%20(05.04.2018).pdf
Rizvi, F., & Lingard, B. (2010). Globalizing education policy. Routledge
Wiseman, A. W. (2013). Policy responses to PISA in comparative perspective. In H. D. Meyer & A. Benavot (Eds.), PISA, power, and policy: The emergence of global educational governance (pp. 303-322). Symposium books.


09. Assessment, Evaluation, Testing and Measurement
Paper

Are Students Underachieving in PISA? The Issue of Test Motivation in Low-Stakes and High-Stakes Tests

Linda Borger1, Stefan Johansson1, Rolf Strietholt2

1University of Gothenburg, Sweden; 2Technische Universität Dortmund, Germany

Presenting Author: Borger, Linda

International large-scale assessments (ILSAs) are playing an increasingly important role in decision-making and reforms, both nationally and internationally (e.g. Grek, 2009; Lindblad et al., 2018). One of the most influential ILSAs is the Programme for International Student Assessment (PISA). Given the impact PISA has on educational debate and policy, it is crucial that results are trustworthy. Yet, parallel to an increase in the number of ILSAs, there has been growing validity concerns regarding for example the content being tested, the influence on national educational systems and potential bias due to lack of sample representativeness (Grek, 2009; Jerrim, 2021; Meyer & Benavot, 2013). Relatively few studies, however, have focused on whether students are motivated to do their best in ILSAs as compared with high-stakes tests.

A motive for our research is international evidence suggesting that tests of low stakes impact student motivation and effort (Finn, 2015; Wise & DeMars, 2005). Whereas the relation between high- and low-stakes testing has been studied previously, findings are inconsistent, and we know little about this relationship in Sweden. Therefore, the following study examines whether there is evidence for the hypothesis that the lack of personal consequences may bias PISA test scores downwards. Indeed, self-reports from Swedish students indicate that they do not do their best in PISA (Eklöf & Hopfenbeck, 2019). However, PISA test scores have not yet been compared to external criteria such as national test scores. The theoretical framework used to interpret the results of the present study is the expectancy-value theory (Eccles & Wigfield, 2002; Wigfield & Eccles, 2000), postulating that test motivation depends on the student's expectations of succeeding at a particular task, the value the student places on the task, and the interaction between the two (Eccles & Wigfield, 2002). The expectancy-value theory has successfully been used in previous studies to explain the test-taking motivation construct (e.g., Eklöf & Knekta, 2017).

Previous research on the relationship between national tests and PISA/TIMSS revealed moderate to high but imperfect correlations (Skolverket, 2022; Wiberg, 2019; Wiberg & Rolfsman, 2019). One possible explanation is that ILSAs have low stakes for students while national tests have high stakes. In order to test the assumption that motivation influences student achievement, we will examine whether test motivation moderates the relationship between PISA scores and the national test scores. Skolverket (2022) found a correlation of .61 between the two measures but our hypothesis is that the relationship is different for different levels of motivation to take the PISA test. With reference to the expectancy-value theory we assume that the average level of test motivation is higher for national tests since this is a test with higher stakes. For students that were particularly unmotivated to do the PISA test, the correlation with their national test scores could therefore be lower. Consequently, the study examines the following research questions: (1) What is the correlation between test motivation in PISA and PISA achievement? and (2) Is the relationship between low-stakes PISA test scores and high-stakes national test scores moderated by students’ test motivation in PISA?


Methodology, Methods, Research Instruments or Sources Used
In the 2018 Swedish PISA test, students’ personal identification numbers were collected, making it possible to link PISA tests scores with register data on students’ national test grades and student background characteristics, collected from Statistics Sweden (SCB). The analyses are based on this combined dataset, including a sample of 5,504 students. The main method used was latent moderated structural equations modelling. The outcome variable is students’ PISA achievement, measured through the ten plausible values by including the type = imputation option in Mplus, the software used. Since reading was the major domain in PISA 2018, the analyses focus on reading. However, robustness checks were conducted using PISA achievement in mathematics and science.

The predictors used are students’ motivation to take the PISA test, formulated as a latent variable and used as moderator in the interaction analysis, and students’ national test grade. The latent variable PISA_Motivation is measured by six statements about students’ motivation in PISA, answered on a four-point Likert scale ranging from “strongly agree” to “strongly disagree” (reverse-coded in the analyses). The scale is provided as a national option in the PISA student questionnaire and contains items intended to measure effort, e.g., “I felt motivated to do my best in the PISA test” and importance, e.g., “Doing well in the PISA test was important to me”. Cronbach's alpha for the PISA motivation scale was .90 for the six items, indicating a high internal consistency. As an indicator of a high-stakes assessment, the students’ national test grade in reading, ranging from A–F and coded numerically, was used as an observed independent variable. Student background characteristics will be used as control variables in further analyses.

In a first step, a measurement model of PISA_Motivation was estimated using confirmatory factor analysis (CFA), and model fit was ensured. Subsequently, structural models were estimated in consecutive steps (Muthén, 2012), starting with models without latent interaction, and then including both main effects and the latent interaction in the final model. The independent observed variable (national test grade) was centered prior to analysis. Model fit was evaluated using commonly used fit indices for structural equation modelling (Marsh et al, 2005). Models were estimated using MLR, and the complex option in Mplus was employed to account for the nested data structure. Analyses were weighted using the final student weight. Missing data was treated under the default method in Mplus (Full Information Maximum Likelihood).

Conclusions, Expected Outcomes or Findings
Results revealed a significant positive correlation between PISA_Motivation and PISA achievement (r = .15), indicating that test motivation predicts achievement. In line with Skolverket (2022), the correlation between PISA achievement in reading and the national test grade in reading was found to be around .6. When controlling for students’ reading ability, in the form of the grade on the high-stakes national test, PISA_Motivation still significantly and positively influenced PISA achievement. In the final model, a significant positive interaction was shown between PISA_Motivation and the national test grade (β = .05, p < .001), indicating that students’ motivation in PISA affects the strength of the relationship between the high-stakes national test grade and the low-stakes PISA achievement.

Graphical analyses of the interaction effects for students with different motivational levels showed that the simple slope differed particularly for students who indicated a low level of motivation in PISA and who received high grades on the high-stakes national test. The students with low motivation in PISA thus had a lower correlation between their PISA test score and their national test grade than the students who reported high motivation. This could be explained, in accordance with the expectancy-value theory, by the fact that these students put in less effort in PISA than on the national test because they do not see PISA as important to them personally. In sum, the study provides some evidence that the low-stakes nature of PISA may bias test scores for certain groups of students, in particular high achievers on the national test with low reported motivation in PISA. In the discussion, other reasons for the discrepancy between PISA test scores and national test grades will be addressed, such as differences in content, format and aims. Additionally, problematic aspects of measuring test effort with self-reported measures are considered.

References
Baumert, J., & Demmrich, A. (2001). Test motivation in the assessment of student skills: The effects of incentives on motivation and performance. European Journal of Psychology of Education, 16(3), 441–62. https://doi.org/10.1007/BF03173192

Eccles, J. S., & Wigfield, A. (2002). Motivational beliefs, values, and goals. Annual Review of Psychology, 53(1), 109-132, https://doi.org/10.1146/annurev.psych.53.100901.135153

Eklöf, H. & Knekta, E. (2017). Using large-scale educational data to test motivation theories: A synthesis of findings from Swedish studies on test-taking motivation. International Journal of Quantitative Research in Education, 4(5), 52-71.

Eklöf, H. & Hopfenbeck, T. (2019). Self-reported effort and motivation in the PISA test. In B. Maddox (Red.), International large-scale assessments in education: insider research perspectives (s. 121–136). Bloomsbury Academic.

Finn, B. (2015). Measuring motivation in low-stakes assessments (Research Report No. RR-15-19). Princeton, NJ: Educational Testing Service. doi:10.1002/ets2.12067

Grek, S. (2009). Governing by numbers: the PISA ‘effect’ in Europe. Journal of Education Policy, 24(1), 23-37. https://doi.org/10.1080/02680930802412669

Jerrim, J. (2021). PISA 2018 in England, Northern Ireland, Scotland and Wales: Is the data really representative of all four corners of the UK?. Review of Education, 9(3). https://doi.org/10.1002/rev3.3270

Lindblad, S., Pettersson D., & Popkewitz, T.S. (2018). Numbers, Education and the Making of Society: International Assessments and Its Expertise. Routledge

Marsh, H. W., Hau, K., & Grayson, D. (2005). Goodness of fit evaluation in structural equation modeling. In A. Maydeu-Olivares and J. McArdle (Eds.), Contemporary Psychometrics (pp. 275–340). Erlbaum.

Meyer, H. D., & Benavot, A. O. (Eds.). (2013). PISA, power, policy. The emergence of global educational governance. Oxford Studies in Comparative Education.

Muthén B. (2012). Latent variable interactions. http://www.statmodel.com/download/LV%20Inter action.pdf

Skolverket. (2022). PISA 2018 och betygen. Analys av sambanden mellan svenska betyg och resultat i PISA 2018 [PISA 2018 and school grades. Analyses of the relationship between Swedish school grades and results in PISA 2018]. Skolverket.

Wiberg, M. (2019). The relationship between TIMSS mathematics achievements, grades, and national test scores. Education Inquiry, 10(4), 328-343. https://doi.org/10.1080/20004508.2019.1579626

Wiberg, M., & Rolfsman, E. (2019). The association between science achievement measures in schools and TIMSS science achievements in Sweden. International Journal of Science Education, 41(16), 2218-2232. doi:10.1080/09500693.2019.1666217

Wigfield, A., & Eccles, J. S. (2000). Expectancy-value theory of achievement motivation. Contemporary Educational Psychology, 25, 68–81. https://10.1016/ceps.1999.1015

Wise, S. L., & DeMars, C. E. (2005). Low examinee effort in low-stakes assessment: Problems and potential solutions. Educational Assessment, 10(1), 1–17. https://10.1207/s15326977ea1001_1


09. Assessment, Evaluation, Testing and Measurement
Paper

A Framework to Estimate and Enhance Effectiveness of Large-scale Assessments in Next Generation Learning Systems

Priyanka Sharma, Amit Kaushik

Australian Council for Educational Research, India

Presenting Author: Sharma, Priyanka

The effectiveness of initiatives in an educational context is often interpreted in terms of their impact on learning outcomes for every unit of investment. Governments invest a significantly high amount of money and effort in large-scale assessments (LSAs) with the intent to provide data-driven evidence to policymakers and researchers. Such evidence indicates the quality parameters of the education system in terms of learning level, equity, sustainability, and other predefined dimensions. Validity of such information is paramount due to its crucial role in decision-making for inputs, functional strategies, and goal setting for intended outputs, which if implemented as intended are most likely to lead to improvement in learning outcomes. Therefore, it is not an exaggeration to say that the effectiveness of LSAs can also be interpreted in terms of gain in learning outcomes or other dimensions, like any other measure. However, measuring and demonstrating effectiveness of LSAs remains a challenge due to multiple reasons besides the complexities involved in efficacy and effectiveness research, like the notion that assessment data themselves offer solutions. Authors make a compelling argument that data including assessment data do not provide solutions, rather they assist the policy and research community by providing valid evidence and insights enabling informed policy formulation and implementation decisions. Rich data and information provided by large-scale assessments can further be used to analyze the impact of those policy decisions.

The paper consists of three major parts. The first part reviews existing initiatives and proposes a logic model based on reasoning to estimate the effectiveness through evidence and/or counterevidence. A logic model depicts how an initiative is expected to make a difference, using explicit statements of the activities that are likely to bring about the intermediate changes and the impact the initiative intends to make. The proposed model postulates that if evidence generated by LSAs at T1 point of time were utilized to make appropriate modifications in policy and interventions regarding inputs, processes, organizational functioning, governance, monitoring mechanism, and outputs can lead to lead to learning gains per unit of investment at T2 point of time.

The second part builds on policy research and secondary analyses of large-scale assessments conducted in India to generate insights into the policy and practice that emerged from large-scale assessments. While the study primarily uses the assessment data and information from the National Achievement Survey (NAS) and the Annual Status of Education Report (ASER), it makes an effort to corroborate the findings with International and national LSAs in a similar context.

The third part recommends a policy implementation framework consisting of a series of steps to design system-specific strategies and monitor efforts. These steps are organized into two main phases: i) a ‘diagnostic’ phase to identify priority areas or enabling outcomes; ii) an ‘action’ stage to devise, implement and evaluate concrete policy interventions. The diagnostic stage mainly consists of cost-effective action-oriented surveys with a tiered approach, while the action stage consists of evidence-driven developmental goals and action plans for various levels of the system, alignment between all actors involved, customized interventions at school level with continuous monitoring in the cycle of 'assess-act-assess'

Education systems around the world have emphasized the need to transform assessments to improve learning. Proposed framework and model may be vital in designing learning systems to improve learning outcomes through effective systemwide assessments. However, there cannot be a one-size-fits-all policy mix. Feasible policy choices depend on contexts, social preferences, and political constraints. A robust and independent institutional framework, stakeholder engagement, and credible communication strategies are vital to enhancing the effectiveness of LSAs and eliminating learning poverty to achieve sustainable development goals.


Methodology, Methods, Research Instruments or Sources Used
The aim of the study was to develop a framework to assess the effectiveness of large-scale assessments, gather evidence of effectiveness, and then recommend an implementation framework to enhance the effectiveness. Accordingly, a mixed research approach was adopted. The methodology of the study has three main components:
1. A literature review of relevant literature on the effectiveness of LSAs, policy initiatives as a result of LSAs, and implementation research in the context of system-level assessments
2. Secondary analysis of ASER and NAS data for the pre-COVID period
3. Drafting a logic model, followed by an implementation framework to utilize the meaningful findings of LSAs to improve quality dimensions of education, based on main findings of the review  
The scope of the literature review was not limited to large-scale assessments in India, but it also covered the role of international LSAs PISA, TIMSS, PIRLS, SACMEQ, PASEC and national LSAs like NAPLAN, and NAEP in educational policies and their impact. Investigators conducted the review along four key components:
• Model of intent and model of change behind LSAs  
• Use of findings of LSAs in the formulation of policy measures
• Framework of planning, implementation and monitoring of the policy initiatives emerged from LSAs  
• Effectiveness studies in LSAs or use of LSAs as a metric in education effectiveness studies
Investigators also undertook secondary analyses of ASER data since 2005 to analyze the cohort relationships associated with learning achievement in basic literacy and numeracy among the learners in the age group corresponding to grades three to eight. ASER is an annual survey report published by the education non-profit Pratham and aims to provide reliable estimates of enrolment and basic learning levels. Basic reading and basic arithmetic abilities are assessed for learners in the age group of 5-16 years. Secondary analyses of NAS data for grade 3, 5 and 8, and learning data of few other countries from UNESCO Institute of Statistics (UIS) were also undertaken. Then triangulation technique was adopted to consolidate the findings.

Conclusions, Expected Outcomes or Findings
The role of LSAs as a tool to improve the quality of education was recognized in 2000 with the launch of the Program for International Student Assessment (PISA) by OECD. This triggered LSAs as policy research in many parts of the world. However, in the past 20 years learning level of students in many countries has either declined or plateaued. Despite spending several years in school, millions of children are unable to achieve basic literacy and numeracy skills (ASER, 2018). More than 50% of primary school children in South Asian nations were in learning poverty even before the COVID-19 pandemic, and this number is projected to be around 80% due to COVID-19- related school closures (World Bank et al., 2022). The report of NAS 2021 has indicated a similar trend (NCERT, 2022).
The review showed that the majority of systems lack a concrete model regarding how LSAs are expected to impact the actions and learning outcomes. Measurement of learning achievement with no follow-up plan of action results in low efficacy of LSA initiatives. Experts have raised an alarm around the deepening learning crisis and recommended three complementary strategies: assess learning in order to measure and track learning better; act on the results or evidence to guide innovation and practice; and, align actors to remove barriers and make the whole system work for learning (World Bank, 2018).  These complementary strategies may be utilized to derive a logic model as a common wireframe for planning, implementation, and monitoring of outcomes.
The proposed tiered approach to assessments to identify priority areas followed by concrete evidence-driven policy interventions and monitoring mechanisms may enable LSAs-driven improvement in learning. The model can assist policymakers and researchers to estimate the impact of stage-specific decisions on outcomes, and disaggregate the impact of individual intermediary enablers on intended outcomes.

References
ASER Centre. (2018). Annual Status of Education Report (Rural) 2018. http://img.asercentre.org/docs/ASER%202018/Release%20Material/aserreport2018.pdf
NCERT (2019). National Achievement Survey 2017. National report to Inform Policy, Practices and Teaching Learning. National Council of Educational Research and Training. Ministry of Education. Government of India. https://nas.gov.in/report-card/2017
NCERT (2022). National Achievement Survey. National Report 2021. National Council of Educational Research and Training. Ministry of Education. Government of India. https://nas.gov.in/report-card/2021
World Bank 2018. World Development Report 2018: Learning to Realize Education’s Promise. Washington, DC: World Bank. doi:10.1596/978-1-4648-1096-1.
World Bank, UNESCO, UNICEF, USAID, FCDO, Bill & Melinda Gates Foundation. (2022). The State of Global Learning Poverty: 2022 Update. https://www.unicef.org/reports/state-global-learning-poverty-2022.
MHRD. (2020). National Education Policy 2020.   https://www.education.gov.in/sites/upload_files/mhrd/files/NEP_Final_English_0.pdf. Ministry of Education (erstwhile Ministry of Human Resource Development). Government of India.


 
Contact and Legal Notice · Contact Address:
Privacy Statement · Conference: ECER 2023
Conference Software: ConfTool Pro 2.6.149+TC
© 2001–2024 by Dr. H. Weinreich, Hamburg, Germany