Conference Agenda

Overview and details of the sessions of this conference. Please select a date or location to show only sessions at that day or location. Please select a single session for detailed view (with abstracts and downloads if available).

Please note that all times are shown in the time zone of the conference. The current conference time is: 17th May 2024, 06:06:54am GMT

 
 
Session Overview
Session
99 ERC SES 04 I: Assessment, Evaluation, Testing and Measurement
Time:
Monday, 21/Aug/2023:
1:30pm - 3:00pm

Session Chair: Klaus Rummler
Location: Wolfson Medical Building, Sem 1 (Yudowitz) [Floor 1]

Capacity: 78 persons

Paper Session

Show help for 'Increase or decrease the abstract text size'
Presentations
99. Emerging Researchers' Group (for presentation at Emerging Researchers' Conference)
Paper

Engaging Diffractive Ethnography To Explore Student and Teacher Perceptions of Collaborative Testing To Enhance Learning and Engage STEM Students.

Helen Bremert

Sydney University, Australia

Presenting Author: Bremert, Helen

Education across the globe, including Europe, the United Kingdom and Australia, has been fuelled by a resurgence of high-stakes testing, exacerbated by outcome-defined policies and competing factors, including the ongoing demands of accountability, performativity, and the rigid alignment between high-stakes exams and curricula (Lingard & Sellar, 2016; Löfstedt et al., 2020; Verger, Parcerisa & Fontdevila, 2019). The performance-based nature of education has turned education into schooling. Kemmis et al. (2014) describe schooling as the opposite of education; it is a ‘technical tool’ (p.25) used to impart and monitor information as outcomes. These influence assessments' objectives, methods, and outcomes, in addition, to teachers' pedagogical assessment practices (Cairns, 2020). Given current political preferences for teachers' accountability and measurable outcomes, students undertake multiple assessments measuring their learning at single points in time, each affecting their depth of understanding, engagement with the subject, and sense of who they are as a student (Andrade & Brookhart, 2020). Further, teachers' pedagogical practices may be modified by the effects of high-stakes testing as they move from collaborative work in the classroom to more 'teach to the test' methods, further reducing student engagement (Resnick & Schantz, 2017). A significant number of studies investigating the advantages of collaborative learning have been undertaken, promoting a constructivist approach to collaboration in the classroom. However, behaviourist methods are still prevalent when assessing students. This discrepancy leads to the question: if collaboration is widely recognized as enhancing learning, why does the reliance on individual summative testing persist? Broader recognition of multiple assessment methods is required to improve engagement in high school STEM classes and abate the documented effects of individual summative testing. Leading to the research questions –

  • How do students perceive collaborative testing?
  • How do students and teachers discern collaborative testing's impact on student learning?
  • What are teachers' perceptions of the utility of collaborative testing?

Perspective

Humans exist in a dynamic world where we are a part of the world and being of that world interact with other humans and non-humans. Participating in the environment entangles or intertwines us with the material and others, all affecting each other as we understand our place, meaning and knowledge of the world (Barad, 2007; Muris, 2022; Plauberg, 2018). Knowledge is not separated from the learner; it is interwoven and interconnected; it affects and is affected by interactions with non-human and human interactions (Barad, 2007; Guillion, 2018).

Teaching science involves a dynamic intra-mingling between students, teachers, surroundings, and our classroom's tools. In this intra-play, students attained their understanding and knowledge of science. They did not do so alone; students did not sit apart, and they did not sit quietly; they participated, interacted and enhanced their understanding of science. These interactions influence our meaning-making (Koro- Ljungberg, 2015). The learner’s thinking and actions act upon the world equally as the thinking of the learner acts on the world. There can be no separation between object or subject, from human and non-human; all are intertwined in the knowing. Barad (2007) termed this interconnecting of theory and knowledge as onto-epistemology.

Onto-epistemology does not separate the object/subject, human/non-human, world/us, knowledge/learning. This theory places equal importance on the material world, ‘matter matters’ (Barad, 2007). Matter is an active participant in the entanglement of meaning, enfolded alongside material and discursive practices; they are constantly reconfiguring to forge reality (Guillion, 2018).


Methodology, Methods, Research Instruments or Sources Used
This paper presents early data from a doctoral study exploring student and teacher perceptions of collaborative testing within collaborative pedagogy and as an addition to current practices of individual, competitive testing. The study design draws on diffractive ethnography to examine (i) teachers' perceptions of the utility of collaborative testing and (ii) students' and teachers' views on the effectiveness of collaborative testing. Additionally, this study examines the effectiveness of assessing students' 21st-century skills while collaboratively testing.
To address this novel approach to testing, the researcher used a multi-phase, collaborative practitioner inquiry method involving eight teachers and the researcher in a reciprocal relationship. The discussion will include qualitative data gathered through interviews, focus groups, audio recordings of student testing groups, observations, and assessment tasks, outlining student and teacher perceptions of the efficacy of this novel assessment method.

Conclusions, Expected Outcomes or Findings
The literature demonstrates a supportive view of the pedagogical value of authentic summative assessments utilising collaborative ideals to benefit student understanding (Reiger & Reiger, 2020). However, almost all research into collaborative testing has been conducted in undergraduate science classes; therefore, this diffractive ethnographic study hopes to broaden understanding and highlight different assessment choices to enhance teachers' pedagogy, practice, and student engagement. Of equal importance to improving teacher assessment strategies, this study will look at the feasibility of using a rubric to assess 21st-century skills such as collaboration, communication and student social skills while students are undertaking collaborative testing.

References
References

Andrade, H. L., & Brookhart, S. M. (2020). Classroom assessment as the co-regulation of learning. Assessment in Education: Principles, Policy & Practice, 27(4), 350-372. doi:10.1080/0969594X.2019.1571992

Barad, K. M. (2007). Meeting the universe halfway quantum physics and the entanglement of matter and meaning. Durham: Duke University Press.

Cairns, R. (2020). Exams tested by Covid-19: An opportunity to rethink standardized senior secondary examinations. PROSPECTS. doi:10.1007/s11125-020-09515-9

Gullion, J. S. (2018). Diffractive ethnography : social sciences and the ontological turn. New York, NY: Routledge.

Kemmis, S., Wilkinson, J., Edwards-Groves, C., Hardy, I., Grootenboer, P., & Bristol, L. (2014). Praxis, Practice and Practice Architectures. In Changing Practices, Changing Education (pp. 25-41). Singapore: Springer Singapore.

Koro-Ljungberg, M. (2016). Reconceptualizing Qualitative Research: Methodologies without Methodology. doi:10.4135/9781071802793

Löfstedt, P., García-Moya, I., Corell, M., Paniagua, C., Samdal, O., Välimaa, R., & Rasmussen, M. (2020). School satisfaction and school pressure in the WHO European region and North America: an analysis of time trends (2002–2018) and patterns of co-occurrence in 32 countries. Journal of adolescent health, 66(6), S59-S69.

Murris, K. (2022). Karen Barad as Educator. In Karen Barad as Educator, Agential Realism and Education (1 ed., pp. XV, 95): Springer Singapore.

Plauborg, H. (2018). Towards an agential realist concept of learning. Subjectivity, 11, 322-338. doi:https://doi.org/10.1057/s41286-018-0059-9

Rieger, G. W., & Rieger, C. L. (2020). Collaborative Assessment That Supports Learning.  J. Mintzes & E. Walter (Eds.), Active Learning in College Science. Springer., https://doi.org/10.1007/978-3-030-33600-4_51

Resnick, L. B., & Schantz, F. (2017). Testing, teaching, learning: who is in charge? Assessment in education: principles, policy & practice, 24(3), 424-432. doi:10.1080/0969594X.2017.1336988


Verger, A., Parcerisa, L., & Fontdevila, C. (2019). The growth and spread of large-scale assessments and test-based accountabilities: A political sociology of global education reforms. Educational Review, 71(1), 5-30.


99. Emerging Researchers' Group (for presentation at Emerging Researchers' Conference)
Paper

Hitting the Mark: Formative Assessment and the Development of Capabilities in a Primary School Context

Valerio Rigo

Libera Università di Bolzano, Italy

Presenting Author: Rigo, Valerio

In December 2020, the Italian Ministry of Education adopted a new assessment modality based on predominantly discursive communication and the abolition of numerical grades. This means was introduced to incentivize greater equality and support students in becoming aware of how they develop their competencies (Girelli, 2022). One of the core elements of this change is the shift from a "normative" way of assessing pupils - that implies the comparison between a single student and the rest of the class - to a criteria-based one. This latter mode of assessment requires the teacher to consider pre-established criteria that help the students better understand their potential (Black, Wiliam, 1998) by indicating the level of achievement of a learning objective. Four fundamental elements must be considered for this purpose: the level of autonomy with which the pupil can achieve the set goals, his or her ability to find resources spontaneously, the fact that learning takes place in an unfamiliar situation, and finally whether the acquisition of the set goals is continuous (i.e. it occurs in more than a single isolated case) or not (MIUR, 2020).

The theoretical framework of this research is the Capability Approach developed by Amartya Sen and Martha Nussbaum (Sen, 1999; Nussbaum, 2006). It is «an intellectual discipline that gives a central role to the evaluation of a person’s achievements and freedoms in terms of his or her actual ability to do the different things a person has reason to value doing or being» (Sen, 2009). The attention devoted to the valorization of non-quantifiable processes in terms of accountability and rethinking the role played by the concept of measurement (Unterhalter, 2019) makes this framework a proper tool for proposing a new look at the evaluation methods currently in use in Italian primary schools. Since the right to education «requires a deep commitment to building human capabilities» and «Education nurtures understandings and builds capabilities that can help to ensure that our futures are more socially inclusive (UNESCO, 2021)», it is necessary to understand whether the learning processes promoted at school can promote the development of such resources.

Therefore, the question from which this research stems is how can formative assessment help students develop their capabilities. To date, few studies within the capability approach have focused on primary schools and formative assessment (Robeyns, 2017; Unterhalter, 2019; Palumbo, Pandolfini, 2019). The Italian context lends itself particularly well to investigating the relationship between these realities since the main document used by teachers to guide their teaching (MIUR, 2012) was formulated by Edgar Morin and the theory of complexity is also a fundamental factor for the capability approach (Szekeley. Mason, 2018).


Methodology, Methods, Research Instruments or Sources Used
The methodology used for the research is based on an action-research approach in three different Italian primary schools. The teachers who will take part in the research were selected within the assessment group of the Movimento di Cooperazione Educativa (MCE-FIMEM). This movement played a crucial role in the formulation of the new assessment system adopted in Italian primary schools (Piscozzo, Stefanel, 2022). The selected teachers will be considered active co-researchers in trying to outline the potential of the capability approach for formative assessment practices, to create a real "community of practice" dedicated to investigating these issues (Wenger, 1998). The pupils taking part in the research project come from second grades of public primary schools in three different regions of Northern Italy. In addition to direct classroom observations and focus groups with the research team of teachers, the research also involves the use of interviews with pupils. This choice was made because both in the capability approach (Robeyns, 2017; Hart, Brando, 2018) and recent literature on formative assessment (Yan, Yang, 2022; Monteiro et al., 2021) the active participation of students appears to be a determining factor.
The research genre adopted is Educational Design-Based Research (McKenney, Reeves, 2020), which aims to develop practical solutions to education problems and theoretical comprehension by cooperating with the stakeholders involved in the research. This genre was chosen because of its versatility in combining theoretical and practical perspectives with a focus on subsequent implementation and spread.

Conclusions, Expected Outcomes or Findings
Considering the early stage of the research, the data collection phase has not yet taken place. One of the main expected outcomes is the rethinking of competence-based teaching. The assessment system adopted in the European Union is strongly oriented towards the development of competencies and skills. The capability approach can make an essential contribution to rethinking assessment processes and practices to make pupils more aware of their potential (Lozano et al., 2012; Palumbo, Pandolfini, 2019; Urbani, 2018).
References
Black, P.J., & Wiliam, D. (1998). Assessment and classroom learning. Assessment in Education: Principles, Policy, and Practice,  5(1), 7-74.
Girelli, C. (2022). Valutare nella scuola primaria. Dal voto al giudizio descrittivo. Roma: Carocci.
Hart, C.S., & Brando, N. (2018). A capability approach to children’s well‐being, agency and participatory rights in education. European Journal of Education, 53, 293-309.
Lozano, J.F., Boni, A., Peris, J., & Hueso, A. (2012). Competencies in Higher Education: A Critical Analysis from the Capabilities Approach. Journal of Philosophy of Education, 46(1), 132-147.
McKenney, S., & Reeves, C. T. (2020). Educational design research: Portraying, conducting, and enhancing productive scholarship. Medical Education, 1-11.
M.I.U.R. (2012). Indicazioni Nazionali. Annali della Pubblica Istruzione, Le Monnier: Firenze.
M.I.U.R. (2020). Linee Guida Ordinanza Ministeriale n. 172/2020.
Monteiro, V., Mata, L., & Santos, N.N. (2021). Assessment Conceptions and Practices: Perspectives of Primary School Teachers and Students. Frontiers in Education, 6.
Otto, H-U., & Ziegler, H. (2006). Capabilities and Education. Social Work & Society, 4(2), 269-287.
Palumbo, M., & Pandolfini, V. (2019). Scuola e disuguaglianze: apprendere dalla valutazione? Sociologia Italiana – AIS Journal of Sociology, 13, 113-132.
Piscozzo, M., & Stefanel, S. (2022). La valutazione nella scuola primaria. Obiettivi, curricoli, scelte. Milano: UTET.
Robeyns, I., (2017). Wellbeing, Freedom and Social Justice: The Capability Approach Re-Examined. Cambridge, UK: Open Book Publishers.
Sen, A. (1999). Development as Freedom, Oxford: Oxford University Press.
Sen, A. (2009). Capability: Reach and Limit. In Debating Global Society: Reach and Limits of the Capability Approach, (pp. 15–28), Milan: Fondazione Giangiacomo Feltrinelli.
Szekely, E., & Mason, M. (2018). Complexity theory, the capability approach, and the sustainability of development initiatives in education, Journal of Education Policy, 1-18.
UNESCO, (2021). Reimagining our Futures Together: A new Social Contract for Education.
Unterhalter, E. (2019). Measuring the Unmeasurable in Education. New York: Routledge.
Urbani, C. (2018). Lo sviluppo professionale docente dalle competenze alla capacitazione. Milano: Franco Angeli.
Wenger, E. (1998). Communities of Practice: Learning, Meaning, and Identity. Cambridge University Press.
Yan, Z., & Yang, L. (2022). Assessment as Learning. Maximizing Opportunities for Student Learning and Achievement. New York: Routledge.


99. Emerging Researchers' Group (for presentation at Emerging Researchers' Conference)
Paper

The Design and Development of a Measurement Scale for Evaluating Teaching Effectiveness from Students' Perspective

Daniel - Emil Iancu

West University of Timișoara, Romania

Presenting Author: Iancu, Daniel - Emil

Student Evaluation of Teaching (SET) is the procedure by which students evaluate and rate teaching performance. Usually, during a SET procedure students complete rating forms or questionnaires about different aspects related to their teachers, but mostly about their teaching practices. Universities or higher education institutions from all over the world implement SET procedures to achieve 3 main purposes. Generally, and from a practical point of view, the main purpose of implementing this type of procedure in most universities is the necessity of reporting SET results to quality assurance agencies. The other main goal of SET procedures, and surely the most important one from an educational perspective, is to provide feedback to academics about their teaching practices and/or to design teacher training programs focused on developing teaching skills. Another important use of SET results is related to evaluating evidence of teaching performance to use the results for academic career advancements or other ways of rewarding teaching effectiveness.

The topic of Student Evaluations of Teaching (SET) is one of the most researched ones in the domain of educational research, with over 2000 studies published in peer-reviewed journals over a period of a little more than 100 years (Spooren et al., 2017). One of the earliest debates in this field of research is about the validity of the SET scales and procedures. The main question was whether the measurement instruments applied to students during these procedures can accurately measure teaching effectiveness. Even if this debate was most active in the 1970s and the evidence was inclining more towards the affirmative answer to the question in case (see reviews from Richardson, 2005 and Marsh, 2007), a recently published meta-analysis (Uttl et al., 2017) presented some evidence which seriously threatens the validity of SET results. The results of the mentioned study strongly suggest that there is no relationship between the SET results of a teacher and the level of their students’ achievement/learning.

Analyzing the SET scales utilized in the studies included in the last-mentioned meta-analysis, we observed that lots of them contain items that are either too vague, either unclear, or even completely unrelated to observable teaching actions and behaviors. Moreover, many SET instruments had only 1 or 2 overall questions about teaching quality.

Some of the main reasons why we chose to develop a new SET scale are the following: (1) most SET scales are not rigorously developed from the theoretical and psychometrical point of view, and those that are, were developed in older times when digital environments were not so common; (2) the majority of SET instruments lack available and published psychometrical evidence for the validity of the results; (3) there are very few instruments that contain questions strictly related to the observable behaviors and actions of teachers; and (4) there are not many SET instruments than can be adapted to different teaching contexts by adapting the selection and number of dimensions to be measured.

Thus, we propose the development and validation of a new multi-dimensional instrument that is based on a systematic mixed-methods process of defining effective teaching in higher education, consists of questions related to the observable actions and behaviors of teachers, and contains a large number of dimensions suitable for several teaching contexts.

Given that the purpose of the study is to develop and validate a SET scale, the research objectives are the following:

1) To perform the theoretical substantiation of the construct and related dimensions.

2) To develop and refine items for each dimension of the scale.

3) To gather evidence in favor of validity (content, construct and criterion) and fidelity (test-retest, internal consistency) of the scale.


Methodology, Methods, Research Instruments or Sources Used
For start, we will carry out an extensive and systematic literature review on the topic of teaching effectiveness and SET measures, to outline a comprehensive definition and dimensionality of the construct of interest. Also to define the construct and identify the dimensions, we will conduct interviews with effective award-winning teachers, with students, and with experts in education, asking them about what an effective teacher does, in terms of actions and behaviors, to facilitate and stimulate learning in higher education.

Once the dimensions are identified, we will define and operationalize them in terms of teaching actions and behaviors that are observable from the student’s perspective. Afterward, we will generate an expanded set of items of 10-15 items for each dimension (considering the dimensions’ operational definitions) and we will also choose an answer scale that reflects the nature of the dimensions and the purpose of the measurement.

Regarding the process of gathering evidence in favor of the new SET scale’s reliability and validity, we will follow the following steps:
1) Assessment of content validity by a panel of 3 experts in teacher training and evaluation, who will evaluate the dimension definitions and the items, for assessing their relevance and coverage of the construct.
2) Pre-piloting the instrument with a small number of participants from the intended population to observe whether the items and answer options are properly interpreted by the respondents and, if necessary, to revise the items according to what was observed.
3) Piloting the instrument on a sample closely representative of the target population and performing descriptive analysis of data, the analysis of the fidelity of the sample (test-retest, internal consistency), and confirmatory factor analysis, to revise and reduce the scale and to refine the items that need refinement.
4) In terms of checking relationships with other constructs, we will use 3 tools that are theoretically related to our construct. To check the convergent validity we will use the ATI (Approaches to Teaching Inventory) (Trigwell, Prosser, and Ginns, 2005) which is a teaching approach evaluation tool. At the same time, we will also use the SEEQ (Students Evaluation of Educational Quality) (Marsh, 1982) which is maybe the most rigorously developed SET scale. In terms of criterion validity, we will analyze the links between our instrument and the R-SPQ-2F (Revised-Study Process Questionnaire-2 Factors) (Biggs, Kember, & Leung, 2001), which refers to the ways in which students approach learning (deep or surface).

Conclusions, Expected Outcomes or Findings
The expected outcome is a new rigorously developed and evidence-based multi-dimensional SET scale for evaluating teaching effectiveness in higher education. We expect to obtain positive evidence of content, construct, and criterion validity and also evidence for the usefulness of the results in improving teaching actions and behaviors. Also, considering the large number of teaching dimensions we want to assess through the new SET scale, we expect that it could be used flexibly in different settings or disciplines, by teachers or SET administrators. In other words, it will offer teachers or SET administrators the opportunity to select and evaluate only the dimensions that are relevant for that specific setting or discipline.

From the perspective of the impact on educational practice and teacher training, the SET scale resulting from this research project will allow the following: (1) giving feedback and support to teachers based on the scores obtained on each dimension; (2) identifying the training needs of teaching staff with the aim of developing training programs aimed at those needs; (3) measuring teaching effectiveness as a result of various initiatives to increase the quality of teaching and reviewing these initiatives based on their effects; and last but not least (4) the use of the results in promotion decisions or for awarding teaching activity, aspects that may lead to greater involvement in teaching improvement initiatives from teachers.

References
American Educational Research Association, American Psychological Association, & National Council on Measurement in Education. (2014). Standards for educational and psychological testing. Washington, DC: American Educational Research Association.

Biggs, J., Kember, D., & Leung, D. Y. (2001). The revised two‐factor study process questionnaire: R‐SPQ‐2F. British Journal of Educational Psychology, 71(1), 133-149.

Marsh, H. W. 1982. “SEEQ: A Reliable, Valid, and Useful Instrument for Collecting Students’ Evaluations of University Teaching.” British Journal of Educational Psychology 52 (1): 77–95.

Marsh, H. W. (2007). Students’ evaluations of university teaching: Dimensionality, reliability, validity, potential biases and usefulness. In P.R., Pintrich & A. Zusho (Coord.), The scholarship of teaching and learning in higher education: An evidence-based perspective (pp. 319-383). Springer, Dordrecht.

Richardson, J. T. (2005). Instruments for obtaining student feedback: A review of the literature. Assessment & evaluation in higher education, 30(4), 387-415.

Spooren, P., Vandermoere, F., Vanderstraeten, R., & Pepermans, K. (2017). Exploring high impact scholarship in research on student's evaluation of teaching (SET). Educational Research Review, 22, 129-141.

Trigwell, K., Prosser, M., & Ginns, P. (2005). Phenomenographic pedagogy and a revised approaches to teaching inventory. Higher Education Research & Development, 24(4), 349-360.

Uttl, B., White, C. A., & Gonzalez, D. W. (2017). Meta-analysis of faculty's teaching effectiveness: Student evaluation of teaching ratings and student learning are not related. Studies in Educational Evaluation, 54, 22-42.

Willis, G. B. (2004). Cognitive interviewing: A tool for improving questionnaire design. Sage Publications.


99. Emerging Researchers' Group (for presentation at Emerging Researchers' Conference)
Paper

Teachers’ Formative Assessment Literacy in the Confucian Culture Context: Design and Validation of an Inventory

Rong Fu

University of Calgary, Canada

Presenting Author: Fu, Rong

Formative assessment places greater emphasis on bidirectional contributions between the students in the assessment process (Yorke, 2003), as well as advocates an equal teacher-student relationship. To promote student learning, both stakeholders are expected to be fully engaged in the learning process (Black & Wiliam, 1998), instead of measuring student performance unilaterally (Wiggins, 1998). Although formative assessment is well established in western classroom practices and is required as one underpinning literacy for teachers (Popham, 2011), when the idea of formative assessment was introduced to Confucian heritage countries in Asia, tensions for K‒12 teachers arose due to the perennial convention of overemphasizing summative assessment in school.

Profoundly influenced by Confucianism, teachers in China may perceive formative assessment as a novel assessment paradigm that contradicts their traditional mindset on assessment. Take Chinese teachers as an example, many of them seem to be reluctant to this reform (Zhan, 2021). Moreover, schooling in these countries has been constantly shifted to virtual online classroom platforms since the COVID-19 pandemic. Given that online classrooms largely constrain the effectiveness of the implementation of assessment (Khan & Khan, 2018), the different classroom modes complicate the teacher formative assessment practices.

However, there is a paucity of research on teachers’ overall literacy of formative assessment in the eastern Confucian heritage countries (Zhan, 2019). There is also a need to design and develop a psychometrically sound and culturally responsive inventory instrument in the field of formative assessment literacy in the Confucian cultural context. Thus, I attempt to design and validate a first easternized online inventory to evaluate teachers’ capacity for formative assessment acquisition and authentic application. The whole process includes four phases under the guide of an exploratory sequential mixed-methods design. Building on the interviews and the prior inventories on relevant domains (e.g., Campbell & Mertler, 2004; Yan & Pastore, 2022), I plan to categorize the dimensions as five scenarios, consisting of 1) face-to-face classroom, 2) online classroom, 3) after-school assignments and tasks, 4) feedback, 5) ethical concerns. Each contains six multiple-choice questions and each question has one best answer. The inventory will be uploaded to www.wjx.cn, a Chinese online open survey website, and automatically code the correct answer as 1 point while the incorrect answers as 0 points. Thus, the total grade for one participant is 30 points and each subsection values 6 points.

In particular, I am fully aware of the unique characteristics of Confucian heritage culture and its ubiquitous impact on teacher assessment practices. Therefore, all the questions are edited in line with the authentic local classroom settings. For example, I will emphasize student dignity considerations to replace the typical western inventory questions about privacy risks.

The expected findings will contribute to the initial investigation of the status of formative assessment literacy for K‒12 teachers in Confucian heritage culture and locate the specific factors influencing the building and practice of teachers in this domain. Eventually, the inventory will be used to gather evidence to inform in-service and preservice teacher professional development programs.


Methodology, Methods, Research Instruments or Sources Used
Given that educational assessment is a complex phenomenon that is difficult to measure solely on quantitative or qualitative information, and Confucian K–12 school contexts are unique through the lens of the social and cultural aspects (Berthrong & Berthrong, 2000). Therefore, we employ an exploratory sequential mixed-method research design (Creswell, & Plano Clark, 2018). The main research thread is:
(1) Qualitative method: we intend to conduct a series of semi-structured interviews with 10 in-service teachers from elementary, junior high, and senior high schools to portray an initial picture of the status of teacher formative assessment literacy in Shanghai, China.
(2) Quantitative method: building on the prior inventories on assessment literacy around the world, and the outcomes from the first-phase interviews, we will vernacularize and design a five-point Likert-type response questionnaire and a multiple-choice scale to explore how teachers perceive, comprehend, and practice formative assessment. About 300 teachers from 12 grades and all major disciplines will be invited as participants. In this phase, we will examine the construct validity of the quantitative data using exploratory factor analysis.
(3) Qualitative method: As a necessity of exploratory sequential design, we will aim to verify the quantitative findings by conducting a new round of qualitative method. Thus, we will invite 3 cohorts of teachers and 1 group of school administrators from three-level schools to participate in the final focus group. Their consensus of formative assessment literacy will conduce to hone the quantitative findings and finally conflate and summarize the overall findings.

Conclusions, Expected Outcomes or Findings
The overarching aim of this study is to design and validate a psychometrically sound and culturally responsive inventory in the Confucian culture context, which can be used as a tool for K–12 teachers’ overall formative assessment literacy. This study can directly depict the current state of Chinese K–12 teachers' formative assessment literacy. I expect to explore the advantages and disadvantages of formative assessment literacy in authentic settings and how Confucianism influences the building of formative assessment literacy for teachers.
Further, it can gather the total grade of a teacher’s general comprehension on formative assessment and five subtotal grades convergent from three dimensions: conceptual, practical and socio-cultural for prospective analysis. Accordingly, the insights from this study could be used as suggestions for ongoing educational practices. By extension, this inventory could be harnessed as an underpinning tool for higher educational institutions to review and redesign pre-service and in-service teacher professional development programs focusing on improving teachers’ formative assessment literacy in Confucian contexts.

References
Black, P., & Wiliam, D. (1998). Inside the black box: Raising standards through classroom assessment. Phi Delta Kappan, 80(2), 139–148.
Berthrong, J. H., & Berthrong, E. N. (2000). Confucianism: A short introduction. Oneworld Publications.
Campbell, C., & Mertler, C. A. (2004). Assessment Literacy Inventory. https://learn.maricopa.edu/courses/811364/quizzes/1226231
Creswell, J. W., & Plano Clark, V. L. (2018). Designing and conducting mixed methods research (3rd ed.). Sage.
Khan, S., & Khan, R. A. (2018). Online assessments: Exploring perspectives of university students. Education and Information Technologies, 24(1), 661–677. https://doi.org/10.1007/s10639-018-9797-0
Popham, W. J. (2011). Assessment literacy overlooked: a teacher educator's confession. The Teacher Educator, 46(4), 265-273. https://doi.org/10.1080/08878730.2011.605048
Wiggins, G. P. (1998). Educative assessment: Designing assessments to inform and improve student performance. Jossey-Bass.
Yan, & Pastore, S. (2022). Are teachers literate in formative assessment? The development and validation of the Teacher Formative Assessment Literacy Scale. Studies in Educational Evaluation, 74, 1–41. https://doi.org/10.1016/j.stueduc.2022.101183
Yorke, M. (2003). Formative assessment in higher education: Moves towards theory and the enhancement of pedagogic practice. Higher Education, 45(4), 477–501. https://doi.org/10.1023/A:1023967026413
Zhan, Y. (2021). What matters in design? Cultivating undergraduates’ critical thinking through online peer assessment in a Confucian heritage context. Assessment & Evaluation in Higher Education, 46(4), 615–630. https://doi.org/10.1080/02602938.2020.1804826
Zhan, Y. (2019). Conventional or sustainable? Chinese university students’ thinking about feedback used in their English lessons. Assessment & Evaluation in Higher Education, 44(7), 973–986. https://doi.org/10.1080/02602938.2018.1557105.


 
Contact and Legal Notice · Contact Address:
Privacy Statement · Conference: ECER 2023
Conference Software: ConfTool Pro 2.6.149+TC
© 2001–2024 by Dr. H. Weinreich, Hamburg, Germany