Swedish and Finnish Pre-Service Teachers’ Perceptions of Summative Assessment Practices
Abstract
:1. Introduction
1.1. Background
1.2. Purpose
- How do pre-service teachers perceive the importance of SAL practices, and how do these values differ by university?
- How do pre-service teachers evaluate their own competence in applying SAL practices, and how do these evaluations differ by university?
2. Literature Review
2.1. Perceptions of Assessment Literacy
2.2. Previous Studies on Pre-Service Teachers’ Assessment Literacy
3. Methods
3.1. Participants
3.2. Data Collection
3.3. Data Analysis
4. Results
4.1. Pre-Service Teachers’ Perceptions of the Importance of SAL Practices
During my first teaching practice [VFU], where there was some kind of a problem, where pupils misunderstood, where I think that they did misunderstand the summative grade they got, the grade, at the end of the term, that’s where I even talked about this in class, about my dilemma, you know, how can the grade be communicated to the children, and what is lost in translation… and through summative, formative process which lead to the summative. So I think, a way I thought to deal with that was to really make the learning intentions and the goal of lessons and even the term, as a whole, make it clear enough to the students, which would in turn lead to them understanding what it ended up in, what that summative thing.(B1:1)
[cooperative assessment] is an important part because you can easily be blinded by what you feel towards a pupil [....] you need to help each other in difficult cases”. [...] I find it absolutely imperative when you are new and inexperienced because….it takes real life experience and you need to practice, practice, practice all the time to become good at assessment. That’s why I need help….(B4: 3–4)
Previously, I didn’t really think about them [pupils] as weak…but now I am more, what should I say… I am more… Now I understand that they also need attention, more than the others. Adaptations…(A3:1)
Well, of course, you should also give them a grade because if you think they have ADHD, then it’s a factor that affects your grading in such a way that they may have difficulty concentrating in the test. It won’t tell you anything about intelligence, but it will tell you [...] that it is difficult to focus.(F4:1)
4.2. Pre-Service Teachers’ Perceptions of Their Competence in Applying SA Practices
- competence in writing multiple choice items: M(A) = 3.50; M(C) = 3.97; F = 3.001; p = 0.053; eta squared = 0.045, small effect size;
- competence in writing true –false items: M(A) = 3.53; M(C) = 4.10; F = 4.782; p = 0.010; eta squared = 0.07, modest effect size;
- competence in using basic statistics: M(A) = 2.30; M(C) = 2.88; F = 4.288; p = 0.016; eta squared = 0.064, modest effect size.
Generally, not competent at all. Or, well I feel that it’s an area where I have limited knowledge and especially with regard to my own ability to make test. I am sure we’ve had something about it in our education but I have difficulties in remembering clearly. I am sure we’ve gone through what you should think about and so on, something I remember is maybe that, that when you make your own tests you should start with easy items and things like that… basic stuff. but generally I feel that I don’t have enough with me there and whatever we had during that course, it is, for my part anyway…. it went too fast, passed by quickly and then we didn’t return to it.(A4:4)
And then, the times when I have tried to construct tests and homework quizzes I realize how, how easily it goes wrong. So I think, I mean I… it’s not that I feel, that I feel that it is impossible to construct different types of tests but, I haven’t practiced. Neither during school practice nor during our education here at the university, so in this respect I do not feel competent, not really.(A4:3)
No, not really. I did some kind of… if it’s like, this online quiz that they can do. I did some kind of test construction to make them test their own… how much they know about things. It’s like—because at the school I was at they didn’t really work with tests like for an assessment-assessment. It’s more for a self-assessment that you know how much you have learnt about something.(B3:1)
5. Discussion
5.1. Perceptions on Conceptual Knowledge, Praxeological Dimension, and Socio-Emotional Issues
5.2. Differences among the Three Teacher Education Programs
5.3. Limitations of the Study
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Conflicts of Interest
Appendix A
Courses | University A Sweden | Courses | University B Sweden | Courses | University C Finland |
Common for all | Assessment, analysis and evaluation of learning and development | Common for all | Language education, curriculum theory, grading and assessment (General Level) | 1 year program Common | Planning, implementation and assessment of teaching |
Long program | English learning, teaching and assessment I–III | Long program | Language education, curriculum theory, grading and assessment (Advanced Level) | Curriculum and development of educational institution | |
Assessment and grading for teachers in secondary and upper secondary school | Language assessment (focus on ie. Swedish as a Second Language) | Teacher as a researcher/ Teacher as a researcher Didactics | |||
Social, cultural and philosophical foundations of education | |||||
School practice | School practice 1–4 | School practice | School practice 1–3 | School practice | School practice basic and advanced |
Short program | Steering, organization and assessment for teachers in secondary school and upper secondary school | Short program | Summative assessment module | ||
School practice | School practice 1–3 |
Appendix B
How Important Do You Think It Is to… | Conceptual Knowledge | Praxeological Dimension | Socio-Emotional Dimension |
inform students about targets and grading criteria | x | ||
use national curriculum/guidelines for summative assessment tasks | x | ||
communicate summative assessment to students and parents | x | ||
cooperate with other teachers in planning/conducting assessment | x | ||
consider individual differences (e.g., special needs, interests) | x | ||
participate in summative assessment in-service training | x | ||
adjust classroom teaching based on results of summative assessments | x | ||
profile language proficiency (e.g., report separately for the “four skills”) | x | ||
use national test results for grading | x | ||
use online testing | x | ||
use ready-made/published tests when assessing students’ performance | x | ||
use self/peer-assessment when grading | x | ||
How competent do you feel to… | |||
write true–false items | x | x | |
write short-answer questions | x | x | |
write multiple-choice items | x | x | |
construct written tests | x | x | |
write open-ended/essay questions | x | x | |
grade standardized items | x | x | |
assess students’ written performance | x | x | |
assess students’ reading skills | x | x | |
grade written tests | x | x | |
assess students’ listening skills | x | x | |
construct oral tests | x | x | |
grade homework/project assignments | x | x | |
assess students’ oral performance | x | x | |
assess students’ interaction | x | x | |
assign reliable grades | x | x | x |
grade open-ended/essay questions | x | x | |
grade oral tests | x | x | |
weight different language skills statistically when grading | x | x | |
use basic statistics (e.g., means, correlations) when analyzing test results | x | x | |
grade portfolios | x | x |
References
- Abrar-ul-Hassan, Shahid, Dan Douglas, and Jean Turner. 2021. Revisiting Second Language Portfolio Assessment in a New Age. System 103: 102652. [Google Scholar] [CrossRef]
- Bachman, Lyle F., and Adrian Palmer. 1996. Language Testing in Practice. Oxford: Oxford University Press. [Google Scholar]
- Bachman, Lyle F., and Adrian Palmer. 2010. Language Assessment in Practice. Developing Language Assessments and Justifying their Use in the Real World. Oxford: Oxford University Press. [Google Scholar]
- Bailey, Kathleen M., and James D. Brown. 1995. Language testing courses: What are they? In Validation in Language Testing. Edited by Alister H. Cumming and Richard Berwick. Clevedon: Multilingual Matters, pp. 236–256. [Google Scholar]
- Barcelos, Ana Maria Ferreira, and Paula Kalaja. 2011. Introduction to beliefs about SLA revisited. System 39: 281–89. [Google Scholar] [CrossRef]
- Borg, Simon. 2006. The distinctive characteristics of foreign language teachers. Language Teaching Research 10: 3–31. [Google Scholar] [CrossRef]
- Borg, Simon. 2011. The Impact of In-Service Teacher Education on Language Teachers’ Beliefs. System: An International Journal of Educational Technology and Applied Linguistics 39: 370. [Google Scholar] [CrossRef]
- Borg, Simon. 2015. Teacher Cognition and Language Education: Research and Practice. Bloomsbury Classics in Linguistics Edition. London: Bloomsbury Academic. [Google Scholar]
- Brindley, Geoff. 2001. Outcomes-based assessment in practice: Some examples and emerging insights. Language Testing 18: 393–407. [Google Scholar] [CrossRef]
- Broadbent, Jaclyn, Ernesto Panadero, and David Boud. 2018. Implementing summative assessment with a formative flavour: A case study in a large class. Assessment & Evaluation in Higher Education 43: 307–22. [Google Scholar]
- Brookhart, Susan M. 2011. Educational Assessment Knowledge and Skills for Teachers. Educational Measurement: Issues and Practice 30: 3–12. [Google Scholar] [CrossRef]
- Caena, Francesca. 2011. Literature Review. Teachers’ Core competences: Requirements and Development. European Commission Thematic Working Group Professional Development of Teachers. Strassbourg: European Commission. [Google Scholar]
- Chapelle, Carol A., Mary K. Enright, and Joan M. Jamieson. 2008. Building a Validity Argument for the Test of English as s Foreign Language. London: Routledge. [Google Scholar]
- Cizek, Gregory J., Heidi L. Andrade, and Randy E. Bennett. 2019. Formative Assessment, History, Definitions, and Progress. In Handbook of Formative Assessment in the Disciplines. Edited by Heidi L. Andrade, Randy E. Bennett and Gregory J. Cizek. New York and London: Routledge, pp. 3–19. [Google Scholar]
- Coombe, Christine, Hossein Vafadar, and Hassan Mohebbi. 2020. Language assessment literacy: What do we need to learn, unlearn, and relearn? Language Testing in Asia 10: 1–16. [Google Scholar] [CrossRef]
- Coombs, Andrew, Christopher DeLuca, and Stephen MacGregor. 2020. A person-centered analysis of teacher candidates’ approaches to assessment. Teaching and Teacher Education 87: 102952. [Google Scholar] [CrossRef]
- Council of Europe. 2001. Common European Framework of Reference for Languages: Learning, Teaching, Assessment. Cambridge: Cambridge University Press. [Google Scholar]
- Davies, Alan. 2008. Textbook trends in teaching language testing. Language Testing 25: 327–47. [Google Scholar] [CrossRef]
- DeLuca, Christopher, and Chi Yan Lam. 2014. Preparing Teachers for Assessment within Diverse Classrooms: An Analysis of Teacher Candidates’ Conceptualizations. Teacher Education Quarterly 41: 3–24. [Google Scholar]
- DeLuca, Christopher, and Don Klinger. 2010. Assessment literacy development: Identifying gaps in teacher candidates’ learning. Assessment in Education 17: 419. [Google Scholar] [CrossRef]
- DeLuca, Christopher, Teresa Chavez, Aarti Bellara, and Cao Chunhua. 2013. Pedagogies for Preservice Assessment Education: Supporting Teacher Candidates’ Assessment Literacy Development. Teacher Educator 48: 128. [Google Scholar] [CrossRef]
- DeLuca, Christopher, Danielle LaPointe-McEwan, and Ulemu Luhanga. 2016. Approaches to Classroom Assessment Inventory: A New Instrument to Support Teacher Assessment Literacy. Educational Assessment 21: 248–66. [Google Scholar] [CrossRef]
- DeLuca, Christopher, Christoph Schneider, Andrew Coombs, Marcela Pozas, and Amirhossein Rasooli. 2020. A cross-cultural comparison of German and Canadian student teachers’ assessment competence. Assessment in Education: Principles, Policy & Practice 27: 26–45. [Google Scholar] [CrossRef]
- Fulcher, Glenn. 2012. Assessment Literacy for the Language Classroom. Language Assessment Quarterly 9: 113–32. [Google Scholar] [CrossRef]
- Giraldo, Frank, and Daniel Murcia. 2018. Language Assessment Literacy for Pre-service Teachers: Course Expectations from Different Stakeholders. GIST Education and Learning Research Journal 16: 56–77. [Google Scholar] [CrossRef] [Green Version]
- Hilden, Raili, and Birgitta Fröjdendahl. 2018. The dawn of assessment literacy—Exploring the conceptions of Finnish student teachers in foreign languages. Apples: Journal of Applied Language Studies 12: 1–24. [Google Scholar]
- Hildén, Raili, and Juhani Rautopuro. 2014. Saksan Kielen A-ja B-Oppimäärän Oppimistulokset Perusopetuksen Päättövaiheessa 2013. Helsinki: Finnish National Board of Education. [Google Scholar]
- Inbar-Lourie, Ofra. 2008. Constructing a language assessment knowledge base: A focus on language assessment courses. Language Testing 25: 385–402. [Google Scholar] [CrossRef]
- Jones, Glyn. 2020. Designing Multiple-Choice Test Items. In The Routledge Handbook of Second Language Acquisition and Language Testing. Edited by Winke Paula Marie and Tineke Brunfaut. New York: Routledge, pp. 90–101. [Google Scholar]
- Kane, Michael T. 2013. Validating the Interpretations and Uses of Test Scores. Journal of Educational Measurement 50: 1–73. [Google Scholar] [CrossRef]
- Klapp Lekholm, Alli. 2011. Effects of School Characteristics on Grades in Compulsory School. Scandinavian Journal of Educational Research 55: 587–608. [Google Scholar] [CrossRef]
- Kremmel, Benjamin, and Luke Harding. 2020. Towards a Comprehensive, Empirical Model of Language Assessment Literacy across Stakeholder Groups: Developing the Language Assessment Literacy Survey. Language Assessment Quarterly 17: 100–20. [Google Scholar] [CrossRef]
- Kunnan, Antony John. 2018. Evaluating Language Assessments. New York: Routledge. [Google Scholar]
- Lam, Ricky. 2015. Language assessment training in Hong Kong: Implications for language assessment literacy. Language Testing 32: 169–97. [Google Scholar] [CrossRef]
- Lau, Alice Man Sze. 2016. Formative good, summative bad?—A review of the dichotomy in assessment literature. Journal of Further and Higher Education 40: 509–25. [Google Scholar] [CrossRef]
- Levi, Tziona, and Ofra Inbar-Lourie. 2020. Assessment Literacy or Language Assessment Literacy: Learning from the Teachers. Language Assessment Quarterly 17: 168–82. [Google Scholar] [CrossRef]
- McDonald, Nora, Sarita Schoenebeck, and Andrea Forte. 2019. Reliability and inter-rater reliability in qualitative research: Norms and guidelines for CSCW and HCI practice. ACM on Human-Computer Interaction 3: 1–23. [Google Scholar]
- Messick, Samuel. 1989. Validity. In Educational Measurement, 3rd ed. Edited by Robert Linn. American Council on Education. Washington, DC: Macmillan. [Google Scholar]
- Oscarson, Mats, and Britt Marie Apelgren. 2011. Mapping language teachers’ conceptions of student assessment procedures in relation to grading: A two-stage empirical inquiry. System 39: 2–16. [Google Scholar] [CrossRef]
- Oxford English Dictionary Online. 2021. Oxford: Oxford University Press.
- Pastore, Serafina, and Heidi L. Andrade. 2019. Teacher assessment literacy: A three-dimensional model. Teaching and Teacher Education 84: 128–38. [Google Scholar] [CrossRef]
- Pellegrino, James W. 2018. Assessment of and for learning. In International Handbook of the Learning Sciences. Edited by Frank Fischer, Susan M. Goldman, Cindy E. Hmelo-Silver and Peter Reiman. New York: Routledge, pp. 410–21. [Google Scholar]
- Pizorn, Karmen, and Ari Huhta. 2016. Assessment in educational settings. In Handbook of Second Language Assessment. Edited by Tsagari Dina and Jay Banerjee. New York: Mouton De Gruyter, pp. 239–254. [Google Scholar]
- Popham, James W. 2011. Assessment literacy overlooked: A teacher educator’s confession. The Teacher Educator 46: 265–73. [Google Scholar] [CrossRef]
- Qiong, Ou. 2017. A brief introduction to perception. Studies in Literature and Language 15: 18–28. [Google Scholar]
- Riksrevisionen. 2011. Lika Betyg, Lika Kunskap? En Uppföljning av Statens Styrning mot en Likvärdig Betygssättning i Grundskolan. Stockholm: Riksrevisionen. [Google Scholar]
- Salkind, Neil J. 2010. Encyclopedia of Research Design. Thousand Oaks: Sage, vol. 1. [Google Scholar]
- Scriven, Michael. 1967. The methodology of evaluation. In Perspectives of Curriculum Evaluation. Edited by Robert Gagne, Ralph W. Tyler and Michael Scriven. Chicago: Rand McNally, p. 39. [Google Scholar]
- Skolverket. 2021a. Available online: https://www.skolverket.se/undervisning/gymnasieskolan/nationella-prov-i-gymnasieskolan/provdatum-i-gymnasieskolan (accessed on 12 December 2021).
- Skolverket. 2021b. Available online: https://www.skolverket.se/undervisning/grundskolan/nationella-prov-i-grundskolan/genomfora-och-bedoma-prov-i-grundskolan#h-Provresultatetsbetydelseforbetyget (accessed on 11 December 2021).
- Skolverket. 2021c. Available online: https://www.skolverket.se/undervisning/gymnasieskolan/nationella-prov-i-gymnasieskolan/genomfora-och-bedoma-prov-i-gymnasieskolan#Provresultatetsbetydelseforbetyget (accessed on 12 December 2021).
- Stiggins, Richard J. 1991. Relevant Classroom Assessment Training for Teachers. Educational Measurement, Issues and Practice 10: 7–12. [Google Scholar] [CrossRef]
- Taras, Maddalena. 2005. Assessment—Summative and Formative—Some Theoretical Reflections. British Journal of Educational Studies 53: 466–78. [Google Scholar] [CrossRef]
- Taras, Maddalena. 2009. Summative assessment: The missing link for formative assessment. Journal of Further and Higher Education 33: 57–69. [Google Scholar] [CrossRef]
- Tavassoli, Kobra, and Hossein Farhadi. 2018. Assessment Knowledge Needs of EFL Teachers. Teaching English Language 12: 45–65. [Google Scholar]
- Taylor, Lynda. 2013. Communicating the Theory, Practice and Principles of Language Testing to Test Stakeholders: Some Reflections. Language Testing 30: 403–12. [Google Scholar] [CrossRef]
- Tenfeldt, Torbjörn. 2021 Många Gymnasieskolor Ratar Nationella Proven. Ämnesläraren, 11 November. Available online: https://www.lararen.se/amneslararen-matte-no/nationella-prov/manga-gymnasieskolor-ratar-nationella-proven (accessed on 10 December 2021).
- Thorsen, Cecilia, and Christina Cliffordson. 2012. Teachers’ grade assignment and the predictive validity of criterion-referenced grades. Educational Research & Evaluation 18: 153–72. [Google Scholar] [CrossRef]
- Ufer, Stefan, and Knut Neumann. 2018. Measuring Competencies. In International Handbook of the Learning Sciences. Edited by Frank Fischer, Cindy E. Hmelo-Silver, Susan R. Goldman and Peter Reimann. Abingdon: Routledge, pp. 433–43. [Google Scholar]
- Vogt, Karin, and Dina Tsagari. 2014. Assessment Literacy of Foreign Language Teachers: Findings of a European Study. Language Assessment Quarterly 11: 374. [Google Scholar] [CrossRef]
- Volante, Louis, and Xavier Fazio. 2007. Exploring Teacher Candidates’ Assessment Literacy: Implications for Teacher Education Reform and Professional Development. Canadian Journal of Education 30: 749–70. [Google Scholar] [CrossRef]
- Xu, Yueting, and Gavin L. T. Brown. 2016. Teacher assessment literacy in practice: A reconceptualization. Teaching and Teacher Education 58: 149–62. [Google Scholar] [CrossRef]
- Yildirim, Ali, Anne Dragemark Oscarson, Raili Hilden, and Birgitta Fröjdendahl. 2021. Teaching summative assessment literacy: A comparative curriculum analysis of three teacher education programs in Sweden and Finland. Paper presented at American Educational Research Association Conference, Orlando, FL, USA, April 8–12. [Google Scholar]
Conceptual Knowledge What Conceptions a Teacher Has of Assessment | Praxeological Dimension How to Monitor, Judge, and Manage the Teaching–Learning Process | Socio-Emotional Dimension How Teachers Manage the Social and Emotional Aspects | |
---|---|---|---|
Fulcher (2012) | * theoretical knowledge | * practical knowledge | * socio-historical understanding of assessment-related activity |
Taylor (2013) | * knowledge of theory principles and concepts | * technical skills, language pedagogy, local practices * scores and decision making. | * sociocultural values, * personal beliefs |
Xu and Brown (2016) | * Disciplinary knowledge and pedagogical content knowledge, * Knowledge of assessment purposes, content and methods | * Knowledge of grading, * Knowledge of feedback * Knowledge of peer and self-assessment * Knowledge of assessment interpretation and communication | * Knowledge of assessment ethics |
DeLuca et al. (2016) | * Assessment purposes, * Measurement theory * Assessment for learning | * Processes * Communication of assessment results * Education and support for teachers | * Fairness * Ethics |
Pastore and Andrade (2019) | * theories * models * purpose, object * methods, * data analysis * reporting and communication | * alignment of learning and assessment, * data-gathering * interpreting evidence * curriculum adaptation * communication, stakeholders engagement, * scaffolding pupils | * collaboration across stakeholder groups * consciousness of one’s role as assessor, * ethical aspects, power and impact issues |
Kremmel and Harding (2020) | * assessment policy and local practices * statistical and research methods * language, structure, use and development | * developing and administering language assessments * assessment in language pedagogy washback and preparation, scoring and rating | * personal beliefs and attitudes * assessment principles and interpretation |
Country | Sweden 61 (47%) | Finland 70 (53%) | |
University | A (Sweden) 30 (23%) | B (Sweden) 31 (24%) | C (Finland) 70 (53%) |
Gender | Male 24 (18%) | Female 104 (79%) | Other 3 (3%) |
Language 1 | Swedish 45 (34%) | Finnish 61 (47%) | Other 25 (19%) |
Self-assessed CEFR level of teaching language | C2–C1 117 (89%) | B2–B1 11 (8%) | A2–A1 3 (3%) |
First teaching Language | English 79 (60%) | Swedish as a second language 37 (28%) | Other 15 (12%) |
N | Mean * | SD | |
---|---|---|---|
informing students about targets and grading criteria | 129 | 4.58 | 0.693 |
using national curriculum/guidelines for summative assessment tasks | 127 | 4.27 | 0.821 |
communicating summative assessment to students and parents | 127 | 4.15 | 0.892 |
cooperating with other teachers in planning/conducting assessment | 129 | 4.14 | 0.808 |
considering individual differences (e.g., special needs, interests) | 129 | 4.14 | 0.817 |
participating in summative assessment in-service training | 126 | 4.07 | 0.869 |
adjusting classroom teaching based on results of summative assessments | 127 | 3.99 | 0.812 |
profiling language proficiency | 126 | 3.69 | 0.784 |
using national test results for grading | 117 | 3.37 | 0.988 |
using online testing | 127 | 3.2 | 1.202 |
using ready-made/published tests when assessing students’ performance | 119 | 3.1 | 1.069 |
using self/peer-assessment when grading | 126 | 3.1 | 1.101 |
Univ A (Swe) | Univ B (Swe) | Univ C (Fin) | Significance Test | |||||||
---|---|---|---|---|---|---|---|---|---|---|
N | Mean | SD | N | Mean | SD | N | Mean | SD | ||
Informing (pupils) about targets and grading criteria | 30 | 4.13 | 0.9 | 31 | 4.61 | 0.667 | 68 | 4.76 | 0.492 | F = 9.9 df = 2 p = 0.0001 etasq = 0.12 |
Communicating SA to (pupils) and parents | 28 | 3.61 | 0.916 | 31 | 3.77 | 1.023 | 68 | 4.54 | 0.584 | F = 11.6 df = 2 p = 0.0001 etasq = 0.23 |
Using self/peer assessment when grading | 29 | 2.55 | 1.152 | 29 | 3.24 | 1.123 | 68 | 3.28 | 1.005 | F = 5.0 df = 2 p = 0.008 etasq = 0.08 |
Profiling language proficiency | 28 | 3.43 | 0.92 | 29 | 3.97 | 0.626 | 69 | 3.68 | 0.757 | F = 3.5 df = 2 p = 0.034 etasq = 0.05 |
Considering individual differences | 30 | 4.27 | 0.691 | 31 | 3.77 | 0.956 | 68 | 4.25 | 0.760 | F = 4,3 df = 2 p = 0.016 etasq = 0.054 |
Using online testing | 29 | 3.00 | 1.254 | 30 | 2.73 | 1.172 | 68 | 3.49 | 1.126 | F = 4.6 df = 2 p = 0.009 etasq = 0.028 |
Inform Students about Targets and Grading Criteria | Communicate Summative Assessment to Students and Parents | Use Online Testing | |||||||
---|---|---|---|---|---|---|---|---|---|
Swedish Unis | Finnish Uni | Swedish Unis | Finnish Uni | Swedish Unis | Finnish Uni | ||||
N | 61 | 68 | 59 | 68 | 59 | 68 | |||
Mean | 4.38 | 4.76 | 3.69 | 4.54 | 2.86 | 3.49 | |||
Std. Deviation | 0.820 | 0.492 | 0.969 | 0.584 | 1.210 | 1.126 | |||
F | 19.028 | 18.914 | 0.075 | ||||||
t | −3.294 | −6.067 | −2.978 | ||||||
df | 127 | 125 | 119.512 | ||||||
Sig. (2–tailed) | 0.00128 | 0.00000001 | 0.004 | ||||||
Cohen’s d | 0.667 | 0.787 | 1.166 | ||||||
effect size | medium | medium | large |
N | Mean | SD | |
---|---|---|---|
Writing true–false items | 131 | 3.92 | 0.865 |
Writing short-answer questions | 130 | 3.85 | 0.836 |
Writing multiple-choice items | 130 | 3.81 | 0.916 |
Constructing written tests | 131 | 3.56 | 0.860 |
Writing open-ended/essay questions | 130 | 3.55 | 0.881 |
Grading standardized items | 128 | 3.52 | 0.972 |
Assessing students’ written performance | 131 | 3.52 | 0.844 |
Assessing students’ reading skills | 130 | 3.51 | 0.865 |
Grading written tests | 131 | 3.46 | 0.806 |
Assessing students’ listening skills - | 130 | 3.38 | 0.893 |
Constructing oral tests | 131 | 3.20 | 0.940 |
Grading homework/project assignments | 127 | 3.19 | 0.932 |
Assessing students’ oral performance | 130 | 3.13 | 0.848 |
Assessing students’ interaction | 130 | 3.11 | 0.856 |
Assigning reliable grades | 131 | 3.11 | 0.834 |
Grading open-ended/essay questions | 131 | 3.08 | 0.771 |
Grading oral tests | 131 | 2.96 | 0.845 |
Weighting different language skills statistically when grading | 124 | 2.77 | 1.052 |
Using basic statistics when analyzing test results | 129 | 2.62 | 1.126 |
Grading portfolios | 127 | 2.57 | 1.005 |
Univ A (Swe) | Univ B (Swe) | Univ C (Fin) | Significance Test | |||||||
---|---|---|---|---|---|---|---|---|---|---|
N | Mean * | SD | N | Mean * | SD | N | Mean * | SD | ||
Writing multiple choice items | 30 | 3.5 | 1.0 | 30 | 3.73 | 0.9 | 70 | 3.97 | 0.8 | F = 3.0 df = 2 p = 0.053 etasq = 0.045 |
Writing true-false -items | 30 | 3.53 | 0.9 | 31 | 3.90 | 0.83 | 70 | 4.10 | 0.78 | F = 4.8 df = 2 p = 0.010 etasq = 0.070 |
Assessing reading skills | 29 | 3.34 | 0.72 | 31 | 3.26 | 0.99 | 70 | 3.69 | 0.83 | F = 3.4 df = 2 p = 0.036 etasq = 0.051 |
Using basic statistics | 30 | 2.30 | 1.21 | 30 | 2.33 | 0.88 | 69 | 2.88 | 1.13 | F = 4.3 df = 2 p = 0.016 etasq = 0.064 |
F | df | Sig. (2–Tailed) | ||||||||
---|---|---|---|---|---|---|---|---|---|---|
N | Mean | Std. Deviation | Std. Error Mean | Cohen’s d | Effect Size | |||||
Write multiple-choice items | Swedish unis | 60 | 3.62 | 0.993 | 0.128 | |||||
Finnish uni | 70 | 3.97 | 0.816 | 0.098 | ||||||
5.126 | 128 | 0.027 | 0.902 | large | ||||||
Write true-false items | Swedish unis | 61 | 3.72 | 0.915 | 0.117 | |||||
Finnish uni | 70 | 4.10 | 0.783 | 0.094 | ||||||
2.481 | 129 | 0.012 | 0.847 | large | ||||||
Write short-answer questions | Swedish unis | 60 | 3.67 | 0.914 | 0.118 | |||||
Finnish uni | 70 | 4.01 | 0.732 | 0.088 | ||||||
7.051 | 128 | 0.018 | 0.821 | large | ||||||
Grade portfolios | Swedish unis | 59 | 2.37 | 1.032 | 0.134 | |||||
Finnish uni | 68 | 2.74 | 0.956 | 0.116 | ||||||
2.699 | 119.27 | 0.043 | 0.992 | large | ||||||
Weight different language skills statistically when grading | Swedish unis | 56 | 2.55 | 1.159 | 0.155 | |||||
Finnish uni | 68 | 2.94 | 0.929 | 0.113 | ||||||
8.865 | 122 | 0.041 | 1.039 | large | ||||||
Assess students’ listening skills | Swedish unis | 60 | 3.22 | 0.865 | 0.112 | |||||
Finnish uni | 70 | 3.53 | 0.896 | 0.107 | ||||||
0.519 | 128 | 0.047 | 0.882 | large | ||||||
Assess students’ reading skills | Swedish unis | 60 | 3.30 | 0.869 | 0.112 | |||||
Finnish uni | 70 | 3.69 | 0.826 | 0.099 | ||||||
0.083 | 126.46 | 0.004 | 1.094 | Large | ||||||
Use basic statistics (e.g., means, correlations) when analyzing test results | Swedish unis | 60 | 2.32 | 1.049 | 0.135 | |||||
Finnish uni | 69 | 2.88 | 1.132 | 0.136 | ||||||
0.030 | 127 | 0.004 | 1.094 | Large |
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations. |
© 2022 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Hilden, R.; Oscarson, A.D.; Yildirim, A.; Fröjdendahl, B. Swedish and Finnish Pre-Service Teachers’ Perceptions of Summative Assessment Practices. Languages 2022, 7, 10. https://doi.org/10.3390/languages7010010
Hilden R, Oscarson AD, Yildirim A, Fröjdendahl B. Swedish and Finnish Pre-Service Teachers’ Perceptions of Summative Assessment Practices. Languages. 2022; 7(1):10. https://doi.org/10.3390/languages7010010
Chicago/Turabian StyleHilden, Raili, Anne Dragemark Oscarson, Ali Yildirim, and Birgitta Fröjdendahl. 2022. "Swedish and Finnish Pre-Service Teachers’ Perceptions of Summative Assessment Practices" Languages 7, no. 1: 10. https://doi.org/10.3390/languages7010010
APA StyleHilden, R., Oscarson, A. D., Yildirim, A., & Fröjdendahl, B. (2022). Swedish and Finnish Pre-Service Teachers’ Perceptions of Summative Assessment Practices. Languages, 7(1), 10. https://doi.org/10.3390/languages7010010