Recent Submissions

  • Exploring the dynamic relationship between Dr. GEPT feedback and learners’ L2 motivation

    Nakatsuhara, Fumiyo; Lam, Daniel M. K.; Jones, Johnathan; Chan, Sathena Hiu Chong; Chen, Sean; Wu, Rachel; The Language Training and Testing Center, Taiwan (The Language Training and Testing Center, Taiwan, 2023-11-03)
    Feedback is an important means to bridge assessment and learning, but its usefulness ultimately depends on whether and how learners engage with and act on the feedback. Learners’ L2 learning motivation may interact with feedback in meaningful and consequential ways, yet there is relatively little research to date that explores such a dynamic relationship, particularly among language learners in secondary education. This study aimed to fill this gap by exploring the relationship between learners’ motivation and assessment feedback offered by Dr. GEPT – automated personalised feedback provided to GEPT each test-taker alongside their test scores, including an analysis of their strengths and weaknesses, learning advice, and vocabulary and sentence patterns for further study. Taking a mixed-methods approach, Phase 1 of this study involved a large-scale questionnaire survey (n = 635) to explore L2 motivation among senior high school learners of English in Taiwan and their general perceptions towards assessment feedback. The questionnaire was developed based on the L2 Motivational Self System model (Dörnyei, 2005, 2009). Phase 2 used learning logs (n = 14) and interviews (n = 10) for an in-depth qualitative inquiry into how learners engaged with Dr. GEPT feedback and how the feedback might have shaped the developments in learners’ learning journeys. The report concludes with a discussion of how Dr. GEPT helps learners develop a positive orientation towards assessments and cultivates learner autonomy, as well as making some suggestions for enhancing the effectiveness of Dr. GEPT feedback.
  • Exploring the speaking construct in academic settings in a digital age

    Nakatsuhara, Fumiyo; May, Lyn; Khabbazbashi, Nahal; British Council; Cambridge Assessment English; IDP: IELTS Australia (British Council, Cambridge Assessment English and IDP: IELTS Australia, 2023-08-16)
    This study explored language functions and skills utilised in technology-mediated academic speaking contexts, which is timely given the increasing prevalence of digitally-mediated communication in higher education settings and the recent introduction of IELTS Indicator featuring a video-call mode in the Speaking Test. Using an embedded mixed-methods approach, the research involved: 1. language function analysis of spoken communication and simultaneous written chat contributions in online taught classes and supervision meetings 2. thematic analysis of students’ and lecturers’ understandings of distinctive features of online academic speaking and what constitutes successful online speaking interaction in those contexts. We analysed a total of over 40 hours of recordings, consisting of 17 video-recorded classes from four undergraduate and postgraduate units in an Australian University, and 23 video/audio recordings of online PhD supervision meetings from a UK university. This was followed by the administration of a questionnaire and semi-structured interviews with selected participants. In order to examine the construct of online academic communication, we adapted O’Sullivan et al.’s (2002) language function checklist for our purposes. Following the identification of language functions and skills observed in real-life online academic settings, we explored the synergy between the functions observed in online teaching and learning contexts and those elicited in the video-call IELTS Speaking Test (Nakatsuhara et al., 2021). Analyses of questionnaire and interview data helped us understand the skills perceived to be important for successful online interaction. The report concludes with a discussion on the multimodal construct of speaking in digitally-mediated academic contexts and the ways in which the findings of this study can be useful in informing the future development of IELTS Speaking Test tasks so that they remain representative of the reality of academic speaking in the digital age.
  • Accommodations in language testing and assessment: Safeguarding equity, access, and inclusion [editorial[

    Taylor, Lynda; Banerjee, Jayanti; University of Bedfordshire; Trinity College, London (SAGE Publications Ltd, 2023-10-07)
  • Language assessment accommodations: issues and challenges for the future [editorial]

    Taylor, Lynda; Banerjee, Jayanti; University of Bedfordshire; Trinity College, London (SAGE Publications Ltd, 2023-10-07)
    In this concluding piece to the special issue, we attempt to tease out and comment on some themes that have emerged from the six published papers. Some of these themes highlight potential avenues for further theoretical and empirical investigation, and may assist in mapping out a coherent research agenda on the topic for language testers and assessment specialists in the future.
  • Relevance and multimodal prosody: implications for L2 teaching and learning

    Madella, Pauline; ; University of Bedfordshire (Frontiers, 2023-12-01)
    In this paper, I build on Scott's relevance-theoretic account of contrastive stress (2021). Contrastive stress works as an extra cue to ostension in altering the salience of a particular constituent in an utterance and, as a result, the salience of one particular interpretation of that utterance. I draw on Scott’s argument that contrastive stress does not encode procedural meaning. Contrastive stress is unpredictable and, as such, it is in confounding the hearer’s expectations that it draws his attention to the accented word and prompt his search for different interpretive effects. I argue that contrastive stress is interpreted purely inferentially precisely because it is one of many pointing devices. It is to be interpreted by virtue of its interaction with other paralinguistic behaviours, all of which being different aspects of the same ostensive act of communication. This leads me to focus on the gestural nature of contrastive stress working as an act of pointing, which, as an ostensive communicative behaviour, conveys that if you look over there, you’ll know what I mean (Tomasello et al., 2007). Finally, I present the implications of analysing contrastive stress in its multimodal context – as prosodic pointing – for the teaching and learning of L2 prosodic pragmatics and the development of interpretive abilities in the L2 hearer’s mind.
  • Exploring open consonantal environments for at-home testing of vowel perception in advanced L2 speakers

    Jones, Johnathan; (Applied Linguistics Review, 2022-11-29)
    Recent work has called for increased investigation into methods used to explore second language (L2) speech perception (Flege 2021). The present study attends to this call, examining a common practice for developing listening prompts in the context of at-home administrations. Vowel perception studies have historically used fixed consonantal frames to determine how well participants can discriminate between target L2 vowels, and the present study compares the effects of employing a fixed consonant-vowel-consonant frame (h-vowel-d) with open (phonologically diverse) consonantal environments using real and nonce words. Thirty-eight Mandarin (n = 31) and English (n = 8) first language speakers participated in a listening experiment and a post-experiment question. Data were framed within Best and Tyler’s (2007) Perceptual Assimilation Model-L2. Internal consistency and proportion correct were calculated and a generalised linear mixed model design was used to investigate how well performance with h-vowel-d prompts predicts performance with the more diverse prompt types. Results suggest an inflation of scores for the fixed frame prompt and support the use of diverse words for listening prompt designs. Findings have implications for vowel perception researchers as well as computer (and mobile) assisted language learning developers wishing to inform their designs with relevant empirical evidence.
  • Non-verbal communication and context: multi-modality in interaction

    Madella, Pauline; Wharton, Tim; Romero-Trillo, Jesús; University of Bedfordshire; University of Brighton; Universidad Autónoma de Madrid (Cambridge University Press, 2023-11-01)
    Traditionally, the study of linguistics has focussed on verbal communication. In the sense that linguistics is the scientific study of language, the approach is perfectly justified. Those working in the sub-discipline of linguistic pragmatics, however, are faced with something of a dilemma. The aim of a pragmatic theory is to explain how utterances are understood, and utterances, of course, have both linguistic and non-linguistic properties. As well as this, current work in pragmatics emphasizes that the affective dimension of a speaker’s meaning is at least as important as the cognitive one and it is often the non-linguistic properties of utterances that convey information relating to this dimension. This paper highlights the major role of non-verbal ‘modes’ of communication (‘multi-modality’) in accounting for how meaning is achieved and explores in particular how the quasi-musical contours we impose on the words we say, as well as the movements of our face and hands that accompany speech, constrain the context and guide the hearer to our intended meaning. We build on previous exploration of the relevance of prosody (Wilson and Wharton 2006) and, crucially, looks at prosody in relation to other non-verbal communicative behaviours from the perspective of relevance theory. In-so-doing, we also hope to shed light on the role of multimodality in both context construction and utterance interpretation and suggest prosody needs to be analysed as one tool in a set of broader gestural ones (Bolinger 1983). Relevance theory is an inferential model, in which human communication revolves around the expression and recognition of the speaker’s intentions in the performance of an ostensive stimulus: an act accompanied by the appropriate combination of intentions. This inferential model is proposed as a replacement for the traditional code-model of communication, according to which a speaker simply encodes into a signal the thought they wish to communicate and the hearer retrieves their meaning by decoding the signal they have provided. We will argue that much existing work on multimodality remains rooted in a code model and show how adopting an inferential model enables us to integrate multimodal behaviours more completely within a theory of utterance interpretation. As ostensive stimuli, utterances are composites of a range of different behaviours, each working together to form a range of contextual cues.
  • L2 writing assessment: an evolutionary perspective

    Green, Anthony (Palgrave Macmillan, 2022-12-08)
    This book tackles three choices that face developers of L2 writing assessments: defining L2 writing abilities; collecting evidence of those abilities (usually by getting L2 writers to write something); and judging their performance (usually by assigning a score or grade to it). It takes a historical view of how assessment developers have made those choices, how contemporary practices emerged, and of alternative techniques that have risen and fallen over time. The three sections each tackle one of these choices. The first considers the social functions that define L2 writing and assessment; the second relates how assessment tasks have adapted to changing conceptions of languages, writing, and assessment; and the third explores how scoring systems have evolved. Each section brings the reader up to date with current issues confronting writing assessment (both in large-scale testing and in language classrooms) before considering the new opportunities and challenges of the digital age. This book will be of interest to students, scholars and practitioners in language assessment, language education, and applied linguistics.
  • The impact of input format on written performance in a listening-into-writing assessment

    Westbrook, Carolyn; British Council; University of Bedfordshire (Elsevier Ltd, 2022-12-06)
    Over the last five decades, research in teaching and testing (academic) listening has investigated different foci. Initially, teaching listening involved bottom-up approaches (Dirven and Oakeshott-Taylor, 1984) then both higher- and lower-level processes were integrated (Voss, 1984). In the early 2000s, different input formats (Read, 2002) and discourse features of lectures (Thompson, 2003) were the subjects of academic listening research. More recently, EAP tests have increasingly taken an integrated approach to reflect real-world tasks, yet few studies have looked at integrated listening-into-writing tasks (Cubilo and Winke, 2013). This counter-balanced measures design study investigates how test taker performance differs on an integrated EAP listening-into-writing task when lecture input is presented as audio only in one half and video in the other half of the input. Two groups of test takers took part in the current study. A Hotelling's T2 test revealed a statistically significant effect on scores when test takers were presented with the audio only input first but there was no significant effect on scores when the video input was presented first. Data on test taker preferences revealed that more people preferred the video input to audio only.
  • Assessing second language pronunciation: a reference guide

    Jones, Johnathan; Isaacs, Talia (Springer, 2022-01-14)
    Pronunciation assessment (PA) is a resurgent subfield within applied linguistics that traverses the domains of psycholinguistics, second language acquisition (SLA), speech sciences, sociolinguistics, and more recently, computational linguistics. Though the terms ‘pronunciation’ and ‘assessment’ are sometimes defined in different ways by different authors, here we regard pronunciation as the vocal articulation of consonants and vowels (segmentals) combined with aspects of oral speech that extend beyond individual sounds, including stress, rhythm and intonation (suprasegmentals).
  • Towards more valid scoring criteria for integrated reading-writing and listening-writing summary tasks

    Chan, Sathena Hiu Chong; May, Lyn; (SAGE, 2022-12-12)
    Despite the increased use of integrated tasks in high-stakes academic writing assessment, research on rating criteria which reflect the unique construct of integrated summary writing skills is comparatively rare. Using a mixed-method approach of expert judgement, text analysis and statistical analysis, the current study examines writing features that discriminate summaries produced by 150 candidates at five levels of proficiency on integrated reading-writing (R-W) and listening-writing (LW) tasks. The expert judgement revealed a wide range of features which discriminated R-W and L-W responses. When responses at five proficiency levels were coded by these features, significant differences were obtained in seven features, including relevance of ideas, paraphrasing skills, accuracy of source information, academic style, language control, coherence and cohesion and task fulfilment across proficiency levels on the R-W task. The same features did not yield significant differences in L-W responses across proficiency levels. The findings have important implications for clarifying the construct of integrated summary writing in different modalities, indicating the possibility of expanding integrated rating categories with some potential for translating the identified criteria into automated rating systems. The results on the L-W indicate the need for developing descriptors which can more effectively discriminate L-W responses.
  • Assessing interactional competence: exploring ratability challenges

    Lam, Daniel M. K.; Galaczi, Evelina D.; Nakatsuhara, Fumiyo; May, Lyn; University of Glasgow; Cambridge University Press and Assessment; University of Bedfordshire; Queensland University of Technology (John Benjamins, 2023-02-13)
    This paper is positioned at the interface of second/foreign language (L2) assessment and Conversation Analysis-Second Language Acquisition (CA-SLA). It explores challenges of ratability in assessing interactional competence (IC) from three dimensions: an overview of the conceptual and terminological convergence/divergence in the CA-SLA and L2 assessment literature, a micro-analytic Conversation Analysis of test-taker interactions, and the operationalisation of IC construct features in rating scales across assessment contexts. It draws insights from these dimensions into a discussion of the nature of the IC construct and the challenges of IC ratability, and concludes with suggestions on ways in which insights from CA research can contribute to addressing these issues.
  • Report to the Nursing and Midwifery Council on language testing policy

    Green, Anthony; Chan, Sathena Hiu Chong; University of Bedfordshire (Nursing and Midwifery Council, 2022-09-28)
    Responding to the NMC’s review of its language testing policy, our project involved: • A review of the extent to which the approach to language testing currently adopted by the NMC is proportionate and appropriate, and • Recommendations for a methodology to investigate whether language tests of interest should be accepted by the NMC to be met.
  • Integrated writing and its correlates: a meta-analysis

    Chan, Sathena Hiu Chong; Yamashita, J. (Elsevier, 2022-07-26)
    Integrated tasks are increasing in popularity, either replacing or complementing writing- only independent tasks in writing assessments. This shift has generated many research interests to investigate the underlying construct and features of integrated writing (IW) performances. However, due to the complexity of the IW construct, there are conflicting findings about whether and the extent to which various language skills and IW text features correlate to IW scores. To understand the construct of IW, we conducted a meta-analysis to synthesize correlation coefficients between scores of IW performances and (1) other language skills and (2) text quality features of IW. We also examined factors that may moderate the correlation of IW scores with these two groups of correlates. Consequently, (1)reading and writing skills showed stronger correlations than listening to IW scores; and (2) text length had a strongest correlation, followed by source integration, organization and syntactic complexity, with a smallest correlation of lexical complexity. Several IW task features affected the magnitude of correlations. The results supported the view that IW is an independent construct, albeit related, from other language skills and IW task features may affect the construct of IW.
  • Book review: Assessing speaking in context: expanding the construct and its applications

    Taylor, Lynda (SAGE, 2022-02-16)
    review of Salaberry MR, Burch AR (2021) Assessing speaking in context: expanding the construct and its applications, Bristol: Multilingual Matters, ISBN 9781788923804
  • Validation of a large-scale task-based test: functional progression in dialogic speaking performance

    Inoue, Chihiro; Nakatsuhara, Fumiyo (Springer Nature, 2022-02-07)
    A list of language functions is usually included in task-based speaking test specifications as a useful tool to describe target output language of test-takers, to define TLU domains, and to specify task demands. Such lists are, however, often constructed intuitively and they also tend to focus solely on the types of function to be elicited and ignore the ways in which each function is realised across different levels of proficiency (Green, 2012). The study reported in this chapter is a part of a larger-scale test revision project for Trinity’s Integrated Skills in English (ISE) spoken examinations. Analysing audio-recordings of 32 performances on the ISE spoken examination both quantitatively and qualitatively, the aims of this study are (a) to empirically validate lists of language functions in the test specifications of the operational, large-scale, task-based examinations, (b) to explore the usefulness and potential of function analysis as a test task validation method, and (c) to contribute to a better understanding of varied test-taker language that is used to generate language functions.
  • Eye-tracking L2 students taking online multiple-choice reading tests: benefits and challenges

    Latimer, Nicola; Chan, Sathena Hiu Chong (Cranmore Publishing, 2022-04-10)
    Recently, there has been a marked increase in language testing research involving eye-tracking. It appears to offer a useful methodology for examining cognitive validity in language tests, i.e., the extent to which the mental processes that a language test elicits from test takers resemble those that they would employ in the target language use domains. This article reports on a recent study which examined reading processes of test takers at different proficiency levels on a reading proficiency test. Using a mixed-methods approach, the study collected cognitive validity evidence through eye-tracking and stimulated recall interviews. The study investigated whether there are differences in reading behaviour among test takers at CEFR B1, B2 and C1 levels on an online reading task. The main findings are reported and the implications of the findings are discussed to reflect on some fundamental questions regarding the use of eye-tracking in language testing research.
  • Assessing speaking

    Nakatsuhara, Fumiyo; Khabbazbashi, Nahal; Inoue, Chihiro; University of Bedfordshire (Routledge, 2021-12-16)
    In this chapter on assessing speaking, the history of speaking assessment is briefly traced in terms of the various ways in which speaking constructs have been defined and diversified over the past century. This is followed by a discussion of elicitation tasks, test delivery modes, rating methods, and scales that offered opportunities and/or presented challenges in operationalising different constructs of speaking and providing feedback. Several methods utilised in researching speaking assessment are then considered. Informed by recent research and advances in technology, the chapter provides recommendations for practice in both high-stakes and low-stakes contexts.
  • Towards the new construct of academic English in the digital age

    Khabbazbashi, Nahal; Chan, Sathena Hiu Chong; Clark, Tony; University of Bedfordshire; Cambridge University Press and Assessment (Oxford University Press, 2022-03-28)
    The increasing use of digital educational technologies in Higher Education (HE) means that the nature of communication may be shifting. Assessments of English for Academic Purposes (EAP) need to be reconceptualised accordingly, to reflect the new and complex ways in which language is used in HE. With a view to inform EAP assessments, our study set out to identify key trends related to Academic English using a scoping review of the literature. Findings revealed two major trends: (a) a shift towards multimodal communication which has in turn resulted in the emergence of new types of academic assignments, multimodal genres, and the need for students to acquire new skills to operate within this multimodal arena; and (b) the limitations of existing skills-based approaches to assessment and the need to move towards integrated skills assessment. We discuss the implications of these findings for EAP assessments.
  • The design and validation of an online speaking test for young learners in Uruguay: challenges and innovations

    Khabbazbashi, Nahal; Nakatsuhara, Fumiyo; Inoue, Chihiro; Kaplan, Gabriela; Green, Anthony; University of Bedfordshire; Plan Ceibal (Cranmore Publishing on behalf of the International TESOL Union, 2022-02-10)
    This research presents the development of an online speaking test of English for students at the end of primary and beginning of secondary school education in state schools in Uruguay. Following the success of the Plan Ceibal one computer-tablet per child initiative, there was a drive to further utilize technology to improve the language ability of students, particularly in speaking, where the majority of students are at CEFR levels pre-A1 and A1. The national concern over a lack of spoken communicative skills amongst students led to a decision to develop a new speaking test, specifically tailored to local needs. This paper provides an overview of the speaking test development and validation project designed with the following objectives in mind: to establish, track, and report annually learners’ achievements against the Common European Framework of Reference for Languages (CEFR) targeting CEFR levels pre-A1 to A2, to inform teaching and learning, and to promote speaking practice in classrooms. Results of a three-phase mixed-methods study involving small-scale and large-scale trials with learners and examiners as well as a CEFRlinking exercise with expert panelists will be reported. Different sources of evidence will be brought together to build a validity argument for the test. The paper will also focus on some of the challenges involved in assessing young learners and discuss how design decisions, local knowledge and expertise, and technological innovations can be used to address such challenges with implications for other similar test development projects.

View more