AAAL 2026: Invited Colloquium
Convened by Andrea Révész and Shungo Suzuki
Artificial Intelligence in Applied Linguistics: Applications, Promises and Challenges
Conveners:
Shungo Suzuki, Lancaster University
Discussant:
Coming Soon
Colloquium Abstract
Recent advancements in artificial intelligence (AI) technologies such as generative AI have sparked significant interest within the field of applied linguistics. Researchers across various subfields are exploring its potential applications, carefully evaluating both the opportunities it offers and the challenges it presents. In this colloquium experts on computational sociolinguistics, corpus linguistics, digital literacies, global Englishes, intercultural communication, language teaching, language assessment, and second language acquisition consider the ethical use and/or potential negative consequences of AI use in their area. Each colloquium paper draws upon the colloquium presenters’ contributions published in the latest issue of the Annual Review of Applied Linguistics (ARAL), focusing on relevant AI technologies and their intersections with applied linguistics. In their work, the authors explored the role of AI through theoretical analysis, empirical research, or systematic review, taking a critical lens to identify constructive and effective pathways for applying AI in applied linguistics.
The colloquium will open with a brief introduction by the conveners, followed by nine 10-minute presentations. Each contribution will conclude with the speakers reflecting on the potential applications, promises, and challenges of AI based on their research. The event will end with an open discussion between the audience and the panelists.
Identity, ideology, and capital: Investing in agentive generative AI practices
Ron Darvin, The University of British Columbia
The proliferation of generative AI (GenAI) has significant implications for language and literacy education, reshaping traditional understandings of authorship, creativity, and learner agency. This paper examines GenAI practices through the conceptual lenses of identity, ideology, and capital as articulated by Darvin and Norton’s (2015) model of investment, and augmented by notions of sociomateriality (Fenwick, 2015). By foregrounding the interaction between human learners and algorithmically driven platforms, this paper highlights the complex distribution of agency and power across human and nonhuman actors. Learners' access to material, cultural, social, linguistic and semiotic resources can shape dispositions towards GenAI and practices. At the same time, platform designs, algorithmic processes, and datasets are not neutral but material manifestations of ideologies and power structures that can steer interactions, where both learners and tools are positioned in different ways. By applying a critical lens to understand human-AI interactions, this paper underscores that learners’ identities and their capacity to exert agency are deeply intertwined with their ability to recognize, interrogate, and navigate these embedded ideologies. To invest in agentive GenAI practices, learners need to develop critical digital literacies that discourage them from passively consuming AI-generated content and instead enable them to recognize how power operates in these interactions. By actively negotiating their own meanings and intentions, learners can transform their GenAI practices into meaningful learning opportunities.
How does interaction with LLM chatbots shape human understanding of culture?
David Wei Dai, University College London
Zhu Hua, University College London
Guanliang Chen, Monash University
Against the proliferation of Large-Language-Models (LLM)-based Artificial Intelligence (AI) products such as ChatGPT and Gemini, and their increasing use in professional communication training, researchers, including applied linguists, have cautioned that these products (re)produce cultural stereotypes due to their training data (Dai & Zhu, 2024). However, there is a limited understanding of how humans navigate the assumptions and biases present in the responses of these LLM-powered systems, described by Jones (2024) as “culture machines” and the role humans play in perpetuating stereotypes during interactions with LLMs. In this article, we use Sequential-Categorial Analysis, which combines Conversation Analysis and Membership Categorization Analysis, to analyse simulated interactions between a human physiotherapist and three LLM-powered chatbot patients of Chinese, Australian, and Indian cultural backgrounds. Coupled with analysis of information elicited from LLM chatbots and the human physiotherapist after each interaction, we demonstrate that users of LLM-powered systems are highly susceptible to becoming interactionally entrenched in culturally essentialized narratives. We use the concepts of interactional instinct and interactional entrenchment to argue that whilst human-AI interaction may be instinctively prosocial, LLM users need to develop Critical Interactional Competence (CritIC) for human-AI interaction through appropriate and targeted training and intervention, especially when LLM-powered tools are used in professional communication training programmes.
Digital accents, homogeneity-by-design, and the evolving social science of written language
AJ Alvero, Cornell University
Quentin Sedlacek, Southern Methodist University
Maricela León, Southern Methodist University
Courtney Peña, Stanford University
Human language is increasingly written rather than just spoken, primarily due to the proliferation of digital technology in modern life. This trend has enabled the creation of generative AI trained on corpora containing trillions of words extracted from text on the internet. However, current language theory inadequately addresses digital text communication's unique characteristics and constraints. This paper systematically analyzes and synthesizes existing literature to map the theoretical landscape of digital language evolution. The evidence demonstrates that, parallel to spoken language, features of written communication are frequently correlated with the socially constructed demographic identities of writers, a phenomenon we refer to as ``digital accents.'' This conceptualization raises complex ontological questions about the nature of digital text and its relationship to identity. The same line of questioning, in conjunction with recent research, shows how generative AI systematically fails to capture the breadth of expression observed in human writing, an outcome we call “homogeneity-by-design.'' By approaching text-based language from this theoretical framework while acknowledging its inherent limitations, social scientists studying language can strengthen their critical analysis of artificial intelligence systems and contribute meaningful insights to their development and improvement.
Bias and stereotyping: Human and artificial Intelligence
Okim Kang, Northern Arizona University
Kevin Hirschi, University of Taxes – San Antonio
Much research has demonstrated listeners’ biases toward L2-accented speech, i.e., perceiving accented utterances as less credible (Lev-Ari & Keysar, 2010), less grammatical (Ruivivar & Collins, 2019) or less acceptable for certain professional positions (Kang et al., 2023), due to their bias and stereotyping issues (Kang & Rubin, 2009). However, AI technologies have emerged as a viable alternative to human perception because they allow for a better understanding of how biases and stereotypes are induced through design, implementation, and training data. For example, in facial recognition, AI recognition has long been known to perform poorly for visual minorities (Hardesty, 2018). In AI systems for speech recognition, AI systems have shown biased and subpar performance in capturing African American Language (Martin & Wright, 2023), and male speakers were more correctly transcribed by YouTube’s automatically generated captions than female speakers (Tatman, 2017). Furthermore, Chinese first language (L1) speakers were less accurately processed than Spanish and Indian L1 speakers of English (Bae & Kang, 2024). The implications of such social biases and their persistence in AI have not yet been fully understood by Applied Linguistics scholars or computer scientists who investigate AI ethics. Accordingly, in this presentation, we will discuss issues in social biases that impact users of different language varieties and relate these to the currently limited research on biases in AI. We will focus on stereotyping research amongst human listeners in various social and educational contexts as well as review research that investigates these biases in AI. We propose that bias found in social perception research can serve as an agenda for how to investigate AI models’ classification, processing, and generation tasks. The presentation will conclude with specific recommendations and future directions for research and pedagogical practices.
Addressing GenAI biases for ELT from a Global Englishes perspective
Seongyong Lee, University of Nottingham Ningbo China
Jaeho Jeon, University of Alabama
Jim McKinley, University College London
Heath Rose, University of Oxford
This study explores the capability of ChatGPT, a large language model (LLM)-based generative AI (GenAI) tool, to support Global Englishes Language Teaching (GELT), addressing the critical gap between GenAI’s potential and its inherent biases that limit representation of diverse English varieties. LLMs typically overrepresent dominant English varieties and favor standardization through algorithmic design, undermining the pluricentric principles of GELT. Thus, it aims to develop effective GenAI-GELT instructional modules using customized ChatGPT models. Through Design and Development Research methodology, we developed three models: a Basic Model using single-step prompting, Refined Model 1 employing multi-step prompting, and Refined Model 2 integrating ELF corpora as external datasets with advanced prompt engineering techniques. The Models were assessed across three pedagogical roles: content curator for World Englishes materials, evaluative feedback provider for ELF writing, and conversation partner facilitating ELF interaction. Findings revealed that while basic prompting defaulted to native-speaker norms, the integration of advanced prompt engineering with external corpus data substantially improved outputs aligned with GELT principles, though standardization biases persisted. The study highlights three critical implications: the importance of teacher-GenAI collaborative design processes, the necessity of teachers' GELT-specific professional knowledge, and their essential role in orchestrating complementary resources to maximize GenAI's pedagogical effectiveness while mitigating inherent biases in language education.
Exploring the dual impact of AI in PELA-Potentials and pitfalls
Tiancheng Zhang, University of Auckland
Rosemary Erlam, University of Auckland
Morena Botelho de Magalhães, University of Auckland
This paper explores the complex dynamics of using AI, particularly generative AI (GenAI), in Post-Entry Language Assessment (PELA) at the tertiary level. Empirical data from trials with a PELA at a large research-led university in New Zealand are presented in two studies. The first study examined the capability of GenAI to generate a reading text and assessment items that might be suitable for use in the PELA. A trial of this GenAI-generated academic reading assessment on a group of target participants (n = 132) further evaluated the suitability of the reading assessment. The second study investigated the use of a fine-tuned GPT-4o model for rating writing tasks and assessing whether Automated Writing Evaluation (AWE) can provide feedback of comparable quality to human raters. Findings indicated that while GenAI shows promise in generating content for reading assessments, expert evaluations reveal a need for refinement in question complexity and targeting specific subskills. In AWE, the fine-tuned GPT-4o model aligns closely with human raters in overall scoring but requires improvement in delivering detailed and actionable feedback.
A subsequent Strengths, Weaknesses, Opportunities, and Threats analysis highlighted AI’s potential to enhance PELA by increasing efficiency, adaptability, and personalization. AI could, for example, extend PELA’s scope to areas such as oral skills and dynamic assessment. However, challenges such as academic integrity and data privacy remain critical concerns. The paper proposes a collaborative model integrating human expertise and AI in PELA, emphasizing the irreplaceable value of human judgment. The need to establish clear guidelines for a human-centred AI approach within PELA is also emphasised so that ethical standards are maintained and assessment integrity upheld.
Tracking the impact of generative AI on EFL students’ motivation, engagement, and writing quality over an academic year
Jerry Huang, Kyoto Sangyo University
Atsushi Mizumoto, Kansai University
Since the introduction of ChatGPT, a widely recognized generative AI chatbot, numerous studies have examined its potential applications in education, especially in language learning, highlighting both advantages and drawbacks. This empirical study tracks a group of university students over one academic year, investigating changes in their motivation, engagement, and writing quality after incorporating ChatGPT into in-class writing workshops. Results showed increases in motivation and engagement, as well as improvements in writing accuracy and cohesion. The findings suggest that educators should take the initiative in introducing GenAI to students and provide guidance on its responsible use. Additionally, educators may benefit from training in designing effective prompts that align with classroom objectives.
On the nascency of ChatGPT in foreign language teaching and learning
Shaohua Fang, Purdue University
ZhaoHong Han, Columbia University
The emergence of ChatGPT as a leading artificial intelligence language model developed by OpenAI has sparked substantial interest in the field of applied linguistics, due to its extraordinary capabilities in natural language processing. Research on its use in service of language learning and teaching is on the horizon and is anticipated to grow rapidly. In this review article, we purport to capture its nascency, drawing on a literature corpus of 71 papers of a variety of genres – empirical studies, reviews, position papers, and commentaries. Our narrative review takes stock of current research on ChatGPT’s application in foreign language learning and teaching, uncovers both conceptual and methodological gaps, and identifies directions for future research.
A question of alignment – AI, GenAI and Applied Linguistics
Niall Curry, Manchester Metropolitan University
Tony McEnery, Lancaster University
Gavin Brookes, Lancaster University
Recent developments in artificial intelligence (AI) in general, and Generative AI (GenAI) in particular, have brought about changes across the academy. In applied linguistics, a growing body of work is emerging dedicated to testing and evaluating the use of AI in a range of subfields, spanning language education, sociolinguistics, translation studies, corpus linguistics, and discourse studies, inter alia. This presentation explores the impact of AI on applied linguistics, reflecting on the alignment of contemporary AI research with the epistemological, ontological, and ethical traditions of applied linguistics. Through this critical appraisal, we identify areas of misalignment regarding perspectives on knowing, being, and evaluating research practices. The question of alignment guides our discussion as we address the potential affordances of AI and GenAI for applied linguistics as well as some of the challenges that we face when employing AI and GenAI as part of applied linguistics research processes. The goal of this presentation is to attempt to align perspectives in these disparate fields and forge a fruitful way ahead for further critical interrogation and integration of AI and GenAI into applied linguistics.