Home Emergent AI-assisted discourse: a case study of a second language writer authoring with ChatGPT
Article Open Access

Emergent AI-assisted discourse: a case study of a second language writer authoring with ChatGPT

  • Sharin R. Jacob

    Sharin R. Jacob is a researcher in Computational Thinking Pathways at Digital Promise Global working in Learning Experience Design. Her research interests examine the linguistic and sociocultural factors that help multilingual students succeed in computing. Sharin’s work has been published broadly in journals including Computer Science Education, ACM Transactions on Computing Education, Teachers College Record, and Journal of Computer Science Integration.

    ORCID logo EMAIL logo
    , Tamara Tate

    Tamara Tate is Associate Director of the Digital Learning Lab at the University of California, Irvine. She leads the Lab’s work on generative AI and writing. As the PI of a NSF-funded grant, she is studying the use of generative AI in undergraduate writing courses. She also studies secondary student writing as a member of the IES-funded national WRITE Center. She received her B.A. in English and her Ph.D. in Education at UC Irvine and her J.D. at U.C. Berkeley.

    and Mark Warschauer

    Mark Warschauer is Distinguished Professor of Education and director of the Digital Learning Lab at UC Irvine with affiliated appointments in Informatics, Language Science, and Psychological Science. His research is currently focusing on generative AI in Education, the use of conversational agents to support children’s learning and on the teaching and learning of computer science for linguistically diverse students. He is a member of the National Academy of Education.

Published/Copyright: November 1, 2024

Abstract

The rapid proliferation of ChatGPT has incited debates regarding its impact on human writing. Amid concerns about declining writing standards, this study investigates the role of ChatGPT in facilitating writing, especially among language learners. Using a case study approach, this study examines the experiences of Kailing, a doctoral student, who integrates ChatGPT throughout their writing process. The study employs activity theory as a lens for understanding writing with generative AI tools and data analyzed includes semi-structured interviews, writing samples, and GPT logs. Results indicate that Kailing effectively collaborates with ChatGPT across various writing stages while preserving her distinct authorial voice and agency. This underscores the potential of AI tools such as ChatGPT to enhance writing for language learners without overshadowing individual authenticity. This case study offers a critical exploration of how ChatGPT is utilized in the writing process and the preservation of a student’s authentic voice when engaging with the tool.

1 Introduction

The release and rapid diffusion of ChatGPT has sparked wildly different views of its relationship to human writing. In light of longstanding concerns about a national writing crisis (Intersegmental Committee of the Academic Senates et al. 2002; The National Commission on Writing 2003, 2004), educators have voiced apprehensions that the teaching of writing and its use as an assessment tool might be seriously compromised. Conversely, the business sector eagerly anticipated the potential for rapidly producing written content at a reduced cost.

Historically, students in the US have grappled with mastering academic writing genres (Graham and Perin 2007). The landscape of academic writing is further complicated when considering the experiences of second language learners (Booth Olson et al. 2023). These students often navigate a dynamic interplay of cultural and social values that may not always align with mainstream academic expectations (Booth Olson et al. 2023; Bunch 2013). This potential misalignment presents an opportunity for deeper exploration of how students with diverse linguistic backgrounds approach academic writing development.

There is a growing body of work on how second language learners navigate the types of writing necessary for academic research (Curtis 2023; Rahman et al. 2023; Steiss et al. 2024; Warschauer et al. 2023). The value of ChatGPT, in offering input, feedback, or scaffolding for writing, is increasingly acknowledged. Yet, a pivotal question arises: Does relying on AI-generated texts equate to academic misconduct, or can AI-assisted writing coexist with the preservation of the human author’s distinct voice and agency?

To explore this issue, we present a case study of a student who thoroughly integrates ChatGPT into all aspects of the authoring process. First, we provide a background on using ChatGPT to facilitate writing for language learners. Next, we describe our participant, Kailing, who uses ChatGPT throughout each stage of the writing process, to write an academic proposal during her doctoral studies. We then describe how activity theory can be employed as a theoretical framework to advance our understanding of how AI tools such as ChatGPT can mediate the writing process. From there, we describe our single-subject case study design and data collection, including semi-structured interviews, analysis of writing artifacts, and GPT logs. Findings from our case study indicate that Kailing applied her expertise to dynamically interact with ChatGPT during the brainstorming, research, text generation, and revision processes. Notably, throughout this process, she retained her own sense of voice and agency. Findings such as these highlight the potential of generative AI tools, such as ChatGPT, to enhance writing for language learners in academic settings.

2 Literature review

In the past 2 years, there has been a rapid increase in studies on the integration of AI in language learning (Bin-Hady et al. 2023; Fitria 2023; Kohnke et al. 2023; Liang et al. 2023; Steiss et al. 2024; Warschauer et al. 2023). Although most findings indicate that language learners could greatly benefit from generative AI tools, they face disproportionate challenges regarding accusations of academic misconduct. Notably, GPT detectors exhibit considerable bias against these students. A study that analyzed the precision of seven widely used GPT detectors on 91 TOEFL essays and 88 US student essays found that while the detectors reliably identified essays from US students, they mistakenly labeled over half of the TOEFL essays as “AI generated,” with an average false-positive rate of 61.3 % (Liang et al. 2023). Building upon this work, Warschauer et al. (2023) explore the affordances and challenges of using AI-based tools for second language learners. They pinpoint the dilemmas faced by these learners as they engage with AI tools. To echo Liang et al. (2023), they underscore the pressure on language learners to mirror the target language. However, when L2 students utilize tools such as ChatGPT, they often face allegations of plagiarism, especially when compared to native speakers (Liang et al. 2023).

In addition to detection bias, generative AI tools, such as ChatGPT, have been shown to exhibit discriminatory bias (Kenthapadi et al. 2023). Since they are trained on large corpora of data, engineers encounter challenges when attempting to audit the data for specific biases and discriminatory elements (Bender et al. 2021). For instance, large language models have exhibited biases and stereotypes related to gender (Bolukbasi et al. 2016), disability (Hutchinson et al. 2020), and language (Lee 2023). Given these findings, it is imperative to approach the use and interpretation of outputs from generative AI tools with caution, ensuring that measures are in place to counteract and mitigate inherent biases.

There is substantial evidence indicating that ChatGPT can perpetuate bias, stereotypes, and inaccuracies. However, despite its recognized limitations, it would be a mistake to overlook its potential benefits for underperforming students, such as emerging writers, students with disabilities, and language learners. Regarding job readiness, Warschauer et al. (2023) highlight the eagerness of corporations to incorporate AI into their systems, yet students designated as English learners have limited access and opportunities to learn how to effectively utilize these tools (Warschauer et al. 2023). Denying language learners the chance to learn and employ generative AI tools can adversely impact their job readiness. A recent study found that using ChatGPT for work-related tasks in areas such as marketing and grant writing reduced the time to complete tasks by 40 % while improving quality by 20 % according to a double-blinded review of work performed. Importantly, low-performing employees improved the most (Noy and Zhang 2023). A similar study from Harvard Business School assigned strategy consultants to conceptualize and develop new product ideas in one of three conditions–without AI, with ChatGPT, and with ChatGPT and an overview of prompt engineering. Again, the two ChatGPT groups improved the most, largely due to a catch-up effect among the below-average writers who improved by 43 % in their writing tasks compared to an improvement of 17 % among above-average writers (Dell’Acqua et al. 2023). Findings such as these suggest that learning tools such as ChatGPT have the potential to level the playing field for struggling writers in the workforce.

Many are concerned that generative AI tools such as ChatGPT could boost productivity at the expense of innovation and creativity. However, empirical evidence suggests that when utilized appropriately, that may not be the case. A study investigating the influence of GenAI on creative content production found that individuals inspired by AI-generated content crafted more innovative narratives compared to those drawing from other sources (Doshi et al. 2023). This impact increased significantly among writers perceived as less creative. Such results highlight ChatGPT’s potential to equalize opportunities for emerging writers.

There are legitimate concerns that technology cannot replicate the nuances of advanced writing, tasks such as these require creativity, innovation, high-quality content, structured organization, and logically sound arguments. However, a recent study investigating ChatGPT’s role in supporting students with argumentative suggested that ChatGPT aids in content generation and streamlines the writing process, including outline preparation, content revision, proofreading, and post-writing reflection (Su et al. 2023). Leveraging generative AI tools for complex writing tasks, such as argumentative writing, is particularly beneficial for language learners, who often grapple more with the higher-order components of argumentative writing compared to their native-speaking counterparts (Booth Olson et al. 2023).

Finally, it has been long argued that human feedback is irreplaceable. However, emerging studies show generative AI tools such as ChatGPT are capable of delivering high-quality feedback for students. For example, a current investigation comparing feedback from humans to that provided by ChatGPT revealed that the practical differences between the two were minimal (Steiss et al. 2024). The consistency between human and AI-generated feedback suggests that ChatGPT could be a reliable and scalable tool, potentially bridging the feedback gap often faced by language learners due to limited human resources.

In conclusion, the rapid integration of AI into language learning has opened new horizons, offering potential benefits, especially for underperforming and L2 students. However, the challenges – ranging from detection and discriminatory biases to issues of academic integrity and the potential dilution of creativity – are undeniable. These concerns underscore the importance of a balanced and cautious approach to the application of tools such as ChatGPT. Notwithstanding its limitations, the evidence suggests that, when used judiciously, ChatGPT can be an asset, providing quality feedback, fostering creativity, and aiding in content generation. Moving forward, as the field continues to evolve, continuous evaluation and mindful implementation of such tools will be crucial to ensure that they serve to augment rather than impede the processes of language learning and advanced writing.

We now turn to the participant in our case study to better understand the conditions under which she employed ChatGPT to produce academic research.

3 Theoretical framework

Activity theory, anchored in Engeström’s (1987, 1999) work and drawing upon the Vygotsky and Cole (1978) sociocultural framework, posits that human interactions are multifaceted, influenced by numerous factors, and inherently dynamic. In aligning this theory with the writing process, we lean on Kessler’s (2020) conceptualization. Within this framework: The subject (or agent) undertakes actions with a specific goal in mind. These actions are facilitated by tools or mediators. In the realm of writing, the end product (the written text) is the goal, while various aids and resources (e.g., computer, chatbot, editing software, thesaurus) facilitate the creation of this text. The interaction is also shaped by three contextual elements: Rules provide guidelines or standards for the subject’s behavior. Community represents the broader context and stakeholders that influence the subject’s actions. Division of labor illustrates how tasks related to the primary goal are distributed among participants. In the context of academic writing, a researcher (subject) may engage with colleagues, field experts, editors, and peers during the writing and revision process. Notably, tensions might arise, such as discrepancies between established writing norms and specific assignment criteria. These tensions, however, can serve as catalysts for the learner’s developmental journey (Fujioka 2014).

We build on this theory by contextualizing it within the use of generative AI tools for writing. In our case study, the subject undertakes the goal of writing a research proposal. The end product, or written text, is mediated by ChatGPT, a generative AI tool that uses language processing to facilitate writing, among other tasks. Within this dynamic interaction between text and tool, our participant is constrained by the rules of academic writing, and the context of her broader community (i.e., the reviewers of the proposal), while engaging in the division of labor by conferring with her colleagues, peers, and scholars on how to improve her writing.

Our study is also based on the cognitive process models of writing (Flower and Hayes 1981), in which writing consists of planning, translating, and reviewing, and revising. Composition is a recursive process (Berninger et al. 1996; McCutchen 1996): writers cycle through the planning, translating, and reviewing multiple times, and these stages all interact with one another throughout the composing process (Flower and Hayes 1981).

4 Methods

4.1 Research questions

This paper addresses the following research questions (RQs):

  • RQ1: How does a graduate student and second language writer in English employ ChatGPT in each phase of her academic writing process?

  • RQ2: To what extent does the student maintain her authentic voice when utilizing ChatGPT in the writing process, and how is her voice characterized within this context?

  • RQ3: How do the students’ perceptions of using ChatGPT for writing evolve?

4.2 Research design

A case study (Mabry 2008) design was employed to delve deeply into the specific experiences of a single participant. This approach aimed to elucidate the ways in which the participant utilizes ChatGPT for academic research, emphasizing her retention of voice and agency.

4.3 Participant

The study centered on Kailing (pseudonym); a PhD student enrolled in a graduate program in Southern California. She was intentionally selected for her intensive engagement with ChatGPT in the context of academic research.

Kailing is an international graduate student from China at a research university in the United States. She has been in the US for 4 years for her master’s studies and the beginning part of her doctoral study. Kailing is moderately proficient in English, especially considering that she didn’t major in English as an undergraduate, but like most international students, she is not yet a polished English language writer. Kailing is a skilled and comfortable user of new technologies and has long used other tools, such as Grammarly, to help her with her writing.

Kailing came to our attention as an extraordinarily capable writer for a learner of English. She received straight As or A + s in her first year of doctoral studies in a writing-intensive field, won a prestigious best paper award at a highly competitive international conference, and has authored grant proposals deemed by faculty to be at a quality level as high or higher than those written by a typical research professor.

It is not only Kailing’s writing that brought her to our attention but also her comprehensive use of ChatGPT at all stages of the writing process. After a brief interview with Kailing, it became clear that she was the most thorough and proficient user of ChatGPT for academic writing that we had met or knew of. We then invited her to be part of this case study by answering our interview questions and sharing with us her prompts and ChatGPT responses as well as the iterative versions of her papers – and generally telling and showing us how she uses the tool for academic writing. We believe that it is important for academics debating the use of generative AI to have a transparent look at how one highly effective student uses the technology before deciding to ban AI in their classes; rather this case study might inspire some use cases to explore for other students.

4.4 Data collection

To explore the strategies employed by Kailing in her academic research with ChatGPT, we utilized semi-structured interviews. These interviews allowed for both predetermined questions and flexibility for follow-up queries based on Kailing’s responses. A total of four interviews, totaling about 2 hours, were conducted. Each interview lasted around 3 minutes. The total word count of the four combined interviews is 8,621 words. The first three interviews were conducted over the summer or 2023 to understand how Kailing used ChatGPT for the purpose of academic writing. We then conducted a second interview during late fall of 2024 to better understand how her use of ChatGPT changed over time. We also analyzed her written artifacts, including her research papers, proposals, and ChatGPT logs.

4.5 Data analysis

The analysis of interview data was undertaken employing inductive qualitative coding approaches (Saldaña 2021). The first cycle of coding involved identifying themes that pertained to the research questions and were rooted in literature to date on the use of ChatGPT for language learners. After assigning initial codes, we developed a codebook to reveal emergent themes from the study (see Table 1 for sample codes). To enhance the validity, we conducted member checks by sharing our interpretations with the participant for their feedback (Lincoln and Guba 1985). This case study focused on a singular participant to explore the nuanced interactions between generative AI and language acquisition processes.

Table 1:

Sample codes, definitions, and examples.

Code Subcode Definition Example
Brainstorming Initial queries Generating broad queries to explore potential ideas. “I kind of realized that if you give a prompt like this to ChatGPT the ideas I received were too generic.”
Refinement Narrowing down queries based on content knowledge and goals. “I refined my prompt ChatGPT and asking how can I apply conversational agent into the storytelling design with children with ASD?”
Research Cross-referencing Ensuring the accuracy and credibility of generated citations. “I need to pause and read more literature to … check [what] ChatGPT comes up with if it is accurate or if there is evidence to support it.”
Text production Language support Using ChatGPT output to provide linguistic scaffolding. “ I always rely on ChatGPT to generate high-quality text instead of using it to generate ideas for me.”
Efficiency Discovering prompting efficiently produces text. “I think ChatGPT always gives me responses that I’m very satisfied with because it writes something decent and good in a short amount of time, which saves me time.”
Revising Initial revisions Reviewing the initial drafts for general improvements. “I would consider this as a skeleton of my writing.”
Adding details Incorporating detailed evidence and supporting information. “And then I would for example go to the New York Department of Education’s website to look at its investment and funds to the math program and then add details into the skeleton to make it more [clear].”
Maintaining voice and agency Voice alignment Revising text to ensure it reflects the writer’s style. “But I wouldn’t necessarily take what ChatGPT suggests because sometimes I feel it doesn’t look like my writing anymore.”
Ethical considerations Ensuring the use of AI tools aligns with academic ethics. “I don’t feel comfortable using something entirely generated by ChatGPT because it wasn’t my writing anymore and so I feel this is against academic ethics if I use something completely generated by ChatGPT.”

5 Results

Section 5.1 below presents the results that address research question number one: How does a graduate student and second language writer in English employ ChatGPT in each phase of her academic writing process? In addressing the first research question, we used the four stages of the writing process: brainstorming, research, text production, and revising to present the results, as these stages comprehensively capture the essential phases of academic writing. This framework aligns with established models of the writing process that emphasize these sequential steps as crucial for effective composition (Flower and Hayes 1981; Scardamalia and Bereiter 1987). Employing ChatGPT at each of these stages highlights its multifaceted role in supporting both the generation of ideas and the refinement of academic texts, particularly for second language writers (Ong and Zhang 2013). By structuring the analysis around these stages, the study offers a clear and systematic exploration of the student’s interaction with GenAI tools, facilitating a deeper understanding of their influence on academic writing.

5.1 The overall writing process

Some people use ChatGPT to brainstorm ideas or outlines for their writing. Others use it as a very capable copy editor. In either case, they claim their writing is authentic, rather than plagiarized, because ChatGPT is only used for 1–2 steps, and not for drafting the content. In contrast, Kailing uses ChatGPT for brainstorming and outlining, for editing, for drafting–and for much else in between. Yet she harnesses such control of every step of the process that we consider what she is doing as authoring, even if it is not writing in the traditional sense. The best way to understand that is by example, so below we present a step-by-step analysis of how Kailing wrote one of her recent outstanding papers, a research proposal.

Kailing’s writing process, assisted by ChatGPT, can be categorized into four distinct stages: brainstorming (planning), research, text production (translating from thought to text), and revision. Within each stage, Kailing engages in a series of intricate decisions, employing advanced cognitive skills to adapt the text in service of her intended communication goals.

5.1.1 Brainstorming

During the brainstorming phase, Kailing progressively refined her queries, moving from broad to specific, based on her existing content area knowledge, knowledge of prompting the AI, and the goals and context of her writing. For instance, while soliciting the chatbot to produce ten instances of how a particular digital technology can be adapted for students with disabilities, she realized that her initial queries were too generic and not feasible for her research context.

I kind of realized that if you give a prompt like this to ChatGPT, the ideas I received were too generic and not very applicable to or not very doable to make it into a proposal because it was too broad and if you think deep about the technical details of these ideas, you realize that it is not very feasible.

To produce more refined ideas, Kailing consulted both her colleagues and existing research literature. In consulting with her peers, Kailing engaged in peer–peer–computer interaction, which benefits language learners by establishing a frame of reference to refine and negotiate meaning in real-time (Bailey and Heritage 2014; Grapin 2020; Ruiz-Primo 2011). Informed by these insights, she adjusted her research query, leading to new discoveries.

And then I refined my prompt [sic] ChatGPT and asking, how can I apply conversational agent into the storytelling design with children with ASD? And then ChatGPT told me something that I didn’t know … there’s this technology called augmentative and alternative communication keyboard for children with autism, which is frequently used for them to help them communicate with people … So I thought was I didn’t know this before, but ChatGPT told me this, which I think is a very useful information.

Drawing on her initial findings, Kailing reviewed scholarly literature to understand the integration of assistive technologies with digital storytelling and conversational agents. This literature review subsequently informed the formulation of her research proposal’s core ideas.

It’s like an interactive storytelling experience with AI where children can use the AC keyboard to … type the starting of the story, and then the system can build upon the children’s story based on the graphs that they use to continue this storytelling process. So this is the process of me using ChatGPT to generate ideas for my proposal.

During the brainstorming phase, Kailing actively refined her prompts through discursive interaction with the chatbot. This discursive interaction involving multiple human participants in conjunction with an AI chatbot empowered Kailing to utilize her own cognitive and social resources to delineate her research (Grapin et al. 2022). However, generative AI has the potential to transcend mere tools for thought, evolving into active participants in the interaction (Grapin et al. 2022). From a human interaction perspective, Kailing applied advanced cognitive skills, such as evaluation and analysis, to synthesize her findings and eventually formulate the primary objective of her research proposal. Conversely, the chatbot performed its functions semi-autonomously while operating in tandem with Kailing, its human guide (Brady and Lehrer 2021). While Kailing primarily drove the prompting process with her intellectual input, the task of text generation was delegated to the AI, capitalizing on its superior English language processing capabilities. As “computer programs essentially automate abstractions,” (Jacob and Warschauer 2018, p. 5), GhatGPT effectively amplified the intellectual capacities of Kailing and her colleagues in order to produce a coherent research topic.

5.1.2 Research

Kailing employed ChatGPT as a tool for conducting preliminary research and consulting the literature to provide evidence for specific claims. She prompted it for lists of prominent researchers on her topic of interest, sought overarching themes pertinent to her research topic, inquired about researchers particularly aligned with those themes, and further investigated funded projects within her target region that pertained to related subjects.

For example, upon developing an outline for her proposal, she asked ChatGPT to generate citations to support her claims. Understanding the intricacies of how the chatbot functions, Kailing was well-aware that ChatGPT does not always produce credible citations, a recognized limitation of the software.

Lots of its arguments cannot be backed up by evidence … This goes to a drawback of ChatGPT which is even though you ask it to include, for example, five or 10 citations to support the literature, it will make up the citations which means it uses fake citations because of how ChatGPT works, which is a probabilistic prediction of the next word based on the previous word. So I think here comes the part where I need to pause and read more literature to … check [what] ChatGPT comes up with if it is accurate or if there is evidence to support it.

The importance of finding credible citations is underscored by research on citation analysis, which reveals rhetorical markers, discourse structures, epistemological foundations, and empirical perspectives that authors assume when providing a discursive framework of the literature (Hu and Wang 2014). As researchers agree that incorporating citations into academic research is an especially complex phenomenon (Barks and Watts 2001; Hu and Wang 2014; Pecorari 2001), language learners can face challenges when they are asked to integrate textual borrowing strategies into their academic writing (Pecorari 2001). Furthermore, the consequences for inaccurately citing borrowed materials frequently take on administrative and legal dimensions, leading to significant repercussions for language learners. Recognizing this, Kailing cross-references the citations provided by ChatGPT with the existing literature to ensure their accuracy. By unpacking the complexities of generating citations, Kailing devised a methodology for using ChatGPT to reference the works of others while maintaining academic integrity and sophistication in textual borrowing strategies.

5.1.3 Text production

In drafting the text, Kailing adopts an iterative approach with the chatbot. She breaks up her writing sessions into manageable chunks and thinks deeply in advance of each one. For the literature review, for example, she considers the points she wants to make, how to organize them coherently, and comes up with bullet points that she inputs to ChatGPT. ChatGPT then does what for Kailing is the “hard work” as a non-native speaker, it rapidly drafts text covering her points. She then pauses and reads more literature, gathering more content knowledge to inform her writing, and supplements the output. Often, Kailing gives ChatGPT such detailed outlines that it does not seem like the AI is adding much value – mostly with transitions between ideas, and maybe also confidence and self-efficacy. She has found that these detailed prompts are more useful than simply a broad query based on the assignment. As an illustration, Kailing reported on the following broad query she made regarding the composition of a research proposal.

Could you write me the research proposal, including the following sections, which are the sections required by the professor? In this assignment, for example, its introduction, which is like background, and conceptual framework and research question and also methodology. I didn’t expect that much from Chat GPT because every time I gave this kind of prompt to ChatGPT, the response was always first of all, very short. And second, especially for the part of the literature review or the conceptual framework. I don’t think ChatGPT could tell a very coherent story.

Kailing recognized that ChatGPT could, at best, supply basic ideas for her proposal. To this end, she drew upon her own expertise to refine and expand on her preliminary draft in order to ensure a more cohesive narrative. In particular, notice her rhetorical awareness of the need to tell a coherent story in order to have a successful proposal. This genre knowledge allows her to successfully note the limitations of ChatGPT’s output and also provide more specific prompting to increase the usefulness of the output.

Interestingly, Kailing believes that ChatGPT excels more as a text generator than as a source of high-quality ideas; however, she acknowledges that she continues to glean new insights from the chatbot. In addition, the underlying large language models are evolving rapidly and improving in significant ways, so this may only be a temporary reality.

To me, because I think ChatGPT is a very high-quality text generator, but it’s … not a very good idea generator. So I always rely on ChatGPT to generate high-quality text instead of using it to generate ideas for me, but on the other hand, I can sometimes get some useful input from chatting material like the AAC keyboard, I wouldn’t have known it if there wasn’t ChatGPT for me.

Kailing used ChatGPT for ideation even though it was not adept at producing unique, creative ideas. Nonetheless, even the fallible tool can provide something to react to or refine, and in some cases even provides usable ideas. She then uses the chatbot for its (currently) more skillful ability to produce fluent text, a labor-intensive endeavor, especially for multilingual writers.

I think ChatGPT always gives me responses that I’m very satisfied with because it writes something decent and good in a short amount of time, which saves me time.

Taken together, these findings suggest that ChatGPT acts as an assistive tool, facilitating language learners such as Kailing in expressing ideas in the target language with a speed and efficiency akin to their native language. In this way, natural language processing facilitates human cognitive processing to achieve desired ends, such as producing robust academic writing.

5.1.4 Revision

Kailing acknowledges ChatGPT’s constraints in delivering the granularity required for scholarly articles. She perceives her initial drafting with ChatGPT as a foundation-building exercise while considering the revision phase as the stage to incorporate essential details and evidentiary support for her assertions.

So going back to the prompts that I just talked about, even though I give ChatGPT a seemingly detailed prompt, but still the response that I got from ChatGPT [is] sometimes still too generic and vague. For example, I just asked Chad GPT to write something about the increasing attention paid by the New York Department of Education to early math learning … There is no detail at all … Some really pretty words? Yes. So I would consider this … if I want to write a paragraph about New York’s investment in math education, I would consider this as a skeleton of my writing. And then I would for example, go to the New York Department of Education’s website to look at its investment and funds to the math program and then add details into the skeleton to make it more [clear].

During the interview, Kailing repeatedly praises the eloquent language, or the “really pretty words” produced by ChatGPT, but also critically assesses the substance of its ideas. Although she concedes having gained some insights from ChatGPT, she believes that the concepts it formulates often fall short of the rigor required for scholarly writing.

Section 5.2 below presents the results that address research question number two: To what extent does the student maintain her authentic voice when utilizing ChatGPT in the writing process, and how is her voice characterized within this context?

5.2 Authorship and voice

The issue of authorship and voice emerges prominently in Kailing’s narrative. While she insists on aligning ChatGPT-generated content with her own voice through revisions, she also values the platform’s ability to produce eloquent text – a feature appreciated by both language learners and native speakers. Nevertheless, her distinct authorial identity is shaped by continuous decisions made during her interactions with ChatGPT, parallel discussions with peers, and her engagement with existing academic literature, all aimed at upholding scholarly integrity.

Kailing possesses a distinct understanding of her writing style and voice; consequently, she meticulously revises her work to ensure it authentically mirrors her authentic voice and writing style.

But I wouldn’t necessarily take what ChatGPT suggests because sometimes I feel it doesn’t look like my writing anymore. So although I know the language is better, or it used more beautiful words, I don’t feel comfortable using something entirely generated by ChatGPT because it wasn’t my writing anymore, and so I feel this is against academic ethics, if I use something completely generated by ChatGPT.

It is interesting that Kailing has such a clear sense of her own voice. Many students have lost their sense of their own voice after high school and the focus on the AP essay genre. Kailing has retained her own voice and guards it fiercely. This is an area that may become more important to teach writers as time goes by and the use of generative AI is commonplace.

Rather than perceiving ChatGPT merely as a writing aid, Kailing regards it as an intellectual collaborator, facilitating her active engagement in the research process.

I definitely enjoyed the process more especially because of this thinking partner role of ChatGPT. I feel like I have like a back and forth interaction with someone who’s always there. And like, and because of this interaction, I can gradually build upon my previous ideas based on my thinking and ChatGPT input and gradually improve the idea.

Kailing employed intricate strategies to maintain her authorial integrity while utilizing ChatGPT for academic research. Initially, she engaged with the chatbot as a collaborative tool for refining her ideas, ensuring she did not lean too heavily on it. She meticulously verified the claims and supported them with evidence anchored in extant academic research. Furthermore, Kailing solicited feedback from peers, using her social resources, and immersed herself in existing scholarly works to hone her arguments. Though she incorporated ChatGPT in the drafting stages, she carefully revised its output to achieve the depth and rigor demanded by academic research. Continuous reflection about her writing approach, coupled with a clear recognition of ChatGPT’s strengths and weaknesses, allowed her to harness the chatbot’s ability to generate high-quality text for her own, specific purposes.

We now turn to Kailing’s follow-up interview to examine how her use of ChatGPT for academic writing changed over time.

Section 5.3 below presents the results that address research question number three: How do the students’ perceptions of using ChatGPT for writing evolve?

5.3 Results from follow-up interview

As Kailing persisted in using ChatGPT, she began to identify predictable patterns in both its content and structure, leading her to decrease her dependence on the program. Specifically, during the brainstorming phase, Kailing observed that regardless of the number of prompts she provided, the software tended to produce the same related topics.

When I first asked ChatGPT to generate ideas, I got pretty excited about these ideas. But then, after … sitting down and thinking about it for a while, I figured … not all of them are realistic or reliable. Even though I asked [ChatGPT] to elaborate on any of the ideas, I figured that it actually gave me pretty similar things. So I figure that is part of ChatGPT’s model. Conversational agents or generative AI is always associated with something like interactive storytelling or adaptive learning. That’s how the model works. I think that the words “interactive storytelling” or “adaptive learning” are more likely to come after “conversational agents.” So, I started to rely less on it.

In Kailing’s experience with ChatGPT, she initially expressed enthusiasm regarding its idea-generation capabilities. However, through her continued use, she began to recognize recurring patterns in the responses. Specifically, she noted that the software frequently presented related suggestions for a given topic. Kailing hypothesized that these patterns are caused by ChatGPT’s design, reflecting a predisposition within the model to connect terms such as “conversational agents” with nearby language such as “interactive storytelling” and “adaptive learning.” Her recognition of these predictable patterns led her to reduce her reliance on ChatGPT, understanding its limitations in diversifying its content-based suggestions.

Kailing discovered a similar limitation to the AI tool at the sentence level, particularly with word choice.

[In the past], I would even use some of the sentences [generated by ChatGPT] in my writing. But … since I’ve used ChatGPT more and more, and also read a lot of text generated by ChatGPT, I’ve noticed some patterns. For example, [words] like “harness,” or “harness AI,” or “we want to address the significant gap between A and B,” and also when discussing culturally relevant pedagogy. “We need to pay attention to the cultural nuances.” So some of these signal words are, I noticed, always generated [by ChatGPT]. So I started to rely less on it.

Kailing initially incorporated ChatGPT’s sentences directly into her work but, over time, recognized repetitive phrasing and vocabulary from the software. These recurring patterns led her to decrease her reliance on the tool for sentence generation.

However, Kailing continued to use ChatGPT to help craft certain textual features at the sentence level that she found to be challenging as a language learner, such as sentence variation.

I started to rely less on ChatGPT in terms of the framing of the exact sentence or word, but I still would put my writing into ChatGPT and ask it to paraphrase it to help me at more of a sentence level … As a non-native speaker, a challenge I came across during academic writing is it’s hard … for me to diversify the sentence structure. It’s very easy for me to say, “Oh, past literature demonstrates this. In addition, previous literature also indicates this.” So you know, always follow this kind of same sentence structure … makes the writing a little bit boring. But by referring to ChatGPT’s paraphrasing, I would know … maybe I can change … the sentence structure a little bit. Like the active tense, the passive tense and how to start the sentence.

Rather than solely relying on ChatGPT for content generation, Kailing began to leverage the tool more specifically for paraphrasing and diversifying her sentences – a challenge she identifies with as a language learner. She acknowledged her tendency to use repetitive sentence structures, which she perceives as monotonous. To this end, ChatGPT represents a valuable tool for language learners, offering targeted support for refining their writing skills, so that these students can better display their content knowledge without being taxed with unnecessary linguistic burdens.

While Kailing still used ChatGPT for multiple purposes in her writing, during the time of the follow-up interview, she found that consulting her colleagues and peers, as well as human writing assistants, was more valuable than using ChatGPT. She explained her reasoning in a joke.

In software engineering, there is a joke that oh, it looks pretty cool that you can use ChatGPT to generate code. So right now your workflow becomes 1) you get the code from ChatGPT immediately, and then 2) you spend 2 h debugging the code generated by chatGPT. I actually feel the same way [with writing].

In the context of academic writing, while tools such as ChatGPT offer immediate support, Kailing’s reflection underscores the importance of human expertise. Her analogy between the time-consuming debugging process and the editing and refinement that is necessary to produce high-quality AI-assisted writing emphasizes the irreplaceable value of human expertise in the writing process.

6 Discussion

In this study, we critically engage with two perspectives: the nature and practice of writing in the age of advanced AI, and the implications for academic writers, particularly those situated in contexts comparable to Kailing’s.

The findings of this study align with activity theory (Engeström 1987, 1999; Vygotsky and Cole 1978), which emphasizes the interaction between subjects, tools, and the community within a sociocultural context. Kailing’s use of ChatGPT across different phases of her writing process exemplifies how generative AI tools can mediate and enhance the writing activities of second language learners. Each stage of her writing – brainstorming, research, text production, and revising – demonstrates how ChatGPT acts as a tool that facilitates cognitive processes and supports Kailing in navigating complex academic tasks, thereby enhancing her writing proficiency and maintaining her authorial voice. This dynamic interaction highlights the tool’s role in not only generating content but also in shaping and refining the user’s ideas through iterative feedback and engagement.

Moreover, activity theory’s focus on the sociocultural context of learning and development is evident in Kailing’s interactions with her academic community (Kessler 2020; Vygotsky and Cole 1978). The rules and expectations of academic writing, as well as the division of labor among her peers and advisors, create a structured environment in which ChatGPT serves as an intermediary tool (Kessler 2020). This framework supports the notion that Kailing’s developmental journey is influenced by the interplay between her cognitive abilities, the affordances of the AI tool, and the academic standards she must meet. These interactions reveal how generative AI tools can be integrated into writing practices to support language learners, reflecting the core principles of activity theory in mediating human actions through cultural and technological artifacts.

At the core of human endeavors lies the role of tools. In examining writing through the lens of activity theory, the introduction of a new tool, such as ChatGPT, recalibrates the dynamic of human-tool interaction, necessitating a comprehensive approach that integrates human, tool, and context (Engeström 1999).

Bateson’s (1972) contemplation of the blind man and his stick offers a deeper reflection on this relationship. He posits the inseparability of the tool (the stick) from the user (the blind man), thereby highlighting the symbiotic relationship between them. Similarly, Postman’s (1993) observation on the transformation of Europe following the printing press’s advent reinforces the transformative potential of tools on societies.

Communication technologies, particularly as highlighted by Ong (1982), possess the capacity to reconfigure cognitive and societal landscapes. The inception of the printing press didn’t merely augment Europe; it reshaped it. Similarly, digital technologies don’t merely enhance our traditional literacies; they redefine them.

The essence of this transformation is evident in the arena of digital literacy. The internet’s emergence was not simply an alternative medium for traditional content delivery. Instead, for many, it heralded a paradigm shift in the art of reading, writing, and research. The challenge and opportunity were to comprehend and adapt to these new digital literacies, which include online research, content creation across multiple media, and writing for online communication to potentially large audiences.

The rise of ChatGPT reinvigorates debates around the nature of writing and the impact of technology on authorship. Does it signify a mere shift in how writing is done, or does it denote a fundamental metamorphosis of the nature of writing? Our exploration of Kailing’s engagement with ChatGPT offers valuable insights into how one adept user of technology navigated this liminal space, authoring through interaction with generative AI.

Our paper highlights questions regarding the nature of writing proficiency and sheds light on the potential of tools such as ChatGPT in academic settings. The case of Kailing serves as an illustrative microcosm of this broader landscape. As demonstrated, while Kailing utilized ChatGPT as a generative tool, she never abdicated her agency, ensuring that the content produced echoed her academic integrity and authorial voice.

This intricate balance between humans and tools, mediated by contextual rules, communities, and divisions of labor (as framed by activity theory), forms the crux of our study. While the “blind man and the stick” analogy underscores the interdependency of humans and tools, Kailing’s narrative reveals a nuanced engagement, where ChatGPT serves both as a collaborator and a tool, aiding but never replacing the human intellect.

To this end, our study highlights Kailing’s changing relationship with ChatGPT in academic writing, focusing on the tool’s strengths and limitations. While Kailing appreciated ChatGPT’s assistance in employing sentence variation, she noted recurring patterns that limited its utility. Her journey underscores the balance between generative AI support and the indispensable value of human expertise in refining and enhancing writing.

The crux of our exploration is not just how AI tools such as ChatGPT can be employed, but more critically, how they reshape the writing process, challenge traditional notions of authorship, and make us rethink the nature of writing pedagogy and assessment in academic settings. At the same time, Kailing’s continued journey underscores the relationship between AI and the human experience, underscoring that while tools such as ChatGPT can offer invaluable support, their limitations necessitate human intervention to ensure high-quality writing.

7 Conclusions

We present a case study of one participant, understanding its limitations. Kailing’s processes and results may well stem largely from her own intellect, personality, or other idiosyncratic characteristics. We do not argue that these results are universal.

However, especially at the emergence of new phenomena, small case studies can help us develop hypotheses to explore through replications and larger studies. Kailing’s example suggests new possibilities for human-AI collaboration that can be of great benefit for second language writing and writers, despite generative AI’s biases, inaccuracies, and limitations.

Further work is warranted to understand how applicable this case is to the broader field of second language writing with technology, and what kinds of pedagogy are needed to help other students learn to critically exploit this powerful new tool.


Corresponding author: Sharin R. Jacob, Digital Promise Global, Redwood City, USA, E-mail:

Funding source: National Science Foundation

Award Identifier / Grant number: 23152984

About the authors

Sharin R. Jacob

Sharin R. Jacob is a researcher in Computational Thinking Pathways at Digital Promise Global working in Learning Experience Design. Her research interests examine the linguistic and sociocultural factors that help multilingual students succeed in computing. Sharin’s work has been published broadly in journals including Computer Science Education, ACM Transactions on Computing Education, Teachers College Record, and Journal of Computer Science Integration.

Tamara Tate

Tamara Tate is Associate Director of the Digital Learning Lab at the University of California, Irvine. She leads the Lab’s work on generative AI and writing. As the PI of a NSF-funded grant, she is studying the use of generative AI in undergraduate writing courses. She also studies secondary student writing as a member of the IES-funded national WRITE Center. She received her B.A. in English and her Ph.D. in Education at UC Irvine and her J.D. at U.C. Berkeley.

Mark Warschauer

Mark Warschauer is Distinguished Professor of Education and director of the Digital Learning Lab at UC Irvine with affiliated appointments in Informatics, Language Science, and Psychological Science. His research is currently focusing on generative AI in Education, the use of conversational agents to support children’s learning and on the teaching and learning of computer science for linguistically diverse students. He is a member of the National Academy of Education.

  1. Research funding: We would like to thank the National Science Foundation (Award No. 23152984) for providing the funding that made this project possible. Findings expressed in this work are those of the authors and do not necessarily reflect the views of the National Science Foundation.

References

Bailey, Alison L. & Margaret. Heritage. 2014. The role of language learning progressions in improved instruction and assessment of English language learners. TESOL Quarterly 48(3). 480–506. https://doi.org/10.1002/tesq.176.Search in Google Scholar

Barks, David & Paula Watts. 2001. Textual borrowing strategies for graduate-level ESL writers. In Deborah Belcher & Alan Hirvela (eds.), Linking literacies: Perspectives on L2 reading-writing connections, 246–267. Ann Arbor, MI: University of Michigan Press.Search in Google Scholar

Bateson, Gregory. 1972. Steps to an ecology of mind: A revolutionary approach to man’s understanding of himself. New York: Ballantine.Search in Google Scholar

Bender, Emily M., Timnit Gebru, Angelina McMillan-Major & Sherri Shmitchell. 2021. On the dangers of stochastic parrots: Can language models be too big? In Proceedings of the 2021 ACM conference on fairness, accountability, and transparency, 610–623.10.1145/3442188.3445922Search in Google Scholar

Berninger, Virginia W., Deborah Whitaker, Yong Feng, H. Lee Swanson & Robert D. Abbott. 1996. Assessment of planning, translating, and revising in junior high writers. Journal of School Psychology 34(1). 23–52. https://doi.org/10.1016/0022-4405(95)00024-0.Search in Google Scholar

Bin-Hady, Wajdi R. A., Abdu Al-Kadi, Abdelaziz Hazaea & Jameela K. M. Ali. 2023. Exploring the dimensions of ChatGPT in English language learning: A global perspective. Library Hi Tech. https://doi.org/10.1108/lht-05-2023-0200.Search in Google Scholar

Bolukbasi, Tolga, Kai-Wei Chang, James Y. Zou, Venkatesh Saligrama & Adam T. Kalai. 2016. Man is to computer programmer as woman is to homemaker? Debiasing word embeddings. 30th Conference on Neural Information Processing Systems (NIPS 2016), Barcelona, Spain., 29. Neural Information Processing Systems Foundation. Inc.Search in Google Scholar

Booth Olson, Carol, Undarmaa Maamuujav, Jenna Steiss & Hyojin Chung. 2023. Examining the impact of a cognitive strategies approach on the argument writing of mainstreamed English learners in secondary school. Written Communication 40(2). 373–416. https://doi.org/10.1177/07410883221148724.Search in Google Scholar

Brady, Corey & Richard Lehrer. 2021. Sweeping across physical and virtual environments. Digital Experiences in Mathematics Education 7. 66–98. https://doi.org/10.1007/s40751-020-00076-2.Search in Google Scholar

Bunch, George C. 2013. Pedagogical language knowledge: Preparing mainstream teachers for English learners in the new standards era. Review of Research in Education 37(1). 298–341. https://doi.org/10.3102/0091732x12461772.Search in Google Scholar

Curtis, Nancy. 2023. To ChatGPT or not to ChatGPT? The impact of artificial intelligence on academic publishing. The Pediatric Infectious Disease Journal 42(4). 275. https://doi.org/10.1097/inf.0000000000003852.Search in Google Scholar

Dell’Acqua, Fabrizio, Edward McFowland, Ethan R. Mollick, Hila Lifshitz-Assaf, Katherine Kellogg, Seshadri Rajendran & Karim R. Lakhani. 2023. Navigating the jagged technological Frontier: Field experimental evidence of the effects of AI on knowledge worker productivity and quality. Harvard Business School Technology & Operations Mgt. Unit Working Paper 24-013.10.2139/ssrn.4573321Search in Google Scholar

Doshi, Anil, R. & Oliver, Hauser. 2023. Generative artificial intelligence enhances creativity. Available at SSRN. https://doi.org/10.2139/ssrn.4535536.Search in Google Scholar

Engeström, Yrjö. 1987. Learning by expanding: An activity theoretical approach to developmental research. Helsinki: Orienta-Konsultit Oy.Search in Google Scholar

Engeström, Yrjö. 1999. Activity theory and individual and social transformation. In Yrjö Engeström, Reijo Miettinen & Raija-Leena Punamaki (eds.), Perspectives on activity theory, 19–38. Cambridge: Cambridge University Press.10.1017/CBO9780511812774.003Search in Google Scholar

Fitria, Tri N. 2023. Artificial intelligence (AI) technology in OpenAI ChatGPT application: A review of ChatGPT in writing English essay. ELT Forum: Journal of English Language Teaching 12(1). 44–58. https://doi.org/10.15294/elt.v12i1.64069.Search in Google Scholar

Flower, Linda & John R. Hayes. 1981. A cognitive process theory of writing. College Composition and Communication 32(4). 365–387. https://doi.org/10.2307/356600.Search in Google Scholar

Fujioka, Mariko. 2014. L2 student–U.S. professor interactions through disciplinary writing assignments: An activity theory perspective. Journal of Second Language Writing 25. 40–58. https://doi.org/10.1016/j.jslw.2014.05.004.Search in Google Scholar

Graham, S. & D. Perin. 2007. A meta-analysis of writing instruction for adolescent students. Journal of Educational Psychology 99(3). 445–476. https://doi.org/10.1037/0022-0663.99.3.445.Search in Google Scholar

Grapin, Scott E. 2020. Multimodal assessment of English learners in science: Expanding what “counts” as evidence of content learning. (Publication No. 27834657) [Doctoral dissertation, New York University] ProQuest Dissertations and Theses Global.Search in Google Scholar

Grapin, Scott E., Lorena Llosa, Anika Haas & Okhee Lee. 2022. Affordances of computational models for English learners in science instruction: Conceptual foundation and initial inquiry. Journal of Science Education and Technology 31. 52–67. https://doi.org/10.1007/s10956-021-09930-3.Search in Google Scholar

Hu, Guangwei & Gui Wang. 2014. Disciplinary and ethnolinguistic influences on citation in research articles. Journal of English for Academic Purposes 14. 14–28. https://doi.org/10.1016/j.jeap.2013.11.001.Search in Google Scholar

Hutchinson, Ben, Vinodkumar Prabhakaran, Emily Denton, Kellie Webster, Yuan Zhong & Sarah Denuyl. 2020. Social biases in NLP models as barriers for persons with disabilities. arXiv preprint arXiv: 2005.00813.10.18653/v1/2020.acl-main.487Search in Google Scholar

Intersegmental Committee of the Academic Senates (ICAS) of the California Community Colleges, the California State University, and the University of California. 2002. Academic literacy: A statement of competencies expected of students entering California public colleges and universities.Search in Google Scholar

Jacob, Sharin R. & Mark Warschauer. 2018. Computational thinking and literacy.” 1, no. 1 (2018). Journal of Computer Science Integration 1(1).10.26716/jcsi.2018.01.1.1Search in Google Scholar

Kenthapadi, Krishnaram, Himabindu Lakkaraju & Nazneen Rajani. 2023. Generative AI meets responsible AI: Practical challenges and opportunities. In Proceedings of the 29th ACM SIGKDD conference on knowledge discovery and data mining, 5805–5806.10.1145/3580305.3599557Search in Google Scholar

Kohnke, Lucas, Benjamin L. Moorhouse & Di. Zou. 2023. ChatGPT for language teaching and learning. RELC Journal 54(2). 537–550. https://doi.org/10.1177/00336882231162868.Search in Google Scholar

Kessler, Mary. 2020. Technology-mediated writing: Exploring incoming graduate students’ L2 writing strategies with Activity Theory. Computers and Composition 55. 102542. https://doi.org/10.1016/j.compcom.2020.102542.Search in Google Scholar

Lee, Meihua. 2023. On the amplification of linguistic bias through unintentional self-reinforcement learning by generative language models—a perspective. arXiv preprint arXiv:2306.07135.Search in Google Scholar

Liang, Wenkun, Meirick Yuksekgonul, Yilun Mao, Eric Wu & James Zou. 2023. GPT detectors are biased against non-native English writers. arXiv preprint arXiv:2304.02819. https://doi.org/10.1016/j.patter.2023.100779.Search in Google Scholar

Lincoln, Yvonna S. & Egon G. Guba. 1985. Naturalistic inquiry. Beverly Hills: Sage.10.1016/0147-1767(85)90062-8Search in Google Scholar

Mabry, Linda. 2008. Case study in social research. In The SAGE handbook of social research methods, 214–227.10.4135/9781446212165.n13Search in Google Scholar

McCutchen, Deborah. 1996. A capacity theory of writing: Working memory in composition. Educational Psychology Review 8. 299–325. https://doi.org/10.1007/bf01464076.Search in Google Scholar

Noy, Shakked & Wen Zhang. 2023. Experimental evidence on the productivity effects of generative artificial intelligence. SSRN 4375283.10.2139/ssrn.4375283Search in Google Scholar

Ong, Walter J. 1982. Orality and literacy: The technologizing of the word. London: Routledge.10.4324/9780203328064Search in Google Scholar

Ong, Justina & Lawrence Zhang. 2013. Effects of the manipulation of cognitive processes on EFL writers’ text quality. TESOL Quarterly 47(2). 375–398. https://doi.org/10.1002/tesq.55.Search in Google Scholar

Pecorari, Diane. 2001. Plagiarism and international students: How the English-speaking university responds. In Deborah Belcher & Alan Hirvela (eds.), Linking literacies: Perspectives on L2 reading-writing connections, 229–245. Ann Arbor, MI: University of Michigan Press.Search in Google Scholar

Postman, Neil. 1993. Technopoly: The surrender of culture to technology. New York: Vintage Books.Search in Google Scholar

Rahman, Mofijur M., Hiroshi J. Terano, Mohammad N. Rahman, Aidin Salamzadeh & Shabib S. Rahaman. 2023. ChatGPT and academic research: A review and recommendations based on practical examples. Journal of Education, Management and Development Studies 3(1). 1–12. https://doi.org/10.52631/jemds.v3i1.175.Search in Google Scholar

Ruiz-Primo, Maria A. 2011. Informal formative assessment: The role of instructional dialogues in assessing students’ learning. Studies in Educational Evaluation 37. 15–24. https://doi.org/10.1016/j.stueduc.2011.04.003.Search in Google Scholar

Saldaña, Johnny. 2021. The coding manual for qualitative researchers, 4th edn. Los Angeles: Sage.Search in Google Scholar

Scardamalia, Marlene & Carl Bereiter. 1987. Knowledge telling and knowledge transforming in written composition. In Sheldon Rosenberg (ed.), Advances in applied psycholinguistics, vol. 2: Reading, writing, and language learning, 142–175. New York: Cambridge University Press.Search in Google Scholar

Steiss, Jenna, Tiffany Tate, Steve Graham, Jesús Cruz, Michael Hebert, Jia Wang & Wen-Chung Tseng. 2024. Comparing the quality of human and ChatGPT feedback on students’ writing. Learning and Instruction 91. 101894.10.1016/j.learninstruc.2024.101894Search in Google Scholar

Su, Yu, Yufen Lin & Chun Lai. 2023. Collaborating with ChatGPT in argumentative writing classrooms. Assessing Writing 57. 100752. https://doi.org/10.1016/j.asw.2023.100752.Search in Google Scholar

The National Commission on Writing for America’s Families, Schools Colleges. 2004. Writing: A Ticket to work. College Entrance Examination Board. http://www.collegeboard.com/prod_downloads/writingcom/writing-ticket-to-work.pdf.Search in Google Scholar

The National Commission on Writing for America’s Schools and Colleges. 2003. The neglected “R”: The need for a writing revolution. College Entrance Examination Board. Available at: http://www.collegeboard.com/prod_downloads/writingcom/neglectedr.pdf.Search in Google Scholar

Vygotsky, Lev S. & Michael Cole. 1978. Mind in society: Development of higher psychological processes. Cambridge, MA: Harvard University Press.Search in Google Scholar

Warschauer, M., W. Tseng, S. Yim, T. Webster, S. Jacob, Q. Du & T. Tate. 2023. The affordances and contradictions of AI-generated text for writers of English as a second or foreign language. Journal of Second Language Writing 62. 100939. https://doi.org/10.1016/j.jslw.2023.101071.Search in Google Scholar

Received: 2024-06-25
Accepted: 2024-09-11
Published Online: 2024-11-01

© 2024 the author(s), published by De Gruyter and FLTRP on behalf of BFSU

This work is licensed under the Creative Commons Attribution 4.0 International License.

Downloaded on 11.9.2025 from https://www.degruyterbrill.com/document/doi/10.1515/jccall-2024-0011/html
Scroll to top button