<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Tamil Co-Writer: Towards inclusive use of generative AI for writing support</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Antonette</forename><surname>Shibani</surname></persName>
							<email>antonette.shibani@uts.edu.au</email>
							<affiliation key="aff0">
								<orgName type="institution">University of Technology Sydney</orgName>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Faerie</forename><surname>Mattins</surname></persName>
							<affiliation key="aff1">
								<orgName type="institution">Vellore Institute of Technology</orgName>
								<address>
									<settlement>Chennai</settlement>
								</address>
							</affiliation>
							<affiliation key="aff2">
								<orgName type="institution">University of Southern California</orgName>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Srivarshan</forename><surname>Selvaraj</surname></persName>
							<affiliation key="aff1">
								<orgName type="institution">Vellore Institute of Technology</orgName>
								<address>
									<settlement>Chennai</settlement>
								</address>
							</affiliation>
							<affiliation key="aff2">
								<orgName type="institution">University of Southern California</orgName>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Ratnavel</forename><surname>Rajalakshmi</surname></persName>
							<email>rajalakshmi.r@vit.ac.in</email>
							<affiliation key="aff1">
								<orgName type="institution">Vellore Institute of Technology</orgName>
								<address>
									<settlement>Chennai</settlement>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Gnana</forename><surname>Bharathy</surname></persName>
							<affiliation key="aff0">
								<orgName type="institution">University of Technology Sydney</orgName>
							</affiliation>
						</author>
						<title level="a" type="main">Tamil Co-Writer: Towards inclusive use of generative AI for writing support</title>
					</analytic>
					<monogr>
						<idno type="ISSN">1613-0073</idno>
					</monogr>
					<idno type="MD5">CDF7D567253CC40BC989D4191C114F5D</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2025-04-23T18:53+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>Large language model</term>
					<term>generative AI</term>
					<term>artificial intelligence</term>
					<term>LLM</term>
					<term>Tamil</term>
					<term>writing</term>
					<term>GPT</term>
					<term>keystroke analysis</term>
					<term>CoAuthorViz</term>
					<term>Tamil Co-Writer</term>
					<term>inclusive AI</term>
					<term>equity</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>The increasing use of generative AI in education highlights its potential for enriching learning experiences. One application for utilising the capabilities of Large Language Models (LLMs) such as the Generative Pretrained Transformer (GPT) is the creation of writing support tools. In particular, tools that can work in partnership with humans to co-write with AI hold great promise and have been tested out for English language writing. However, the adaptation of such tools to languages other than English is limited, presenting a disadvantage for learners from linguistically diverse backgrounds. In the current study, we extend previous works in English to develop a writing aid prototype for the low-resource Indian regional language Tamil for co-writing with AI called Tamil Co-Writer. The tool additionally provides a visual summary of user interaction and co-authorship metrics for each writing session for users to reflect on their usage of AI in their own writing. We posit that such interactive tools using the latest generative AI technologies can help writers improve their writing skills and productivity in their own regional languages supporting inclusive AI for education.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">Introduction</head><p>Large Language Models (LLM) are sophisticated artificial intelligence (AI) systems trained on massive amounts of textual data to generate text. LLMs such as Generative Pre-trained Transformer (GPT) are capable of producing language that is grammatically correct and appears human-written. They are seen to be performing well in a variety of language-related tasks such as translating, summarizing, responding to questions, and creating new content, and are increasingly employed across many sectors. Although relatively new, LLMs are starting to be deployed in intelligent support systems for learners and writers <ref type="bibr" target="#b0">[1,</ref><ref type="bibr" target="#b1">2]</ref>. For writing support, a tool that uses GPT-3 called CoAuthor was used to collect a dataset of collaborative writing between humans and AI <ref type="bibr" target="#b2">[3]</ref>. The technology allowed users to write freely, solicit suggestions from GPT-3, accept or reject those suggestions, and alter previously written texts or accepted suggestions in any sequence they desired. Past research also presented a visual representation (CoAuthorViz) of user interactions with the tool to study AI-dependency behaviours of users <ref type="bibr" target="#b4">[5]</ref>. This demonstrates the usefulness of AI-based tools in enhancing learner capabilities; however, they only support English language users and do not address the needs of diverse groups of learners to promote inclusivity in education.</p><p>The objective of the current study is to introduce a working prototype of an AI co-authoring tool for the regional language 'Tamil' <ref type="bibr" target="#b3">[4]</ref>. Tamil is a Dravidian language primarily spoken in the Indian state of Tamil Nadu and northeast Sri Lanka. While estimated to be spoken by over 80 million native speakers worldwide, Tamil falls under the category of low-resource languages for Natural Language Processing research, characterized by limited tools and datasets. Recent studies in under-resourced languages aim to bridge this gap by specifically targeting the creation of additional resources <ref type="bibr" target="#b5">[6,</ref><ref type="bibr" target="#b6">7]</ref>. Our prototype tool called 'Tamil Co-Writer' can help Tamil language learners engage in interactive writing sessions with AI support in the form of auto-generated suggestions. The tool incorporates speech to text input using Automatic Speech Recognition (ASR) for increased accessibility. The learner also receives a summary statistic and a visual graph at the end of each writing session using the tool to reflect on their dependence on AI suggestions in their writing. Our study demonstrates how tools such as this can cater to linguistically and culturally diverse groups of users to aid their writing for equitable use of AI. The contributions of this paper are as follows:</p><p>• Development of a novel writing aid prototype for Tamil using the open-source GPT-2 model called Tamil Co-Writer, incorporating an ASR model for text-to-speech input and visual statistics of AI usage. • The conceptualisation and evaluation of how LLMs can be effectively embedded in writing tools for improved accessibility and language support for diverse users whose first language is not English.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">Background</head><p>Automated writing analysis, feedback, and digital tools have long been used to support writers in their writing process <ref type="bibr" target="#b7">[8,</ref><ref type="bibr" target="#b8">9]</ref>. In learning analytics, the sub-field of writing analytics has examined how tools and analytics techniques can be used to help learners with their writing processes and products <ref type="bibr" target="#b9">[10,</ref><ref type="bibr" target="#b10">11]</ref>.</p><p>With the advent of advanced technologies such as LLMs, more sophisticated tools are now being developed that invite writers to co-write with AI as an active companion. Most well-documented among AI co-creation tools is CoAuthor <ref type="bibr" target="#b11">[12]</ref>, which is a combination of an interface, dataset, and experiment all in one. The tool was evaluated for use by over 60 people for creative and argumentative writing tasks where they co-wrote with GPT-3 generated text suggestions. The interactions between the writer and the GPT-3 suggestions were also captured using keystroke logging, which has further led to the study of human-AI interaction behaviours <ref type="bibr" target="#b4">[5]</ref>. Another such tool produced 'sparks', which were new sentences generated by the AI to inspire users to write scientific content <ref type="bibr" target="#b0">[1]</ref>. Here, the inspirations helped writers with elaborating sentences with detail, providing interesting angles to engage readers, and showing potential reader perspectives. A web application called Wordcraft provided multimodal machine intelligence to help writers make integrative leaps in creative writing as they wrote stories <ref type="bibr" target="#b12">[13]</ref>. A similar co-creative story authoring tool utilizes large language models and story grammars, allowing authors to easily engineer text generation to meet their expectations <ref type="bibr" target="#b13">[14]</ref>. While most tools support the drafting process of writing, other tools that support the revision process not fully handing over the creative process to AI are also being developed. This includes a human-in-the-loop iterative text revision system called Read, Revise, Repeat (R3) where writers interacted with model-generated revisions for deeper edits <ref type="bibr" target="#b14">[15]</ref>. Commercial tools such as Grammarly, ProWritingAid, Quillbot, and Scribbr that previously focused on grammar and style also appear to have been transformed by LLMs into co-writing tools in English, providing feedback on content relevance, tone, cohesion. They enhance contents upon request, and correct errors <ref type="bibr" target="#b15">[16]</ref>. Being commercial tools, their exact capabilities and limitations have not been well documented in the research space and constantly evolve over time. In the new world, plagiarism checkers are also being incorporated into tools such as Quillbot <ref type="bibr" target="#b18">[19]</ref>, which, while suggesting improvements or providing feedback, also check for potential plagiarism in content.</p><p>A plethora of products facilitate co-authoring, including content generators, programming automation, and AI-based virtual assistants propelled by the increasing availability of generative AI. Co-pilots and assistants where writers can obtain suggestions "as-they-type" are touted as the future of writing tools. However, all tools discussed above have been developed for writing in English. There are no such authoring tools created for low-resource languages such as the regional Indian languages, e.g. Tamil. By not focusing on languages other than English, we miss the rich cultural context and nuances in regional languages that are not well-resourced for NLP tasks. Fortunately, this is beginning to change for Indian languages as fine-tuned LLMs are starting to emerge <ref type="bibr" target="#b16">[17]</ref>. Our research aims to contribute to the growing space for Tamil learning technologies and data to cater to linguistically diverse user groups and research areas. This paper explains the technical components in building such an AI cowriting system using open-source technology that can be generalized for other languages for writing support.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">Development and evaluation of Tamil Co-Writer</head><p>Our novel collaborative writing tool 'Tamil Co-Writer' consists of a simple user interface for writing Tamil text and several underlying technical components that facilitate AI support and speech recognition. Figure <ref type="figure" target="#fig_0">1</ref> shows a flow chart of the individual components, which are explained below. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1.">Tool functionality</head><p>The front end of Tamil Co-Writer contains an input text field that allows users to start writing their text in a writing session. The user has an additional option to provide input in speech format, which is processed by a fine-tuned ASR model to convert speech-to-text. This option can improve the accessibility of writing support tools by providing users with the option to write text as they speak their language. The use of non-standard vocabulary, accent differences, and background noise make ASR models difficult to accurately identify speech <ref type="bibr" target="#b17">[18]</ref>, with increased challenges in low-resource languages such as Tamil having insufficient training data. When using the tool, the user first enters the text in Tamil or speaks through the ASR system in Tamil. Upon asking for a GPT-2 suggestion (by clicking the 'Get suggestion' button), this Tamil text is translated to English and sent out for text generation using GPT-2. Five suggestions automatically generated using GPT-2 are displayed. The writer can accept a suggestion as is, reject a suggestion, or accept the suggestion and then modify it. At the end of the writing session, the writer obtains summary statistics and graphs that show their level of collaborative writing with the aid of GPT-2. The website is built using Django 2 and locally tested for prototyping. The tool thus consists of the three main components below at the back end, described next:</p><p>1. The input module processes the written text or speech from the user. 2. The automated text generation module then suggests new text using the GPT-2 model. 3. The metrics generation module populates the visual graph and key metrics of AI usage.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2.">Input module</head><p>The input module processes the written text entered by the user directly in Tamil or the entered speech for conversion to text. A screenshot of the user interface is in Figure <ref type="figure" target="#fig_1">2</ref>. For ASR, we use a fine-tuned XLSR Wav2Vec2.0 model using the Connectionist Temporal Classification (CTC) algorithm <ref type="bibr" target="#b19">[20,</ref><ref type="bibr" target="#b20">21]</ref>. CTC operates by introducing a unique 'blank' symbol that may be introduced in the target sequence between any two consecutive output symbols. The method then converts each potential output symbol, including the blank symbol, into a sequence of probabilities. Blank symbols were employed during decoding to calculate the most likely output sequence given the input sequence. The ability to accommodate the input and output sequences of varying lengths is one of the CTC's main features, making it ideal for voice recognition jobs where the input and output sequences' lengths might change. Moreover, CTC is capable of handling situations where the alignment between the input and output sequences is not known beforehand, such as when the input voice signal could have pauses or other disturbances.</p><p>For fine-tuning the XLSR Wav2Vec2.0 model for better performance for Tamil, we use the common voice open dataset <ref type="bibr" target="#b21">[22]</ref>. This dataset has over 850 voices and 7.83GB worth of data. For training data that is this big, a powerful machine having high GB and RAM for running non-stop for a minimum of 40 hours is required. In this research, we employ Google Colab Pro with 100 computer units to train the data. The dataset was cleaned by removing special characters. The audio data used to train the XLSR-Wav2Vec2 model was captured at 48 kHz for Babel, Multilingual LibriSpeech (MLS), and Common Voice and then down-sampled to 16 kHz. Common Speech data needed to be down-sampled to 16 kHz for training as it was initially captured at 48 kHz. The Wav2Vec2FeatureExtractor feature extractor was used with a feature size of 1, sampling rate of 16000, and a padding value of zero. Batchwise padding was performed on the training samples to obtain the longest sample. The feature vector size, which is the combined dimension of all the features derived from the speech representations, is 1. Table <ref type="table" target="#tab_0">1</ref> shows the hyperparameters of the model. The input Tamil text is then converted to English for processing using the existing Google translate API <ref type="bibr" target="#b22">[23]</ref>, google trans of version 4.0. A web service enables programmers to include translation capabilities into their programs by querying the API built by Google. This Google Cloud Platformprovided API offers a straightforward REST-based user interface for translating text between languages. The query to the Google Translate API contains the text to be translated and the destination language, and the translated text in the intended language is returned by the API. The API offers access to both neural and statistical machine translation models and covers over 100 languages, including some less widely spoken languages. By selecting glossary words or offering their own translation models, developers may also alter the translation results. After translating the input Tamil text to English, Tamil Co-Writer sends it to the GPT-2 text generator to generate suggestions for the writer.</p><p>The accuracy of ASR or machine translation was gauged using the word error rate (WER) measure. As compared to the total number of words in the reference text, WER reflects the proportion of words that the system erroneously recognizes or translates <ref type="bibr" target="#b23">[24]</ref>. A lower WER score generally denotes a higher degree of system accuracy with respect to speech recognition or machine translation. However, it is important to note that it is not always a true indicator of the system's performance as there might be other contextual factors in the language that are better assessed by a human native speaker. During the manual evaluation of some examples of our Tamil ASR, we observed errors in a few characters and in some cases, the affix and suffix. Many of the predictions are close to ground truth phonetically even though there are errors in the written script, and Tamil is a phonetic language. An additional metric could be introduced in the future to measure the phonetic proximity, that is comparing the phonetic content in addition to than written content. In the fine-tuned ASR model we used, we achieve a WER of 60%, which is good for basic applications in low-resource languages such as Tamil that lack huge training sets. While noting the limitations, we proceed to use this model for translation for the current prototypical version of our tool, as the main objective of this research is not to enhance ASR in Tamil, but rather to showcase how it can be used in an AI-based writing aid.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.3.">Automated text generation module</head><p>The automated text generation module in Tamil Co-Writer generates Tamil text suggestions to the writer using the open-source large language model GPT-2. The large-scale neural language model Generative Pre-trained Transformer 2 (GPT-2) was created by OpenAI as an extension of the original GPT model and performed better across a range of NLP tasks due to its substantially greater size <ref type="bibr" target="#b24">[25]</ref>. The Transformer architecture, a kind of neural network that processes input sequences via self-attention techniques serves as the foundation for GPT-2. The model is pretrained on a sizable corpus of text data using an unsupervised learning technique, allowing it to pick up on the linguistic patterns and structures without direct supervision. Once trained, GPT-2 produces writing that resembles that of a human being by predicting the following word in a series based on the preceding ones using probabilities. It can also be applied for specialized jobs such as text categorization, question-answering, and language translation. The capability of GPT-2 to produce cohesive fluid language that is difficult to distinguish from human written material is one of its standout characteristics. Table <ref type="table" target="#tab_1">2</ref> shows the parameters used for the GPT-2 model to evaluate and generate text suggestions for the writer. While newer models can provide much better performance (and are discussed as part of future work), GPT-2 provides a baseline model for testing that is free to access and deploy. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.4.">Metrics generation module</head><p>The tool additionally logs the keystroke-level action of users to analyse and provide summaries of their interactions with AI suggestions using the metrics generation module. This kind of learning analytics from user logs of AI interaction helps users reflect on their process of writing and reliance on AI, which is deemed to be the future of assessment <ref type="bibr" target="#b24">[25]</ref>. The visualization follows CoAuthorViz, which is a graphical representation introduced by recent work <ref type="bibr" target="#b4">[5]</ref>, to represent co-authorship behaviors during a writing session when users interact with AI suggestions at the sentence level. The CoAuthorViz visual representation and its interaction metrics provide insight into how writers utilize AI writing assistants and can help writers reflect on their dependence on AI suggestions in their writing practices. This was used to analyze human-AI collaborative writing for English, and how the graph can be read is explained in detail in the previous paper with examples <ref type="bibr" target="#b4">[5]</ref>. Here we illustrate CoAuthorViz for our Tamil Co-Writer tool in Figure <ref type="figure" target="#fig_2">3</ref> using an example from a test writer. In this case, majority of the writing was done independently by the writer (sentences 1, 3, 5, 8, 10, 11, 13, 15, 16, and 18 with black squares), and even when text from GPT-2 was obtained (sentences 2, 4, 6, 7, 9, 12, and 18 with GPT-2 written text), the writer still added more text. Additionally, we can identify instances when a GPT-2 call was made, but the writer disregarded it (white triangles in sentences 6 and 10 empty GPT-2 calls). We also see instances where the writer made changes to the phrases recommended by GPT-2, suggesting partial satisfaction with the suggestions and further editing (squares enclosing gray triangles in sentences 7 and 9).</p><p>In addition to the creation of CoAuthorViz for Tamil, we present a summary of the most important events noted in each writing session, which offers concrete metrics that authors may use with visuals to learn more about their writing patterns, as demonstrated in a previous work <ref type="bibr" target="#b4">[5]</ref>. This summary includes three metrics: sentence, API, and ratio. The sentence metrics define co-authorship across all sentences in the writing session, , the API metrics keep track of the number of calls made to obtain GPT-2 suggestions and the Ratio metrics consolidate them to study co-authorship behaviours in relation to the total number of sentences generated in a writing session. Table <ref type="table">3</ref> shows key co-authorship metrics for the example writing session discussed in Figure <ref type="figure" target="#fig_2">3</ref>. The author has written independently for most of the writing session as indicated by their Autonomous Writing Indicator (RB = 0.61). The total usage of GPT-2 in a sentence being low (RC = 0.38) indicates that the user has a more independent writing style; or is not too satisfied with the AI-offered suggestions. In none of the sentences was the user completely dependent on the GPT-2 suggestion (SC = 0). These metrics can help writers understand how much of their writing is produced by the model and how much their own work is at the end of a writing session. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Table 3 Co-authorship metrics for the sample writing session in</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.">Discussion and future work</head><p>Despite the increasing popularity of large language models in NLP based applications, low-resource languages such as the regional Indian language Tamil have a lower adoption rate. In this work, we introduced a novel human-AI collaborative writing tool prototype for Tamil writing called Tamil Co-Writer that used GPT-2 to auto-generate text suggestions to the user when they require help in their writing session. Additionally, we introduced a fine-tuned ASR model for Tamil with 60% WER, which enables the to provide Tamil speech as an input for improved accessibility. We also demonstrated the use of a visual representation called CoAuthorViz and its co-authorship metrics applied through our tool for writers to improve their understanding of AI usage when co-writing with AI. We posit that AIgenerated suggestions can not only help language writers obtain new ideas when writing in Tamil, for instance, in coming up with characters and plots for creative writing outputs such as stories, but they can also provide them with exposure to new vocabulary and different ways of structuring sentences.</p><p>In the current prototypical version of our tool, we employed the GPT-2 model for equitable access, because it is an available open-source for deployment. However, newer models created with larger training data, such as higher versions of GPT, can provide better text suggestions for the user (with the downside of costs). We plan to use open-source versions of large language models such as LlaMA or Falcon LLM in later versions of the tool when deploying to users. Exciting new developments such as the introduction of an early-stage LLM for Tamil (fine-tuned from LlaMA with 16,000 Tamil tokens using the LoRA methodology) are starting to occur <ref type="bibr" target="#b16">[17]</ref>. LoRA introduces trainable low-rank matrices into specific layers of a pre-trained model, reducing the need for pre-training a large number of weight parameters directly and achieving higher training efficiency. We see this as a synergistic development and a potential opening for an NLP ecosystem in Tamil. If open-source LLMs such as these are developed for global regional languages, tools such as Tamil Co-Writer would be further augmented for supporting learners. Together, they could add to a more inclusive and diverse linguistic AI landscape and ecosystem.</p><p>Currently, the Tamil Co-Writer tool also uses Google API for translation purposes with trial access; in future versions, we would like to use better and open-source translation models specifically tailored for Tamil to deploy the tool without paid access to API. This would also improve the user experience for writers with faster processing times, as there would not be any time lost in the back-and-forth translation through APIs. There is also scope to create a better-fine-tuned model for ASR by increasing the run time, input data, and training on heavy resources. Future work will involve additional support for code-switching between languages <ref type="bibr" target="#b27">[28]</ref> for multi-lingual usage and improved user interaction features.</p><p>Learning analytics offers new opportunities to support effective human-AI collaboration by helping educators and students become aware of the processes involved in learning, in addition to the final products. In current work, in addition to using AI to support their writing, users can also reflect on their usage of AI-generated suggestions using the CoAuthorViz visualisation and metrics. The creation of visual representations to study co-authorship behaviours and associated metrics opens up new avenues to investigate writing processes, such as studying user characteristics and collaboration dynamics among writers <ref type="bibr" target="#b4">[5]</ref>. Analytics from generative AI can thus help close the LA cycle by facilitating personalised and adaptive interventions <ref type="bibr" target="#b26">[27]</ref>. Writers may use this data to spot patterns and trends in their writing processes, such as if they frequently employ AI for particular kinds of material. This language-agnostic approach can help improve students' writing practices and feedback-seeing behaviors when engaging with AI by understanding their own and AI's respective roles in the writing process, and aids researchers to study these processes.</p><p>Future work can also inform feedback mechanisms to provide feedback to the user when overreliance on AI is observed in their writing and develop effective models for optimal human-AI collaboration for writing. The research thus provides useful insights from our prototype evaluation that can be extended to other languages to cater to diverse groups of audiences and their writing needs for language development and can pave the way for more inclusive AI tools for education in the future.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Figure 1 :</head><label>1</label><figDesc>Figure 1: Workflow and components of Tamil Co-Writer</figDesc><graphic coords="3,154.25,94.65,286.39,161.10" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head>Figure 2 :</head><label>2</label><figDesc>Figure 2: Tamil Co-Writer user interface for input</figDesc><graphic coords="3,187.10,624.27,235.00,110.00" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_2"><head>Figure 3 :</head><label>3</label><figDesc>Figure 3: Example of a CoAuthorViz writing session</figDesc><graphic coords="6,190.07,56.70,229.05,187.60" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_3"><head>Figure 3</head><label>3</label><figDesc>Parameters Value Total number of sentences (SA) 21 Number of sentences completely authored by the writer (SB) 13 Number of sentences completely authored by GPT-3 (SC) 0 Number of sentences co-authored by GPT-3 and writer (SD) 8 GPT-3 dependence indicator (RA) 0 Autonomous writing indicator (RB) 0.61 Total GPT-3 usage in sentences (RC) 0.38</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head>Table 1 Hyperparameters for the fine-tuned Tamil ASR model</head><label>1</label><figDesc></figDesc><table><row><cell>Parameters</cell><cell>Value</cell></row><row><cell>Epochs</cell><cell>20</cell></row><row><cell>Batch size</cell><cell>16</cell></row><row><cell>Gradient accumulation steps</cell><cell>2</cell></row><row><cell>Evaluation strategy</cell><cell>Steps</cell></row><row><cell>Half-precision floating point format (FP16)</cell><cell>True</cell></row><row><cell>Save strategy</cell><cell>Epoch</cell></row><row><cell>Evaluation steps</cell><cell>100</cell></row><row><cell>Logging steps</cell><cell>10</cell></row><row><cell>Learning rate</cell><cell>0.0001</cell></row><row><cell>Attention dropout</cell><cell>0.1</cell></row><row><cell>Hidden dropout</cell><cell>0.1</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_1"><head>Table 2 Parameter set for text generation in GPT-2 model</head><label>2</label><figDesc></figDesc><table><row><cell>Parameters</cell><cell>Value</cell></row><row><cell>Maximum length</cell><cell>30</cell></row><row><cell>Number of return sequence</cell><cell>5</cell></row><row><cell>Temperature</cell><cell>0.3</cell></row><row><cell>Top-p</cell><cell>1</cell></row><row><cell>No repeat ngram size</cell><cell>2</cell></row></table></figure>
		</body>
		<back>
			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">Sparks: Inspiration for Science Writing using Language Models</title>
		<author>
			<persName><forename type="first">K</forename><forename type="middle">I</forename><surname>Gero</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Chilton</surname></persName>
		</author>
		<idno type="DOI">10.1145/3532106.3533533</idno>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2022 ACM Designing Interactive Systems Conference, DIS &apos;22</title>
				<meeting>the 2022 ACM Designing Interactive Systems Conference, DIS &apos;22<address><addrLine>New York, NY, USA</addrLine></address></meeting>
		<imprint>
			<publisher>Association for Computing Machinery</publisher>
			<date type="published" when="2022">2022</date>
			<biblScope unit="page" from="1002" to="1019" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">Metaphorian: Leveraging Large Language Models to Support Extended Metaphor Creation for Science Writing</title>
		<author>
			<persName><forename type="first">J</forename><surname>Kim</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Suh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Chilton</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Xia</surname></persName>
		</author>
		<idno type="DOI">10.1145/3563657.3595996</idno>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2023 ACM Designing Interactive Systems Conference, DIS &apos;23</title>
				<meeting>the 2023 ACM Designing Interactive Systems Conference, DIS &apos;23<address><addrLine>New York, NY, USA</addrLine></address></meeting>
		<imprint>
			<publisher>Association for Computing Machinery</publisher>
			<date type="published" when="2023">2023</date>
			<biblScope unit="page" from="115" to="135" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Coauthor: Designing a human-ai collaborative writing dataset for exploring language model capabilities</title>
		<author>
			<persName><forename type="first">M</forename><surname>Lee</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Liang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Q</forename><surname>Yang</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2022 CHI conference on human factors in computing systems</title>
				<meeting>the 2022 CHI conference on human factors in computing systems</meeting>
		<imprint>
			<date type="published" when="2022">2022</date>
			<biblScope unit="page" from="1" to="19" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<monogr>
		<title level="m" type="main">Tamil language</title>
		<ptr target="https://en.wikipedia.org/wiki/Tamil_language" />
		<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
		<respStmt>
			<orgName>Wikipedia</orgName>
		</respStmt>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">Visual representation of coauthorship with GPT-3: Studying human-machine interaction for effective writing</title>
		<author>
			<persName><forename type="first">A</forename><surname>Shibani</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Rajalakshmi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Mattins</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Selvaraj</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Knight</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">16th International Conference on Educational Data Mining</title>
				<editor>
			<persName><forename type="first">M</forename><surname>Feng</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">T</forename><surname>K¨aser</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">P</forename><surname>Talukdar</surname></persName>
		</editor>
		<meeting><address><addrLine>Bengaluru, India</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">Corpus Creation for Sentiment Analysis in Code-Mixed Tamil-English Text</title>
		<author>
			<persName><forename type="first">B</forename><forename type="middle">R</forename><surname>Chakravarthi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Muralidaran</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">P</forename><surname>Priyadharshini</surname></persName>
		</author>
		<author>
			<persName><surname>Mccrae</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 1st Joint Workshop on Spoken Language Technologies for Under-resourced languages (SLTU) and Collaboration and Computing for Under-Resourced Languages (CCURL)</title>
				<meeting>the 1st Joint Workshop on Spoken Language Technologies for Under-resourced languages (SLTU) and Collaboration and Computing for Under-Resourced Languages (CCURL)<address><addrLine>Marseille, France</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<monogr>
		<author>
			<persName><forename type="first">S</forename><surname>Khanuja</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Bansal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Mehtani</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Khosla</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Dey</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Gopalan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">K</forename><surname>Margam</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Aggarwal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><forename type="middle">T</forename><surname>Nagipogu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Dave</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Gupta</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2103.10730</idno>
		<title level="m">Muril: Multilingual representations for indian languages</title>
				<imprint>
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b7">
	<monogr>
		<title level="m" type="main">AcaWriter: A learning analytics tool for formative feedback on academic writing</title>
		<author>
			<persName><forename type="first">S</forename><surname>Knight</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Shibani</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Abel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Gibson</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Ryan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Sutton</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Wight</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Lucas</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Sandor</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Kitto</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Liu</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<monogr>
		<title level="m" type="main">Digital writing technologies in higher education: theory, research, and practice</title>
		<author>
			<persName><forename type="first">O</forename><surname>Kruse</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Rapp</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><forename type="middle">M</forename><surname>Anson</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Benetos</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Cotos</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Devitt</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Shibani</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2023">2023</date>
			<publisher>Springer Nature</publisher>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<monogr>
		<title level="m" type="main">Natural Language Processing-Writing Analytics</title>
		<author>
			<persName><forename type="first">A</forename><surname>Gibson</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Shibani</surname></persName>
		</author>
		<ptr target="https://solaresearch.org/wp-content/uploads/hla22/HLA22_Chapter_10_Gibson.pdf" />
		<editor>Charles Lang, George Siemens, Alyssa Friend Wise, Dragan Gaševic, and Agathe Merceron</editor>
		<imprint>
			<date type="published" when="2022">2022</date>
			<publisher>SoLAR</publisher>
			<biblScope unit="page" from="96" to="104" />
			<pubPlace>Vancouver, Canada</pubPlace>
		</imprint>
	</monogr>
	<note>2nd ed</note>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">Analytic Techniques for Automated Analysis of Writing, Digital Writing Technologies in Higher Education: Theory</title>
		<author>
			<persName><forename type="first">A</forename><surname>Shibani</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Research, and Practice</title>
		<imprint>
			<biblScope unit="page" from="317" to="331" />
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<analytic>
		<title level="a" type="main">Coauthor: Designing a human-ai collaborative writing dataset for exploring language model capabilities</title>
		<author>
			<persName><forename type="first">M</forename><surname>Lee</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Liang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Q</forename><surname>Yang</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2022 CHI conference on human factors in computing systems</title>
				<meeting>the 2022 CHI conference on human factors in computing systems</meeting>
		<imprint>
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<analytic>
		<title level="a" type="main">Where to hide a stolen elephant: Leaps in creative writing with multimodal machine intelligence</title>
		<author>
			<persName><forename type="first">N</forename><surname>Singh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Bernal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Savchenko</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><forename type="middle">L</forename><surname>Glassman</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">ACM Transactions on Computer-Human Interaction</title>
		<imprint>
			<biblScope unit="volume">30</biblScope>
			<biblScope unit="issue">5</biblScope>
			<biblScope unit="page" from="1" to="57" />
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b13">
	<analytic>
		<title level="a" type="main">A hybrid approach to co-creative story authoring using grammars and language models</title>
		<author>
			<persName><forename type="first">A</forename><surname>Riddle</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the AAAI Conference on Artificial Intelligence and Interactive Digital Entertainment</title>
				<meeting>the AAAI Conference on Artificial Intelligence and Interactive Digital Entertainment</meeting>
		<imprint>
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b14">
	<monogr>
		<author>
			<persName><forename type="first">W</forename><surname>Du</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><forename type="middle">M</forename><surname>Kim</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Raheja</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Kumar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Kang</surname></persName>
		</author>
		<author>
			<persName><surname>Read</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2204.03685</idno>
		<title level="m">revise, repeat: A system demonstration for human-in-the-loop iterative text revision</title>
				<imprint>
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b15">
	<monogr>
		<title level="m" type="main">How we use ai to enhance your writing, Grammarly Spotlight</title>
		<author>
			<persName><surname>Grammarly</surname></persName>
		</author>
		<ptr target="https://www.grammarly.com/blog/how-grammarly-uses-ai/" />
		<imprint>
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b16">
	<monogr>
		<author>
			<persName><forename type="first">A</forename><surname>Balachandran</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2311.05845</idno>
		<title level="m">Tamil-Llama: A New Tamil Language Model Based on Llama 2</title>
				<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b17">
	<analytic>
		<title level="a" type="main">Automatic speech recognition errors detection and correction: A review</title>
		<author>
			<persName><forename type="first">R</forename><surname>Errattahi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">E</forename><surname>Hannani</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Ouahmane</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Procedia Computer Science</title>
		<imprint>
			<biblScope unit="volume">128</biblScope>
			<biblScope unit="page" from="32" to="37" />
			<date type="published" when="2018">2018</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b18">
	<monogr>
		<title/>
		<author>
			<persName><surname>Quillbot</surname></persName>
		</author>
		<ptr target="https://quillbot.com/" />
		<imprint/>
	</monogr>
</biblStruct>

<biblStruct xml:id="b19">
	<analytic>
		<title level="a" type="main">wav2vec 2.0: A framework for self-supervised learning of speech representations</title>
		<author>
			<persName><forename type="first">A</forename><surname>Baevski</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Zhou</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Mohamed</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Auli</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Advances in neural information processing systems</title>
		<imprint>
			<biblScope unit="volume">33</biblScope>
			<biblScope unit="page" from="12449" to="12460" />
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b20">
	<analytic>
		<title level="a" type="main">Connectionist temporal classification: labelling unsegmented sequence data with recurrent neural networks</title>
		<author>
			<persName><forename type="first">A</forename><surname>Graves</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Fernández</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Gomez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Schmidhuber</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 23rd international conference on Machine learning</title>
				<meeting>the 23rd international conference on Machine learning</meeting>
		<imprint>
			<date type="published" when="2006">2006</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b21">
	<monogr>
		<ptr target="https://commonvoice.mozilla.org/en" />
		<title level="m">Common Voice Mozilla</title>
				<imprint/>
	</monogr>
</biblStruct>

<biblStruct xml:id="b22">
	<monogr>
		<ptr target="https://cloud.google.com/translate?hl=en" />
		<title level="m">Google Cloud Translation AI</title>
				<imprint/>
	</monogr>
</biblStruct>

<biblStruct xml:id="b23">
	<analytic>
		<title level="a" type="main">Word error rate estimation for speech recognition: e-WER</title>
		<author>
			<persName><forename type="first">A</forename><surname>Ali</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Renals</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics</title>
				<meeting>the 56th Annual Meeting of the Association for Computational Linguistics</meeting>
		<imprint>
			<date type="published" when="2018">2018</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b24">
	<analytic>
		<title level="a" type="main">Language models are unsupervised multitask learners</title>
		<author>
			<persName><forename type="first">A</forename><surname>Radford</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Wu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Child</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Luan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Amodei</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Sutskever</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">OpenAI blog</title>
		<imprint>
			<biblScope unit="volume">1</biblScope>
			<biblScope unit="issue">8</biblScope>
			<biblScope unit="page">9</biblScope>
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b25">
	<monogr>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">M</forename><surname>Lodge</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Howard</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Bearman</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Dawson</surname></persName>
		</author>
		<author>
			<persName><surname>Associates</surname></persName>
		</author>
		<ptr target="https://www.teqsa.gov.au/guides-resources/resources/corporate-publications/assessment-reform-age-artificial-intelligence" />
		<title level="m">Assessment reform for the age of artificial intelligence</title>
				<imprint>
			<publisher>A. G. Tertiary Education Quality and Standards Agency</publisher>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b26">
	<monogr>
		<author>
			<persName><forename type="first">L</forename><surname>Yan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Martinez-Maldonado</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Gašević</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2312.00087</idno>
		<title level="m">Generative Artificial Intelligence in Learning Analytics: Contextualising Opportunities and Challenges through the Learning Analytics Cycle</title>
				<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b27">
	<monogr>
		<author>
			<persName><forename type="first">N</forename><surname>Jose</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><forename type="middle">R</forename><surname>Chakravarthi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Suryawanshi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Sherly</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">P</forename><surname>Mccrae</surname></persName>
		</author>
		<title level="m">2020 6th international conference on advanced computing and communication systems (ICACCS)</title>
				<imprint>
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
	<note>A survey of current datasets for code-switching research</note>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
