Albuquerque, New Mexico, USA
Computational Approaches to Linguistic Code-Switching, CALCS 2025
Albuquerque, New Mexico, USA

First Call for Papers


This edition will be the seventh edition of the workshop collocated with NAACL 2025.

Bilingual and multilingual speakers often engage in code-switching (CS), mixing languages within a conversation, influenced by cultural nuances. CS can occur at inter-sentential, intra-sentential, and morphological levels, posing challenges for language understanding and generation. Models trained for a single language often struggle with mixed-language input. Despite advances in multilingual pre-trained language models (LMs), they may still perform poorly on CS data. Research on LMs' ability to process CS data, considering cultural nuances, reasoning, coverage, and performance biases, remains underexplored.

As CS becomes more common in informal communication like newsgroups, tweets, and social media, research on LMs processing mixed-language data is urgently needed. This workshop aims to unite researchers working on spoken and written CS technologies, promoting collaboration to improve AI's handling of CS across diverse linguistic contexts.

Topics of Interest


The workshop will invite contributions from researchers working in NLP and speech approaches for the analysis and processing of mixed-language data. Topics of relevance to the workshop include the following:

  1. Development of data and model resources to support research on CS data
  2. New data augmentation techniques for improving robustness on CS data
  3. New approaches for NLP downstream tasks: question answering, conversational agents, named entity recognition, sentiment analysis, machine translation, language generation, and ASR in CS data
  4. NLP techniques for the syntactic analysis of CS data
  5. Domain, dialect, genre adaptation techniques applied to CS data processing
  6. Language modeling approaches to CS data processing
  7. Sociolinguistic and/or sociopragmatic aspects of CS
  8. Techniques and metrics for automatically evaluating synthetically generated CS text
  9. Utilization of LLMs and assessment of their performance on NLP tasks for CS data
  10. Survey and position papers discussing the challenges of CS data to NLP techniques
  11. Ethical issues and consideration on CS applications.

Submissions


The workshop accepts three categories of papers: regular workshop papers, non-archival and cross-submissions. Only regular workshop papers will be included in the proceedings as archival publications. The regular workshop papers are eligible for the best paper award. All three categories of papers may be long (maximum 8 pages plus references) or short (maximum 4 pages plus references), with unlimited additional pages for references, following the ARR formatting requirements. The reported research should be substantially original. Accepted papers will be presented as posters and orals. Reviewing will be double-blind, and thus, no author information should be included in the papers; self-reference that identifies the authors should be avoided or anonymized. Accepted regular workshop papers will appear in the workshop proceedings. We welcome papers with a maximum of 2 pages for non-archival submission. Please send us an email if you are submitting the non-archival submission. The limitation section is optional and will not be counted in the page limit. The submission portal is open on OpenReview.

Shared Task on Automatic Evaluation for Code-Switched Text Generation


This shared task focuses on developing automatic evaluation metrics for code-switched (CS) text generation. Participants are tasked with creating systems that can accurately assess the quality of synthetically generated CS text, considering both fluency and accuracy. This is crucial because:

  • Scarcity of CS Data: CS text data is limited, making automatic generation vital for data augmentation and improving model performance.
  • Growing Demand: The need for CS text is increasing, particularly in dialogue systems and chatbots, to enable more natural and inclusive interactions.
  • Lack of Robust Evaluation: Current methods for evaluating CS text are insufficient, hindering progress in this field.

This shared task aims to address this gap and drive further research in automatic evaluation metrics for CS text generation.

Languages Supported:

  • Public Leaderboard: English-Hindi, English-Tamil, English-Malayalam
  • Private Leaderboard: English-Indonesian, Indonesian-Javanese, Singlish (English-Chinese)

Metric:

Accuracy: Systems will be evaluated based on their accuracy in predicting human preferences for CS text. This will be measured by comparing the system's ranking of generated sentences (Sent 1 vs. Sent 2) with human annotations in the CSPref dataset.

Dataset:

The CSPref dataset will be used for this task. It contains:

  • Original L1: English sentences
  • Original L2: Hindi, Tamil, or Malayalam sentences
  • Sent 1, Sent 2: Two different CS generations based on the original sentences.
  • Chosen: Human annotation indicating the preferred sentence (Sent 1, Sent 2, or Tie).
  • Lang: Language pair

Data is available here: https://huggingface.co/datasets/garrykuwanto/cspref

Evaluation:

  • Systems will be ranked on a public leaderboard based on their accuracy in predicting human preferences on the English-Hindi, English-Tamil, and English-Malayalam language pairs.
  • A private leaderboard will evaluate system performance on unseen language pairs (English-Indonesian, Indonesian-Javanese, Singlish) to assess generalization ability.

Submission:

  • Participants will submit their system's predictions for each instance in the test set, indicating their preferred sentence (Sent 1, Sent 2, or Tie).

Goal:

The goal of this shared task is to encourage the development of robust and reliable automatic evaluation metrics for CS text generation, ultimately leading to more fluent and accurate CS language models.

Competition Page:

For more information about the competition, please visit the competition page: https://eval.ai/web/challenges/challenge-page/2437/overview

Important Dates


Paper Submission

  • Workshop submission deadline (regular and non-archival submissions): 21 February 2025
  • Notification of acceptance: 8 March 2025
  • Camera ready papers due: 17 March 2025
  • Workshop date: 3/4 May 2025

Shared Task Submission

  • Train release and platform: 23 January 2025
  • Test release: 14 February 2025
  • Results submission: 21 February 2025
  • Paper submission: 28 February 2025
  • Notification deadline: 8 March 2025

All deadlines are 11.59 pm UTC -12h (“anywhere on Earth”).

Program Committee


Ana Valeria González    Novo Nordisk
Astik Biswas    Oracle
Alexander Gelbukh    ICM
Barbara Bullock    UT Austin
Constantine Lignos    Brandeis University
Costin-Gabriel Chiru    Ghent University
Els Lefever    Ghent University
Emre Yılmaz    University of Houston-Downtown
Emily Ahn    University of Washington
Grandee Lee    SUSS
Helena Gómez Adorno    UAM
Jacqueline Toribio    UT Austin
Julia Hirschberg    Columbia University
Khyathi Chandu    Allen AI
Parth Patwa    AWS
Pastor López Monroy    Centro de Investigaciones Matemáticas Avanzadas
Manuel Mager    Amazon
Samson Tan    Amazon AWS
Segun Taofeek Aroyehun    University of Konstanz
Steven Abney    University of Michigan
Yang Liu    Amazon
Yerbolat Kassanov    ByteDance

Organizers


Genta Indra Winata,
Senior Applied Scientist, Capital One AI Foundations, USA
Sudipta Kar,
Senior Applied Scientist, Amazon Alexa AI, USA
Marina Zhukova,
Ph.D. Candidate, Department of Linguistics, University of California, Santa Barbara, USA
Injy Hamed,
Postdoctoral Associate, MBZUAI, UAE
Garry Kuwanto,
Ph.D. Student, Boston University, USA
Mahardika Krisna Ihsani,
M.Sc. Candidate, MBZUAI, UAE
Barid Xi Ai,
Postdoctoral Researcher, NUS, Singapore
Derry Tanti Wijaya,
Associate Professor, Monash University Indonesia, Indonesia
Adjunct Faculty, Boston University, USA
Thamar Solorio,
Professor, MBZUAI, UAE
Professor, University of Houston, USA