Programme
8:45-9:00
Doors open
9:00-10:00
Keynote 1: The INCREC project: Creativity and Technology in Translation
Ana Guerberof Arenas
In this talk, I will give an overview of how creativity is conceptualised in the social sciences, mainly psychology and sociology, including different frameworks that facilitate analysing creativity. I will also touch upon how the technological field has presented and has studied creativity, and how translation in combination with technology can be explored.
With this aim in mind, I will also present results from the CREAMT project (2020-2022) that explored creativity in literary texts in different translation modalities: translation by professional literary translators, machine translation using a customized neural engine, and post-edition. Further it looked at the impact on readers by looking at narrative engagement, enjoyment and translation reception.
Finally, I will describe the new ERC CoG INCREC research project (2023-2028) that looks to uncover the creative process in professional literary and audiovisual translators in order to create specific frameworks, and how and when technology, e.g. machine translation, can be used to enhance rather than constrain creativity. But also if the intended audiences, readers and viewers, appreciate, not only cognitively but also emotionally, creative shifts in translated content and why this might be.
10:00-10:30
Machine Translation Meets Large Language Models: Evaluating ChatGPT's Ability to Automatically Post-Edit Literary Texts
Lieve Macken
Large language models such as GPT-4 have been trained on vast corpora, giving them excellent language understanding. This study explores the use of ChatGPT for post-editing machine translations of literary texts. Three short stories, machine translated from English into Dutch, were post-edited by 7-8 professional translators and ChatGPT. Automatic metrics were used to evaluate the number and type of edits made, and semantic and syntactic similarity between the machine translation and the corresponding post-edited versions. A manual analysis classified errors in the machine translation and changes made by the post-editors. The results show that ChatGPT made more changes than the average post-editor. ChatGPT improved lexical richness over machine translation for all texts. The analysis of editing types showed that ChatGPT replaced more words with synonyms, corrected fewer machine errors and introduced more problems than professionals.
10:30-11:00
Coffee break
11:00-11:30
Prompting Large Language Models for Idiomatic Translation
Antonio Castaldo, Johanna Monti
Large Language Models (LLMs) have demonstrated impressive performance in translating content across different languages and genres. Yet, their potential in the creative aspects of machine translation has not been fully explored. In this paper, we seek to identify the strengths and weaknesses inherent in different LLMs when applied to one of the most prominent features of creative works: the translation of idiomatic expressions. We present an overview of their performance in the EN IT language pair, a context characterized by an evident lack of bilingual data tailored for idiomatic translation. Lastly, we investigate the impact of prompt design on the quality of machine translation, drawing on recent findings which indicate a substantial variation in the performance of LLMs depending on the prompts utilized.
11:30-12:00
‘Can Make Mistakes’. Prompting ChatGPT to Enhance Literary MT output
Gys-Walt Van Egdom, Christophe Declercq, Onno Kosters
Operating at the intersection of generative AI (artificial intelligence), machine transla-tion (MT), and literary translation, this paper examines to what extent prompt-driven post-editing (PE) can enhance the quality of machine-translated literary texts. We assess how different types of instruction influence PE performance, particularly focusing on literary nuances and author-specific styles. Situated within posthumanist translation theory, which often challenges traditional notions of human intervention in translation processes, the study explores the practical implementation of generative AI in multilingual workflows. While the findings suggest that prompted PE can improve translation output to some extent, its effectiveness varies, especially in literary contexts. This highlights the need for a critical review of prompt engineering approaches and emphasizes the importance of further research to navigate the complexities of integrating AI into creative translation workflows effectively.
12:00-12:30
Impact of Translation Workflows with and without MT on Textual Characteristics in Literary Translation
Joke Daems, Paola Ruffo, Lieve Macken
The use of machine translation is increasingly being explored for the translation of literary texts, but there is still a lot of uncertainty about the optimal translation workflow in these scenarios. While overall quality is quite good, certain textual characteristics can be different in a human translated text and a text produced by means of machine translation post-editing, which has been shown to potentially have an impact on reader perceptions and experience as well. In this study, we look at textual characteristics from short story translations from B.J. Novak's One more thing into Dutch. Twenty-three professional literary translators translated three short stories, in three different conditions: using Word, using the classic CAT tool Trados, and using a machine translation post-editing platform specifically designed for literary translation. We look at overall text characteristics (sentence length, type-token ratio, stylistic differences) to establish whether translation workflow has an impact on these features, and whether the three workflows lead to very different final translations or not.
12:30-13:30
Lunch
13:30-14:30
Keynote 2: CAT, TM, NMT, and AI: A Literary Translator’s Dream Team
Andrew Rothwell
Since the public release of ChatGPT in November 2022, there has been an explosion of interest in the remarkable text-production capabilities (including paraphrasing, summarizing and translation) of such generative AI tools. Whether they can be of assistance to the literary translator, and if so, how they can best be made to interoperate with existing CAT and MT environments, remains, however, a largely moot question.
This paper will describe my developing use of different technologies, over almost a decade, to produce English translations of classic novels in French by Emile Zola and Marcel Proust. Acknowledging the documented reticence of literary translators to adopt computerised tools, I will nevertheless argue for the practical benefits of using:
an electronic ST
aligned bilingual editor (aka CAT tool)
translation memory
termbase
online dictionaries and thesauri
NMT (free-standing and CAT-integrated)
generative AI.
The core of the paper will be a presentation of how these technologies are now being combined in a single interface, taking as an example the recently AI-enhanced CAT tool Wordscope. Wordscope offers an integration of translation memory, machine translation from several providers, and ChatGPT as a research and paraphrasing tool, in a de-cluttered online environment.
The paper will describe different options for using the tool for literary translation, and discuss some theoretical implications of doing so in a Cognitive Translation Studies framework. In conclusion, I will argue that Lommel’s (2018) notion that the translator’s creative capacity is ‘augmented’ rather than inhibited by computer technologies applies no less to literary than to ‘commercial’ translation, albeit in significantly different ways.
14:30-15:00
Using a Multilingual Literary Parallel Corpus to Train NMT Systems
Bojana Mikelenić, Antoni Oliver
This article presents an application of a multilingual and multidirectional parallel corpus composed of literary texts in five Romance languages (Spanish, French, Italian, Portuguese, Romanian) and a Slavic language (Croatian), with a total of 142,000 segments and 15.7 million words. After combining it with very large freely available parallel corpora, this resource is used to train NMT systems tailored to literature. A total of five NMT systems have been trained: Spanish-French, Spanish-Italian, Spanish-Portuguese, Spanish-Romanian and Spanish-Croatian. The trained systems were evaluated using automatic metrics (BLEU, chrF2 and TER) and a comparison with a rule-based MT system (Apertium) and a neural system (Google Translate) is presented. As a main conclusion, we can highlight that the use of this literary corpus has been very productive, as the majority of the trained systems achieve comparable, and in some cases even better, values of the automatic quality metrics than a widely used commercial NMT system.
15:00-15:30
Coffee break
15:30-16:00
LitPC: A Set of Tools for Building Parallel Corpora from Literary Works
Antoni Oliver, Sergi Alvarez-Vidal
In this paper, we describe the LitPC toolkit, a variety of tools and methods designed for the quick and effective creation of parallel corpora derived from literary works. This toolkit can be a useful resource due to the scarcity of curated parallel texts for this domain. We also feature a case study describing the creation of a Russian-English parallel corpus based on the literary works by Leo Tolstoy. Furthermore, an augmented version of this corpus is used to both train and assess neural machine translation systems specifically adapted to the author’s style.
16:00-16:30
An Analysis of Surprisal Uniformity in Machine and Human Translations
Josef Jon, Ondřej Bojar
This study examines neural machine translation (NMT) and its performance on texts that diverege from typical standards, focusing on how information is organized within sentences. We analyze surprisal distributions in source texts, human translations, and machine translations across several datasets to determine if NMT systems naturally promote a uniform density of surprisal in their translations, even when the original texts do not adhere to this principle. The findings reveal that NMT tends to align more closely with source texts in terms of surprisal uniformity compared to human translations. We analyzed absolute values of the surprisal uniformity measures as well, expecting that human translations will be less uniform. In contradiction to our initial hypothesis, we did not find comprehensive evidence for this claim, with some results suggesting this might be the case for very diverse texts, like poetry.
16:30-16:45
Closing