Abstracts

Nicolas Bacaër - Institut de Recherche pour le Développement

Feedback on the post-edited machine translation of a popular science book 

Although machine translators are still of very limited use for the translation of literary texts, they can nevertheless be of some service for the translation of less subtle texts such as popular science books. In this presentation, I will present some remarks on an experiment carried out with the automatic translator DeepL to translate and then post-edit in several languages a book initially published in French. Using the English translation (done manually) as a pivot, the book was translated into 12 languages: Italian, Spanish, Portuguese, Romanian, German, Dutch, Russian, Polish, Czech, Bulgarian, Hungarian and Chinese. With the Yandex machine translator, the book was also translated from Russian into Ukrainian. Other translations are in preparation: into Japanese, Swedish, Catalan, Greek... These translations raise the question of linguistic diversity in a field characterised by the virtual monopoly of a single language and by a high concentration of profits in a very small number of publishing houses.

 

Carole Birkan-Berz - Université de la Sorbonne nouvelle - Paris III : EA4398

On the Frontiers of Literature and Translation : Facebook neuron machine translation of literary texts in troubled times 

Facebook/Meta is a social network that systematically offers the translation into various languages of a variety of original content posted by Internet users, thanks to a convolutional neural network (CNN). Like many of its competitors, Meta highlights the results produced by its model, namely the dissemination of information in real time throughout the world, as well as its accessibility and democratisation, by sharing it since 2022 in open source.

In times of conflict, journalists and citizens have been turning to social media to obtain information on theatres of operation that are not well covered by news agencies. In 2017, Facebook launched the option to read these posts in automatic translation. Although the risks of misinformation are numerous, they are offset by the supposed reliability of the source, which is often picked up a few days later in the traditional media, then translated by a human translator, professional or not. The content often comprises war literature or a form of poetry that has been termed 'poetry of extremity' -- eyewitness accounts, pleas, poetry of despair or resilience. These texts are literary because they express a singular voice - often that of righteous indignation - and imbued with many erudite reference. They demand to be read but seem to resist translation, at least by the Facebook algorithm. This raises the question of whether this device works as well as the social network seems to claim. This qualitative study proposes to examine this phenomenon located at the border of literature and translation. We will look in turn at a testimony from China in 2022 to commemorate a dissident, then at a pamphleteering post condemning non-interventionism, and at poems written at the beginning of the war in Ukraine.

 

Juliette Bourget - Université Sorbonne Nouvelle - EA 4398 PRISMES - ED 625 MAGIIE

Does a machine have a singular voice? Towards an investigation of the style of automatic translators

Literary translation is often perceived as difficult to handle with technology, hence a growing number of studies that seek to evaluate the performance of machine translators for this particular type of translation. As Youdale (2020) notes, the difficulty of literary texts lies not only in problems of lexicon, as the lexicon in literature is often simpler and less varied than in technical texts, but in the unusual ways in which authors manipulate language. Moreover, a literary translation is expected to respect the meaning of the original, but also to offer a comparable reading experience, which must involve preserving the stylistic characteristics of the source text. For example, in a novel like Patricia Highsmith's The Talented Mr Ripley (1955), the empathy that the reader feels for the hero, who is a criminal devoid of any moral sense, is made possible by the spectral presence of the narrator whose voice, which is supposed to guide the reader, is constantly contaminated by interventions from the character's thoughts.

We therefore propose to analyse how different neural machine translators approach the translation of the statements in this novel that present this mixture of voices, as they contain stylistic elements known to be potentially problematic for the machine, such as the length of sentences, the mixture of registers, and the presence of free indirect speech, in addition to the linguistic complexity linked to the association of the two voices. Beyond a simple identification of the errors produced by NAT, we wish to use the tools of corpus stylistics in order to determine the potential existence of a personal strategy for each translator: is it possible, as Mona Baker (2000) does for biotranslators, to identify recurrent linguistic patterns, which would show a particular use of language? The strategies of different NAT systems will be compared to those of a human translator, in order to test a new measure of coherence: the aim is not to study referential cohesion and anaphoric networks at work in a literary text, but to understand whether the machine is capable of providing a consistent voice.

 

Helene Buzelin - Université de Montréal

Publishing translation technologies - Survey of Quebec professionals

In the last few decades, the book and translation industries have undergone upheaval. In the 1990s, translation memories made their (lasting) entry into translation agencies and university courses. At the same time, the book industry began its 'digital revolution'. Finally, recent advances in NMT have struck a chord in the imagination, reactivating old myths and sometimes creating a 'sense of dread' (Cronin 2013) in the profession.

How do translators working for publishing houses relate to the technologies of translation and to the changing and multiple materialities of the book? How do they experience these changes? What tools and media do they use? Which ones do they prefer? Have they changed their habits? To what extent and how have they appropriated the technologies that are now an integral part of the 'workstation' of any 'pragmatic' translator?

While theoretical reflections on the materiality of translation abound, empirical research giving voice to those who live by translation is rare. Research that questions publishing translators on this subject is even rarer, as if the use of certain technologies in the literary field (the publishing window) were taboo, inconceivable. It is in this spirit that we conducted a survey between 2018 and 2020 among translation professionals in Quebec working in different publishing sectors: from "legitimate" literature to didactic books, including essays, children's books and cookbooks.

In this paper, we propose to share the results of the twenty or so semi-structured interviews conducted with these professionals. The reflections will cover several of the issues outlined in the call: the transferability of pragmatic translation practices in literature, the way in which MT and CAT tools change the relationship to the text, the collaboration of human and machine, and the attitudes of publishers, as perceived by these translators, towards these translation technologies. We will see that the working experiences are varied and the realities complex, challenging conventional wisdom and sometimes even the most rational assumptions.

 

Isabelle Chauveau, Loïc De Faria Pires - Université de Mons

Post-editing NMT in the classroom: the antithesis of feminist literary translation?

This proposal concerns a multidisciplinary study linking post-education teaching and feminist literary translation. Literature can be seen as a vehicle for women's identity construction (Fourgnaud, 2017: 5). This seems to be particularly true for fairy tales (Schanoes, 2014: 1). Therefore, the gendered translation of such works should enable the receivers of the target text to fully benefit from the works in the target language. Gender bias seems to be a frequent MT problem (Rabinovich et al., 2017: 1074; Prates et al., 2019: 1), regardless of language pairs and engines (Stanovsky et al., 2019: 1679). Thus, gender biases, when not properly post-edited, lead to the discrimination of women (Vanmassenhove et al., 2018: 3003).

We conducted a practical study with 60 male and female first-year translation students at the University of Mons. We provided them with a 300-word excerpt from Angela Carter's (1979) feminist rewriting of the Bluebeard tale, entitled The Bloody Chamber. This extract was automatically translated into French by two commercial engines: Google Translate and DeepL. Half of the cohort were asked to post-edit the DeepL version and the other half the Google Translate version. Beforehand, we compiled a list of relevant gender biases produced by the two engines. We then analysed how the students post-edited these items. Finally, the students filled in a questionnaire related to the EP strategies used.

During our presentation, we will present the data obtained from these questionnaires, which we will put into perspective with the post-edited texts, in order to determine whether the content produced by the students corresponds to their declared strategies. We will also determine whether one of the two engines leads students to produce better texts (in terms of genre) than the other. Finally, we will conclude with a series of future perspectives related to gender issues in the field of post-publishing NAT.

 

Isabelle Collombat, Hanna Martikainen - Université Sorbonne Nouvelle - Clesthia, Langage, systèmes, discours EA 7345

What is the place of tool augmented translation in the literary sphere?

The advent of machine translation (MT), especially neural translation, and the spectre of its application to the literary sphere give rise to a number of 'passionate diatribes' and fuel the fear that computers will 'take our jobs away' (Hadley 2020: 14). Literary translators fear that their profession will be as profoundly transformed as that of their colleagues working in pragmatic translation, and that commercial imperatives and a lack of understanding of the role of translators will affect the quality of translated works, with the risk of a widening gap in prestige between the different literary genres.

We shall see that it is necessary to go beyond the simple reactions of neophobia leading to the instinctive rejection of novelty and the fear of change in order to expose the reasons which militate in favour of the defence of biotranslation in the literary sphere; While MT can be successfully used for the translation of informative texts, it seems illusory to apply it to the translation of expressive texts, since such texts involve sensitivity from transmission to reception, as well as a certain vision of creativity and intentionality, which we will develop.

This raises the question of so-called "tool-augmented" translation, i.e., using the digital ecosystem familiar to pragmatic translators, in literary translation: while some tools can be useful and increase creative potential, particularly in terms of refining style, others present a cognitive ergonomics that is not very conducive to interpretative freedom. At present, the intervention of the machine in the creative translation process is felt to be a constraint. In order to overcome the current dynamic, we call for a rethinking of the tool-based approach to literary translation, in the image of what is done in other fields, so that the tool is at the service of creativity in translation and not an obstacle to it.

 

Manon Hayette - Université de Mons - ChinEAsT, FTI-EII

Do new technologies simply 'add legs to a snake'? The contribution of CAT and MT to the translation of Chengyu from Chinese

Phraseological units (PUs), known to be strongly linked to their language-culture of origin, are a stumbling block for literary translators. Since these difficulties are due in part to the lack of optimal tools, one may wonder whether new techniques using AI could solve this dilemma.

Our paper will discuss chengyu, the most distinctive PUs in Chinese (Conti, 2019; 2020, p. 412), whose inherent characteristics - relative figment, quadrisyllabic pattern, functional fluency - make translation singularly difficult.

In our study, we will make use of a parallel literary corpus (ZH-FR) previously compiled and analysed by Henry (2016) in his thesis and including some works by contemporary Chinese authors 高行健 Gao Xingjian, 莫言 Mo Yan, 苏童 Su Tong and 余华 Yu Hua. Since the automatic extraction of Chinese phrasal phrases is particularly difficult (segmentation problems), we will take up the chengyu previously isolated manually by Henry. First, we will observe the translation of chengyu in parallel corpora (ZH-FR) such as The Lancaster Corpus of Mandarin Chinese (Xiao and McEnery) and OPUS2 (SketchEngine). Then, we will examine the translations of the same UPs screened by various machine translation software (GoogleTranslate, DeepL, Bing Microsoft Translator, 百度翻译 Baidu Fanyi, 云译 Cloud Translation). Finally, we will show how we can also make the best use of online bilingual dictionaries (such as Pleco, 法语助手 frdic.net, 萌典 moedict.tw), by cross-referencing them with the data from the corpora, in order to propose improvements in translation.

These analyses, which we will compare with the model of chengyu translation criticism applied by Henry (and whose limitations we will mention), will contribute to the reflection on the possible contribution of CAT and MT to ZH-FR literary translation and its criticism.

 

Daniel Henkel - Université Paris 8 Vincennes St. Denis - TransCrit EA1569

A Corpus-based Study of Target-language Norms/Deviation in Human and Machine Translation 

This is the first large-scale corpus-based study of both human and machine literary translation.

Original works and translations which were both in the public domain (late 19th-early 20th c., ca.1865-1930) were collected on a 1-work/author-translator basis so as to produce a “bidirectional” (Johansson 2007) corpus consisting of 4 subcorpora: Original English (35 authors, 3.5 m. words), Original French (35 authors, 3.4 m. words), English-translated-from-French (35 translators, 3.3 m. words) and French-translated-from-English (35 translators, 3.7 m. words). All source/target pairs were aligned at sentence-level and tagged for part-of-speech and lemma with TreeTagger. At 140 authors/translators and over 14 m. words total, this is currently the largest bidirectional, aligned, linguistically-annotated corpus of English and French and the largest English/French literary corpus. All original texts were translated a second time in late 2021-early 2022 using the latest neural-MT version of DeepL Pro.

The bi-directional model allows comparisons to be made, not just between source- and target-texts, but between Original English and English-translated-from-French. The statistical comparison between authors and translators in the same language has demonstrated that, in many respects, human-translated texts display a combination of interlinguistic interference or “translationese” and interlinguistic influence or “shining-through” and thus deviate from target-language norms enough to be considered as a distinct “sub-species”.

The inclusion of machine-translated target-texts then allows further comparisons to be made between human translators and MT. Whenever deviation from target-language norms is observed in the corpus of human-translated target-texts, it can be seen that MT likewise has learned to imitate human translators, rather than authors, or is at least subject to the same interlinguistic interference and influence. Moreover, basic metrics such as source/target-length ratio show that MT exhibits less variation than human translation, which may be construed as a rough indicator of the level of creativity or “freedom” with respect to the source-text.

Daniel Henkel is an Associate Professor at Paris 8 University, where he teaches Computer-Assisted Translation and Terminology, and translator of scientific articles. His research in Contrastive Linguistics and Translation makes use of corpora and statistical methods to evaluate interlinguistic influence and interference in translation and assist human translation through electronic resources.

 

Bartholomew Hulley - Université de Lorraine : EA2338 - Interdisciplinarité dans les Etudes Anglophones - Interdisciplinarity in English Studies

Using NMT to measure bio-translator style

In the past, texts output by machine translation engines (SMT and NMT based) have been evaluated by comparing them with ‘gold standard' bio-translations. This practise has led to the formulation of numerous metrics that purport to reflect the accuracy of the machine translated text, by quantifying the number of disparities, or ‘errors', between the two. Acronyms such as WER, TER, HTER, BLEU and NIST have therefore become familiar to those researching machine translation tools. However, it is worth recognising that such metrics are not wholly objective because they rely upon human subjectivity to define what constitutes an error-free translation. Until a truly objective way of automatically measuring translation quality is devised, any metric involving human judgement will inevitably be imperfect. Any system based on the notion of a singular ‘gold-standard' translation therefore risks classifying perfectly acceptable translations as erroneous.

In this context and given the widely recognised and increasing reliability of NMT, it seems logical that a certain amount of objectivity would be restored into the practise of textual analysis if the subjects concerned were reversed. That is, instead of using a bio-translation to measure a machine translation, to do precisely the opposite: to use a machine translation to measure a human generated TT. Indeed, considering that an NMT engine will translate according to ‘learned' paradigms with complete disregard for extra-linguistic contexts (such as style, purpose and function - meaning it will tackle every ST in precisely the same manner), it thereby offers a dynamically ‘fixed' baseline from which to measure one or many human generated TTs.

In a case study analysing English translations of awards-nominated French comics titles, I demonstrate how applying this rationale can offer the researcher a measure of the translator's approach to any given text and enable the objective comparison of one translator with another.

 

Miguel A. Jimenez-Crespo - Rutgers University System

General features of human translation in the age of machine translation: an additional perspective on what makes human translations highly creative

Recent research into creativity and machine translation in literary texts (e.g., Moorkens et al 2018; Guerberof and 2020, 2022) has opened up the field to the study of specific features of human translations that are not shared with machine translated texts. Among the different features that can have an impact on creativity, this paper focuses on the impact of “translations universals” (Baker 1993, 1995) or “general features of translation” (Chesterman 2004). For years, research into these “general features of translation” have provided insights into features of translated texts independent of the language direction that separate them from non-translational or original texts. The origins of this research area go back to the study of translated literary texts, such a children's stories by Gellerstam (1986) or literary translations in the Translational English Corpus (Baker 1995, 1996; Kenny 1998). With an eye on discovering how MT or post-edited texts differ from those translated by humans, studies have described how these texts lack certain features associated to lower levels of creativity, such as interference (Toral 2019; Popovic 2019; Kuo 2019), simplification/ normalization (Lipshinova-Koltunski 2015) or low lexical diversity (Brglez and Vintar 2021). In addition, NMT also produces highly literal translations that lack the lexical or referential explicitation (Lipshinova-Koltunski 2015; Ahrenberg 2017; Kuo 2019; Kruger 2021, 2022), one of the most common features of human translations and a potentially creative strategy. This paper critically reviews existing contrastive studies on the general features of translated language to put a focus on the “added value of human translation” (Massey and Ehrensberger-Dow 2017: 308) and how Corpus-Based Translation Studies can shed light on the debate on creativity in this area. This represents an additional perspective to fully what make human translation different, unique and highly creative. Examples will be provided from the multitranslational corpus of translated creative published in digital media, alongside the Google NMT “See Translation” integration in Twitter tweets (Jimenez-Crespo 2021, 2022). In doing so, the paper will offer insights into distinguishing features of creativity in published translations that are absent machine translated texts.

 

Waltraud Kolb - University of Vienna

"Quite puzzling when I first read it”: Is reading for literary translation different from reading for literary post-editing? 

Translating and post-editing literary texts are widely considered to be two quite different translatorial activities, not least by literary translators themselves who tend to feel constrained by machine translation output (e.g., Moorkens et al. 2018) or see their “artistic autonomy” at risk when post-editing (Oeser 2020, 23).

One of the most obvious differences is that in the case of post-editing the MT system provides the first draft, while in human translation translators have to invest considerable effort in drawing up a first version. This (human) drafting phase (Mossop 2000) is usually the phase in which the bulk of research occurs and the translator's (creative) engagement with the source text is the most intense. One significant part of this engagement with the source text is the act of reading. Is “reading for translation” different from “reading for post-editing”?

In my presentation, I will use translation process data (think-aloud protocols and keylogs) from an empirical study in which five professional literary translators translated a short story by Ernest Hemingway into German and five different literary translators post-edited a first draft generated by DeepL. I will first outline differences in task organization and workflow patterns between the two groups of participants (some results relating to workflow patterns of the group of translators have been published in Kolb 2017) and then zoom in on the participants' reading processes. Drawing on cognitive stylistics and narratology frameworks (Herman 2003; Boase-Beier 2019), I will show how translators and post-editors develop narrative understanding of the story, the characters, and the worlds evoked by the source text; how they infer meaning from the text and its style and, for example, fill in blanks and gaps during their readings; and how differences between reading processes may be reflected in the final target texts.

 

Philippe Lacour - Campus Universitário Darcy Ribeiro - Universidade de Brasilia

The three structural limits of Artificial Intelligence

I try to identify and distinguish three structural limits of Artificial Intelligence, respectively related to computation, probabilistic approach and modelling. Firstly, I will show that the notion of computation is difficult to apply to the domain of meanings, because it comes up against certain "internal limitations" of formal symbolisms (Ladrière); in other words, if computation is indeed a thought, not all thought takes the form of a computation. Then, I will insist on the difference, in the sciences of randomness, between probability and frequency (statistics) - a difference often implicitly reduced, if not denied, in conjectural approaches, when the reasoning surreptitiously passes from what has actually taken place (in the past) to what is likely to happen (Granger). Finally, I would stress that any modelling of a signifying phenomenon is both legitimate and strictly framed by a certain number of constraints, whether in the forward movement of constituting data from the signifying real or in the backward movement of applying them to this same real.
These three remarks remind us first of all that, according to its initial project (now largely forgotten), AI aims above all to better understand human intelligence and not to replace it. They also point, at a deeper level, towards a tenacious presupposition that makes automation an equivalent of technology, to the detriment of other possibilities, numerous and rich, of interaction with machines (Simondon).

Finally, I draw some conclusions about the future of literary translation in the age of neural machine translation and post-editing. It is doubtful that it will be able to replace human translators, for the three reasons mentioned above. It is much more likely that new types of collaboration between machines and human collectives will emerge, based on processes of assistance, suggestion, comparison and interpretation.

 

Claire Larsonneur - Université Paris 8 – Vincennes – Saint-Denis - TransCrit 

Can an AI package be a literary author? The case of language models

The beauty of artificial intelligence technologies is that they can be applied to all sorts of domains, once they are converted to data. Neural translation is part of a continuum of linguistic tools: text-to-speech and vice versa, predictive typing and so on. All these facets of machine language processing are mobilised in what could be the next stage of linguistic artificial intelligence: language models. Although the names BERT or ELLMO are not well known to the general public, the launch of GPT-3 in the summer of 2020 has made it the talk of the town in the press, because it is a joint project between Microsoft and Elon Musk, but also because of the power of the tool. Language models are distinguished by their versatility. From a few initial elements provided by humans, they can autonomously write all kinds of content: summaries, medical diagnoses, philosophical essays, news, computer programs. This is made possible by the computing power currently reserved for a few very large groups, and by the complexity of the model, which goes as far as processing several hundred million parameters. After a brief review of the history of these tools, I would like to analyse in more detail the type of literary production they generate by looking at several examples in the fields of journalism, online role-playing games, and theatre. I will then look at the reception of this type of work and ask how it is positioned in the economy of culture, between performance and commercialisation.

 

Bruno Poncharal - Université Sorbonne nouvelle - EA 4398

Testing NMT for discourse coherence

Despite the spectacular progress made in recent years, current NMT systems are limited by the fact that they generally translate at the sentence level, i.e. without being able to take into account the co(n)text, which will lead to problems of textual consistency. Furthermore, professional translators and translatologists agree that we do not translate "sentences" - or "words" - but "texts". A text could be defined, in the first instance, as a coherent sequence of sentences, this discursive coherence itself depending largely on complex, often language-specific anaphoric relations. The literature on the functioning of anaphoric relations in a variety of languages is abundant, but there has been little contrastivist research in this area. My own practice of translation (from English into French), particularly of humanities and social sciences (non-fiction) texts, has taught me that the way in which anaphoric relations are organised at the text level is often not transposable from one language to another. This phenomenon is all the more noticeable in HSS texts where argumentation or demonstration play a central role.

On the other hand, certain phenomena characteristic of literary texts such as free indirect speech and represented perception often pose translation problems as they result from an unstable combination of linguistic markers involving complex temporal and (pro)nominal anaphora. The fact that free indirect speech passages can escape the untrained eye suggests that a machine will have even more difficulty in identifying and translating them accurately.

Our study will therefore draw on a few examples from the two types of text we have discussed, in order to show precisely where machine translators stumble in these texts.

We will also show how the comparison of NMT with bio-translation can help us to better understand what it really means to read and understand a "text", even though the specialists themselves admit that the machine does not understand.

 

Mehmet Şahin, Tunga Güngör, Ena Hodzik, Sabri Gürses , Zeynep Yirmibeşoğlu, Harun Dallı, Olgun Dursun - Bogazici University

Literary Machine Translation to Produce Translations that Reflect Translators' Style

In our presentation, we will be reporting on a research study* that involves building an English-Turkish machine translation (MT) model to translate literary texts whose style has been defined using corpus tools. We aim to determine the extent to which an MT system can reproduce a translator's style if it is trained on a corpus of previous translations by the same translator. Our presentation will focus on the following questions: 

Does corpus stylistics allow us to better study the style of the translator?  

Can we teach machines to replicate a translator's style?  

To answer the first question, we adapted Leech and Short's (2007) methodology of literary style analysis to Turkish and used AntConc (Anthony, 2022) and WordSmith (Scott, 2021) as corpus visualization tools to analyze the style of a particular English-Turkish translator both qualitatively and quantitatively (Youdale, 2019).  Our preliminary findings suggest consistent use of particular lexical and morphological forms as discourse markers. Our corpus analysis findings will be presented against a reference corpus representing Turkish literary translations between 1946 and 2015 in order to determine whether the stylistic features we identified are exclusive to our translator or general linguistic phenomena.  

We have also applied text classification methods used in authorship attribution to identify different translators. One key question here is how these classifications are affected when two variables, namely author and translator, are involved instead of only the author.  

To replicate a translator's style in machine translation, we built several machine translation models based on both LSTMs and transformers. We are using our own datasets consisting of chosen translators alongside publicly available bilingual corpora. We will conclude our presentation with the initial findings and implications for our future work.

*The study is part of a scientific research project funded by TÜBİTAK (the Scientific and Technological Research Council of Turkey). The title of the project is “Literary Machine Translation to Produce Translations that Reflect Translators' Style and Generate Retranslations” and its grant number is 121K221.

 

Anna Schewelew - University of California, Santa Barbara

“Pushkin Need Not Shudder”: Machine Translation and Poetic Language

In his 1949 memorandum on Machine Translation, Warren Weaver famously outlined the idea that computers could be used to translate. Yet Weaver explicitly excluded literary texts as a use case for machine translation because literature is the locus of the “alogical elements in language” (Weaver 1949) that cannot be transformed into mathematical terms and thus defy automation. Weaver's “reverence” for literary translation is remarkable because it suggests that understanding literary texts as art, that is beyond their merely propositional content, is out of reach for “intelligent” language processing machines. While machine translation has come a long way since 1949, I argue that it is still ill equipped to tackle poetic language because poetic language transforms linguistic structures into aesthetic objects that cannot simply be reduced to a statistical distribution of words because their meanings are also derived from their sound, graphic appearance, their “mouthfeel”. To understand and translate literature, it is not enough to understand and translate a distribution of words. It is equally important to understand and translate a distribution of words as aesthetic objects. In order to translate literature properly a machine translation system would thus need to be able to switch from a “logic” mode to a “sense” mode and vice versa – something that is very much out of reach for contemporary machine translation systems that can process quantifiable data only. Finally, by looking at contemporary neural machine translation systems through the lens of Schleiermacher's theory of translation, I show that – if literary texts are to encode novel, multilayered ways of speaking about the world, as Schleiermacher suggests – machine translation alone must fail to recreate this novelty because due to its reliance on distribution probabilities within large corpora of texts it is can reproduce what has already been said over and over again.


Kristiina Taivalkoski-Shilov, Paola Brusasco - Université de Turku

MT as a Re-editing Aid? Revising the Finnish and Italian translations of Silent Spring and The Job with the help of DeepL 

Neural networks and deep learning applied to translation software have been yielding more and more adequate outputs, characterized by such fluency that even literary texts are being experimented with as potentially machine translatable (cf. Toral & Way 2018; Kuzman et al. 2019; Oliver, Toral & Guerberof 2019; Toral et al. 2020, among others). Fascinating as it is, the possibility of literary texts translated by software does not sit comfortably because of the threat is poses to translators and to one of the most typically human characteristics – creativity. It seems reasonable, therefore, to try and avail oneself of the possibilities offered by machine translation (MT) in a way that would still leave the translator in control of the process. To do so, we envisage MT as a tool to support the translator in the very specific case of revision, i.e. “editing, correcting or modernizing a previously existing translation for re-publication” (Koskinen & Paloposki 2010/2016).

Our contribution focuses on two texts originally written in English – R. Carson's Silent Spring (1962) and S. Lewis' The Job (1917) – translated into Finnish and Italian in 1963 and 1955 respectively. Since the topics – environmentalism and working women's difficulties – are still relevant, while the language may have aged or, as is the case with the Finnish version, may not have succeeded in recreating the style of the source text, we are envisaging new translations carried out with the aid of MT. Our study is an experiment to ascertain a) whether selected strings of the machine output can be effectively incorporated in the unsatisfactory parts of the first translated versions and b) if not, what kind of editing would be required. Two chapters have been selected from each text to be translated – Silent Spring into Finnish and The Job into Italian – with a generic web-based system, DeepL, and the resulting outputs are discussed from the point of view of their usability for the purpose mentioned above.

 

Samuel Trainor - Université de Lille - Centre d\'Études en Civilisations Langues et Lettres Étrangères (CECILLE) - EA 4074

Mixing Human and Nonhuman Translation: Neurodiversity and the Sociology of Inclusion in Contrapuntal Translation Practice

In Bruno Latour's essay "Mixing Humans and Nonhumans Together: The Sociology of a Door-Closer” (1988), he proposes the term “transcription” to refer to the “translation of any script from one repertoire to a more durable one”, such as the replacement of a police officer with a set of traffic lights.

This paper calls into question this paradigm of social replacement as it relates to NMT. Latour's metaphor derives from an instrumentalising misrepresentation of translation. It is hardly surprising that current debates surrounding machine translation are haunted by its spectre. The majority of NMT platforms are designed to work autonomously, even when they incorporate responsive learning. They are also invariably ‘domesticating'. Their designers want them to pass a Turing Test: producing simulacra of human translations capable of convincing a reader they were written by an accomplished native speaker. This is classic translatorial ‘invisibility'. Anyone with a grounding in translation theory can unpick the basic premise. However, this paper suggests that translation professionals feeling threatened by NMT should avoid abstruse theoretical objections, or appeals to human creativity, and promote the ethical case for a more inclusive, ‘neurodiverse' translation practice.

What separates ‘inclusion' from ‘accessibility' in social policy is the involvement of a variety of target users at an early stage of conception. The growing influence of artificial neural networks in our societies is redefining the concept of neurodiversity, potentially rendering all humans neurodivergent. So the argument for early-stage ‘biotranslation' (not merely in post-editing) has become an ethical question of inclusion. A key contention is that NMT cannot be discussed in isolation from other technologies and their social and ecological impacts. Inclusive working methods are significantly boosted by technological platforms for collaboration. Meanwhile, similar technologies are making literary publishing increasingly multimodal and interactive. But all of this comes at a cost. If NMT could be moved away from current paradigms of replacement, made ecologically more sustainable and methodologically more open and plural – in particular if it could be integrated into technologies that enable interaction, working polyphonically in collaboration with biotranslators – then it could become genuinely valuable in literary translation.

 

Jean-Louis Vaxelaire - Université de Namur

Translating text or sequences of phrases? Machine translation between Turkish and French

Ray Kurzweil, one of the leading figures in the computer world, announced in 2011 that there would be no need for translators by 2029. However, he made an exception for literary translation, which he said was already difficult for humans to do. Since then, with the advent of neural MT, research teams (at Microsoft and Google for example) claim to have surpassed the quality of human translations. From my linguistic point of view, this kind of triumphant discourse mainly shows a lack of understanding of languages and the translation process.

An analysis of the examples of MT presented in the research literature shows that, while they are sometimes excellent, they always involve English as source or target and, secondly, the textual genre of the translated passage is always close to the training material. As soon as one translates between languages that do not include English and in genres that are far from the original corpus (the range is wide, from Facebook's oral writing to literary texts), the results are much less satisfactory.

We will begin by observing the theoretical problems that explain the difficulties encountered by MT and then secondly, we will analyse the results of some excerpts of Turkish poetry and prose texts (among others Karaosmanoğlu, Hikmet, Dağlarca, Baykurt and Pamuk) in French and compare them with the official translations. The results vary according to the four translation tools used (Google Translation is generally superior to Webtran), but pose different types of problems that can often be explained by the fact that, unlike the human translator, the software does not have a global vision of the text and treats it as a series of unrelated sentences. The question of genre, which is essential in literature, is also never taken into account.

Online user: 1 Privacy
Loading...