Академический Документы
Профессиональный Документы
Культура Документы
1
See pp. 33-36 for a select list of Arnt Lykke Jakobsen’s publications.
x Contents
Uncovering the ‘lost’ structure of translations with parallel treebanks .... 199
Matthias Kromann, Iørn Korzen & Henrik Høeg Müller
Triangulating product and process data: quantifying alignment units
with keystroke data.................................................................................... 225
Michael Carl
CBS is unique among the thousands of business schools around the world.
During the last two decades, it has developed a broad range of disciplines
within the social sciences and the humanities, emphasising cross-
disciplinary collaboration in order to meet the challenges from a fast
changing and complex global knowledge society. Language, culture and
communication are now recognised as core areas of its profile.
Within translation and translation technology, Arnt Lykke Jakobsen
has been an invaluable academic entrepreneur, who has founded a research
centre, implemented EU projects and established international net-
works. He has also been a key figure in one of the six world-class research
environments started up at CBS in 2008. In addition, he has been actively
involved in setting research agendas, developing strategies and facilitating
collaboration across research areas both in the Academic Council and in
countless everyday CBS activities.
Arnt has shown himself to have an open mind, boundless curiosity
and proved in all ways a highly valued colleague. It has been a privilege to
work with him.
Finn Junge-Jensen
President of CBS, 1987–2009
INTRODUCTION
Anthony Pym shows how translation process research can benefit the
training of translators. Nineteen students from different language programs
Introduction 3
The second section brings together various contributions dealing with ways
in which computers can assist translation process researchers in their
endeavours.
Michael Carl proposes a strategy and a set of tools for cross-validating and
triangulating Translog product and process data. Translation process
scholars are familiar with translation units (cognitive entities in the process
data), but Carl now introduces the notion of an alignment unit (AU), which
refers to translation equivalences in the source and target texts in the
product data. Once the source and target texts have been fragmented into
AUs, all the keystroke data obtained from Translog can be allocated to
them.
participants, ethics, data explosion and validity), some of which have been
further subdivided. Each of the issues is discussed followed by possible
solutions to the problems.
order had to be reversed even though there was no significant word order
effect on pupil dilation. The study suggests possible interpretations for this
difference.
illustrates how various scholars have attempted to come to grips with the
notion by means of various categorisations (textual and procedural; local
and global; product- and process-oriented). The paper concludes with a
map showing how the different strategy notions might be related to each
other.
Inger M. Mees
When, in 1985, Arnt Lykke Jakobsen moved from his post at Copenhagen
University to accept an associate professorship at the Copenhagen Business
School (CBS), his father jumped for joy. Now the grocer‟s son was back
where he belonged. He‟d had the good sense to return to a business
environment with every promise of a secure and successful future career.
And today, almost twenty-five years on, were he still alive, Jakobsen père
would have every reason to say “I told you so”. At the time, our protagonist
might have had his doubts about his decision. Coming as he did from a
traditional arts faculty with an emphasis on core university subjects such as
literature, history and philosophy, and having as his main interest English
literature, the move to the Business School was a major step, and he must
have felt uncertain as to what the future held in store. His research had so
far largely focussed on literary theory and analysis, notably the works of
Malcolm Lowry (Jakobsen 1980, 1981) and D. H. Lawrence (Einersen and
Jakobsen 1984). In Jakobsen (1985a) he had attempted to explain why
Lawrence still appeals so much to modern readers – even though the
reasons for reading him are so different from the situation in the thirties.
Jakobsen (1985b) is a meticulous and captivating account of how
Lawrence‟s works were received and interpreted in Denmark in the half-
century spanning 1932 to 1982. Incidentally, we may also note Jakobsen
(2007), where Arnt returns to his early literary interests. But although the
main focus was on literature, one can also trace an early fascination with
text linguistics, pragmatics and semantics, all of which formed a basis for
10 Inger M. Mees
his literary analyses, and which would now prove invaluable to him in his
new role of making the teaching of translation at CBS more theoretically
informed. His faculty for critical thinking and the intellectual curiosity that
drove his research were clearly present right from the outset of his career.
Arnt Lykke Jakobsen was born on 27 December 1944, and grew up
in the little town of Skjern in western Jutland – far from the bustle of the
metropolis, and where at the time a large percentage of the population had
only limited education with very little chance of acquiring more. His all-
round talents were evident from an early age, both to his parents and his
schoolteachers. It was therefore no surprise that he became, as his mother
respectfully used to put it, “the first academic in the family”. After
completing secondary school in Tarm, a nearby town, at the age of 18 he
moved to the capital to read English at Copenhagen University, obtaining
his MA in 1972. He was appointed Assistant Professor at the University in
the same year, and Associate Professor in 1978. In 1973/74 he spent a year
in the USA as Visiting Lecturer at Tufts University, Medford/
Massachusetts.
Coming from a somewhat remote area had given him an insatiable
curiosity about the wider world and one of the ways he satisfied this was
through reading the classics of world literature. Above all, he devoured
works in English by famous authors – Laurence Sterne, Joseph Fielding,
George Eliot, Joseph Conrad, W.B. Yeats, James Joyce, Graham Greene,
John Fowles, Salman Rushdie, to name just a few. Even today, his desire to
obtain information about anything and everything remains undiminished. I
have often seen him grab a dictionary, or start googling, right in the middle
of a conversation if a topic crops up where he feels he needs to boost his
knowledge. It could be anything – translating an English word into
German, the rendering in English of a Danish idiomatic expression, the
geographical location of a town or country (Google Earth is a favourite!), a
composer‟s date of birth, the Latin name of a flower or bird, the workings
of software programs, or even the rules of cricket – I could go on. His
genuine need to soak up all kinds of information is doubtless the
explanation for his encyclopaedic knowledge covering a wide range of
topics. And it is also why he is such a fascinating conversationalist.
Arnt Lykke Jakobsen: portrait of an innovator 11
As mentioned above, 1985 saw him make the move from Copenhagen
University to the Department of English at CBS. His new CBS colleagues
had a great deal of experience in teaching practical translation skills, but
the subject was greatly in need of a person who could provide a theoretical
foundation.
Before Arnt began to concentrate on domain-specific translation,
there was a transitional period in which his original interest in literature
was still clearly apparent, but in which one could already see in which
direction he was heading. In his last year at the university he had
contributed numerous brief entries on British and American literature to the
supplementary volumes of a Danish encyclopaedia, Gyldendals Leksikon.
Although he was rapidly adjusting to his new situation, for a time Arnt
remained much preoccupied with literary matters; he was still working on
D. H. Lawrence‟s reception in Denmark, and was also invited to talk on the
main Danish radio channel on various authors (for example, on Lawrence
Arnt Lykke Jakobsen: portrait of an innovator 13
and his wife Frieda von Richthofen, and on John Fowles). Arnt still retains
his love of literature, one manifestation of this being his personal library
containing many thousands of books (including many first editions) for
which he is ever trying to find new storage space. It is a joy to watch him
picking up a book, gently removing the dust jacket, fingering the covers,
studying the title page, and then slowly turning the pages.
In the year of his CBS appointment, he published a paper in Danish
(Jakobsen 1985c) on translation and localisation (in its original broad sense
of adapting a text to a local culture, rather than the current more specialised
meaning of translating software packages, websites and other products that
need to be adapted for international markets). In this piece, his literary
roots are still clearly to be seen. The examples are from older Danish and
English literary works, including a translation into Danish of Alexander
Pope‟s Rape of the Lock, where, among other things, the translator has
localised the Thames as a Danish equivalent, Øresund („the Sound‟). In an
attempt to define the difference between translation and localisation, he
explains (1985c: 12) that translation involves equivalence between two
semantic structures while localisation consists in equivalence between two
semiotic structures, namely the complete meaning structures of the two
linguistic communities involved.
Jakobsen (1988) provides a detailed overview of the earliest
translations from English into Danish; these appeared in the middle of the
seventeenth century – the very first translated writings having come via
Latin. Translations of English literature into other modern European
languages emerged at the beginning of the seventeenth century but in
Denmark things moved somewhat more slowly. At that time, English was
little known, and the availability of German translations had made
translations from English into Danish seem superfluous (pp. 367ff.). The
vast amount of work involved in tracing these early specimens, and the
accurate biographical information provided on the translators, are typical of
Arnt‟s approach to research. Like his previously mentioned study of
Lawrence (1985b), this research is characterised by the same patient,
conscientious and meticulous search for facts, precise details and evidence
of extensive reading. Arnt never goes for the easy way out. Every line he
produces is considered and reconsidered, written and rewritten, and then
checked and double-checked.
14 Inger M. Mees
In the mid and late 1980s Arnt became fascinated by the new possibilities
available through the introduction of word processors. I remember his talk
to his colleagues at CBS in the mid-80s when I had helped him carry up an
incredibly heavy, monstrously large word processor from his car parked in
the basement area to a room on the third floor where he was demonstrating
this new toy. This was at a time when floppy disks really were floppy and
the whole apparatus was slow and clumsy – the computer groaned while
going about its tasks. The rest of us had just about come round to accepting
the electric typewriter. Some of us had even grown adept at switching the
golf balls to obtain different fonts, but very few of us were ready for these
new electronic wonders. Eventually we all did come to terms with them of
course, and even managed to give up hitting the return key every time we
reached the end of a line. But Arnt was way ahead of his colleagues. I
remember his eyes shining and the fervour with which he elaborated on all
the features and advantages of word processing. His enthusiasm made a
deep impression on us all.
Not surprisingly, the computer age (this was before we started
calling it the “information age”) did indeed present new avenues for
linguistic research, and Arnt quickly saw the potential it offered for storing
and processing large quantities of texts which would then enable the study
of translation phenomena. He was keen on using the LOB (Lancaster-Oslo-
Bergen) text corpus and even wrote a short manual for it in Danish. By the
late 1980s Jakobsen had built up his own Danish electronic corpus
(OVSDA), which was intended as a complement to an existing corpus of
original Danish novels. Remember that this was in the era before it was
possible to scan text and he had himself typed out no fewer than a thousand
extracts, each approximately 250 words in length, culled from fiction
translated from English into Danish between 1970 and 1975. OVSDA was
indexed using Word Cruncher. Word frequency and word collocation lists
Arnt Lykke Jakobsen: portrait of an innovator 15
were produced, forming the basis for certain of his publications on particles
and compounds (see below), all of which are replete with entertaining and
illustrative examples.
One publication arising from this painstaking work was Jakobsen
(1987), a significant corpus study illustrating differences in the frequency
of occurrence of a selection of words. An interesting example is provided
by the difficulties encountered when attempting to translate Danish modal
particles. An item like jo (in the sense “you know”, “you see”) is noticeably
more frequent in the corpus of novels written originally in Danish than in
the corpus of translated fiction. This far more restricted usage obviously
has an impact on the translations of source texts in which modality plays an
important role, and if one resorts to a translation using modal verbs or
adverbials, this results in a “distinct awkwardness of style” (1987: 104).
Arnt‟s work on corpora provided hard evidence for the frequent lack of
correspondence between the way meanings are mapped in original and
translated texts. The combination of working on word frequency lists and
his involvement in teaching courses on translation possibly sparked
Jakobsen‟s interest in lexicography. He wrote several reviews of
dictionaries, for instance, the NTC’s American Idioms Dictionary
(Jakobsen 1990); the very popular Danish–English dictionaries compiled
by Jens Axelsen, published by Gyldendal, commonly termed “Gyldendal‟s
red dictionaries” (Jakobsen 1991); and Politikens Visuelle Ordbog
(Jakobsen 1993). Consequently, the move to the study of compounds in the
early 1990s turned out to be not such a big step. Incidentally, the beginning
of this decade also marked the point at which Arnt switched to publishing
consistently in English, thus making his work accessible to a much wider
audience. A selection of his most important post-1992 publications on
translation written in English has been listed separately on pp. 33 to 36
(these are shown henceforth with an asterisk).
Arnt published several articles on the translation of compounds in
language for special purpose (LSP) texts, e.g. Jakobsen (1992a), and on
teaching translation of technical compounds, e.g. Jakobsen (*1992). For
these papers, he used for exemplification a comparison of the lexical
inventories of two Danish corpora, both of one million words, one being a
general Danish corpus and the other a biotechnological corpus. Arnt
retained his interest not only in the translation of compounds, but also of
16 Inger M. Mees
Arnt was now very conscious that Copenhagen Business School was
offering a degree in translating and interpreting. At Copenhagen University
he had been used to dealing with the translation of literary works, but now
legal, technical and economic language had to become an essential part of
his everyday life. He recognised that it was necessary for CBS to indicate
how different it was from the University, and from then on Arnt made it his
mission to emphasise the significance of LSP; perhaps not whole-heartedly
at first (he might be forgiven for remaining more enamoured of literature
than of LSP texts) – but with the realisation that this was what he had been
appointed to do and that there would be no future unless he accepted the
necessity of focussing on LSP. In 1990 he put in his first major application
for a project to be called Oversættelse af Fagsproglige Tekster”
(“Translating LSP Texts”) or OFT.
In that era, the pressure of having to obtain external funding was not
as significant as it is today but, characteristically, Arnt was ahead of his
time and could obviously sense the way things were developing. He spent
hours and hours writing the application, all the while downing gallons of
his favourite drink, coffee. His indefatigable efforts were rewarded: the
Danish Research Council for the Humanities allocated a huge grant (4.7
million Danish kroner) to OFT, and Arnt was appointed to act as its leader;
see Jakobsen (1992b) for an outline of its aims and objectives. The OFT
project ran from 1990 to 1994, involving 35 researchers from five different
Danish universities, and six different languages (Danish, English, German,
French, Spanish and Russian). Much thought had gone into its formulation.
Arnt‟s cultural background with its focus on literature and the
interpretation of texts forced him to realise that studying words in isolation
Arnt Lykke Jakobsen: portrait of an innovator 17
was not a fruitful enterprise. He remained faithful to his earlier view that
texts should be treated holistically and as culture-bound entities (semiotic
structures). This explains his insistence on focussing on translating LSP
texts and not merely translating LSP and so his overriding concern became
how language was actually used in the translation of authentic texts. The
LSP areas covered were IT, biotechnology and medicine. There were five
subprojects: (1) LSP translation theory in eastern Europe from 1970 to
1990; (2) translation theory of expert human translation of LSP texts; (3)
translation theory in connection with machine translation of LSP texts; (4)
meta-lexicography (specifically how translators draw on special-purpose
dictionaries); and (5) an applied project on the use of LSP contrastive
grammar for the translation of agro-industrial texts between Danish/English
and Russian (Jakobsen 1992b, *1994a). Arnt himself was involved in the
second of those listed above – one of his concerns being how translation
could be defined.
Apart from being engaged in research and heading the project, Arnt
continued to teach translation, and his experience as a teacher formed the
basis for insightful reflections on how translation skills could best be
acquired. In a talk given at a seminar on teaching methods organised by
CBS colleagues in the early 90s, Arnt adeptly transformed the Danish
saying Man må krybe før man kan gå (“You have to learn to walk before
you can run”) into “Man må skrive før man kan oversætte” (“You have to
learn to write before you can translate”). In Jakobsen (*1994d) he pursued
this line of thought, stating that translation is a special type of text
production. Inspired by the strong emphasis at the time on the function of
the target text, he observed that translation begins with the need for a target
text, and with all the usual questions we ask ourselves when we sit down to
write a text (p. 145). He argued that writing is a skill which precedes the art
of translation and that translation students should therefore first of all
concentrate on acquiring writing skills. To ensure this, a more process-
18 Inger M. Mees
Either the translator leaves the author in peace, as much as is possible, and moves the
reader towards him: or he leaves the reader in peace, as much as possible, and moves
the author towards him.
(Friedrich Schleiermacher)
In the 1980s one of the main discussions in the field of translation studies
centred around pinning down the concept of translation. I remember Arnt
being enchanted by Skopos theory and the function of the text, and how he
was absorbing the work of scholars such as Hans Hönig, Paul Kußmaul,
Hans Vermeer, Katharina Reiß, Christiane Nord and Justa Holz-Mänttäri.
Later on, he read Ernst-August Gutt‟s application of relevance theory to
translation with equal enthusiasm. He became ever more convinced that
translation should be seen as cross-cultural transfer and as the production
(or “creation” as he is wont to call it) of a text. In addition, the 1980s saw
the advent of cognitive empirical research. Ericsson and Simon‟s
introspective methods and Hans Krings‟s application of these awakened a
novel interest in what actually goes on during the process of translation. In
many ways, the OFT project united the issues Arnt himself found most
interesting, namely focus on complete texts, theoretical considerations, and
the integration of computer technology. Furthermore, the emphasis placed
on investigating the actual translation of LSP texts – as opposed to
focussing on terminology – was the main impetus for kindling his later
passion for the cognitive processes involved in translation.
All the above-mentioned translation scholars were trying to define
the concept of translation, for instance by debating the level of importance
of source as opposed to target texts. This discussion led to a paper entitled
“Translation as textual (re)production” (Jakobsen *1993) in which he
argued that it was fruitless to proclaim either source-text or target-text
hegemony. Translations are interlingual equivalents of existing texts with
which they have an intertextual relationship, but despite this obvious
dependence they nevertheless lead a life of their own. Thus translating
“always involves both textual production and textual reproduction” (1993:
74).
20 Inger M. Mees
By the time the OFT project had finally been wrapped up, Arnt had got a
taste for co-ordinating people, teaching and research, and took on
leadership responsibilities as head of the English department from 1996 to
1998. Even though he fulfilled this role excellently, I think he was
somewhat disappointed at having to deal with so many routine matters that
were totally unrelated to research. Rather than the excitement of initiating
new projects, choosing new staff, organising conferences and PhD courses,
and providing inspiration to colleagues and younger researchers, he found
himself caught up in a multitude of run-of-the mill administrative problems
thrown up by the everyday running of the department. He longed to lead a
new research project and consequently decided against putting his name
forward for a second term as departmental head. Nevertheless, he
continued to assume many administrative duties and sat on numerous
boards, committees, working groups and task forces. He was, for example,
a member of the IT committee, the Faculty Board and the Academic
Council, and was appointed Director of Studies for the Open University
(Faculty of Language, Communication and Culture). He also undertook a
number of consultancy tasks, such as chairman of a committee assessing
the level of English in Upper Secondary Education for the Danish
Evaluation Institute, and consultant on the translation and interpreting
degrees in Estonia and Lithuania for the Danish Ministry of Education.
22 Inger M. Mees
Translation processes
Whilst the OFT project was running, Arnt was still much preoccupied with
the concept of translation and attempting to find a definition of what it
implied. The next step was moving from the actual translation of texts to
defining and describing the mental activity underlying translation – in other
words, a study of the cognitive processes. The first major project at CBS
focussing on these mental processes was TRAP (“Translation process”),
headed by Professor Gyde Hansen. This was once again a project funded
by the National Research Council for the Humanities; it began in 1996 and
ran for a period of five years.
No truly empirical research into translation processes took place until
the mid-1980s, largely owing to a lack of appropriate methodology.
Introspective methods had been rejected in the behaviourist paradigm and
consequently translation researchers were unwilling to utilise them.
However, Ericsson and Simon‟s (1980) article “Verbal reports as data” and
their later (1984) book Protocol Analysis distinguished different types of
introspection such as “think aloud” (as concurrent verbalisation) and
retrospection (both immediate and delayed). They assessed the data that
can be elicited by these methods against the background of an information-
processing model. As a consequence, researchers – particularly those
working in the field of translation studies – felt empowered to use
introspective methods once again, the first being Krings (1986); Göpferich
(2008: 16ff) provides further detail.
Although “think aloud” was intended to capture what was held in
short-term memory, so that we could get closer to understanding what goes
on in the translator‟s mind, it was difficult to analyse the data obtained in
this way. As numerous scholars have pointed out, there are several
disadvantages to this method. Firstly, the cognitive load may be too high
(thinking aloud is an additional activity superimposed on the activity of
translating a text and may interfere with the actual translation processes).
Secondly, it may be difficult for translators to put into words what they are
actually doing so that by no means everything is verbalised. Finally, some
processes may have become automatised so that they will not be
accompanied by any verbalisation. There seems to be universal consensus
Arnt Lykke Jakobsen: portrait of an innovator 23
Translog
Arnt designed the program, but most of the actual programming was
carried out by his son, Lasse Schou. The first full Translog documentation
appeared a year later (Jakobsen and Schou *1999); for an account of the
development of the program in its many incarnations, and its uses see
Schou et al. (this volume). A description of how the data are recorded, and
how they can be represented and analysed, can be found in Jakobsen
(*1999). In a manner characteristic of his intellectual generosity, Arnt has
made the program freely available to the scientific community. Many of the
contributors to this volume (including myself) were among the earliest
users of the original DOS version, for example Fabio Alves, Susanne
Göpferich, Gyde Hansen, Brenda Malkiel, Ricardo Muñoz and Anthony
Pym. The Windows version together with the creation of a homepage
(http://Translog.dk) further expanded the employment of the technology
worldwide. Jakobsen (*2006) provides a summary of what had so far been
achieved with Translog. It discusses the development of the program and
how this resource can be used for translation experiments – either on its
own or together with other data elicitation methods. Although Translog was
24 Inger M. Mees
Arnt grew more and more fascinated by the ways in which computers could
assist translators. It is therefore not surprising that he produced a review
(Jakobsen *2003a) of a work which had greatly impressed him, namely,
26 Inger M. Mees
Expert performance
On the one hand, the centre‟s research looks outward at the transformational
processes in organisations and societies that are significant features of
globalisation, and, on the other hand, it looks inward at the mental and language
processes which constitute translational expertise.
As has been outlined above, since the launch of CRITT work has
concentrated on integrating eye-tracking research with Translog. The first
studies are already in existence and are collected in Göpferich et al.
(*2008d); for a brief description, see Göpferich et al. (*2008c). Jakobsen
and Jensen (*2008a) describes an experiment in which eye movements
were tracked while 12 translators (six professionals and six students) read
four almost identical texts with different purposes in mind: (1)
comprehension; (2) with the instruction of having to translate the text
afterwards; (3) in combination with oral translation; and (4) in combination
with written translation. Task times, fixation frequency, gaze times and
duration of fixations were all measured, indicating that there was a linear
progression in attention from the first task to the last. The results for Task 4
revealed that students directed more visual attention to the source text than
to the target text, while the reverse was true for the professionals.
Sharmin et al. (*2008b), a development of a poster session (Sharmin
et al. *2008e), also examines eye movement. Three different texts (one
easy and two difficult) were read by 18 translation students under three
different time constraints. When less time was available, the average
duration of fixations on the source text was reduced, whilst making the text
more complex had the effect of increasing the number of fixations – though
not their duration. In addition, it was found that average fixation duration
was consistently longer in the TT area than in the ST area.
The latest developments are directed towards establishing yet
stronger ties with computational linguists, and have led to ideas for new
research methods for the study of human translation processes. Eye-
tracking and keystroke logging will be employed to monitor subjects while
they read or translate a text. Using the data thus collected, it is hoped to
Arnt Lykke Jakobsen: portrait of an innovator 31
Say what we may of the inadequacy of translation, yet the work is and will always be
one of the weightiest and worthiest undertakings in the general concerns of the world.
(Johann Goethe)
he has had no hand in this particular volume (CSL 38), which is a labour of
love, compiled as a special tribute to him from his colleagues and many
friends at home and abroad.
References
Jakobsen, A. L. 1988. The earliest translations from English into Danish (c.
1640–1666). In G.D. Caie and H. Nørgaard (eds) A Literary Miscellany,
presented to Eric Jacobsen. Publications of the Department of English 16.
367–387.
Jakobsen, A. L. 1990. Review of NTC’s American Idioms Dictionary (ed. R.A.
Spears). Sprint 2: 74–79.
Jakobsen, A. L. 1991. Review of Jens Axelsen. Dansk–engelsk. Undervisning.
Gyldendals røde ordbøger. Copenhagen: Gyldendal. Hermes: Journal of
Linguistics 7: 117–121.
Jakobsen, A. L. 1992a. Oversættelse af komposita i fagsproglige tekster (Eng–
da/da–eng). Hermes: Journal of Linguistics 8: 29–42.
Jakobsen, A. L. 1992b. Oversættelse af fagsproglige tekster [Translation of LSP
texts]. In Jakobsen (ed.) Oversættelse af fagsproglige tekster. Indlæg fra
Sandbjerg-konferencen, 21.–22. november 1991 [Translation of LSP texts.
Papers at the Sandbjerg conference, 21–22 November 1991]. (ARK 65).
Copenhagen: Copenhagen Business School. 1–23.
Jakobsen, A. L. 1993. Review of Politikens Visuelle Ordbog. Sprint 1: 58–63.
Jakobsen, A. L. 2007. The reception of D. H. Lawrence in Denmark. In C.
Jansohn and D. Mehl (eds). The Reception of D. H. Lawrence in Europe.
London: Continuum. 255–264.
Jarvella, R. J., Bang, E., Jakobsen, A. L. & Mees, I. M. 2001. Of mouths and
men: non-native listeners‟ identification and evaluation of varieties of
English. International Journal of Applied Linguistics 11 (1): 37-56.
Kiraly, D. 2000: A Social Constructivist Approach to Translator Education:
Empowerment from Theory to Practice. Manchester (UK): Northampton
(MA): St. Jerome.
Krings, H. P. (1986) Was in den Köpfen von Übersetzern vorgeht. Eine
empirische Untersuchung zur Struktur der Übersetzungsprozesses bei
fortgeschrittenen Französischlernern. Tübingen: Narr.
Selected publications1
1
Only published work in the field of translation is included. Papers in Danish and
publications on topics other than translation have been omitted.
34 Inger M. Mees
1994d Starting from the (other) end: integrating translation with text
production. In C. Dollerup & A. Lindegaard (eds). Teaching
Translation and Interpreting 2: Insights, Aims, Visions. Amsterdam:
Benjamins. 143–150.
1998 Logging time delay in translation. In G. Hansen (ed.). LSP Texts and
the Process of Translation. (Copenhagen Working Papers in LSP 1)
Copenhagen: Copenhagen Business School. 73–101.
2007a (with K.T.H. Jensen and I.M. Mees) Comparing modalities: idioms
as a case in point. In F. Pöchhacker, A.L. Jakobsen and I.M. Mees.
Interpreting Studies and Beyond. A Tribute to Miriam Shlesinger.
(Copenhagen Studies in Language 35) Copenhagen :
Samfundslitteratur. 217–249.
36 Inger M. Mees
2007b (eds) (with F. Pöchhacker and I.M. Mees). Interpreting Studies and
Beyond. A Tribute to Miriam Shlesinger. (Copenhagen Studies in
Language 35) Copenhagen: Samfundslitteratur.
2008b (With S. Sharmin, O. Špakov & K.-J. Räihä) Where on the screen
do translation students look while translating, and for how long? In
S. Göpferich, A.L. Jakobsen & I.M. Mees (eds) 2008d. 31–51.
Translog data do not speak until we begin to theorise about them. [...] Only if we
succeed in developing ideas of how translation comes into being, can we begin to
make sense of Translog data. (Jakobsen 1999: 15)
Abstract
2009 is not only the year of Arnt’s 65th birthday; it is also the 10th
anniversary of one of his greatest accomplishments: the creation and
documentation of Translog. The basic idea of developing a tool for logging the
translator’s keystrokes was conceived by Arnt back in 1995, while the first
comprehensive description of its potential, together with a manual, appeared
ten years ago in an edited volume (Hansen 1999). This paper gives a review of
how Translog has evolved, where Translog is used and what the main areas of
research are. The programming and development of Translog has taken place
in close collaboration with Arnt’s eldest son, Lasse Schou. The first section is
his account of how it all started and how Translog has developed over the past
ten years. The second and third sections give an overview of where and how
Translog is used.
The Translog project was initiated in 1995 one morning when Arnt came to
me in the dining room and showed me a couple of handwritten notes. He was
very excited about this new idea he had of developing a computer program
that would revolutionise the world of translation research by recording every
keystroke in a text production session, and afterwards replaying the whole
thing on screen. Another important function, the linear representation of
events in the production process, was also already on the drawing board.
38 Lasse Schou, Barbara Dragsted & Michael Carl
Translog 1.0 had two main components: a component for recording sessions
(called Writelog) and a component for playing back and displaying the linear
representation (called Translog). Mouse movements were not part of the set-
up back then, but many of the basic features known from the latest Translog
version were actually present: displaying a source text with different timing
possibilities, dictionary look-ups, the use of different time values for the red
asterisk representing a pause, and replaying the log file at different speeds.
Timing, however, was a bit of a problem in version 1 because of the coarse-
grained resolution of the DOS clock that ticked away at a rate of 55
milliseconds. Although it may not sound all that bad, the playback was
noticeably affected when replaying fast typing.
The timing issue, among many others, was addressed in the next version,
Translog 2000, which was programmed in the Windows environment for the
first time, using Delphi as development tool. The user experience was
enhanced a great deal, the tasks were generally easier for the users to perform,
and the mouse was finally recognised as a user input tool worth logging. The
program now looked like a professional market product and was distributed at
a quicker pace than its predecessor because it was no longer necessary to
Ten years of Translog 39
throw a floppy disk in an envelope and send it via snail mail – you could just
email it around the globe. As for timing, the new component had, with its
overkill of 0.5 microsecond precision, an unfortunate impact on the processor,
which was jammed most of the time with 100 % usage, causing other running
software to protest fiercely.
My work started with a third complete rewrite of the source code. The latest
craze in the Microsoft development world was WinForms and C# 2.0, which I
had gained some experience in, and which was a natural choice of platform.
The new version was, imaginatively, called Translog 2006, and came with
many new features, including Unicode support (a feature widely requested in
the countries with non-Latin alphabets), XML support, multiple log file
support and richer source text/stimulus functions. But the hottest and most
eagerly anticipated new thing was of course the integration with eye-tracking
software and the ability to both record and play back eye movements and
words in focus through Translog.
During the development of the 2006 version, the Translog website
(www.translog.dk) was equipped with a rather advanced ordering and
activation system where the software could be freely downloaded for
academic purposes, but where users had to activate the software and thereby
40 Lasse Schou, Barbara Dragsted & Michael Carl
lock the license to their computer. This mechanism ensured that all versions
were downloaded from a central place with the latest patches instead of
spreading in the wild, allowing us to keep track of downloads around the
world; cf. below.
The Eye-to-IT project was successfully wrapped up in the spring of
2009, and current development is focusing on creating a new ready-to-ship
version where audio recording, eye tracking and keystroke logging are
integrated into one software bundle. So after 14 years of development the idea
is still alive and kicking, and the software has proved its value to researchers
all over the world.
Since the first version of Translog was released and described in 1999, the tool
has gradually found its way to research and teaching institutions throughout
the world. The list of Translog users ten years after its launch includes 97
universities and four private/individual users in 45 different countries in all
continents all over the world.
14
12
10
8
6
4
2
0
1999 2000 2001 2002 2003 2004 2005 2006 2007 2008 2009
Figure 1. Number of scientific publications using Translog as data acquisition
technology
Peaks in publications can be seen in 1999, 2002, 2003, and over the past five
years, there has been an (almost) steady increase. The early peaks are mainly
due to the two volumes edited by Gyde Hansen (1999 and 2002) in which
much of the early work on Translog is accumulated. Another volume, edited
by Fabio Alves, appeared in 2003 and also incorporates a number of
publications drawing on Translog data. In the wake of the Eye-to-IT project,
there was another boom of Translog- and eye-tracking-related publications in
2008 and 2009, with several edited volumes in the Copenhagen Studies in
Language series (Pöchhacker et al. 2007, Göpferich et al. 2008, Göpferich et
al. 2009).
42 Lasse Schou, Barbara Dragsted & Michael Carl
4. Baseline
Translog has had an immense impact on translation process research over the
past decade. What has been provided here is a snapshot of the development
and dissemination of the program and the publications drawing on Translog
data. The number of users is still increasing, emphasising the immense and
growing value of the tool. The program is continually being further developed,
most recently with the integration of eye-tracking analysis and new features
such as XML support. Work is being done to design more sophisticated
methods of analysis (Carl and Jakobsen 2009a; Carl and Jakobsen 2009b;
Carl, this volume) which will help us make better sense of the data we obtain
from Translog (and other sources), keeping in mind that “Translog data [...]
44 Lasse Schou, Barbara Dragsted & Michael Carl
References1
1 List of publications involving Translog data. The list may not be exhaustive, but this is
what we have been able to find. If relevant references have been left out, we apologise
for the omission.
Ten years of Translog 45
Zwischenbilanz und Perspektiven aus der Außensicht. Berlin: Frank & Timme.
95–114.
Göpferich, S. 2008. Empirische Methoden zur Bestimmung der Translatqualität im
funktionalen Paradigma. In I. Bartoszewicz, J. Szczek & A. Tworek (eds).
Linguistica et res cotidianae. (Linguistische Treffen in Wroclaw 2). Wroclaw,
Dresden: Atut-Neisse. 393–410.
Göpferich, S. 2009. Towards a model of translation competence and its acquisition:
the longitudinal study TransComp. In Göpferich et al. (eds). 11–37.
Göpferich, S., Jakobsen, A. L. and Mees, I. M. (eds.) 2008. Looking at Eyes. Eye-
Tracking Studies of Reading and Translation Processes. (Copenhagen Studies
in Language 36). Copenhagen: Samfundslitteratur.
Göpferich, S., Jakobsen, A. L. and Mees, I. M. (eds.) 2009. Behind the Mind.
Methods, Models and Results in Translation Process Research. (Copenhagen
Studies in Language 37). Copenhagen: Samfundslitteratur.
Hansen, G. 1999. Das kritische Bewußtsein beim Übersetzen: eine Analyse des
Übersetzungsprozesses mit Hilfe von Translog und Retrospektion. In Hansen
(ed). 43–68.
Hansen, G. (ed.) 1999. Probing the Process in Translation: Methods and Results.
(Copenhagen Studies in Language 24). Copenhagen: Samfundslitteratur.
Hansen, G. 2002. Interferenz bei Referenz im Übersetzungsprozess. In Van
Vaerenbergh (ed.). 303–326.
Hansen, G. 2006. Retrospection methods in translator training and translation
research. JoSTrans. 2–41.
Heiden, T. 2005. Blick in die Black Box: kreative Momente im
Übersetzungsprozess. Eine experimentelle Studie mit Translog. In H. Lee-
Jahnke (ed). Processus et cheminements en traduction et interprétation 50(2):
448–472.
Hjort-Pedersen, M. & Faber, D. 2010 (forthcoming). Explicitation and implicita-tion
in legal translation – a process study of trainee translators. Meta 55 (2).
Hurtado Albir, A. and Alves, F. (2009). Translation as a cognitive activity. In J.
Munday. (ed.). The Routledge Companion to Translation Studies. London:
Routledge. 210-234.
Immonen, S. 2006. Translation as a writing process: pauses in translation versus
monolingual text production. Target 18 (2): 313–336.
Jakobsen, A. L. and Schou, L. 1999. Translog documentation. In Hansen (ed.)
Appendix 1, 1–36.
Jakobsen, A. L. 1999. Logging target text production with Translog. In Hansen (ed).
9–20.
Jakobsen, A. L. 2002. Translation drafting by professional translators and by
translation students. Traducción & Comunicación v.3, 2002, 89–103. Also in
G. Hansen (ed.). 2002. Empirical Translation Studies. (Copenhagen Studies in
Language 27). Copenhagen: Samfundslitteratur. 191–204.
Ten years of Translog 47
Jakobsen, A.L. 2003. Effects of think aloud on translation speed, revision and
segmentation. In Alves (ed). 69–95.
Jakobsen, A. L. 2005. Instances of peak performance in translation. Lebende
Sprachen 50 (3): 111–116.
Jakobsen, A. L., Jensen, K. T. H., and Mees, I. M. 2007. Comparing modalities:
idioms as a case in point. In Pöchhacker et al. (eds). 217–249.
Jakobsen, A. L. and Jensen, K. T. H. 2008. Eye movement behaviour across four
reading tasks. In Göpferich et al. (eds). 103–124.
Jakobsen, A. L. 2009. Looking at the eyes of translators. Paper for the Ist
International Research Workshop on Methodology in Translation Process
Research. University of Graz, April 6–8, 2009, Graz, Austria.
Jarvella. R. J., Jensen, A., Jensen, E. H., Andersen, M. S. 2002. Towards
characterizing translator expertise, knowledge and know-how: some findings
using TAP and experimental methods. In A. Riccardi (ed). Translation Studies.
Perspectives on an Emerging Discipline. Cambridge: Cambridge University
Press. 172–197.
Jensen, A. 1999. Time pressure in translation. In Hansen (ed). 103–119.
Jensen, A. 2001. The Effects of Time on Cognitive Processes and Strategies in
Translation. (Copenhagen Working Papers in LSP 2 – 2001 (PhD thesis))
Jensen, K. T. H. & Pavlović N. (2009). Eye tracking translation directionality.
Proceedings of New Research in Translation and Interpreting Studies 2007,
Universitat Rovira i Virgili 18–20 October 2007, Tarragona, Spain.
Jensen, K. T. H. 2009. Shifts in source text and target text attention during
translation. Paper given at the IATIS conference, Monash University 8-10 July
2009, Melbourne, Australia.
Lee-Jahnke, H. 2007. Food for thought in translation training. In B. Lewandowska-
Tomaszczyk & M. Thelen (eds). Translation and Meaning 7: 261–282.
Livbjerg, I. and Mees, I. M. 1999. A study of the use of dictionaries in Danish-
English translation. In Hansen (ed). 135–149.
Lorenzo, M. P. 1999. Apuntes para una discusión sobre métodos de estudio del
proceso de traducción. In Hansen (ed). 21–42.
Lorenzo, M. P. 2002. ¿Es posible la traducción inversa? Resultados de un
experimento sobre traducción profesional a una lengua extranjera. In Hansen
(ed). 85–124.
Magalhães, C. and Alves, F. 2006. Investigando o papel do monitoramento
cognitivo-discursivo e da meta-reflexão na formação de tradutores. Cadernos
de Tradução 17 (2): 71-128.
Malkiel, B. 2006. The effect of translator training on interference and difficulty.
Target 18 (2): 337–366.
Malkiel, B. 2009. From Ántonia to My Ántonia: tracking self-corrections with
Translog. In Göpferich et al. (eds). 149–166.
48 Lasse Schou, Barbara Dragsted & Michael Carl
METHODOLOGICAL
ISSUES
What can translation research learn from Darwin?
Andrew Chesterman
Abstract
Charles Darwin can be a model for any empirical researcher. This essay
describes his attitudes to his work, the kind of background knowledge he
brought to his major project, his working method, and the way he presented
his ideas in writing. Particular attention is given to Darwin’s way of
arguing, his consideration of possible counter-evidence, and the dialogic
nature of his rhetorical strategy. I also note some problems in Darwin’s
marketing of his theory via translation, and mention some recent research
on how he was motivated by the moral cause of the anti-slavery movement.
1. Introduction
This year, 2009, is not just Arnt Lykke Jakobsen‟s anniversary year. It is
also 200 years since Darwin‟s birth, and 150 years since the publication of
The Origin of Species. It is perhaps a good moment to see what we in
Translation Studies might learn from Darwin‟s achievement. I am not
thinking here of the biological content of his theoretical claims, although
Translation Studies has already shown some interest in applying the notion
of the meme, as a cultural analogy of the gene, in explaining how
translation contributes to cultural evolution (see e.g. Chesterman 1997,
Vermeer 1997). What I wish to describe is Darwin‟s research methodology,
and his way of thinking and writing. I readily confess that my interest in
this subject has been directly inspired by my recent reading of The Origin
of Species. There is much here we can learn from, as translation scholars.
In presenting my response to Darwin‟s book, I also wish to honour
Arnt Lykke Jakobsen‟s achievements as a leading empirical scholar in
Translation Studies. In particular, I am thinking of his emphasis on
methodology in all his work: the way he encourages both qualitative and
quantitative methods, the use of triangulation, and the careful combination
52 Andrew Chesterman
of detailed analysis and a broader vision. Much of what I say here will be
familiar to him! In the conclusion to one of his first Translog articles, he
wrote:
[M]ulti-methodological approaches [are] a challenge we should take up. And an
obvious place to begin would be to try to identify converging observations in
qualitative and quantitative data about the same translation events. By
triangulating on the basis both of qualitative and quantitative data, by testing
hypotheses based on qualitative data against quantitative data, and vice versa, I
believe we can soon begin to make stronger and more informed guesses about
translation. (Jakobsen 1999: 19)
2. Attitudes
3. Background knowledge
Darwin was a polymath, and had read widely both within and outside his
own field. He drew inspiration from other fields, and was able to see
54 Andrew Chesterman
relations and connections that all fed into his Big Idea. I already mentioned
his early studies of geology. During his five years on the Beagle, he read
the very latest geological theories by Charles Lyell, one of the founders of
modern geology. This work had drawn his attention to the vast age of the
Earth, and the slow processes by which mountains and continents were
formed. Geology had already begun to produce empirical palaeontological
evidence for the idea of slow change. When Darwin came to look at the
question of biological evolution, he was thus already attuned to a very
different time-scale than most of his contemporaries. (A whole chapter of
The Origin of Species deals exclusively with geology.) Darwin had also
read Malthus, and was aware of the economic theories about the
consequences of the struggle for limited resources. He was of course
familiar with previous work on evolution in his own field. – As an
interdiscipline, translation research is increasingly benefiting from input
from outside traditional textual studies. But do we always have adequate
expertise in these other fields?
In one sense, his special gift was to refine and combine a variety of
ideas which were already “in the air”, as it were, in order to produce a new
whole. In other words, he was a great unifier, colligating a huge variety of
facts under a single uniting principle. Salmon (1998) in fact takes Darwin
as an example of what he calls explanation as unification, an alternative to
explanation in terms of causality. – In current Translation Studies there
seems to be some disagreement on the value of seeking grand abstract
principles which might unify a wide range of phenomena. On one hand,
scholars such as Hermans (most recently 2007) have argued for strongly
contextualized case studies; on the other, e.g. Pym (2008) has sought to
link Toury‟s two “laws” of growing standardization and interference under
the single more general principle of risk avoidance. Perhaps both
tendencies are needed; but presumably there should be some connection
between them eventually.
Darwin‟s familiarity with what was going on in other fields was
partly a result of his extensive correspondence. His network of international
contacts was extraordinarily wide, and he exploited this network most
effectively.
In counterbalance to this breadth of background knowledge we also
find a willingness to specialize. Darwin went into some topics in great
What can translation research learn from Darwin? 55
4. Method
The voyage of the Beagle has been by far the most important event in my life
and has determined my whole career [...]. I have always felt that I owe to the
voyage the first real training or education of my mind. I was led to attend closely
to several branches of natural history, and thus my powers of observation were
improved, though they were already fairly developed.
The investigation of the geology of all the places visited was far more
important, as reasoning here comes into play. On first examining a new district
nothing can appear more hopeless than the chaos of rocks; but by recording the
stratification and nature of the rocks and fossils at many points, always
reasoning and predicting what will be found elsewhere, light soon begins to
dawn on the district, and the structure of the whole becomes more or less
intelligible. [...]
During some part of the day I wrote in my Journal, and took much pains in
describing carefully and vividly all that I had seen; and this was good practice.
[...]
The above various special studies [e.g. reading Lyell] were, however, of no
importance compared with the habit of energetic industry and of concentrated
attention to whatever I was engaged in, which I then acquired. Everything about
which I thought or read was made to bear directly on what I had seen or was
likely to see; and this habit of mind was continued during the five years of the
voyage. I feel sure that it was this training which has enabled me to do whatever
I have done in science. (2002: 42–43)
Note: I suspect that some of the cases, and likewise some other facts... First
the caution, then the extension to other data, then the unifying claim.
Another aspect of Darwin‟s method is the attention he pays to
taxonomy. At many points in The Origin of Species he discusses the idea of
classes themselves, the problems of establishing the right categories, the
way classes form a hierarchy, and how this hierarchy can then be explained
What can translation research learn from Darwin? 57
I.e. first state the clear facts, then the possible hypothesis, then the crucial
condition which would test the hypothesis, then the implication, if the
possibility becomes a probability of increasing strength. Testing the
hypotheses is of course vital. Darwin himself carried out, and reported in
detail, a large number of often laborious tests to test his points. He wants,
say, to check an idea about the instincts of bees. “Following the example of
Mr Tegetmeier, I separated two combs, and put between them a long, thick,
square strip of wax: the bees instantly began to excavate minute circular
pits in it [...]” (1968: 250).
Darwin‟s caution, and his honesty, are also seen when he openly
states that a given explanation is only a limited one. “It must, however, be
confessed that we cannot understand, excepting on vague hypotheses,
several facts with respect to the sterility of hybrids” (1968: 280).
One of the most striking aspects is the way in which Darwin is so
careful to specify the kind of evidence that would count against his
argument. Sometimes it is as if he is really inviting critics to shoot his
58 Andrew Chesterman
And again:
That many and grave objections may be advanced against the theory of descent
with modification through natural selection, I do not deny. I have endeavoured
to give them their full force. (1968: 435)
I had, also, during many years, followed a golden rule, namely, that whenever a
published fact, a new observation or thought came across me, which was
opposed to my general results, to make a memorandum of it without fail and at
once; for I had found by experience that such facts and thoughts were far more
apt to escape from the memory than favourable ones. (2002: 75)
– Note: anything that opposed his theory, not anything that supported it!
One of the reasons for his admiration of the geologist Lyell was that
whenever Darwin made some remark to him, “[h]e would advance all
possible objections to my suggestion, and even after these were exhausted
would long remain dubious” (2002: 58). Another colleague is criticized for
“his excessive fear of ever making a mistake” (2002: 60). Darwin writes: “I
have steadily endeavoured to keep my mind free, so as to give up any
hypothesis, however much beloved (and I cannot help forming one on
every subject), as soon as facts are shown to be opposed to it” (2002: 86).
(Do we in Translation Studies spend too little time exposing our hypotheses
and arguments to counter-evidence?)
Darwin has a Popperian attitude towards errors: they must be
acknowledged, and are there to be learned from. In Notebook M (one of his
vital research diaries, this one was for thoughts on “Metaphysics with
Morals & Speculations on Expression”) he writes in 1838: “Probably some
error in the argument here. Should be grateful if pointed out” (cited in
Padel 2009: 52). At one point in his recollections, he recalls how one of his
geological papers had been “a great failure” because he had falsely
assumed that no other explanation was possible for a particular
phenomenon than the one he proposed, which must therefore be correct.
Later research had brought a better explanation, however. Darwin notes:
“my error has been a good lesson to me never to trust in science to the
principle of exclusion” (2002: 48). He is aware of the gaps in his own
research skills, too: he knows he is bad at drawing and anatomical
dissection.
There is likewise humility in Darwin‟s attitude of respectful
disagreement towards geologists who supported the idea of immutable
species. “I feel how rash it is to differ from these great authorities, to
whom, with others, we owe all our knowledge” (1968: 316). But...
I look at the natural geological record, as a history of the world imperfectly kept,
and written in a changing dialect; of this history we possess the last volume
alone, relating only to two or three countries. Of this volume, only here and there
60 Andrew Chesterman
a short chapter has been preserved; and of each page, only here and there a few
lines. (1968: 316)
5. Rhetoric
Let us now pick out some other salient features of Darwin‟s way of
presenting his ideas, in terms of communicating with his readers. Some of
these may be related to the fact that The Origin of Species was a much
shortened version of Darwin‟s original manuscript, and was prepared rather
hastily for publication after he had read Wallace‟s paper proposing a similar
theory. So Darwin needed to get his ideas out in the public arena quickly, or
risk being overtaken.
He writes for the general intelligent layman, not exclusively for an
elite specialist readership. There are few technical terms, and they are
carefully explained; and there is almost no formalization. The Big Idea
comes early, but is gently introduced by a chapter on the kind of normal
breeding and domestication carried out by farmers, pigeon-fanciers or
florists, as they seek to maximize certain characteristics of an organism.
This will be familiar stuff to his readers. The theme then shifts from
breeding by human agents to breeding by natural selection; and away we
go.
Part of his readability stems from his use of analogies and similes –
like the analogy between domestic breeding and natural selection. Breeds
themselves are compared to dialects (1968: 97): neither can be said to have
62 Andrew Chesterman
He is not only arguing for a given position, then, but also against an
opposing one. The rhetorical strategy throughout the text is that of a
dialogue or debate, with opponents sometimes explicitly present,
sometimes only implicit. (Are we sometimes too monologic, as if talking to
ourselves?)
Darwin gets purely conceptual problems out of the way early. Many
of the key concepts he uses have no adequate definition, he admits.
Naturalists disagree about how to define „species‟ or „subspecies‟ or
„variety‟, and it has not proved possible to state where the borderline
between them might run. Yet “every naturalist knows vaguely what he
means when he speaks of a species” (1968: 101). This is enough. Darwin is
quite happy to work with poorly defined concepts. (After all, one might
say, most definitions too evolve, they are rarely born perfect!) But Darwin
is aware of the crucial criteria that gradually refine definitions: it is a
question of the degree of variation within a given category (1968: 106–
108). Deciding that a given degree of variation marks the range within a
What can translation research learn from Darwin? 63
6. A note on marketing
It‟s not enough to have good ideas. You also have to sell them, to market
them. In Darwin‟s case, some of the marketing was easy. The first and
second editions of The Origin of Species sold out rapidly. It is of interest to
translation scholars to note, however, that the business of organizing
translations into other major languages was sometimes problematic. On
Darwin‟s initiative, the first translation appeared a year later (1860) in
German. But there were many problems with the first French translation,
discussed in a fascinating article by Brisset (2002).
After problems finding a translator, the work had been entrusted to
Clémence Royer, a largely self-educated and supremely self-confident
intellectual lady who agreed so strongly with Darwin‟s views that she
produced a French version that was strongly coloured by her own
supportive position, with abundant footnotes explaining the original,
defining terms, and with critical notes on Darwin‟s argument, including
references to her own research. Of particular interest is the observation (by
Miles 1989, cited by Brisset) that many stylistic changes were evidently
motivated by the translator‟s wish to domesticate Darwin‟s mode of
argument in accordance with the French tradition of scientific writing,
which at that time was more positivist, more inductive, than the British one.
She therefore transformed many of Darwin‟s hypothetical conjectures into
firm facts and thus changed the form of some of his arguments. Moreover,
she also extended Darwin‟s argument into the social sphere – something
that Darwin had carefully avoided in this book. After his first surprise at
the translator‟s audacity, Darwin was horrified by the translation, and the
following years saw an increasingly frustrating correspondence between
him and the French publisher. That “verdammte Mlle Royer whose errors
are endless” (letter to Henrietta Darwin, June 1865, cited in Brisset 2002:
195) cast a long shadow over the years of Darwin‟s illness as he struggled
with revisions of later editions. – Moral: it is good that important works are
translated, but woe betide the scholar if the translator tries to take over the
responsibility of the original author! In this case, the translator‟s
manipulation was not welcome. Evidently, there should have been much
closer control during the translating process, as in the case of the German
What can translation research learn from Darwin? 65
translator, who had been in regular contact with Darwin. Later editions had
a different French translator and publisher.
7. A note on motivation
References
Jakobsen, A.L. 1999. Logging target text production with Translog. In G. Hansen
(ed.). Probing the Process in Translation: Methods and Results.
(Copenhagen Studies in Language 24) Copenhagen: Samfundslitteratur. 9–
20.
Miles, S.J. 1989. Clémence Royau et De l’origine des espèces: traductrice ou
traîtresse? Revue de Synthèse 4: 61-83.
Padel, Ruth 2009. Darwin. A Life in Poems. London: Chatto & Windus.
Pym, A. 2008. On Toury‟s laws of how translators translate. In A. Pym, M.
Shlesinger and D. Simeoni (eds). Beyond Descriptive Translation Studies.
Amsterdam: Benjamins. 311–328.
Salmon, W.C. 1998. Causality and Explanation. New York: Oxford University
Press.
Vermeer, H.J. 1997. Translation and the „Meme‟. Target 9(1): 155–166.
Eye to IS: on qualitative research in interpreting studies
Franz Pöchhacker
Abstract
Prolog
The debate over valid scientific methods and findings in interpreting re-
search can be traced back at least to the early 1970s, when Henri Barik
disseminated his work in translation journals such as Babel and Meta, and
representatives of the interpreting profession, with some reason, showed
little appreciation for the findings from his experiments. Ecological valid-
ity, though not referred to as such at the time, was the bone of contention,
and the field came to split into controlled experimenters, on the one hand,
and champions of studying authentic data from real-life professional
practice, on the other. “Gerver vs. Seleskovitch” could be shorthand for the
methodological divide, though the two, appearing side by side in Richard
Brislin’s (1976) anthology, could easily be shown to share a profound
interest in the cognitive process of interpreting in either mode.
In the late 1980s, when the budding field entered the “Trieste era”,
the enduring distinction between experimental and authentic data was
overlaid by a more fundamental division, that is, “scientific research vs.
personal theories”, as coined by Daniel Gile (1990). Gile’s charge against
On qualitative research in IS 69
1.3 Paradigms
The controversy over the appropriate way(s) of dealing with empirical data
in research on conference interpreting found its clearest expression in
Barbara Moser-Mercer’s (1994) essay on the division of the field into a
“liberal arts community”, on the one hand, and a “natural science para-
digm”, on the other. What is interesting here about this much-discussed
manifesto for the scientific method in interpreting research are the criteria
by which the two camps are distinguished, namely, logical precision and
quantification. The liberal arts community is said to be less logically
rigorous, and the natural science camp bent on quantification. While
“general theorizing” is mentioned only in connection with the liberal arts
group, either would presumably be seen as engaged in theory-building and
involve a certain amount, or type, of logic.
This is where one might expect the division to be explained with
reference to such notions as deduction vs. induction, or abductive reasoning
as a “less logical” alternative, but this is not the case. Indeed, the
fundamental issue that I would like to highlight in this debate remains
unaddressed. A small pointer to it may be identified in Gile’s paper when
he refers to the fundamental role of “facts”, which come in two forms
(1990: 29): “facts collected through systematic observation” and “facts en-
countered in the daily, personal and subjective experience of the inter-
preter”. Though I realize I am stretching this quote beyond the author’s
On qualitative research in IS 71
original intention, its perfect fit to introduce the core issues of ontology and
epistemology seems too hard to resist: facts, reality, reliable knowledge –
do they “exist”? Are they accessible? And if so, how? Oppositions such as
“objective facts” vs. subjective understanding; (scientific) knowledge vs.
beliefs; and systematic theory vs. individual experience can be seen as
reflecting different worldviews on how to gain knowledge by “doing
science”. And yet, the term “epistemology” is rarely encountered in the
interpreting literature. The debate over paradigms recalled above does
address key issues in the philosophy of science, but it hardly engages
directly with the nature of epistemological positions. The following
paragraphs will therefore introduce some of the themes relevant to the
subsequent discussion of qualitative research in IS.
2. Notes on epistemology
In the human quest for knowledge, the central status of ideas going back to
Platonic-Aristotelian philosophy is beyond doubt. And so is, at least for
several centuries now, the value of engaging with empirically accessible
phenomena. Pure rationalism, uncoupled from sensory experience, is of
limited value in gaining knowledge about the world, and equally futile
would be the attempt to limit what can be known to what can be captured in
measurements. The latter positivist stance is nevertheless what has tradi-
tionally been associated with the natural sciences, which are increasingly
dominant also in the study of human beings and their pursuits.
The broadly antithetical positions of rationalism and positivism, or
empiricism, have major implications for the way a researcher would treat
data, if at all. Most importantly, this concerns the relationship between the
observer and the object of study: Is there a reality “out there”, independent
of the person(s) wishing to study it? Or is it only an appearance or re-
72 Franz Pöchhacker
4. Qualitative research in IS
4.3 Balance?
This rather even balance over several decades does not support the
idea of a trend toward more qualitative research; however, the anthology
covers the conference-interpreting classics rather more fully than
community-based domains, the literature on which began to grow strongly
only in the late 1990s.
For the purpose of capturing the development of IS literature from
the early 1990s up to the present, the Benjamins Translation Library (BTL)
as the most important book series in the field can be regarded as a
comprehensive and relevant data set. As of mid-2009, a total of 85 titles
were listed on the publisher’s website (Benjamins 2009). Out of these, 21
can be classified as books devoted mainly to interpreting, from Vol. 3 on
Empirical Research in Simultaneous Interpretation (Lambert & Moser-
Mercer 1994) to Vol. 84, by Claudia Monacelli, on Self-preservation in
Simultaneous Interpreting.
Taking a look at the very first and at the last volume currently
available to me – Vol. 83 by Kumiko Torikai (2009), the contrast could
hardly be more striking: the former a collection of 17 papers on “empirical
research” of which not a single one is based on qualitative data; the latter a
monograph based on five life-story interviews with interpreters, and as far
removed from quantification and numerical significance as one could
imagine.
But picking the first and last clearly does not amount to a represen-
tative sample, so a more systematic account of the 21 interpreting volumes
in the BTL series seems desirable. As can be seen in the above examples,
some of these books are collections of papers by different contributors
whereas others, ten to be precise, were written by single authors. The
eleven collective volumes, in turn, are either based on conference
proceedings (e.g. of the four Critical Link conferences) or, in four cases,
compiled with a focus on a particular theme (empirical research, signed-
language interpreting), purpose (research training), or individual (Daniel
Gile).
Whereas examining the individual contributions to the eleven
collective volumes is beyond the scope of this paper, it is feasible here to
classify the ten single-author monographs with regard to their empirical
component. The result is relatively even: Two of the books were not
designed as empirical studies, and the remaining eight include three quali-
On qualitative research in IS 81
6%
23%
29 6 65
qual.
quant.
both
fieldwork survey experiment
71%
33% 30%
qual. 30 26 44
quant.
both
fieldwork survey experiment
37%
With regard to the question driving this analysis, the most noteworthy
finding to be gleaned from the relationships depicted in Figure 1 may well
be the more than five-fold increase in the percentage of studies reporting
qualitative as well as quantitative data. If these are combined with the
number of purely qualitative studies, the share of research papers pre-
senting qualitative data has in fact more than doubled when comparing the
earlier period (late 1990s) to the more recent volumes of the journal (2004–
2009). In short, the shift reflected in this analysis is not so much an
increase in qualitative research at the expense of studies relying on quanti-
fication, but a trend toward empirical studies combining these two types of
data. (Ideally, the present analysis should do the same – and there are
On qualitative research in IS 83
5. IS as interpretive studies?
Judging from the examples and quantitative data presented above, it seems
safe to argue that interpreting research has come a long way since its initial
fixation on experimenting in the psycho-statistical paradigm. Even then, in
the 1970s, an alternative approach consisting in the use of naturalistic data
in close proximity to the research participants had been advocated, though
the Paris School’s “interpretive approach” to the analysis of qualitative data
was not yet founded on an interpretive (constructivist) epistemology nor on
the principles of what has come to be known as the paradigm of qualitative
research.
Empirical research relying on qualitative data such as interviews,
field notes, retrospective reports and interpreters’ renditions as such has
been quite visible in IS in recent decades, and by some indicators may be
said to have increased in relation to quantification-based work. Most signi-
ficantly, though, there appears to have been a growth in the number of
studies that combine quantitative and qualitative data and methods. Such
multi-method designs, typically involving the so-called triangulation of
data from different sources, are often seen as a typical feature of qualitative
research, but the correspondence need not be so straightforward, and little
may be gained by expecting researchers to pledge allegiance to either the
quantitative or the qualitative paradigm.
As a complex cognitive as well as social process, and a multifaceted
cultural phenomenon, interpreting is open to and in need of research across
a broad range of methods, and stands to gain from diversity rather than
uniformity of methodological approach. Where consensus does seem
desirable, if not even required, in order to cope with multiple methodo-
logies, is the field’s epistemological underpinning. And what could be
more appropriate to the study of interpreting – which is, first and foremost,
understanding – than an interpretive, constructivist theory of knowledge?
In this sense of IS as “interpretive studies”, understanding how others could
understand something differently would not only be the hallmark of an ex-
84 Franz Pöchhacker
References
Abstract
1. Introduction
1996; Kußmaul 1998; Orozco 1999; Lauffer 2002; Neunzig 2002; Williams
& Chesterman 2002; Alves 2003b; Hansen 2003, 2006; Lörscher 2005;
O’Brien 2005; PACTE 2005; Campbell & Wakim 2007; Bayer-
Hohenwarter 2008; Göpferich 2008; Muñoz 2009). However, some of the
problems faced by our forerunners in the eighties and the nineties remain
unsolved.
Some TPR problems seem associated with the need for a consistent
theoretical framework to interpret data1 (cf. Risku 2000; Rydning 2001;
Muñoz 2006), others are related to the nature of our research designs.
There is still enormous variation between TPR projects in terms of research
foci, research protocols, ecological validity, subjects’ categorization, source
texts and data interpretation (cf. Fraser 1996; Rodrigues 2002). While it is
true that some of these factors may be irrelevant for certain research goals,
others are pervasive and threaten to maintain current endeavors as a
heterogeneous set of juxtaposed attempts which cannot be interconnected,
nor built upon. In short, progress in methodology has not lived up to
improvements in data collection procedures.
Following Göpferich’s (2008: 9–16) classification criteria, we might
describe most current TPR efforts as qualitative endeavors focusing
(mainly) on online processes by means of methodologies inspired by
psycholinguistics and collecting data at one point in time only, although
there are some longitudinal studies underway (e.g. Hansen, on sources of
interference, TransComp and PACTE, on the acquisition of translation
competence). A different, popular classification distinguishes between pre-
experimental, quasi-experimental, and true experimental designs (Campbell
& Stanley 1963; Cook & Campbell 1979). Pre-experimental designs adhere
to the scientific method, but they do not make use of control groups. Quasi-
experimental designs employ diverse means of comparing groups, but they
fail to randomize the choice of subjects or their assignment to control and
study groups. True experimental designs satisfy both conditions. Thus, the
main difference between these categories relates to the way subjects are
dealt with.
1
In an intelligent move which temporarily sets aside the lack of an uncontested
framework, some research teams, such as LETRA and TransComp, have developed
storage and codification systems which will make data available for new analyses in
the future.
Subject profiling in translation process research 89
2
Gould (1981: 24–25) argued that psychometrics is a form of scientific racism since
“the abstraction of intelligence as a single entity, its location within the brain, its
quantification as one number for each individual, and the use of these numbers to
rank people in a single series of worthiness, invariably find that oppressed and
disadvantaged groups – races, classes, or sexes – are innately inferior and deserve their
status.” The second edition of his book, The Mismeasure of Man (1996), was enlarged to
include an explicit rejection of The Bell Curve (Herrnstein & Murray 1994), which
argued that intelligence is a better predictor than educational level and parents’
socioeconomic status of factors such as financial income, job performance, and crime.
Subject profiling in translation process research 91
Two composite intelligence tests stand out from the rest as far as
reliability and popularity are concerned: the Stanford-Binet Intelligence
Scale and the Wechsler Adult Intelligence Scale (WAIS). Stanford-Binet
measures working memory, visual-spatial processing, knowledge,
quantitative reasoning, and fluid reasoning in subjects between 2 and 23
years old, and it is better than WAIS at ferreting out extremes (geniuses,
mental retardation). WAIS measures working memory, processing speed,
verbal comprehension, and perceptual organization in adults between 16
and 89 years old, and its reliability is outstanding (Kaufman &
Lichtenberger 1999). Since many subjects in TPR are older than 23, WAIS
seems better suited for subject profiling.
The Perceptual Organization Index (POI) in the WAIS test measures
fluid reasoning (matrix reasoning subtest), spatial processing, attentiveness
to detail and visual-motor integration. It is primarily a measure of perceptual
(visual) processing, and it is not clear whether it is important for our
purposes, perhaps except when eye tracking is used. Since POI subtests
have been thoroughly revised in the WAIS4,3 and since indexing subjects
according to mental abilities should not be confused with IQ testing, it is
perhaps best to omit this section (see Section 6). Several research projects,
such as the TOP Ullevål 600 study of the University of Oslo, make use of a
limited number of the WAIS3 subtests only to measure specific abilities.
Let us have a glimpse at working memory (WM) and processing speed (PS).
3
The WAIS test has been subjected to several revisions. The latest version, WAIS4,
was released in 2008. The version discussed here is WAIS3 (1997), for which
standardized versions are already available for many major languages, including
Chinese (cf. Yao et al. 2007), French (EIWA-3, Wechsler 2000), German (HAWIE-3,
cf. Von Aster et al. 2006), Portuguese (cf. do Nascimento & de Figueiredo 2002), and
Spanish (EIWA-3, 1999; cf. García et al. 2003).
92 Ricardo Muñoz Martín
study of more than 3,000 schoolchildren (ages 5–11), Alloway et al. (2009)
found that roughly 10 % had low WM scores which might explain
problems such as deficient monitoring of the quality of their work, and
difficulties in generating new solutions to problems.
WM seems to be able to predict some translation-relevant skills, such
as L1 fluency (Daneman 1991) and recall (Braga Tomitch 1999). WM also
interacts with foreign language skills (Van den Noort et al. 2006). In L2,
for example, WM seems to correlate positively4 with L2 fluency, accuracy,
and complexity (Mota 2003). WM also seems to be a good predictor of
success in mathematical problem solving (Holmes & Adams 2006; De
Smedt et al. 2009) and insight problem-solving, i.e., the phenomenon of
suddenly solving an apparently difficult problem (Murray & Byrne 2005),
and in reasoning (Süß et al. 2002). Furthermore, WM may also be used to
predict multitasking performance (Konig et al. 2005), where it can predict
errors of commission, i.e., errors related to addressing a task demand
incorrectly, as opposed to errors of omission, i.e., allowing a task demand
to lapse (Oberlander et al. 2007). The fragmentary way of reading while
translating might also be influenced by WM, for high-span readers’
performance on hypertexts seems to equal that of low-span readers on
linear texts (Fontanini 2005).
Bajo et al. (2001), Rothe-Neves (2003), Liu et al. (2004), Dragsted
(2005), Macizo & Bajo (2005), Mizuno (2005), Padilla et al. (2007: 59–79)
and Timarová (2008), among others, have already underscored the
importance of WM in translation and interpreting processes. Since, in
general, “working memory may even be the primary determinant of
proficiency in cognitive domains, at least when the influence of prior
knowledge and experience is minimal” (Hambrick & Engle 2003: 190), the
above-mentioned factors suggest that WM scores might have a strong
explanatory power in TPR as well, where they might become a predictor of
translation problem-solving success.
4
Fortunately, like other cognitive abilities, WM develops over time and can also be
improved under certain circumstances. For example, Klein & Boals (2001) suggest
that expressive writing can increase WM capacity, and Ransdell et al. (2001) show
that fluency in another language may confer long-term working memory benefits.
Subject profiling in translation process research 93
Processing speed (PS) in the WAIS3 test refers to the speed at which
cognitive processes can be carried out, and it relates to the ability to
automatically perform easy or overlearned cognitive tasks. PS does not
seem related to successful problem solving (Welsh et al. 1999), although it
does correlate with errors caused by omissions (Oberlander et al. 2007).
Nevertheless, PS is interesting for other reasons. For instance, it seems to
be related to WM efficiency. A higher PS might reduce demands on WM,
thereby fostering reasoning and learning (Zhu & Weiss 2005: 315); a lower
PS, on the other hand, may degrade cognitive performance because the
result of previous processing may have faded away when needed in
subsequent mental operations (Salthouse 1996). So, PS seems to have an
indirect effect on the way the mind works, and not only on its pace, as
experimental results indicate for aspects such as verbal ability (Wiedel &
Schwartz 1982), creativity (Rindermann & Neubauer 2004), abstract
reasoning and, indeed, problem solving in tasks where subjects do not need
to retrieve information from their long-term memories (“fluid intelligence;”
Sheppard & Vernon 2008).
From a developmental point of view, PS becomes faster until ages 25
to 30, and then starts a slow, steady decline (McGrew & Woodcock 2001).
Expertise may reduce this decline (Nunes & Kramer 2009), so that abilities
of older high-level experts may exceed those of younger persons with
lower levels of expertise (Horn & McArdle 2007). Since expertise levels
are relevant for many TPR projects, and Horn & McArdle (2007: 241)
found that expertise speed and cognitive speed were collinear, PS might
become an important touchstone against which to contrast translation
expertise levels.
There are other reasons why PS could be important for TPR studies.
Some TPR projects reach conclusions said to be related to translation speed
(e.g. automation, expertise level) but which have been arrived at through
typing speed. However, the translation behavior of (mainly trainee)
subjects in our studies is often hindered by their lack of typing skills (that
94 Ricardo Muñoz Martín
is, by slower movement times5), which does not necessarily correlate with
lower PS scores but, so far, we have no way to account for this difference.
Indexing subjects as to their mental PS might do the job, and thus this
might have been an interesting supplement to studies where timing was
important, such as De Rooze (2003), Rothe-Neves (2003), Dimitrova
(2005), Dragsted (2005), Jakobsen (2005), and Sharmin et al. (2008).
Furthermore, in TPR we use the terms planning and reviewing to
broadly refer to macrotextual phases of text production, as usual in writing
research (Jakobsen 2002). In psychology, Hayes & Flower (1986) also
identified planning, actual text generation, and reviewing as the three key
processes in L1 writing, but at a microtextual level. Levy & Ransdell
(1995) found that the median time for text generation was 7.5 seconds,
whereas planning, and reviewing/revising took about 2.5 seconds each (that
is, activities alternated at these rates). Figures should be different when
translating, due to the interplay with the original text, but one of the
characteristics associated to better performance when translating is routine
interiorization, i.e. procedural automation as evidenced by uneventful, fast
renderings. Differences between monolingual writing and the automated
translation of segments might yield information as to the specifics of
translating. Hence, this microtextual approach might also be very
informative in segmentation and pause analyses. Under this scope, PS
might turn out to be a reliable index to triangulate microtextual translation
planning, generating, and reviewing/revising phases in subjects as well.
5
Ericsson (2003: 57–58) summarizes research findings which point to reading in
advance and preparing movements ahead as the main means of distinguishing expert
from regular typists. Both characteristics may improve with higher PS.
Subject profiling in translation process research 95
6
Verbal comprehension subtests should not be used to test foreign language skills,
since psychologists do not consider it appropriate to administer the test in languages
other than the subjects’ own. See, however, results for subject X below.
7
This is the case for Cambridge exams for English, CIEP/Alliance Française for
French, Goethe Institut/TELC for German, and DELE for Spanish.
8
The Educational Testing Service (ETS), which administers the TOEFL, states on its
web page that more than 6000 colleges, universities, and licensing agencies in 136
countries use it as indicator of the English skills of their applicants. More than 20
million test takers (cumulative figure) let ETS publish statistics summarizing results
by study level, gender, native tongue, and country.
96 Ricardo Muñoz Martín
4. Case study
9
Anyway, Dunkel & Davies (1994: 65) have shown that TOEFL scores do not capture
the real abilities in oral comprehension in ESL University students.
Subject profiling in translation process research 97
realistic. Translations were performed against the clock: subjects had to try
to complete them in one hour, although they were told in advance they
should finish their translations even if they took longer. Subjects’ times to
complete the four texts were averaged and are displayed in Table 1 under
translation time (TT).
Translations were revised by a translation teacher, a professional
translator with more than 10 years of in-house experience in a translation
and localization company, and a translation graduate/PhD student who had
started to work as a freelance translator. Quality evaluation followed
guidelines summarized in Muñoz & Conde (2007). Evaluators were
requested to perform the task to the best of their knowledge and without
any specific instructions except for the requirement to later classify each
translation into one of four categories: very bad, bad, good, and very good.
These categories were then converted into numerical scores and results for
the four texts were averaged, converted into a 100 scale, and rounded. In
Table 1, they are shown under average grade (AG).
In order to triangulate data, reformulations were defined as “changes
in the copy which include either syntactic modifications or a different lexical
choice for at least one item,” and repetitions were defined as “deletions
followed by the typing of exactly the same string,” so that they would also
comprise those of just one character, including punctuation marks, but not
blank spaces. These parameters were chosen because of their objective
nature and their potential relationship with problem solving in translation.
Two evaluators independently computed reformulations and repetitions in
the texts. No significant differences were found, and results were averaged.
Two of the subjects had technical problems and did not manage to
record all translations, and a third one was ill for a short period of time and
was therefore not able to translate all four texts. Subject X was a foreign
language native speaker, so her data were not computed, although they are
shown in Table 1 below to illustrate the contrast with native speakers.
The results for 18 students are summarized in Table 1. Scores for the
TOEFL and grades have been converted to a 100 point scale, to ease the
analysis. WAIS3 scores are already averaged on a 100 point scale from the
98 Ricardo Muñoz Martín
start, with an SD of 15, so scores below 85 and above 115 may safely be
considered meaningful.
Table 1. Subjects’ ages (Age) and scores for working memory (WM), processing
speed (PS), verbal comprehension (VC), TOEFL sections 2 and 3 (T2 and T3),
average translation time (TT), and average grade (AG); subjects in each group
arranged according to their WM scores
Subj A B C D E F G I H J
Age 21 21 21 34 21 21 23 22 22 21
WM 117 118 122 124 125 127 129 132 134 139
PS 123 114 125 132 123 93 132 134 140 117
VC 111 123 110 108 114 103 100 121 106 110
T2 92 88 57 85 83 72 82 87 63 82
T3 95 93 85 87 90 81 88 93 90 100
TT 48:43 48:21 40:28 48:03 51:42 48:42 51:41 45:01 43:55 48:57
AG 95 95 75 60 80 90 85 85 70 80
Subj K L O M N P Q X
Age 23 23 21 21 22 21 23 26
WM 81 83 100 100 102 107 108 129
PS 98 117 112 117 125 118 106 101
VC 102 110 104 110 102 119 104 98
T2 78 83 83 85 75 92 83 85
T3 50 92 90 87 93 95 88 88
TT 58:30 53:47 44:32 46:18 46:07 45:47 49:23 50:22
AG 75 75 80 70 75 85 50 60
From the point of view of screening out informants (our first goal), subjects
K and L showed significant low scores for WM. When all subjects are
considered, this amounts to roughly 10 %, the same rate Alloway et al.
(2009) found in children. These subjects probably never had special
training to improve their WM, so they might run into some of the
aforementioned troubles found in children as well. Although the average
grade for their translations did not seem particularly affected, they were the
ones who took longest to perform the tasks, owing to both more frequent
and longer pauses. Hence, when behavior is to be observed and time is a
relevant factor, subjects with WM scores below 85, such as K and L,
Subject profiling in translation process research 99
should probably be omitted from the study,10 for their behavior might be
affected by this special circumstance, as evidenced by the low English
verbal comprehension scores (T3) in subject K – whose scores were low
for both WM and PS – though higher than those of subjects C, F, H, and N
in formal English command (T2).
Subject X took the WAIS3 subtest battery in Spanish as well, where
she had standard scores in PS, high scores in WM, and slightly low scores
in VC. Her scores for T2 were moderately high (sample av. 80.5) and
average in T3 (sample av. 88). She also took long to translate the texts, and
her results were poor, probably due to her lacking Spanish skills.
Table 2. Average scores for the whole sample (17) and groups with low, standard
and high WM capacities
10
Subjects F, J, L, N and X showed a difference higher than 20 points between WM and
PS scores. Dehn (2006: 107) considers this gap both significant and infrequent but
Binder et al. (2009) think of it as normal variations within healthy adults.
100 Ricardo Muñoz Martín
5. Conclusion
6. Post Scriptum
Further work seems necessary to ease both testing and the interpretation of
results, and also to reduce the time they take. The standard administration
of the full WAIS3 takes more than 90 minutes (Ryan et al. 1998), which is
very long.11 WAIS3 tests are designed to be administered orally and
individually, so its application seems clearly impractical. As for the TOEFL
test, the administration of the reading comprehension section takes about
one hour and the section on structure and written expression, about half an
hour. When added to the time WAIS3 subtests take, it means that subjects
11
This has prompted the design of many short forms, either by leaving out the easiest
and most difficult items in each subtest or by reducing the number of subtests.
However, short forms of the WAIS3 are, simply, less reliable (Jeyakumar et al.
2004).
102 Ricardo Muñoz Martín
will have to spend three additional hours with the researchers. This is a
grave disadvantage. Furthermore, administering and scoring a battery of
WAIS3 subtests is not an easy task and may result in variations (Ryan &
Schnakenberg-Ott 2003; Kaufman & Lichtenberger 2005: 197–202). But
there may be other strategies to ease the task, apart from abandoning POI
altogether.
Braden (1999: 2) states that evaluators can change the timing, the
setting, and the formats of both test presentation and response to
accommodate clients with disabilities, and illustrates this with a version of
WAIS3 translated into American Sign Language for the deaf. Also, Rossi
et al. (2007) found no significant differences between individual and
collective administration of the WAIS3. Thus, there exists the possibility of
standardizing a written presentation and response format of the WAIS3 test
set as a web page.12 This strategy might make administration easier and,
crucially, homogeneous across trials, even when applied to several subjects
at once. The scoring of results can also be automated, in such a way that the
researcher may only need to choose some statistical variables, as in the
example provided in Crawford et al. (2008).
So, there may be ways to shorten and homogenize test administration
and interpretation, in such a way that profiling subjects becomes both more
feasible and more manageable. It will still be time-consuming. However,
this should not be seen as an impediment, but just as both a disadvantage
and a challenge. Plants and stones don’t move, gases and stars don’t think,
machines and materials don’t feel. In the social sciences, the quest for
validity demands additional work, but that is exactly what makes it so
fascinating. And in the last decades, Arnt has taught us that it is worth the
effort.
References
12
In this project, a PowerPoint and paper answer-sheet formats were used for both
WAIS3 and TOEFL, which allowed for the testing to be carried out collectively.
Testing was broken down into four sessions, which preceded each translation task.
Subject profiling in translation process research 103
Hambrick, D. Z., & Engle, R. W. 2003. The role of working memory in problem
solving. In J. E. Davidson & R. J. Sternberg (eds.). The Psychology of
Problem Solving. New York: Cambridge University Press. 207–232.
Hansen, G. 2003. Controlling the process. Theoretical and methodological
reflections on research in translation processes. In F. Alves (ed). 25–42.
Hansen, G. 2006. Retrospection methods in translator training and translation
research. JoSTrans 5: 2–40.
Hayes, J. R. & Chenoweth, N. A. 2006. Is working memory involved in the
transcribing and editing of texts? Written Communication 23/2: 135–149.
Hayes, J. R. & Flower, L. S. 1986. Writing research and the writer. American
Psychologist 41: 1106–1113.
Herrnstein, R. J. & Murray, C. 1994. The Bell Curve: Intelligence and Class
Structure in American Life. New York: Touchstone Books (Free Press).
Holmes, J. & Adams, J. W. 2006. How are pupils’ working memory skills linked to
their mathematical abilities? Educational Psychology 26/3: 339–366.
Horn, J. L. & McArdle, J. J. 2007. Understanding human intelligence since
Spearman. In R. Cudeck & R. C. MacCallum (eds). Factor Analysis at 100.
Mahwah, N.J.: Erlbaum. 205–247
Jakobsen, A. L. 1998. Logging time delay in translation. In G. Hansen (ed.). LSP
Texts and the Process of Translation. Copenhagen: CBS. 73–101.
Jakobsen, A. L. 2002. Translation drafting by professional translators and by
translation students. In G. Hansen (ed.). Empirical Translation Studies:
Process and Product. Copenhagen: Samfundslitteratur. 191–204.
Jakobsen, A. L. 2003. Effects of think aloud on translation speed, revision and
segmentation. In F. Alves (ed.). 69–95.
Jakobsen, A. L. 2005. Investigating expert translators’ processing knowledge. In
H. V. Dam, J. Engberg & H. Gerzymisch-Arbogast (eds). Knowledge
Systems and Translation. Berlin: Walter de Gruyter. 173–189.
Jakobsen, A. L. 2006. Research methods in translation–Translog. In K. P. H.
Sullivan & E. Lindgren (eds). Computer Keystroke Logging and Writing:
Methods and Applications. Oxford: Elsevier. 95–105.
Jeyakumar, S. L. E., Warriner, E. M., Raval, V. V. & Ahmad, S. A. 2004.
Balancing the need for reliability and time efficiency: short forms of the
Wechsler Adult Intelligence Scale-III. Educational and Psychological
Measurement 64/1: 71–87.
Kaufman, A. S. & Lichtenberger, E. O. 1999. Essentials of WAIS-III Assessment.
New York: Wiley.
Kaufman, A. S. & Lichtenberger, E. O. 2005. Assessing Adolescent and Adult
Intelligence. New York: Wiley.
Kellogg, R. T. 2001. Competition for working memory among writing processes.
The American Journal of Psychology 114/2: 175–191.
Klein, K. & Boals, A. 2001. Expressive writing can increase working memory
capacity. Journal of Experimental Psychology: General, 130: 520–533.
Konig, C. J., Buhner, M. & Murling, G. 2005. Working memory, fluid
intelligence, and attention are predictors of multitasking performance, but
106 Ricardo Muñoz Martín
Abstract
This is an exploratory study of retrospection as a research method for
investigating problems and strategies both in simultaneous interpreting and
in translation. It is based on data from six subjects, three performing a
translation task, three an interpreting task, with retrospection cued by the
source text/source speech. In the coding of the retrospective protocols, the
categories devised by Ivanova (1999) for interpreting were applied and
found to be applicable to translation as well. Retrospection yielded similar
data for both groups, in terms of amount of data and reported instances of
problems, monitoring and strategies.
1. Introduction
less suitable for translation, which often requires quite a long time. In
addition, translation tasks are characterized by a recursive mode of working
with the same textual material (the source text, the growing target text), e.g.
revision carried out in several run-throughs of the text. This may make it
difficult or impossible for subjects to recall accurately at what point during
the task a certain problem was processed.1
Software for computer logging with a replay function, such as
Translog, has given an upsurge to retrospection for studying the translation
process. Here, retrospection is cued by the source text (ST) and the replay
of the process of target text (TT) writing, both of which are shown on the
computer screen. Aspects studied have for instance been the effect of time
pressure on the process, and sources of disturbance in the process (see, e.g.,
Hansen 1999, 2006a, b).
1
This is confirmed by Levy et al. (1996), who found the correspondence between data
from introspection and retrospection obtained from the same subjects during a writing
task to be quite low: in the best case, 75 % of the retrospective responses matched the
ones given concurrently with task performance; whereas in the worst case, such a
match was found only in 10.5 % of the responses (Levy et al. 1996: 553).
112 Birgitta Englund Dimitrova & Elisabet Tiselius
2
It should be noted that all researchers above point out that they instructed their
informants to comment only on the process.
Exploring retrospection as a research method 113
In our study, we follow Ivanova’s (1999) design for the interpreting task,
slightly adapting it for the translation task. Thus we use only a written
transcript of the ST, not allowing our subjects to listen to their own
interpreting or see their own TT while doing their retrospection.
The interpreting task was recorded and the translation task logged
with Translog. Since the process data were not used as a cue for
retrospection, it can be triangulated with the retrospective data.
Furthermore, by presenting the ST sentences one by one in the translation
task we tried to minimise the above-mentioned recursiveness and ensure
that the processing relevant to the translation of a particular sentence was
done while the subject was still working on that sentence. This should
guarantee that processes recalled really did occur at that point in time, and
will facilitate triangulation with the Translog data.
We thus have two principal types of data tapping into the cognitive
processes: on-line and off-line. Importantly, the off-line data are generated
independently of the on-line data. This is shown schematically in Figure 1,
where the arrows indicate the cueing for the tasks:
will lead to richer retrospective data from the interpreting than the
translation task. On the other hand, the longer duration of the translation
task can be expected to result in more processing of problems, which could
produce richer retrospective data from this task than from the interpreting
task. Since translating a text (even under time pressure) takes much longer
than interpreting the same text when delivered as a speech, the exact same
conditions do not prevail for both tasks. To make them as similar as
possible, we put time pressure on the translation subjects and gave the two
groups the same conditions regarding the use of aids.
Although this strategy notion is process-based, the very fact that strategies
are plans for solving problems in the process implies that they will often
116 Birgitta Englund Dimitrova & Elisabet Tiselius
3. Research questions
4.1 Participants
Table 1. Subjects
Female Male Years at university Age
Translating subjects: 2 1 3–4 20–30 (n=2)
Therese, Josephine and Felix 40–50 (n=1)
Interpreting subjects: 3 4–5 20–30
Amadeus,3 Kristina and Lisa
The cue for the first task was an English ten-minute speech (09:35) from
the European Parliament. All subjects interpreted or translated it into
Swedish. For the translation task, a transcription of the speech, with
normalized orthography and punctuation, was used. The number of words
was 1093.
Before executing the task, the subjects were given some background
information on the speech and a word list. They were given a few minutes
to consult the information, and decided themselves when to start
interpreting or translating. TSs were told that the ST comprised 52
sentences and were asked to finish the task as quickly as possible. The ST
was presented on the computer screen, one sentence at a time. They were
instructed to translate sentence by sentence and to avoid going back to
sentences that had already been translated for revision.4 Their only aid was
the word list. The writing process was logged with Translog. The
3
Subjects were asked to choose their own fictitious name for the study, and one of our
female subjects chose the male-sounding name Amadeus.
4
To give some ecological validity to this set-up, the subjects were told to imagine that
they were EU translators and that the speech constituting the ST was given and
concurrently transcribed at the very moment of the translation event, the sentences
being delivered to the translator one by one as they were transcribed. The time
pressure was thus imposed by the research situation and the translation brief, but at
the same time, it was self-regulated by the subjects.
118 Birgitta Englund Dimitrova & Elisabet Tiselius
researcher was present during the task. The original speech and the
interpreting of the ISs were recorded on a double-track tape recorder.
5. Results
Below we present some quantitative data from the analysis of the protocols.
Although our coding categories (see Appendix) are taken from Ivanova
(1999), no comparison will be made here with her results since none of her
subject categories are quite comparable to ours.5
5
Cf. however an earlier study with other subjects (Tiselius 2008).
Exploring retrospection as a research method 119
The number of words in the interpretings (see Table 2) is much lower than
in the source speech (1093). Kristina’s and Amadeus’ interpretings
comprise 66.0 % and 68.4 %, respectively, of the number of words in the
source speech, whereas Lisa’s word count is as low as 45.0 % of the source
speech. These figures and the differences between the three subjects are a
first indication of the application of different strategies in answer to
processing problems.
For the TTs produced by TSs (see Table 3), there is very small
variability in the word count, ranging between 94.0 % and 97.1 % of the
ST. These high figures and the narrow range of variation are probably due
to the imposed way of translating sentence by sentence and the instruction
not to go back to earlier sentences for revision.
In time for task, the differences between ISs are very small, as
expected, since the interpreting is done simultaneously and cued by the
recorded speech. More unexpected, given the relatively long ST and the
subjects’ self-regulated time pressure, is the low variability between TSs.
In time for retrospection, there is little difference between the two
groups in the raw data, the time ranging between 25 and 36 minutes for
interpreting subjects and between 26 and 45 for translating subjects. This
measure also includes the researcher’s utterances, so the figures cannot be
taken at face value. The next measure, number of words in retrospection, is
a better indicator of the amount of data from retrospection, since here,
researcher utterances which were longer than feedback signals were
excluded.6 The mean figure for number of words in retrospection after the
interpreting task is slightly lower than after translation, but there is quite a
lot of variation between subjects: the longest retrospection, in terms of
words, is from a TS, the shortest one is from an IS. Of the four
6
When looking at the number of words, it should be noted that most subjects read the
ST aloud during retrospection, but with varying degree of completeness.
120 Birgitta Englund Dimitrova & Elisabet Tiselius
retrospections between those extremes, the longer ones are by ISs and the
shorter ones by TSs.
7
For a description of the categories and examples, see Table 1 in the Appendix.
Exploring retrospection as a research method 121
The frequencies for the subcategories differ a great deal between the two
groups, most noticeably in the categories Monitoring/Translation and
Monitoring/Mood. Monitoring/Translation contains reports of ascertaining
8
For a description of the categories and examples, see Table 2 in the Appendix.
Exploring retrospection as a research method 123
5.1.4 Strategies
The next step in the analysis involved identifying those strategies which are
mentioned in the protocols as responses to reported problems (outlined in
Table 3 in the Appendix). Furthermore, in contrast to Ivanova (1999), we
also show the numbers of what we call isolated strategies, which we define
as strategies not related to a report of a processing problem or instance of
monitoring. Tables 10 and 11 present the figures for all reported strategies.
Furthermore, macro strategies (Section 2.7) are shown in Tables 14 and 15.
Table 10. Strategies with reported processing problems (PP) or monitoring (M)
and isolated strategies (IS)
Table 11. Strategies with reported processing problems (PP) or monitoring (M)
and isolated strategies (TS)
There is a difference in raw figures between TSs and ISs. However, the
high figure for TSs is due to the extensive reporting of only one TS.
Reports on strategies for monitoring are very similar in the groups. When it
comes to isolated strategies, ISs report more. This may again be due to the
cognitive load of the task, resulting in recall of the strategy but not of the
processing problem leading to the strategy.
124 Birgitta Englund Dimitrova & Elisabet Tiselius
9
For a description of the categories, explanations of abbreviations, and examples of the
strategies, see Table 3 in the Appendix.
Exploring retrospection as a research method 125
In Tables 12 and 13, only those strategies that are reported in relation to the
problem category to which they are applied as a solution are presented. All
participants report a wide array of strategies. Again, TSs seem to recall a
larger number of strategies related to processing problems. As can be
expected, the most frequently reported strategy for ISs is Deletion whereas
for the TSs, it is Compromise, probably due to the imposed time pressure
of the task and the lack of aids.
10
For a description of the categories, explanations of the abbreviations, and examples of
the macro strategies, see Table 4 in the Appendix.
126 Birgitta Englund Dimitrova & Elisabet Tiselius
Table 16. Words in retrospection and number of reported category instances per
100 words, IS
Participant Number of words Total number of Number of reported category
in retrospection reported instances per 100 words of
instances retrospection
Amadeus/IS 2808 80 2.84
Kristina/IS 4381 107 2.44
Lisa/IS 4230 81 1.91
Mean 3806 89 2.39
Table 17. Words in retrospection and number of reported category instances per
100 words, TS
Participant Number of words Total number of Number of reported category
in retrospection reported instances per 100 words of
instances retrospection
Therese/TS 5376 149 2.78
Josephine/TS 4171 88 2.10
Felix/TS 3738 114 3.05
Mean 4428 117 2.64
There are important individual differences here. The small amount of data
does not permit us to draw any definite conclusions. Interestingly, the two
protocols with the smallest number of words (Amadeus and Felix) are the
most informative, in this sense. However, it could be that subjects’ own
retrospective style also determines to what extent the retrospection
becomes informative.
Exploring retrospection as a research method 127
The subjects studied here on the whole all report the same types of
processing problems although some of the problems are more frequent in
either TSs or ISs. Furthermore, the subjects make use of the same
strategies, although with different frequencies.
The next step in the analysis will involve triangulating data from the
protocols with process data. Evidence of processing problems are found in
the interpretings, for instance in the form of pauses, omissions in the TT,
repairs, hesitations, false starts, and talking speed. In the log files, long
pauses and revisions potentially indicate problems in the translation
process. Evidence of application of strategies is found in the textual
patterns of the TT.
The strategies reported will be compared with the actual textual
solutions. Some preliminary analyses show certain discrepancies between
reported strategies and actual solutions in the text or in the utterances, in
the data from all six subjects. We have found instances of Reporting a
strategy vs. no evidence of strategy found in the TT, and of Reporting one
strategy vs. TT gives evidence of another strategy. There are also examples
where the subject reports a strategy for a particular sentence, but the TT
shows that it was actually applied in another sentence.
6. Discussion
References
Appendix
Perception (P) Problems with I didn’t have time to catch In our midst – now that I
hearing that it was tobacco related read it, I did not recognize
deaths. (8/Amadeus) it, so I don’t know
(9/Therese)
Lexical access Failure to access I said tb but I’m not sure I don’t know what tb is, so
in SL (L) meaning of a SL what it means (8/Amadeus) then – (9/Therese)
chunk, which has
been identified as
familiar
Syntactic Failure to This is difficult, when they Or maybe the whole
processing recognize syntax start with that they say sentence was a little bit
(Syn) patterns ‘which’, which, like you complicated, that you had
start having those to – change around a little
subordinate phrases and bit (2/Therese)
Comprehension
Text integration Difficulties in It’s like you understand I didn’t know that ‘unsafe’
(TC/integ/) constructing a what it’s all about, but you – what they were refering
coherent miss certain details (21- to, if it is the cigarrettes
representa-tion for 22/Amadeus) are dangerous or that,
SL chunks well. (26/Therese)
Text Comprehension Yes, ‘community’ again I And I thought, as I read
comprehension difficulties due to became like confused over ‘framework convention’ I
(TC/bgkn) lack of background what is com- which thought that it was some
knowledge community (47/Amadeus) kind of meeting or almost
conference. But then I
started thinking, con-
vention, that is more like –
it’s about something else
and then maybe ‘ongoing’
aren’t the right words
(47/Therese)
TL retrieval Problems in I got stuck in this, that is ‘I ‘Public health’ I don’t
(TLr) rendering a SL must here compliment’, know – I didn’t write
chunk in TL you know, I thought, oh!, ‘folkhälsan’ (public
how should you put that health) I think – well,
anyway, I’m a bit
Translation (Tr/)
(10/Amadeus)
uncertain of how it should
be translated. (3/Therese)
Equivalent Problems in I was thinking about And a couple of times I
(eqv) selecting an whatever that is called in hesitated – ‘tobacco
appropriate Swedish, I considered products’, if it was better
equivalent when guidelines or descriptions. to write tobacco products
there is a choice (37/Amadeus) or tobacco goods
(35/Therese)
132 Birgitta Englund Dimitrova & Elisabet Tiselius
(SL, TL) Problems due to Then it was too many Since I didn’t want to take
high SL input rate words, sort of, like terms, too much time I think that
in relation to ‘tobacco advertising’ I did it fairly simply.
interpreter’s own ‘sponsorship’, so I don’t
Simultaneity of tasks (Sim/)
(31/Therese)
output rate know what I said.
(45/Amadeus)
TL delays (TL Delays in TL You know, I heard this but And that is such a thing
delays) product due to I didn’t get it out, where you could have
translation. everything that is written come up with something
here. It was really difficult better, if you had reflected
to, you know, use these two some more on it.
processes at the same time, (9/Therese)
to listen and understand
and then speak, it was like
a catch or an obstacle.
(10/Amadeus)
Anthony Pym
Abstract
1. Introduction
processes; our role is merely to suggest one way in which process research
might find a use in the training of translators.
Empirical work can be applied to the training situation in several
flavors. An exemplary mode would be the TransComp project to study
students’ progress longitudinally (Göpferich 2009), which will give
information on long-term learning processes and might thus justify a
developmental model of competencies. A simpler and more direct mode of
attack is to review all experiments comparing novices and professionals on a
wide range of performance variables (the body of research is already
considerable), to collect the significant differences, and to organize those
differences in terms of a set of learning objectives or competencies. After all,
if we can say what the differences are between novices’ and professionals’
performances, then we can presumably say what students need to be trained
in, albeit without information on sequencing. A third procedure, far more
direct and perhaps of more limited applicability, is to have students engage in
rough experiments as part of the training process itself, both as a means of
self-discovery and as an approach to learning about research. This third mode
of application (there are no doubt others) is what we report on here.
We are by no means the first to venture down this path. Juliane House
(1986, 2000) long ago pointed out that having students translate in pairs or
small groups (―translation in and as interaction‖) was not just good for think-
aloud protocol research – it was good for the students’ learning processes as
well. Fabio Alves (2005) reports on the use of Translog in the classroom,
detailing a series of activities where each cycle of translation–analysis–
discussion accounted for some 10 credit hours. And then there are many
authors who point out the benefits of process-based activities, focusing not
just on what students write but also on how they go about making decisions.
The kinds of experiments that we report on here are different from those
antecedents in several important respects. Like House, we wanted to
encourage talk between students, but we were not prepared to devote class
time to recording and analyzing protocols – the talking and reporting had to be
one-off and quick. Like Alves, we wanted students to look at their translating
critically and to develop a meta-discourse of some kind, but we were not able
Using process studies in translator training 137
to spend 10 hours putting students at ease with Translog, nor another 10 hours
per translation plus analysis. Finally, perhaps like all good teachers, we could
have mobilized a battery of tricks to raise awareness of complex decision-
making processes, but in this case we especially wanted the tasks to look and
feel like empirical research. Slick, quick, and yet empirical. The reasons for
these desiderata lie in our particular institutional setting.
2. Three experiments
The experiments we are about to present come from the 2008/09 Translation
Practicum course for second-year Masters students at the Graduate School of
Translation, Interpretation and Language Education at the Monterey Institute
of International Studies in California. The Monterey Masters programs are
highly focused on the training of top-flight professionals, and they are very
expensive, as some students do not hesitate to point out. On the basis of
economics alone, each class has to be justified as progress toward a
professional goal. Any methodology that requires 10 credit hours per
translation would obviously be a very hard sell. Hence our interest in
relatively quick tasks, with a special emphasis on students being able to
discover something as a result of each class. But why then our concern with
interaction? And why the empirical?
The general aims of the Practicum course, described in the handbook,
were to make the students 1) self-critical of their translation processes, and 2)
aware of the contributions of new technologies to the actual act of translating.
However, the non-official aims were perhaps more interesting.
The Monterey Masters courses are structured around separate language
programs, which means that the students in different groups rarely have the
chance to interact with each other in translation or interpreting classes. For
example, the Korean students (all from Korea) tend not to find out how any of
the other students translate, nor vice versa. One of the social and pedagogical
aims of the course was thus to make students aware of their different cultural
backgrounds and their quite different approaches to translation. Students had
138 Anthony Pym
and their diversity might yet be representative of the many kinds of situations
that translator training can involve.
Following the first contact hour (described below), the students selected
the aspects of translation processes that they wanted to work on. The course
was then structured around that selection, with experiments on machine
translation, translation memories (Wordfast, Trados), reviewer styles,
translator styles, spoken vs. written translation, and translation procedure
analysis, along with non-experimental topics such as ethical problem-solving
and a survey of pay scales in different countries. The second half of the
Practicum, not described here, looked at translation products, with a particular
emphasis on corpus analysis.
Below we present reports on three of the experiments: the very first one,
then one from Week 7, and another from Week 8.
The first experiment was conducted in the first hour of the first class, with all
attendant confusion and technological hitches. The aim was basically to
demonstrate to students that 1) these classes were going to be practical but
quite different from their other translation classes, and 2) there were things
that they could discover about themselves and about technology. Each
language group was divided into two (giving a multilingual Group A and a
multilingual Group B). The class instructions were as follows:
Please translate the following text for publication in an online dictionary of technical terms,
to be used as a general reference source.
Since we are interested in how long these processes take, please work at what you consider
your normal pace. You will be stopped after 12 minutes.
3D printing
The making of parts and products using a computer-driven, additive process, one layer at a
time. 3D printing builds plastic and metal parts directly from CAD drawings that have been
cross sectioned into thousands of layers. It provides a faster and less costly alternative to
machining (cutting, turning, grinding and drilling solid materials).
Used for making both prototypes as well as final products, 3D printing evolved from the
“rapid prototyping” industry, pioneered by Chuck Hull of 3D Systems in the mid-1980s.
Capable of making a part from scratch in just hours, 3D printing is used to create models
to determine if a design meets the customer’s concept and expectations. It is also used to
create prototypes of parts to test their form, fit and function with other parts in an
assembly.
The time limit was stipulated as 12 minutes in order to make students work at
a brisk pace, but repeated public extensions were then given so that all
subjects except three finished the text within 25 minutes (presumably under
roughly the same pressure). The only quantitative variable sought was the total
time taken to produce the final translation (i.e. drafting plus self-review). All
the French-program students and two Chinese-program students were working
into their L2, so there was no level playing field. The text was selected so as to
ensure that all students would both understand the subject-matter quickly and
yet need to search for some terminology. Following the translation exercise,
students formed language-specific groups of two or three in which to compare
each MT-based translation with a non-MT based translation into the same
language. Each group then reported orally on the differences between the two
translations.
The general result of this simple experiment was that there was no
significant difference in the time taken with MT and without it, no significant
difference between the language groups (see Table 1), and no systematic
difference between the qualities of the translations as assessed by the students.
The general quantitative message was thus that students were not going to lose
Using process studies in translator training 141
any time by working with MT – so perhaps they should consider how this
mode of work might affect them in the future.
The qualitative findings, reflected in the class discussion, were that 1)
translators into Chinese and Korean (our one Japanese translator escaped from
this activity) generally reacted negatively to the use of MT, even though this
difference is not reflected in the quantitative results (we expected Asian
languages to give worse results because the MT databases were presumably
smaller for those languages), 2) the translators into French appreciated some
of the terminology proposed by Google Translate, 3) all translators
complained about the MT misreading syntax and were generally appalled by
the resulting wild mistranslations, even though this might be one of the
advantages of MT with respect to TM – better a clear mistake than a series of
fuzzy matches you have to think about for a while (see Guerberof 2009).
This experiment would not hold up to any sort of analysis if judged as a piece
of empirical science. The standard deviations are enormous: the fastest
subject, a high-tech man working into L1 Korean with MT, took 10 minutes;
two translators, into French and Chinese, took 30 minutes for no obvious
reason, and the all-time slowest, a perfectionist into L2 Chinese with no MT,
took 38 (a long story typical of real life – he had trouble with his laptop, went
to the library to finish the translation, and thus avoided the external pressure
and saved all kinds of face). These individual differences in turn explain the
different times for the various language groups. Further, this was the first time
most of these students had reviewed MT output. With greater experience and
familiarity, their performances in this mode would presumably become
quicker and more standardized. In short, this is not something we would like
to publish as research.
142 Anthony Pym
All students did all of that in 100 minutes (these are intelligent students). Only
a few had significant technical problems, so those times have been edited out
of the data (percentages have been calculated on the basis of the other tasks
only). The class group on that day included our Japanese colleague but was
missing one Korean and three students from the French program. Irregular
attendance is also a fact of life (and would mess things up for the following
week’s activity, but on we go.) The results are shown in Table 2.
A major difficulty in this experiment was the attempt to define the separate
tasks in a clear way. As predicted by Mossop’s categories, some students
found that they were revising fragments as they went along, at the end of each
paragraph or as part of the general drafting process. In those cases we simply
asked them to count the total review time as best they could, without insisting
on the ―post-draft‖ criterion. If we were looking for a scientific result, some
special count would have to be made of these intermediary stages of review. In
the class situation, though, it is enough that each student be able to apply their
own consistent criteria. The difficulty of quantifying the review stage actually
became a major topic of discussion at the end of the lesson.
Any good translation teacher will object here that we should not use
newspaper texts, since they are poor representatives of most real-world
translation situations. We are indeed embarrassed by this recourse to
convenience. There were, however, several logics in play: 1) in order to have
all students going into their L1, texts had to be selected in various source
languages, including some unknown to the instructor, 2) these students mainly
find employment in government departments where current events are indeed
material for translation, 3) the use of very recent news would ideally minimize
the differences in prior knowledge of the subject matter – fresh news is
presumably fresh for all, 4) we also reduced the risk that a translation of the
text was already available on the web (as happens with disconcerting
frequency). The text choice was thus supposed to reduce the variables, and
might yet be justified on that score.
As it happened, the results shown in Table 2 betray the weaknesses of
both the classification and the text-selection criteria. Two outliers are shaded
(K2 and F2) because their scores are indeed shady: the screen recording shows
that K2 did a great deal of reviewing while drafting but did not count it at all
(she would change criteria the following week), and F2 confessed that he
chose a subject with which he was exceptionally familiar, and thus ostensibly
required no reviewing and only trivial documentation. Both these cases would
create problems down the line, as we shall see below.
Using process studies in translator training 145
K1: I just assumed I was ―watercolorist‖. The result shows that I am.
K2: I found it interesting to see that I spent so much time revising my text. It was not like I
translated the whole text and then revised it. I revised the text as I was translating.
K3: I think I am an oil painter kind of translator. Actually, I am always trying to plan first
and then start translating, which is what I learned to do. However, I always fail to plan
perfectly.
K4: I read briefly to get the summary of the source text. And then get right into the
translation, even before looking up the words. And then I look up the words and then
revise.
C1: Based on the recording, apparently I don’t plan before translating. I read the text as
quickly as I can, then start to translate segment by segment if I can comprehend.
C2: I think I am an ―architect‖ because I spent the majority of my time on pre-drafting and
minimal time on post-revising.
C3: I plan first and do the task. Before this experiment, I didn’t notice that I spend a lot of
time on comprehending and don’t go back to make changes often. This surprised me a lot.
C4: I’m more of an ―oil painter‖ translator. I typically skim an article for context prior to
translating then end up doing a good deal of the research/glossary building during
translation and a lot of editing after translation.
C5: I prefer to get into the translation right away without previewing the whole article first
and then do the revisions. I can translate quite quickly with the pressure of limited time.
C6: I always do the task first while making changes at the same time. I was a little surprised
that I spent only 37 percent of the time on translating.
C7: I found the watercolorist style best describes the translation job I recorded myself
doing.
146 Anthony Pym
F1: Unless it’s a technical text, I tend to do the task, then make changes.
F2: For this particular text, I was able to translate first, then make changes after because it’s
a topic with which I am somewhat familiar in both languages.
F3: It appears that I am an oil painter at heart, putting minimal work into the planning
stages and significant effort into the revision stage at the end.
When you go from these self-descriptions to the data in Table 2, the concept of
translator styles does make some sense. However, when I try to identify the
styles on the basis of the data alone, nothing statistically significant leaps out
(although I hope someone will prove me wrong). This could indicate that the
time-on-task categories would have to be far more sensitive if they are to
correspond to any ―styles‖ with pretensions to psychological reality. Mossop’s
categories are great for getting students to talk about how they translate, but
they would appear not to have much empirical virtue beyond that.
The students’ comments also contained some revealing self-criticisms:
K1: It was interesting to see how little time I use for comprehending and revising. I get
comments from professors that my accuracy is not good enough that they ask me if I am
paying enough attention to reading. They were right! I should do that more!
K3: I did not know that I spend such a long time on web searching. But I don’t spend
enough time on revising.
C2: My way of translating seems too slow. I should spend more time on revising.
C4: I think increased use of Multiterm will help speed up the process, especially for
automatically inputting proper names.
F1: I was surprised at how long it took me to do it. And I don’t think my translation is
sufficiently colloquial.
F2: What surprised me most was my typing speed. I know that I do not generally type very
fast because I was never trained in typing properly.
F 3: One basic thing that became obvious was that I make a good number of typos (about 3-
5 every sentence), which I fix immediately.
Using process studies in translator training 147
C4: Because of the ordering of information in a Chinese sentence, I have found that it is
safer to get it down in awkward English first then worry about restructuring the English
sentence later, otherwise I risk omitting details at the beginning of long Chinese sentences.
The data do not allow us to confirm any correlation between language pair and
translator style, but the hypothesis remains of interest. Another student found
time for a basic segmentation analysis and a small self-improvement program:
F3: I identified a fairly consistent pattern as I translate: I look at the text (5 seconds), type
the first part of the sentence (10-15 seconds), read the rest of the sentence (5-10 seconds),
and then translate the remainder of the sentence (10-30 seconds). I did not realize that I
start translating before I even read/understand the whole sentence. Sometimes I would have
to retranslate the first part of the sentence, but not all that often. As a side note, I think the
aspect of where the eyes are looking and what the brain is doing during the typing phase is
quite an interesting question too. It seemed that in class I would look at what I was typing
on the screen and think about that before moving on to the next part of the sentence. A very
segmented approach. After class, I wanted to work on a translation while keeping in mind
the issues that came up during this exercise, and I tried focusing on reading ahead and
understanding the second part of the sentence while still typing my translation to the first
part – all while making an effort to eliminate typos too. It was a challenge (which is another
reason I’m pretty sure I hadn’t been doing it much before) but it also felt like my translating
was much more fluid and somewhat faster.
If we are looking for the pedagogical virtues of process analysis, this student’s
comments are about as good as it gets. In fact, all the students’ responses were
generally much better than our research design.
A week after the above experiment, the group was asked to perform exactly
the same task, but faster and in pairs (two students had to translate the same
text, then compare results in order to get some rough assessment of translation
quality). On the basis of previous research on the time variable, we believed
that most students would be able to work about 25 % or 30 % faster without a
great loss in translation quality. As in the previous experiment, we advertised a
148 Anthony Pym
time limit (actually set at 35 % faster) and then we did not apply it rigorously
(to allow the translations to be completed). Here are the instructions:
Today’s task is the same as last week’s except that you have to go 35 % faster and you will
work in pairs (i.e. two of you will translate the same text). Your mean time last week was
19.2 minutes, so your time for this week is 12.48 minutes. Here are the steps:
1. Form a pair with someone going into your preferred target language.
2. Both of you should find a respectable online daily newspaper in your L2 (the language
you want to translate from). Select a lead (i.e. front-page) story for today. Cut about 200
words, or a rough equivalent in characters, from the beginning of that story. (Use Word
Count to get the number of words. Select fewer words if necessary. Select something of
similar conceptual length if you are dealing with characters rather than words.) DO ALL
OF THIS VERY QUICKLY.
3. Paste those 200 or so words into a Word document, open a TM, and get ready to translate
them using the TM. There is no real need to use a special memory or glossary for this
exercise.
4. Start BB Flashback. Select RECORD.
5. Translate the text into your L1, doing web searches and revising as necessary. Aim to
complete the translation in 12.48 minutes, since that is the time after which you will be
stopped.
6. Have a break.
7. Play back your screen recording. Try to keep a track of how many seconds you spend on
the following tasks: a) technical problems, b) reading, comprehending, c) documentation
(web searches), d) translating, drafting, e) reviewing after the drafting (not including the
correction of typos as you type).
8. Upload your translations and the analysis of your time-on-tasks, plus brief answers to the
following questions: a) Did the faster translator make more mistakes than the slower
translator? b) What differences can you see with respect to the time-on-task analysis you
did last week? c) What recommendations would you give to a translator who wanted to
know how to translate faster without making excessive mistakes?
The mean time for these translations was actually 35.38 % faster than for the
previous exercise, with the slowest translator (C7, a self-described ―water-
colorist‖) showing a gain of 16 %, and the fastest claiming a gain of 58 % (C4,
an ―oil painter‖). The general shifts in percentages per task are shown in Table
3 (student K4 has disappeared because she was not in class).
Using process studies in translator training 149
The totals for the percentage shifts suggest that speed was achieved by cutting
down on documentation, initial reading, and reviewing. The variation between
the scores for each task is nevertheless extreme. Further, student F2 was
actually slower in this ―fast‖ experiment, since on this occasion he was not
able to choose his favorite subject matter. As mentioned, K2 also messed
things up by deciding to count the reviewing time in a different way. This is
why these two students (K2 and F2) were marked as outliers in the previous
experiment and they have been removed from our comparative analysis of
percentages of time on task (Tables 4 and 5) – their scores do not tell us
anything significant about translation processes as such.
The reduced magnitudes of the shifts and the relative standard deviations now
indicate that the students generally gained time by cutting down on
documentation and reviewing, but not so much on reading. The students with
150 Anthony Pym
the highest time gains tend to be those who originally spent the most time on
reading and documentation (the relative ―architects‖), but this is only because
they were the slowest in the first translation. None of this is anything to write
home about. More importantly, none of the students reported that the faster
translation was of significantly worse quality. They were thus generally
prepared to accept that more time does not automatically equal higher quality,
and that all of them could speed up their normal translation processes.
The comments were once again more revealing than the numbers. Here
we give the main observations by each student, indicating in square brackets
those whose previous self-descriptions used one of Mossop’s metaphors:
K1: I spent a lot less time researching. When you have a time limit, your efficiency goes up
and you just focus on the text.
K2: I spent more time on reading and translation compared to last week.
K3 [―oil painter‖] In the timed translation tasks, I focused on understanding the context and
tried to make sentences simple and clear, which in turn helps readers understand with ease.
C1: Generally, both of us spent more time on translating and no time on revising.
C2: This time I spent less time on reading and translating [he increased the percentage of
reviewing].
C3: With time pressure, I didn’t have much time to consider if I had used the right
characters or phrases and some spelling mistakes might be found in this translation.
C4: [―oil painter‖] More [percentage of] time spent on documentation and translation;
significantly less [percentage of] time on revision.
C5: [―oil painter‖] I just caught the rough meaning of the sentences and got to the
translation right away in order to finish the translation before the time limit.
C6: [―oil painter‖] First, I had to speed up the process of translation by reducing the time
spent on documentation and revision. Second, more multitasking work is required in speed
translation, which involves reading, comprehension, web searching, translating and revising
in a single process.
J1: The difference was that I spent a higher percentage of time on ―documentation‖ (word
research) than ―reading‖, compared to last week. What it meant was that, instead of
understanding and translating the text as a part of co-text and context, I use a more word-
for-word strategy. But the text I used this time was on the stock market, and I just did not
know exactly how to phrase terms in my language, which in turn meant looking up a lot of
words in dictionaries and spending a lot of time on documentation.
F1: I spent much less time on documentation this week, and revised while I translated. I
would have liked to have time to finish and go back to some parts that I didn’t feel like I
rendered accurately.
F2: I think I would tell a translator to do as I say and not as I do, as it were, meaning spend
less time researching and possibly less time reading, and just dive right into the text and
start translating it and then make sure to leave some time for revisions.
F3: [―oil painter‖] Looking just at percentages, the proportion of time I spent on translation
in the second week almost doubled, the reading time was nearly halved and revision time
decreased somewhat. In absolute time, though, I think it is interesting that the number of
seconds I spent translating was practically the same in both exercises. That would suggest
that even without a time constraint, I work about as fast as I can during the translation
phase and then use the revision phase to clean everything up. There is, though, the factor of
the reading and comprehension, which during the first week seemed like discrete units but
under time pressure blended much more into the translation phase.
Once again, the students’ assessments seem more astute than our
research design. As C6 and F3 both suggested, the blending of categories
under time pressure does indeed show in the screen recordings but not in our
numbers. Worse, since the in-class discussion was based on percentages rather
than absolute times, there was no real chance to talk about the possibility that
drafting times remained fairly constant, as predicted by F3.
Some students also made a series of recommendations based on their
analysis:
C5: I think the secret of translating faster is not to hesitate for too long before starting to
translate and make revisions afterward so you won’t waste too much time getting stuck on
the phrases you have problems with.
C6: If they really want to speed up their translating process without major meaning shift or
error, they should focus more on reading comprehension and try to be more linguistically
flexible.
152 Anthony Pym
C7: What I recommend is try to cut from the documentation part and the revision part, but
not the reading and comprehending part if the translator wants to make as few mistakes as
possible.
F1: For a person trying to translate quickly, I would say to avoid documentation if at all
possible, and do some revising at the end to revisit those sections that were difficult or
about which there were some uncertainties.
F3: Translate quickly and save time at the end to revise – but set some sort of deadline so
that not too much time is spent revising.
J1: My recommendation is [usually] to do the research ―in the middle of or together with
the translating process‖. But as the results show today, this strategy did not really work out,
so probably translating first and going over your own translation (reviewing afterwards,
including doing some research on uncertain terms) may be better.
3. Conclusions
Although we have presented numbers that look like empirical research, our
purpose has not been to formulate any findings about the translation process.
The only conclusions we could risk are so general as to be banal: the use of
Using process studies in translator training 153
terms of risk, the use of external resources (both written and human), and the
key role of reviewing. The obvious problem, however, is that any findings that
have been justified on the basis of a fully controlled experimental situation,
with a homogeneous student group, wonderful p-values and clear causal
logics, may then be applied in pedagogical situations that can be extremely
diverse and relatively uncontrolled, both within and between different class
groups. The use of experiments in class might nevertheless allow solid science
(plus a few of the current idealistic pre-set full-control lists of competencies)
to be circumvented to a certain extent: students draw their own conclusions
about their own developing competencies, and the group is relatively free to
chart its own sequence of short-term learning objectives. In short, the use of
experiments can be used to empower the group (as constructivists love to say)
and humanize education (as we prefer to say), rather than impose plans and
rules that we are far from certain about.
On the basis of the above, the direct use of process experiments may be
recommended as a pedagogical exercise, at least in advanced classes where
the various technologies actually work. Although we look forward to the day
when process research will provide a solid basis for our institutionally guiding
lists of competencies (such things cannot be avoided), there is no real need to
wait: much of the exciting stuff can be done right now.
* The author would like to thank Serafima Khalzanova for her perceptive
comments on a previous version of this text, and the translation students at the
Monterey Institute of International Studies for their participation in the
activities.
References
Alves, F. 2005. Bridging the gap between declarative and procedural knowledge in
the training of translators: meta-reflection under scrutiny. Meta 50/4. Available
at http://www.erudit.org/revue/meta/2005/v50/n4/019861ar.pdf. Accessed
August 2009.
Göpferich, S. 2009. Towards a model of translational competence and its
acquisition: the longitudinal study TransComp. In S. Göpferich, A. L. Jakobsen
& Inger M. Mees (eds). Behind the Mind. Methods, Models and Results in
Translation Process Research. Copenhagen: Samfundslitteratur. 11-37.
Using process studies in translator training 155
COMPUTER
ASSISTANCE IN
PROCESS RESEARCH
Adding value to data in translation process research:
1
the TransComp Asset Management System
Susanne Göpferich
Abstract
easily if they are provided with detailed meta-data which can be used as
search criteria.
In an effort to solve the problems mentioned above, all materials
used in the longitudinal study TransComp, such as the source texts, the
translation assignments and the model translations, as well as all data
2
obtained in the experiments, such as the translation process protocols , the
log files, the screen records, and the subjects‟ target texts without and with
evaluation mark-up will be made available to the scientific community in
an Asset Management System (AMS), an open-source-based storage,
administration and retrieval system for digital resources. For the translation
process protocols, XML-based transcription conventions have been
developed on the basis of the Guidelines for Electronic Text Encoding and
Interchange (TEI Consortium 2007) of the Text Encoding Initiative (TEI).
They are described in detail in Göpferich (forthcoming a). In this way the
problems pointed out by Englund Dimitrova (2005: 82 f.) are addressed.
She states that so far “no single, widely accepted model for coding and
analysis” has been developed and that “there does not yet seem to be an
established way of reporting protocol data”. The AMS will contribute to
the solution of this problem and facilitate future multi-centre studies, in
which, for instance, source texts and assignments can be downloaded from
the system and used with subjects from other translation-oriented
programmes and with other language combinations; these data can then
flow into the system and be compared with the ones from our own and
other studies.
Ideally, such an AMS could form part of an Internet portal that
provides access to one or several archives of data collected in
investigations into translation processes. These could be stored in such a
way that by applying certain search criteria, specific types of data could be
retrieved (e.g., all the data resulting from a specific project or from
3
experiments in which professional translators took part). Translation
process researchers could then use these data as a corpus of reference with
2 The term translation process protocol refers to the transcript of what has been said
(e.g. think aloud) but also of other actions that have occurred during the translation
process, such as the consultation of dictionaries and the refitting of the headset.
3 For such an archive with regard to writing process research, cf. Strömqvist et al.
(2006: 71) and Sullivan/Lindgren (2006b: 210 f.).
162 Susanne Göpferich
which to compare their own data and findings, which could then also be
uploaded into the archive, thereby enabling several smaller-scale studies to
become extended into a larger-scale one on a cooperative basis.
Furthermore, such an Internet portal could also form a valuable
resource of material for translation pedagogy, in which a process-oriented
approach has now been advocated for about a decade (cf. Gile 1994, 2004;
Kußmaul 1995, 2000, 2007), especially in “[c]ourses for experienced
translators, who want to improve their methods rather than acquire basic
experience, which they already have” (Gile 1994: 112), in “[c]ourses
involving source and target languages that the teacher does not know”
(Gile 1994: 112), and in courses for students at the beginning of their
training in translation (Gile 2004). Samples from the process data stored in
the Internet portal could be used in the translation classroom not only to
give students insight into typical shortcomings in problem-solving
processes but also into strategies that have led to successful or particularly
creative translation solutions (cf. the numerous examples in Kußmaul
1995, 2000, and 2007).
Asset Management Systems provide the type of functionality needed
for these purposes. These electronic systems allow us to import and export
digital resources of any type, such as texts, graphics, videos, and sound
files (including format conversions), to annotate them using meta-data, to
retrieve data and analyse them, to view and play sound and video files, to
bundle files which belong together, and to archive and version them.
By providing easily accessible information on the Internet, an AMS
supports collaborative work beyond the boundaries of departments,
institutions, and even countries. Another advantage of an AMS is the
possibility of connecting files (e.g., transcripts and the screen-recording
files on which they are based) via hyperlinks and thus reducing the amount
of work involved in data transcription as will be explained below (cf. also
Stigler 2008). Additionally, Asset Management Systems allow various
display options. Whereas some researchers want to use specific
interpretation categories that have been tagged into a corpus of process
data, others prefer a clean corpus without such tagging and do not want to
be influenced by other researchers‟ interpretation mark-up. As will be
shown, the display options in an AMS can be designed in such a way that
Adding value to data in translation process research 163
users can choose what types of information in the transcripts they wish to
display. Furthermore, Asset Management Systems also allow storing
documents in different versions, for example, in the form of „pure‟
transcripts without interpretation mark-up as a resource for various types
of research, and as transcripts with such tagging for purposes of
4
verifiability.
In the following, the AMS developed for the longitudinal study
TransComp (Göpferich et al. 2008 ff.) will be described.
5 The Translog project specification describes how the source text will be displayed.
166 Susanne Göpferich
All data collected in the experiments can be accessed in two ways: either
via the “Experimental Waves”, in which the data were collected, or via the
“Subjects”, who took part in the experiments. Fig. 4 shows the list of all
experiments conducted in the first experimental wave (“Wave 1”) with the
student subjects.
Figure 4. Experiments conducted in the first experimental wave with the student
subjects
Fig. 5 below shows the list of all experiments in which the student subject
BKR has taken part so far.
Adding value to data in translation process research 167
This display can be subdivided into three areas: (1) an area (“more”),
where the meta-data can be displayed; (2) the window on the left, where
the transcript can be displayed; and (3) the window on the right, where the
screen record can be replayed.
6 Problem No. 1 (“speaks vs. talks”) refers to the following passage of the source text,
which had to be translated from English into German: “Moja, together with a dozen
or so other chimps and one gorilla in the United States, talks. She doesn‟t speak – she
talks. She communicates with her fingers in American Sign Language, devised for,
and used by, hundreds of thousands of deaf Americans.”
Adding value to data in translation process research 173
7 This may take a few seconds because the screen recording file (in flash format) has to
be downloaded first.
174 Susanne Göpferich
The other links on the left of Fig. 1 provide access to the “Publications”
resulting from the research project TransComp, to the documentation of the
XML schema file developed for TransComp (“Schema documentation”; cf.
Göpferich forthcoming a), the “Bibliographic Database TransPro”, which
contains an online bibliography on translation process research,
information on a “Mailing List” for people interested in translation process
research, and to the links of “Other Research Groups” active in the field of
translation process research.
The materials used and all the data collected in the TransComp
experiments conducted so far have been uploaded into the AMS described
above (Göpferich et al. 2008 ff.). At the moment, these materials and data
are password-protected because the source texts will also be used in future
test waves of TransComp, and we have to make sure that our subjects do
not have access to them until the last test wave has been completed (in
August 2011). After this, password protection will be removed and the
data can be accessed freely.
a last step, each object has to be attributed to the „data containers‟ to which
it belongs. For each experiment, these are the containers of the
experimental wave and of the subject involved. This ensures that the data
can be accessed in the AMS both via the link “Experimental Waves” and
via the link “Subjects”.
4. Conclusion
References
Appendix A
</div>
<div type="main-phase">
<u start="00:03:39" end="01:02:27">ok wa:s das en jetzt?<pause
dur="3"/>vorhin war schon besser.<vocal>sighs</vocal>
<incident type="reads ST"><shift loud="p">we don’t want a thing
because we found a reason for it we find reasons for it because we want
it</shift></incident><pause dur="2"/>ok
<incident type="self-dictates">wa:rum? men:schen rauchen
</incident><pause dur="7"/> <vocal>sniffs</vocal>wir wollen etwas
NIcht weil wir einen grund dafür gefunden haben<pause dur="1"/>wir
finden grün:de dafür weil wir<pause dur="1"/>es wollen<pause
dur="6"/>wir wollen ein ding nicht<pause dur="5"/>
<incident type="self-dictates">nicht, weil wir einen grund dafür
<vocal>sniffs</vocal>gefunden haben<pause dur="2"/>wir finden
gründe dafür, weil wir es wollen</incident>stimmt das?
<incident type="reads TT"><shift loud="p">wir wollen ein ding <pause
dur="1"/>nicht, weil wir einen grund dafür gefunden haben <pause
dur="1"/>wir finden gründe da:für,<pause dur="2"/>weil wir es
wollen<pause dur="3"/></shift></incident>ja, stimmt
<vocal>sniffs</vocal><pause dur="10"/>
<incident type="types"/><vocal>sniffs</vocal>
<incident type="reads ST"><shift loud="p">inspite of the importance of
psychological<pause dur="1"/>or conditioning factors in
addiction</shift></incident>
<incident xml:id="1" type="problem" subtype="conditioning"
start="00:06:16" end="00:06:17">häh:::? </incident>
<incident type="reads ST"><shift loud="p">it is the craving that <pause
dur="2"/>most often causes the addict to fail in any attempt to
quit</shift></incident><vocal>sighs</vocal>inspite <pause
dur="1"/>trotz der<vocal>smacks</vocal>importance ähm: ja:
wichtigkeit<pause dur="1"/>von psycho cholo:gischen: und conditioning
<incident xml:id="1.2" type="problem" subtype="conditioning"
start="00:06:47" end="00:06:54">ähm: ja das weiß ich immer noch nicht
so recht wie ich das übersetzen soll</incident><pause dur="1"/>
factors,<pause dur="1"/>hm:<pause dur="1"/>addiction abhängigkeit
und gewohnheit<pause dur="1"/>fü:r<pause dur="2">physi<pause
dur="1"/>physische<vocal>sniffs</vocal> ähm: substanzen <pause
dur="7"/>.h conditionin ja,<pause dur="2"/>ja, ok. .h trotz der
wichtigkeit<pause dur="1"/>von: <pause dur="2"/>psychologischen und
conditioning [...]
</u>
</div>
<div type="post-phase">
<u start="01:02:27" end="01:09:17">
<incident type="reads TT">warum menschen rauchen. .hhh wir wollen
ein ding nicht weil wir einen grund dafür gefunden haben wir finden
gründe dafür weil wir es wolln:</incident><pause dur="5"/>ja.
<incident type="self-dictates">will:</incident>[…
</u>
</div>
</body>
</text>
</TEI>
</teiCorpus>
182 Susanne Göpferich
Appendix B
The summers spent with Arnt and colleagues from the Copenhagen
Business School at the retreat in Skagen were as close to heaven as I’ll ever
get. In his enviably laid-back way, Arnt managed to mold us into an
irrepressible think tank. We wrote articles, he produced grant proposals,
and all of us reveled in long walks along the magnificent beaches of that
picturesque town on the tip of Jutland. Our explorations of the “migratory
dune,” our visits to the “sunken church,” and the wonderful evenings spent
lolling in the spacious parlor of the mansion that became our home during
those magical weeks – are etched in my memory and in my heart. Back in
Copenhagen, it was Arnt who introduced me to Karen Blixen’s home, and
all the stories that go with it, and Arnt – never rushed, never appearing
pressured – who took me to see the opera building and other highlights of
Copenhagen architecture. It was also Arnt who set up CRITT and so many
other cornerstones of collaborative research at CBS. And on an even more
personal level, it was Arnt who did me the unforgettable honor of co-
editing and publishing Interpreting Studies and Beyond – a very special
volume of the Copenhagen Studies in Language. I feel privileged to have
been given this opportunity to express my appreciation, in some small way,
for the gift of having Arnt as my friend. (And I am grateful to my co-
authors, who have not had the good fortune of meeting Arnt, for agreeing
to dedicate our first collaborative article to Arnt Lykke Jakobsen.)
Miriam Shlesinger
1
This research was partially supported by the Israel Science Foundation, grant no.
1180/06.
2
We thank our colleague Jonathan Fine for his valuable input on an earlier version of
this article.
184 Miriam Shlesinger, Moshe Koppel, Noam Ordan & Brenda Malkiel
Abstract
1. Introduction
In the Book of Judges we are told of a battle between two peoples, the
Gileadites and Ephraimites. When the defeated Ephraimites tried to cross
the Jordan River in an effort to return to their own territory, the Gileadites
blocked the passage. Being unable to spot an Ephraimite by appearance –
for all practical purposes, the Ephraimites resembled members of the
neighboring tribe of Menashe – the Gileadites devised a method of
identifying them by their pronunciation: capitalizing on the Ephraimites‟
well-known inability to distinguish between [s] and [š]. Every person who
wished to cross the river was required to pronounce the telltale [šɪbələθ],
and whoever pronounced it [sɪbələθ] met his demise. The biblical story
(Judges, 12) is among the earliest descriptions of a categorization task. It
includes two important concepts – classes and features – with the
Markers of translator gender: Do they really matter? 185
Ephraimites and non-Ephraimites as classes and the minimal pair [s] and
[š] a phonetic feature.
In many ways, the story of the Ephraimites is reminiscent of
contemporary work on text categorization, in which one aims to assign an
anonymous text to a particular class. Early work in authorship attribution
used a variety of statistical methods to identify what are known as stylistic
discriminators – characteristics that were relatively consistent in the oeuvre
of a particular author, but varied from one author to the next.
Mathematicians and linguists saw the potential of machine learning for
extending the scope of text categorization, and attempted to bridge between
early work in stylometrics and contemporary computer-based
methodologies. One of the main by-products of this interface, the study of
author attribution, is a productive field of inquiry in computer science to
this day (see e.g. Joula 2008, Koppel et al. 2008, Stamatatos 2009). Indeed,
the growing popularity of machine-learning techniques at the turn of the
millennium has allowed for a sophisticated, robust and accurate approach
to this endeavor. As in the case of our opening example, typical features of
language, such as the distribution patterns of function words or content
words, n-grams, lexical categories (technically part-of-speech tags), and
even character n-grams have been used to categorize items by gender, age
group, personality, mother tongue, etc.
There are two obvious advantages to automated text categorization
over the standard (i.e. social science) statistical methods: (a) it is
incomparably faster; and (b) it is bias-free, as evidenced by the fact that its
analyses are completely reproducible. Among other applications, it can be
used to examine problems of interest to scholars in the humanities, and can
be tailored to a broad range of languages, text types, and research
questions. Thus, for example, Koppel et al. (2006) were able to identify the
most likely author of a document of unknown provenance and Strous et al.
(forthcoming) succeeded in discriminating between texts produced by those
diagnosed with schizophrenia and those produced by non-psychiatrically ill
individuals.
In the present study, we set out to explore another categorization task
– the classification of translated texts, focusing on translator-gender
attribution. We will first consider the matter from a standard social-science
perspective. We will base our study on a corpus of documents some of
186 Miriam Shlesinger, Moshe Koppel, Noam Ordan & Brenda Malkiel
3
Elraz found clear differences between male and female interpreters as well, but the
number of participants – 2 males and 2 females – did not allow for any clear
conclusions.
4
Elraz – citing Vásquez-Ayora ( – draws a clear distinction between explicitation,
on the one hand, and addition or amplification, on the other, with the former being a
broader, more “generic” term relating to the need to convey meaning whereas the
latter refers simply to the grammatical constraints of the target language. Her
implication, therefore, is that translations by women reflect a greater concern for the
transfer of meaning.
5
To quote the website: “Words Without Borders (WWB) opens doors to international
exchange through translation, publication, and promotion of the world‟s best writing.
WWB publishes selected prose and poetry on the web […]. Monthly issues of its
online magazine feature new selections of contemporary world literature, most of
which would never have been accessible to English-speaking readers without WWB.”
188 Miriam Shlesinger, Moshe Koppel, Noam Ordan & Brenda Malkiel
male-authored texts are marked with one plus-sign (+). Words relevant
only for female-authored texts are marked with a double plus-sign (++).
The items within each group are sorted according to their frequency in the
texts in descending order.
The mean sentence length in translations by men in our corpus was found
to be much higher than in those by women: 19.18 words per sentence as
opposed to 16.93. This complements the finding that men make greater use
of the logical connectives and, or, but, yet, however. We also found that the
TTR in men‟s translation was higher than in women‟s: 0.052 vs. 0.048,
though this difference is not significant.
To summarize then, we have found a number of indicators of
translator gender. Taken together, these markers seem to suggest that there
are distinctive male and female translator styles.
A
(x1,x2,...,xN)=A
Text (x1,x2,...,xN)=A
(x1,x2,...,xN)=A
Cleaning+ : Learning
Feature :
Extraction (x1,x2,...,xN)=B
Algorithm
B (x1,x2,...,xN)=B
(x1,x2,...,xN)=B
Text
Model for
A vs. B
of points for the particular document is the one to which that document is
assigned.
K-fold cross-validation is used to assess the reliability of the system.
Ten-fold cross-validation is a common choice. It is illustrated in Figure 2:
described above. The second author and his colleagues have performed
classification tasks on sections of the British National Corpus and the
International Corpus of Learner English as well as on e-mail messages,
blogs, nineteenth-century novels, rabbinic literature, and commissioned
essays.
Linguists and computer scientists working in the field of author
attribution have also used machine-learning techniques to discriminate
between texts written by men and those written by women. Recent
publications in this direction include Koppel et al. (2002), and Argamon et
al. (2003). Among other findings, the authors showed that based solely on
function words and parts of speech, they could predict which texts had been
written by men and which by women with accuracy of about 80% in ten-
fold cross-validation tests. In related work (Argamon et al. 2007), similar
features were used to determine an author‟s gender, age, mother tongue,
and personality.
4. Findings
5. Discussion
our ideas and learn to set more specific and local agendas” (Baker 2004:
29), with the influence of gender on translations figuring as one such
specific agenda – one which was the focus of the present study. As Baker
(2004) points out, the models available for systematic analysis do not cover
the full range of features that are proving to be of interest to translation
scholars, in terms of analyzing differences between translated and non-
translated text, with translations being treated here as a kind of genre or
text type. Based on previous studies in both computer science and
Translation Studies, we had expected ten-fold cross-validation to bring us
closer to devising such a model, and discriminating between translations by
men and by women, but found that there is no reliable model available that
will allow us to examine and account for the uneven distribution of these
features across categories. Thus, although we – like Saldanha (2003), Elraz
(2004) and Leonardi (2007) – did find certain features to be significantly
more common in the translations by men or by women, these findings fell
short of providing a profile that might serve for making reliable predictions
about translator gender.6
Our data appear to indicate that whatever the differences between
male- and female-translated documents, they are less robust than those
found in original writing – so much so that even when we controlled for
author gender (i.e. even when comparing male-translated vs. female-
translated texts written solely by males or solely by females), our attempts
to reproduce a more complete profile of the gender distinction failed.
In our view, the primary importance of these findings lies in their
methodological implications. The failure of cross validation to produce a
reliable means of predicting translator gender may call into question the
significance of traditional statistical methods, in which individual
phenomena are selected for testing, and isolated instances of significant
differences are assumed to add up to a means of distinguishing male- from
female-translated documents. More broadly, if indeed the use of cross-
6
Some of our findings, moreover, run counter to theirs; e.g. Saldanha (2003) found
that females use longer sentences than males, whereas our findings point in the other
direction. Whether the different results derive from the methodologies used or from
the corpora is not clear. To resolve the issue we would presumably need to enlarge
and refine our corpora as well as our methods and to include not only lexical features
but others (e.g. word order, morphology, indicators of lexical variety, subtle stylistic
markers etc.) as well.
196 Miriam Shlesinger, Moshe Koppel, Noam Ordan & Brenda Malkiel
6. Conclusions
Both the biblical story and modern author attribution studies present us
with ways of looking at the interface between classes and features, between
the status and social situation of the pre-given social categories and the
ostensibly neutral, indeed technical, elements used to distinguish between
them. The ethnic identity of the Ephraimites was a social fact, a structural
category imposed on individuals by society, a politicized category, with
political consequences. Although the consequences are not a direct result of
the means used to distinguish between them and non-Ephraimites, the
means of categorization proved effective, and may well teach us something
about classes and about identities. The means used in the present study fell
short of providing such a profile. Fortunately, the stakes are less fateful.
7
The present study focused on translation as a product. Triangulation with process-
oriented methodologies may be useful as well in homing in on differences in micro-
level decisions of male and female translators, as revealed in the introspective meta-
translational discourse (e.g. think-aloud protocols), online tracking (through such
programs as Translog), or even eye-tracking (see Göpferich et al. 2008). We hope to
be able to revisit process-oriented studies from the translator-attribution perspective,
and to pinpoint junctures at which the translator's decision-making process may show
consistent gender-specific patterns.
Markers of translator gender: Do they really matter? 197
References
Argamon, S., Koppel, M., Fine, J., & Shimoni, A. 2003. Gender, genre, and
writing style in formal written texts. Text 23 (3): 401–412.
Argamon, S., Koppel, M., Pennebaker, J. & Schler, J. 2007. Automatically
profiling the author of an anonymous text. Communications of the ACM:
119–123.
Baker, M. 1996. Corpus-based Translation Studies: the challenges that lie ahead.
In H. L. Somers (ed.) Terminology, LSP and Translation: Studies in
Language Engineering in Honour of Juan C. Sager. Amsterdam/
Philadelphia: John Benjamins. 175–186.
Baker, M. 2004. The treatment of variation in corpus-based translation studies. In
A. Kruger (ed.). Language Matters: Studies in the Languages of Africa 35
(1): 28–38.
Baroni, M. & Bernardini, S. 2006. A new approach to the study of translationese:
machine-learning the difference between original & translated text. Literary
and Linguistic Computing 21 (3): 259–274.
Biber, D. 1995. Dimensions of Register Variation: A Cross-linguistic Com-
parison. Cambridge: Cambridge University Press.
Chung, C. & Pennebaker, J. 2007. The psychological function of function words.
In K. Fiedler (ed.). Social Communication. New York: Psychology Press.
243–259.
Elraz, I. 2004. His vs. Hers: Does Gender Shape One’s Translation? M.A. thesis.
Department of Translation and Interpreting Studies. Bar-Ilan University.
(Unpublished).
http://www.biu.ac.il/HU/tr/stud-pub/tr-pub/takzir/Elraz,%20Inbal%20.htm.
Göpferich, S., Jakobsen, A. L. & Mees, I. M. (eds). 2008. Looking at Eyes: Eye-
Tracking Studies of Reading and Translation Processing. (Copenhagen
Studies in Language 36). Copenhagen: Samfundslitteratur.
Halliday, M. A. K. and Matthiessen, C. M. I. M. 2003. An Introduction to
Functional Grammar. Hodder Arnold.
Herring, S. 1996. Two variants of an electronic message schema. In S. Herring
(ed.). Computer-Mediated Communication: Linguistic, Social & Cross-
Cultural Perspectives. Amsterdam: John Benjamins. 81–106.
Holmes, J. 1990. Hedges and boosters in women‟s and men‟s speech. Language
and Communication 10 (3): 185–205.
Joula, P. 2008. Author attribution. Foundations and Trends in Information
Retrieval 1 (3): 233–334.
Koppel, M., Argamon, S. & Shimoni, A. 2002. Automatically categorizing
written texts by author gender. Literary and Linguistic Computing 17 (4):
401–412.
Koppel, M., Mughaz, D. & Akiva, N. 2006. New methods for attribution of
rabbinic literature. Hebrew Linguistics: A Journal for Hebrew Descriptive,
Computational and Applied Linguistics 57: 5–18.
198 Miriam Shlesinger, Moshe Koppel, Noam Ordan & Brenda Malkiel
Abstract
1. Introduction
1
The translation segments may be short word spans in the two texts that correspond
closely in meaning, long word spans which correspond only loosely in meaning (for
instance, when content has been shifted because of cultural adaptations), or word
spans in one text which have no natural counterpart in the other text (explicitations
and implicitations). Our notion of translation segments even includes translation
errors whose corresponding text segments are easy to identify in the two texts, but
where the two segments unintentionally differ in meaning or function. Although our
annotation scheme makes it possible to annotate explicitations, implicitations,
translation errors and shifts in meaning, and the resulting annotation can be used to
easily identify these phenomena when they occur, the annotation does not in itself
explain the deeper semantic and pragmatic mechanisms that are responsible for these
phenomena, and these mechanisms are not the focus of our investigation.
200 Matthias Buch-Kromann, Iørn Korzen & Henrik Høeg Müller
2
Because of memory limitations, the brains of real translators are unlikely to store all
the processed parts of a source and target text along with a corresponding bilingual
linguistic analysis. However, we do believe that the brains of real translators are
likely to store locally coherent fragments of bilingual linguistic analyses at each
instance of time, and that, taken together, these locally coherent fragments will
usually (but not always) constitute a globally coherent bilingual linguistic analysis,
which functions as the cognitive justification for our idealised translator.
Uncovering the ‘lost’ structure of translations with parallel treebanks 201
The CDT treebanks are based on the theoretical assumption that human
language users interpret a text compositionally by linking all the
morphemes, words, and discourse segments in a text or spoken discourse
by means of a large inventory of dependency relations, including
complement and adjunct relations, supplemented by a set of secondary
202 Matthias Buch-Kromann, Iørn Korzen & Henrik Høeg Müller
(1) Mary was furious. John’s missing repayment had forced her to give up all she had
worked for. She pulled the trigger. It was payback time.
Figure 1. A complete CDT analysis of the entire discourse (1). The labels are
explained in the following sections.
Uncovering the ‘lost’ structure of translations with parallel treebanks 203
(2) The two convicted executives, chairman Niels Jensen and director Peter Hansen,
appealed their judgment with a demand for acquittal. The chairman received a
year and a half in jail. The bank’s director received 6 months in jail.
204 Matthias Buch-Kromann, Iørn Korzen & Henrik Høeg Müller
The exclusive reliance on a single linguistic level may also blur the insight
that many constructions at one linguistic level have almost identical
counterparts at other linguistic levels, and that the inventory of linguistic
relations should be very similar across the different levels.
Finally, there is a plethora of linguistic theories that could be used to
describe the different linguistic levels. Although these theories tend to be
based on the same underlying principles, these principles are frequently
formulated in ways that may seem almost entirely incompatible with each
other, as dictated by the technical machinery that the theory uses to encode
complex phenomena such as discontinuous word order and secondary
dependencies. While it is obviously impossible to eliminate the reliance on
linguistic theory, the use of a unified annotation scheme for morphology,
syntax and discourse greatly reduces the number of potential
incompatibilities between the different linguistic levels because the
annotation is based on a single, coherent underlying linguistic theory.
In the Copenhagen Dependency Treebanks, the smallest segments in
the text or spoken discourse are morphemes which are linked into larger
units of morphology, syntax and discourse by means of a primary tree
structure supplemented by an inventory of secondary relations. Our claim is
that this inventory of mechanisms is sufficient to give us a unified account
of morphology, syntax and discourse which is theoretically appealing while
providing an excellent basis for time efficient large-scale linguistic
annotation. In the following sections, we will describe our annotation
scheme in more detail.
2.2. Syntax
The arrows point from governor to dependent, with the relation name
written either at the arrow tip, or in the middle of the arrow if a word has
more than one incoming arrow. For example, the arrow from “had2” to “It1”
identifies “It” as the subject (“subj”) of “had”, and the arrow from
“forced3” to “to5” identifies the phrase headed by “to” as the prepositional
object (“pobj”) of “forced”. The primary dependency relations are
organised in a relation hierarchy in which the relations are classified as
either complement or adjunct relations. The syntactic annotation scheme
206 Matthias Buch-Kromann, Iørn Korzen & Henrik Høeg Müller
Table 1. The fifteen most frequent complement and adjunct relations in the
syntactic annotation with examples
tree in Figure 3 in which the relative clause headed by “was7” has been
extraposed from the direct object and placed after the time adverbial
“today5”.
2.3. Morphology
In (3), the Spanish word “antihéroe” is constructed from the root “héroe”
by attaching the prefix “anti” as a “NEG:oppo” dependent of the root. The
NEG:oppo relation indicates that “anti” negates the meaning of “héroe” so
that the new word acquires the opposite meaning of the base.
In (4), the word “repayment” is constructed from the root “pay” by first
indicating iterative aspect by attaching the prefix “re” to “pay”, and then
transforming “repay” into a predicative eventive core noun by means of the
transformative suffix “ment” which takes “repay” as its dependent.
210 Matthias Buch-Kromann, Iørn Korzen & Henrik Høeg Müller
Table 2. The main relation types in the morphological annotation (relation types
with head-switching are italicised)
In (5), the Danish word “arbejdsgiver” is constructed from the verb “give”
(“give”) by turning it into the agent nominalisation (PRED:agent) “giver”,
headed by the transformative suffix “r”, and then specifying the direct
object (patient) role of “giver” by means of the prefix dependent
“arbejde@N” with interfix “s”, where “arbejde@N” denotes the noun
reading of “arbejde”. (In Danish “arbejde” exists both as a noun and as a
verb.)
2.4. Discourse
Table 3. The main relation types in the discourse annotation with examples
(dependent sentence is underlined)
CAUSE:* (cause: We should be ashamed of the state of our schools. We are one of
the world’s richest countries.)
CONC (concession: I know we haven’t known each other for long. Will you marry
me?)
COND (condition: OK, I will help you. On the condition that this never happens
again.)
CONS:* (consequence: In three years many of the DSB ferries will have to seek new
waters. An exciting story will be history.)
CONSOL:* (consolidation: One of the country’s legal experts in cooperative
housing is Svend Trangeled. According to him the greatest problems lie in
partnership arrangements.)
DESCR:* (description: Let’s go to your place, she said. She was different than I
thought.)
ELAB:* (elaboration: She was different than I thought. In the taxi she put her head
on my shoulder.)
INTACT:* (interactional signals: – There is something I would like to ask you. –
Yes?)
PREPAR (preparation, headline: A permanent job. I am so happy that I do not have
to go on welfare again, says Lisa…)
TIME:* (temporal: The Ministry of the Interior expects this part of the job to be
concluded before 2010. Then the Ministry and the national board of health will
consider whether the work is to be conducted further.)
CONJ:* (conjunction: The bank’s director was sentenced to six months in prison.
Two board members were acquitted.)
CONTR:* (contrast: The societal savings are not immediately visible. But it is a fact
that new drugs will minimise the overall health cost.)
DISJ (disjunction: Are you coming up now? Or are you going to stay in bed all day?)
JOINT (no clear discourse relation: We are hungry and are refused by all
restaurants. On a side street a façade is blinking in green neon.)
QUEST:* (question/answer sequence: Why did he say that? I have no idea.)
form node represents the position associated with the node in numerical
contexts and the string associated with node in string contexts, and the
feature node[name] represents the value of attribute name at node.
Uncovering the ‘lost’ structure of translations with parallel treebanks 217
finds all occurrences of the expletive word $X with lemma “there” which
has been analysed as an “expl” dependent of a preceding word $Y. The
DTAG query language makes it easy to search a parallel treebank for
particular phenomena, allowing linguists to find quantitative and qualitative
corpus-based evidence for linguistic hypotheses that would be difficult to
investigate otherwise.
By means of the DTAG query language, the CDT treebanks can be used to
prove or disprove a wide range of linguistic hypotheses. The following two
sections outline such hypotheses, one for NP structure and one for
discourse.
4.1. NP structure
Derivation → compound
(7a) escritorio – skrivebord ‘writing desk’ (escribir: ‘write’)
(7b) dentadura – tandsæt ‘set of teeth’ (diente: ‘tooth’)
(7c) petrolero – olietankskib/oliehandler ‘oil dealer/oil tanker’ (petróleo: ‘oil’)
4.2. Discourse
(12) Da han kom til Rom, blev han syg – Arrivato a Roma, si ammalò. [participle]
‘When he came to Rome, he became ill’ – ‘[Having] arrived in Rome, he
became ill.
(13) Nato nel 1947, ... – Han blev født i 1947, og … [finite verb]
‘Born in 1947, …’ – ‘He was born in 1947, and …’.
Uncovering the ‘lost’ structure of translations with parallel treebanks 221
The CDT annotation will provide very precise answers regarding the
frequency of such changes between verb forms (as well as changes between
word classes) and the contexts in which they occur, and will also inform us
whether particular verb types, text types or discourse relations occur more
frequently than others in specific morphological data.
Another important difference between the Scandinavian and
Romance languages, linked to differences in sentence length, is the
different use of paragraph change and punctuation marks. Together with
longer sentences, there seems to be a tendency, at least in some text types,
also to use longer paragraphs in Romance languages while Scandinavian
texts tend to subdivide the content into smaller units. Furthermore, the
higher proportion of long sentences can be linked – among many other
things – to a much more frequent use (especially in French and Italian) of
the colon and semicolon in cases where a Danish text would select a full
stop. But these phenomena need much more investigation, and therefore
the CDT discourse annotation distinguishes between different kinds of
sentence and main clause adjoining. If a full stop coincides with a change
of paragraph, the sign + is added to the name of the discourse relation. A
very frequent relation is the simple conjunction, cf. Table 3, so that a
relatively common discourse relation label is “+CONJ/(and)”. Usually a
connective is not rendered explicit in these cases, hence the parentheses
around “and”.
Differences in the use of punctuation marks will automatically be
detected in CDT, as the system includes an individual annotation of
punctuation marks at the syntactic level, cf. Figure 1 (units 4, 20, 25 and
30), and the cross-linguistic alignment in Figure 5 (units 5/15 and 10/21).
But in addition to that, since the use of colon and semicolon varies
particularly between the Scandinavian and Romance languages, CDT
discourse annotation adds a colon or semicolon to the relation in cases
where two main clauses are separated by a colon or a semicolon, e.g.
“:CONJ/(and)”, “;CONJ/(and)”.
A final issue concerns differences in word order, where a particularly
interesting phenomenon is the Romance tendency to begin sentences with a
time, place or modal adverbial vs. the Danish tendency to begin sentences
with the subject – a phenomenon which to our knowledge has not yet been
investigated. Such cases can be detected effortlessly by means of the CDT
222 Matthias Buch-Kromann, Iørn Korzen & Henrik Høeg Müller
annotation and will therefore be easy to search for and analyse both
quantitatively and qualitatively once the treebank annotation is complete.
5. Conclusion
In this paper, we have described the main aspects of the annotation scheme
in the Copenhagen Dependency Treebanks, and demonstrated that a unified
scheme solves many of the consistency problems that arise if the
annotations are performed separately. We have argued that treebanks are an
important resource for research in linguistics, psycholinguistics and
translation studies, and outlined several linguistic hypotheses about
translations and contrastive linguistics that can be explored with the CDT
treebanks.
6. Acknowledgments
This work was supported by grants from the Danish Research Council for
the Humanities and the Copenhagen Business School. The translation work
has been kindly co-sponsored by the Danish translation company Lingtech.
We thank Lotte Jelsbech Knudsen, Morten Gylling, and the other CDT
annotators for many fruitful discussions.
References
Michael Carl
[B]y testing hypotheses based on qualitative data against quantitative data, and vice
versa, I believe we can soon begin to make stronger and more informed guesses about
translation. (Jakobsen 1999, 19)
Abstract
1 Introduction
225
226 Michael Carl
Figure 1. The Figure represents the dimensions of the User Activity Data (UAD): on the left
the product data and on the right side the process data. Both types of data provide different
aspects of the translation session.
represents the relation between process and product data as it is currently pro-
duced by Translog. The left side shows the static part of the UAD, the source
text (ST) and the target text (TT), together with their correspondence links.
The right side plots the dynamic process data which consists of keyboard
actions and gaze behaviour. According to Jakobsen (2006, 104):
The keystroke data can be interpreted quantitatively and unambiguously in terms of
either text production, text elimination or cursor navigation [. . . ] actions. [. . . ] This
means that the keystroke record is also immediately interpretable in terms of linguistic
units (letters, words, clauses, etc.) giving the researcher full access to studying the
typing process by which a text is produced.
and AUs can be directly observed, discussed and in most cases agreed upon
between different annotators. But most importantly, AUs do not change in
time and do not have the dynamic aspect that TUs have.
A large number of trainable alignment programs exist (e.g. Dagan et al.
1993, Och and Ney 2000, Kromann 2003, to name a few) which can be used
to (semi-) automatically align segments of a ST with their correspondences in
a TT. An isomorphism between AUs and TUs is often assumed in a machine
translation context, where aligned texts serve as a basis for the extraction of
TUs to be subsequently reused in the machine translation system.
It is, however, unclear whether and to what extent such an isomorphism
also exists in the case of human translation. One attempt to define TUs in hu-
man translation process data has been undertaken by Dragsted (2004). With
the intention to detect TUs on the basis of long pauses between keystrokes,
Dragsted (2004, 142) states that “[i]ntuitively, the TU must contain both SL
and TL elements”. She comes to a number of conclusions about the length (in
words) of TUs under different conditions. However, a differentiation between
TUs and AUs might be instrumental in dealing with inconsistent segmenta-
tion in the product and process data. For instance, the Danish-English AU:
“er det planen ↔ we expect” shows a long translation pause in the process
data between “we” and “expect”, which indicates incompatible segmenta-
tion boundaries in both types of data. A distinction into separate TUs on the
process side and AUs on the product side would allow us to correlate and
“triangulate” these units without the need to give up or change definitions.
A systematic triangulation of TUs and AUs is beyond the scope of this
study. Rather, we shall prepare the ground for such an investigation and
correlate and quantify AUs as detected in the product data with keystroke
data. In order to do so in a systematic and automatic manner, we need to:
1. add further alignment information to define the AUs. We use the DTAG
software package2 which allows us to semi-automatically align the ST
and TT;
2. transform the Translog product and process data into a representation
which better reflects the phenomena which we want to investigate;
3. use a query language to retrieve all keystroke activities for each AU.
2
DTAG http://code.google.com/p/copenhagen-dependency-treebank/ is a project
aimed at creating linguistically annotated text collections (treebanks) on the basis of the dependency-based
grammar formalism Discontinuous Grammar. A number of mono- and bilingual semi-automatic annotation
and visualisation tools can be downloaded from the web site.
Triangulating product and process data 229
Figure 2. The ST representation includes word information (i.e. the surface form of the
word), information on word location on the screen (i.e. the top-left and bottom-right pixel
positions of the box containing the word) and the word position in the text (i.e. the cursor
position of the word’s first letter).
<W cur="0" top="511" btm="544" lft="21" rgt="69">Selv</W>
<W cur="5" top="511" btm="544" lft="77" rgt="115">om</W>
<W cur="8" top="511" btm="544" lft="123" rgt="332">udviklingslandene</W>
<W cur="26" top="511" btm="544" lft="340" rgt="459">forståeligt</W>
<W cur="38" top="511" btm="544" lft="467" rgt="510">nok</W>
<W cur="42" top="511" btm="544" lft="518" rgt="542">er</W>
An exhaustive fragmentation of the product data into AUs would then also
provide a framework for the study of the process data, since each keystroke
may be considered to contribute to one (or more) AU(s). The number and
distribution of keystrokes which fall into each AU constitute a novel way of
classification and a yet unstudied field of research.
In Section 2, we describe how we transform the Translog product data
into a format that makes it better suitable for the triangulation we are aiming
at. Section 3 describes how we transform the Translog keystroke data into
a format for better interrogation. The aim is to represent UAD in a kind of
relational data structure which allows us to automatically detect and correlate
the various dimensions of the product and process data. Section 4 outlines a
query language that allows us to relate the different dimensions of the data
in a formalised and exhaustive manner. Our particular aim is to correlate and
quantify AUs with keystroke data. Section 5 describes a special operator of
the query language which makes it possible to retrieve all keyboard activities
which belong to a given TT string. With these concepts we are ready to
discuss the triangulation of AUs and keystroke data in Section 6.
2 Product data
UAD into a representation that can be more easily correlated. The product
data consists of the ST and the TT, which can be extracted from the Translog
data.3 The representation we are aiming at is shown in Figures 2 and 3. In
addition we need alignment information to know which pieces in the SL text
correspond to which pieces of the TL text. We represent alignments as shown
in Figure 4.
A2 K M S
Number of words in TT 102 101 112 101
Aligned words in TT 97 92 106 98
Aligned words in ST 100 97 98 96
Number of AUs 82 71 80 69
Table 1. Summary of properties of product data: number of words in the translations, number
of aligned words and number of AUs. The source text has 100 words, see Appendix.
3
The word location information in Figures 2 and 3 is only available in the Translog ‘premium’ version
(http://www.translog.dk/), which is also able to collect gaze data. In the ‘ordinary’ version, only the
cursor information is available which has to be extracted post-hoc from the log file.
4
The ST is reproduced in the Appendix
Triangulating product and process data 231
The translations were produced by professional translators and the texts were
manually aligned by three translation students with no particular instruction
as to how alignment should be done. The alignments were not analysed in
detail. The difference in number of AUs reflects their average number of
words, and may indicate a different perception of compositionality. Each
translation alignment took approximately 30 minutes and was in most cases
straightforward.
3 Keystroke data
As outlined in Jakobsen (1999, 2006) Translog logs all keyboard and mouse
activities. Different keyboard and/or mouse actions can be used for the inser-
tion and deletion of text, and for cursor navigation. For instance, deletions
can be achieved through backspace, or by pressing the delete key. A piece
of text can also be deleted by first highlighting it and then hitting the dele-
tion or backspace key or even by overwriting the highlighted sequence with
other characters. Highlighting of the text can, again, be achieved by various
means, e.g. by using the mouse button or by a combination of using shift and
left, right, up, or down keys, or even pressing shift control and left, right, up,
or down keys, etc. All these events are coded differently in Translog output.
The full information is needed to replay the translation session in a natural
way.
For an analysis of how a translation emerges, a more reduced representa-
tion is sufficient. We are likely to be not so much interested in how the cursor
was moved to a particular position in the text, but rather that it was there at a
particular point in time. We may also be not so much interested in knowing
whether a sequence was deleted by first highlighting it using the mouse or a
combination of keystrokes, or whether it was deleted using repeated delete
or backspace keys. Rather, for our intended analysis it is sufficient to ob-
serve that the sequence in question was deleted. For knowing what happens
to the text, we are basically only interested in knowing the insertions and the
deletions that take place at any point in time.
Figures 5, 6, and 7 show examples of the simplified key logging in-
formation: each line represents either an insertion or a deletion at a given
time (in ms, from the start of the translation session) and a cursor position in
the text where the action takes place. In the sequence of keystrokes in Fig-
ure 5, first the word “bør ” was written, then it was deleted and replaced by
232 Michael Carl
Figure 6. The example shows a correction pattern “ m[ der]å der” to insert an omitted
“å”. Single keystrokes are used rather than an entire block as in Figure 5. Curly brackets
around “{å der}” as well as the interruption symbol in an immediate correction may be left
out if they do not increase readability.
Figure 7. This is a delete & paste operation in which the word “nødt ” is moved from
cursor position 6 to cursor position 11.
Triangulating product and process data 233
“må ”. The caption of the figures also show the writing progression in the
S-notation as introduced by Kollberg (1998) and Perrin (2003). Notice that
the cursor positions for “b” and “m” are identical and the deletion operation
returns the cursor to the position where it was previously. Figure 6 shows
an example where an omitted “å” is inserted after “m” to produce “må” by
deleting the first four letters (including a blank space) of the following word
“derfor” that had already been typed.
Translog also allows the translator to use the clipboard, i.e. to copy a text
fragment into a buffer, and to paste it somewhere else in the text later. If the
copy operation is linked to a deletion of the fragment from the text (i.e. using
ctrl X), the action is represented as a deletion operation, while the pasted
fragment is represented as an insertion operation, as shown in Figure 7. In
order to keep track of the text’s origin, additional information is required with
the pasted segment. The time stamp of the copy operation (time=52551)
serves as a reference together with its starting and end cursor positions. This
allows us to trace back keystrokes of the moved segments, as outlined in
Section 6.
A2 K M S
insertions 1219 1361 1316 1579
deletions 118 112 261 204
Table 2: Summary of properties for keystroke data, number of insertions and deletions
4 A query language
Table 3. The rule prints out keystroke patterns with less than 500ms between successive
keystrokes.
A rule may consists of several successive patterns. In this way several dimen-
sions of the UAD can be correlated and triangulated. Basically, each pattern
matches successively every node in the data dimension and instantiates the
variables with the retrieved values. If a pattern successfully matches a se-
quence of nodes, the rule switches to the next line until either a pattern fails
or the end of the data was reached. The rule then backtracks and searches for
the next pattern in the previous line.
Thus, the rule in Table 4 retrieves all keystrokes that are related to the
production of the translation for “developing”. The pattern in line 1 itera-
tively matches all instances of “developing” in the ST and instantiates the
variable $A1 with the cursor positions. The pattern in line 2 retrieves the
cursor position of the translation via the alignment data (A) and the pattern in
line 3 retrieves the word form (e.g. “udviklingslandene”) of the translation
from the TT (T) and stores it in the variable $V1. The variables $A1 and $A2
Triangulating product and process data 235
1 S>[str.eq.developing,cur.V.$A1]
2 A>[src.eq.$A1,tgt.V.$A2]
3 T>[cur.eq.$A2,str.V.$V1]
4 !:InitKeyRange(’Key’, $A2-1, $A2+length($V1));
5 K<kp[cur.val.Key]
6 !:PrintPattern(’kp:time,type,cur,str’);
Table 4: The rule prints all keyboard activities related to the translation of “developing”.
thus contain the ST and TT cursor positions of an AU, while the variable $V1
contains the TT part of the AU.5
The keystroke pattern (K) in line 5 looks backwards into the keyboard
data and marks all keystroke nodes with the marker kp which are part of the
AU sought. A special operator val is used to keep track of the beginning
and end cursor positions in the TT segment. The beginning and end cursor
positions of the TT segment are initialised with a function:
InitKeyRange(’Key’,$A2-1, $A2+length($V1))
PrintPattern(’kp:time,type,cur,str’);
in line 6 prints out the marked keyboard nodes. In this way, the rule retrieves
all keyboard activities which belong to the production of the translation for
“developing”.
AUs are the smallest units which connect the ST and the TT, and keyboard
logging data reveals how AUs emerge in time. In Section 5 we have discussed
how process data can be assigned to TT segments C. In this section we look
at triangulation of AUs with keyboard data. Table 1 in Section 2 shows that
almost the entire ST and TT are fragmented into AUs for the four transla-
tions, except for a small number of words which remain unaligned. Since we
assume that every keyboard activity is driven by the goal to produce a trans-
lation of the ST, we can now break the set of keystrokes down into smaller
parts and investigate the new properties of AUs independently.
Figures 8 and 9 show how AUs emerge in time. The graphs reveal where
there are pauses and deletions, and how keystrokes and gaze activities are dis-
tributed over time. It gives a general picture of how the translation develops
by relating each activity to the ST unit which it translates. We consider the
following keystrokes to be part of an AU:
Figure 8. The graphs represent translation progression patterns of 2 subjects, A2 (left) and M
(right). On the Y-axis the words of the ST are enumerated (1 to 100). The Y-axis gives pro-
duction time of the translation in ms. The dots indicate different kinds of keyboard activities
(insertions and deletions). The line represents gaze behaviour on the ST words. Fixations on
the TT window are indicated with a triangle symbol on line 100. Notice that M needs more
than twice as much time as A2, half of it due to revision. The dots on the 0-line indicate
keyboard actions which could not be attributed to any AU.
Triangulating product and process data 239
Figure 9. The graphs represent translation progression for K (left) and S (right). Both
needed approximately the same time to translate the text. There are long stretches of time in
S’s translation progression graph where no keyboard and gaze activities are recorded. It is
unclear what happened during this period of time.
240 Michael Carl
A) [udvi] ud[ivk]vi[lk][l]klingslandene
Table 5. The example plots correction patterns and writing progression of AU2 .
A) shows the S-notation and B) illustrates 9 correction steps in the development of
“udviklingslandene”.
This definition does not distinguish between types of keyboard actions such
as correcting typos, changing inflection or substituting lexemes or even
changes as a consequence of reconsidering the syntactic structure of the trans-
lation. While such a classification is beyond the scope of this paper, this def-
inition has the advantage that (almost) all keystrokes can be associated with
particular AUs.
The example in Table 5 shows the keystrokes involved in the production
of the AU2 : “developing countries ↔ udviklingslandene”.
The example represents words 2 and 3 of the ST (see Appendix) and the
product data is represented in Figures 2, 3 and 4. Table 5 shows a number of
(immediate) corrections and their representation in S-notation as outlined in
Figure 5: characters in square brackets represent deletions and the underscore
“_” represents a blank space. Indexes represent temporal order of corrections
and identical indexes indicate closeness in time.
Table 5 shows the development of “udviklingslandene”, looking
backwards into the keyboard data. It consists of 9 steps, of which the last
corresponds to the final form as observed in the final translation. In steps 5 to
8 the characters “kl” are accidentally reversed and corrected. Steps 3 and 4
show a similar procedure with correction of reversed letters “vi”. At step 2,
the entire sequence “udvi” is deleted, in order to insert a blank space before
the word which had been omitted in step 1.
We incorporate steps 2 and 1 into AU2 , since we consider the keystroke
immediately preceding the TT sequence to be part of the AU. Otherwise
these keystrokes could not be assigned to any AU. For the same reason
all the keystrokes in Figure 5 are part of the same AU, and the keystrokes
“_m[_der]å” in Figure 6 are part of another AU.
A more complex keyboard pattern can be observed during the produc-
tion of AU “go ↔ at iværksætte” for word number 38 of the ST in
the Appendix. The example is taken from translator M (Figure 8) where 36
Triangulating product and process data 241
Figure 10. These graphs represent properties of the AUs and their associated keyboard ac-
tivities for the four subjects {A2, K, M, S}. The horizontal X-axis lists the words in the
ST. The vertical Y-axis in the top graph shows the number of keystrokes/length of the AU
translation. A value of 1 means that the length of the AU translation equals the number of
keystrokes (i.e. no corrections take place during the production of that AU). A value > 1 in-
dicates deletions. For instance, at word number 21 (translation of “on”), translator M needs
17 times more keystrokes than the length of the translation.
The bottom figure plots the length of the translation for each ST word. It shows, for
instance, that the translation of word 21 (in the M data) is one character long. It also shows
that translations of most words have similar length, across the four translations. Note that
some of the divergences might be due to inconsistencies in alignment strategies, e.g. words
7 to 11 and 95 to 99.
AUs can be recognised through the same length and/or number of keystrokes. For
instance, ST words 14 to 16 are part of the same AU. In each translation they have the same
length (bottom) and the same number of keystrokes (top), which indicates that these words
are clustered into one AU.
242 Michael Carl
For instance, the following keystrokes are part of such unclassified activities.
The segment was first inserted at the end of the text, then modified several
Triangulating product and process data 243
times (as can be seen through the square brackets) and finally replaced by
“følges ad”. It is the last AU from the data of M in Figure 8:
[bliver derfor nødt til[at ] a gå hånd i hån]
Figure 11. These graphs show the time delay between successive keyboard activities of the
four subjects. The X-axis plots the words in the ST. The Y-axis in the top Figure shows aver-
age intra-word keystroke time span between two successive keystrokes within the production
of one AU. The bottom Figure shows inter-word keystroke time delay between the end of
an AU and the start of typing of the next AUs. Notice that the time scale is one magnitude
higher for inter-AU pauses than for intra-AU pauses.
keystrokes of the length of the final translation. It is obvious that there are
parts which are easy for all translators, i.e. where the keystoke/length ratio
is 1 for everyone, and that there are parts which are ‘difficult’ for everyone.
A similar, although not so clear picture emerges in the inter- and intra-word
keystroke time span in Figure 11. Also here, much time is spent during the
translation of words 19 to 30, 71 to 85 and 91 to 99, particularly on inter-word
time spans as shown in the lower graph, but other factors seem to interfere.
For instance, much time is also spent around the translations of word 45, for
both inter- and intra-word keystrokes, but this obviously did not have a high
impact on correction patterns (Figure 11). We hope to investigate this further
and to relate these quantitative findings to the linguistic properties of the AU.
7 Discussion
TT. The notion of Alignment Unit (AU) allows us to fragment the product
data into smaller segments which can be quantified and triangulated with
process data. While a general theory of human translation founded on trans-
lation process data may perhaps not be reached in the near future, we show
a way of correlating and modelling the data in a quantitative manner. At this
point we are still far from being able to formulate conditional probabilities
over Process Data (PD) and AUs (and maybe also Translation Units) which
would answer questions such as: given a history of PD what is the probability
of the next AUi to be deleted, inserted or modified? The extent to which we
can answer this and related questions will ultimately determine the success
of integrating advanced translation aids with human translation activities.
Flick (2008, 20) points out that triangulation is to be understood as a
“strategy on the way to a deeper understanding of the investigated object and
thus a step towards more knowledge and to a lesser extent to validity and ob-
jectivity of the interpretation” (my trans.). And so we hope that triangulating
product and process data may contribute to a deeper understanding of human
translation processes.
246 Triangulating product and process data
References
Alves, F. and Vale D. C. 2009. Probing the Unit of Translation in time: aspects of
the design and development of a web application for storing, annotating, and
querying translation process data. Across Languages and Cultures 10(2): 251-
273.
Carl, M., and Jakobsen A. L. Forthcoming. Towards statistical modelling of
translators’ activity data. International Journal of Speech Technology.
Carl, M., Jakobsen, A. L. and Jensen, K. T. H. 2008. Studying human translator
behavior with user-activity data. In Proceedings of the Natural Language
Processing and Cognitive Science (NLPCS) at the International Conference of
Enterprise Information Systems (ICEIS), 12–16 June 2008, Barcelona, Spain.
114–123.
Dagan, I., Church, K. W. and Gale, W. A. 1993. Robust bilingual word alignment
for machine aided translation. In Proceedings of the Workshop on Very Large
Corpora: Academic and Industrial Perspectives, 22 June 1993, Columbus,
Ohio. 1–8.
Dragsted, B. 2004. Segmentation in Translation and Translation Memory Systems.
Dissertation, Copenhagen Business School.
Flick, U. 2008. Triangulation: Eine Einführung. Wiesbaden: VS Verlag für
Sozialwissenschaften.
Jakobsen, A. L. 1999. Logging target text production with Translog. In G. Hansen
(ed.) Probing the Process in Translation: Methods and Results (Copenhagen
Studies in Language 24). Copenhagen: Samfundslitteratur. 9–20.
Jakobsen, A. L. 2006. Research methods in translation – Translog. In K. P. H.
Sullivan, and E. Lindgren (eds). 2006. Computer Keystroke Logging and
Writing: Methods and Applications. Oxford/Amsterdam: Elsevier. 95-105.
Kollberg, P. 1998. S-notation – A Computer-based Method for Studying and
Representing Text Composition. Licenciate thesis, Royal Institute of Techno-
logy, Stockholm.
Krings, H. P. 2005. Wege ins Labyrinth — Fragestellungen und Methoden der
Übersetzungsprozessforschung im Überblick. Meta 50(2): 342–358.
Kromann, M.T. 2003. The Danish dependency treebank and the DTAG treebank
tool. In Proceedings of the 2nd Workshop on Treebanks and Linguistic
Theories, 14–15 November 2003, Växjö, Sweden. 217–220.
Och, F. J. and Ney, H. 2000. Improved Statistical Alignment Models. In
Proceedings of the 38th Annual Meeting of the Association for Computational
Linguistics (ACL), Hongkong, China. 440–447.
Perrin, D. 2003. Progression analysis (PA): investigating writing strategies at the
workplace. Pragmatics 35: 907–921.
Triangulating product and process data 247
Appendix
The ST consists of the following 100 words. For technical reasons punctua-
tion marks were deleted from the text but are shown here in brackets (·) to
facilitate reading:
EYE-TRACKING
STUDIES
Eye tracking in translation process research:
methodological challenges and solutions
Sharon O’Brien
Abstract
1. Methodological challenges
Equipment
At the time of writing, the Tobii brand of eye tracker (1750, T60 or T120
models) was the most popular model in Translation Studies. This eye
tracker is a desktop eye tracker with an in-built camera and infrared diodes.
Its advantage is that it resembles a normal computer monitor and is
therefore quite “uninvasive” for research participants, who are physically
unaware that their eyes are being tracked (the participants would of course
be made aware in advance that their eye movements were being tracked,
see “Ethics” below). This type of eye tracker increases the ecological
validity for the type of research projects conducted in Translation Studies
and it is generally deemed to be more advantageous than the head-mounted
eye tracker1 because it does not burden the participant with physical
equipment on the body. However, there is a trade-off between accuracy and
validity here since head-mounted eye trackers are generally considered to
be more accurate than desktop ones.2
Some of the challenges associated with using an eye tracker include
(i) cost, (ii) rapid technological redundancy and (iii) learning curve. To date
some translation research centres have managed to acquire funding to
purchase an eye tracker. However, this technology is relatively expensive
and funding is difficult to attain. The analysis software, which is used to
analyse the data generated during eye-tracking sessions, may be frequently
updated or, indeed, completely redesigned and there is sometimes a lack of
backwards compatibility, meaning that eye-tracking sessions recorded in
new versions cannot be viewed or analysed in older versions and vice
1
Examples of current head-mounted eye trackers are the Arrington Research Mounted
Eye Tracking System (http://www.arringtonresearch.com/headmountframe.html) or
the ASL EYETRAC®6 head mounted optics (http://www.a-s-
l.com/site/Products/EYETRAC6Series/HeadMounted/tabid/57/Default.aspx) [Last
Accessed 12/05/09].
2
For example, the Tobii 120 model can collect data at a rate of 120 Hz whereas the
ASL head-mount model can collect data at 120, 240 and 360 Hz, according to the
technical specifications on each company’s website.
Eye tracking in translation process research 253
versa. A steep learning curve can also be associated with the use of the
analysis software.
Accommodation
Once an eye tracker has been acquired, it needs to be housed somewhere.
Eye trackers do not require vast amounts of space: a normal desk will
suffice. The challenge is to find a space which meets specific requirements
for conducting eye-tracking research. In particular, consideration must be
given to light, sound and familiarity of the surroundings. If, as is the case
for many translation process researchers, one is interested in measuring
cognitive effort, then recording pupil dilations will be of particular interest.
However, pupil dilation is known to be influenced by many factors, such as
changes in light intensity, sound, caffeine, drugs, substance abuse, eye
colour and even heavy eye make-up. It is highly recommended, therefore,
that the lighting in the room where the eye tracking research takes place is
controlled. At a minimum, blackout blinds and a consistent source of
lighting are recommended. It has also been shown that the pupil reacts to
sound so it is logical to control for this. Since soundproofing a room is
beyond the means of most translation research groups, one should attempt
to house the eye tracker in a quiet location and/or to conduct sessions when
noise interruptions are least likely. Other factors, such as caffeine levels,
substance abuse, eye make-up etc. are more difficult to control but
participants can be made aware that these factors can influence
measurements and can be asked to avoid caffeine prior to sessions or not to
self-select if they have a record of substance abuse, for example.
Familiarity
Translation process studies, including those using eye-tracking equipment,
frequently use professional translators as “research participants” in a bid to
uncover professional work practices, strategies and differences between
professionals and novices, among other things. This raises a challenge for
ecological validity: we wish to observe what professional translators
“normally” do, but we remove them from their “normal” work
environments in order to do so. The fact that the eye-tracking monitor is
most likely different in shape and size from their usual monitor, or that the
operating system, software, version numbers, language packs, screen layout
254 Sharon O’Brien
or even keyboard type differ from their usual work environment may have
an impact on their performance. The research community should not
abandon research because of these challenges, but, where a “normal work
environment” is an important factor, the eye-tracking environment ought to
be set up in such a way that the participant is familiar and comfortable with
it.
Terminological considerations
Different terms are used in translation process research to designate people
who earn their living by translating and people who are learning to do so.
The first category is typically termed “professional”, a wide variety of
names has been attributed to the second category, e.g. semi-professional,
novice, student. Although the term “professional” is used consistently to
refer to somebody who is a practising translator and not a student, the
scope of the term varies. A professional could be a person who has more
than 10 years of experience working full-time as a freelance translator or it
could mean a graduate who has worked for less than one year in a
translation agency. Therefore, when we make claims about how
“professionals” operate, we are making claims about people with very
different levels of experience. It should be acknowledged that even if the
behaviour of two professional translators each with 10 years of experience
was compared, it would not be surprising to see evidence of quite different
behaviour. We are, after all, observing humans. Therefore, the research
domain would benefit from agreeing criteria for the term “professional” so
that we can make valid cross-study comparisons.
As mentioned, the term used for people who are studying to become
translators also varies. The validity of the term “semi-professional” to
Eye tracking in translation process research 255
Numbers
Since translation process research relies on human participants, one of the
challenges is including an adequate number of participants who fit the
specific profile we are hoping to investigate in order to enable researchers
to make generalisable claims. If we wish to include professional translators
in a study, then funding is required to pay such participants. Some
professional translators will give freely and generously of their time.
However, this raises questions about ecological validity: if somebody is
donating their time freely in the name of research, one cannot be certain
that they will behave “normally”, i.e. as one does in everyday translation
work. Admittedly, it is also difficult to prove “normal behaviour” even
when a person is being paid for their participation. However, it is
reasonable to expect that they would behave normally if they are being
paid. A certain aspect of the “normality” is of course compromised because
the translator knows that the researcher is not a “real” client and that they
have little to lose if they do not perform well. This is a methodological
concern which has no easy solution.
Given limited funding, it is difficult to recruit adequate numbers of
participants to make generalisable claims. Some studies have included five
subjects, others ten, others 20. For example, in the research reported in a
recent volume of the Copenhagen Studies in Language (Göpferich et al.
2008), the average number of participants in eye-tracking studies of
translation was 12. Making valid generalisations on the basis of such
numbers is questionable. Nonetheless, these studies have significant value.
First, they are pioneering in their use of eye tracking in translation studies.
Researchers are testing equipment, research designs, and methodologies.
Secondly, such studies are valuable for generating hypotheses using small
communities which can then be tested on larger communities.
256 Sharon O’Brien
Competence
Let us assume for one moment that research funding is abundant and we
can afford to recruit, say, 50 professional translators and 50 student
translators for a comparative study of translation strategies, cognitive load
or translation competence. Let us also assume that we have agreed rigid
definitions for our categories of “professional” (e.g. must have worked full-
time as a translator in a specific specialised field for a minimum of 10
years) and “student” (must be in their second semester, at post-graduate
level, with an average score in economic translation of 65 % and no fails in
any subject). Can we assume that the 50 professional translators have equal
competence or that the 50 students have equal competence? Our experience
as translator trainers allows us to reasonably expect diverse levels of
competence in different aspects of translation (e.g. SL comprehension, TL
production, specialised domain knowledge, etc.) within each of these
groups. The first challenge for translation process researchers is to
acknowledge this diversity. We cannot claim that our participants are all
equal. At the same time, having diversity within one group does not
invalidate our studies. By having greater numbers, we can both expose and
explore the diverse behaviour and competences and look for commonalities
among the members of specific groups. The commonalities then allow us to
make general claims about the nature of the translation process.
Eye tracking in translation process research 257
When it comes to using eye trackers, this diversity is all the more
apparent. Even if we can build a group of research participants using tight
control criteria, we are still faced with individual differences that have an
impact on the study. Participants may have individual eye conditions,
which reduce the quality of the data collected by the eye tracker. For
example, it is claimed by the manufacturers of eye-tracking equipment that
spectacles and contact lenses do not hinder eye tracking. In reality, the
quality of eye data can be better for one person when wearing spectacles
and for another when wearing contact lenses. Dark-coloured eyes may also
have an impact on the ability of the eye tracker to differentiate the pupil
from the iris. On some occasions, the quality of eye-gaze data might be so
low that participants’ data cannot be used in a study.
3
What qualifies as “significantly high” is not yet agreed among translation-process
researchers. The threshold set for data quality will depend on many factors, such as
the general objective of the study, how many participants one has, other measurement
criteria being used, etc. As a very general and somewhat simplistic guideline, if less
than 70 % of the time is spent looking at the screen, the researcher might want to
consider removing that participant from the study. However, this could be seen as a
lowest limit threshold.
258 Sharon O’Brien
that participant’s data has been analysed and it becomes clear that
something went wrong during the session. Ironically, in this author’s
experience, this happens less with students and more with professionals.
1.3 Ethics
It is more and more the case that researchers wishing to recruit human
participants for research purposes, even in Translation Studies, are required
to obtain ethical approval from their organisation’s ethics committee. While
this might initially seem unnecessary, it is a very important step in research
preparation. The ethics approval procedure requires the researcher to state
in a comprehensible way the purpose of the research and precisely what is
required from participants. A researcher could be tempted not to inform
participants that their eyes are being tracked, in order to stave off the
“white coat effect” mentioned above, but this would be considered
unethical by many.
The ethics approval documentation also sets out participants’ rights,
i.e. (usually) that they will not be named, that their data will be protected
and stored securely and that they can withdraw from the study at any stage
without penalty. The latter right is of special importance for the two groups
of participants that are commonly used in translation process research. In
the case of student translators, especially when the researcher is also a
teacher of the participant, it is important for the participant to know that
removing themselves from the study will have no impact on other student
activities.
Professional translators may also wish to remove themselves from a
study for a number of reasons. If the professional’s relationship to the
researcher is one-to-one, this is not such a problem (except for the
researcher who has lost a precious participant). However, this raises an
interesting dilemma if the professional translator is a sub-contractor to an
industrial sponsor of the research and the industrial sponsor is paying for
the professional’s time. Under normal working conditions, there might be a
penalty if translators removed themselves from a commercial project prior
to its completion. On the other hand, when that “project” is research being
paid for commercially and the researcher is obliged to say that the
participant can withdraw at any stage without penalty, then there is clearly
260 Sharon O’Brien
produces a veritable sea of data: there are gaze plots showing the number
and sequence of fixations, hotspots showing the areas on the screen that
were most frequently fixated, video files showing the eye gaze, reading and
text production data which are sometimes overlaid with concurrent or
retrospective protocols, and each eye-tracking recording also produces a
very large data file with millisecond-based data on the position of the eyes
according to the X, Y coordinates of the screen, the left and right pupil
sizes, the validity of the data at any point in time, the fixation number, etc.
If a keyboard-logging tool such as Translog is used in conjunction with the
eye tracker, then researchers also have at their disposal a Translog replay
file and a log file containing information about all the keyboarding and
pause activity. All of this is produced per person, per recording. On the
positive side, researchers have a rich set of data at their disposal from
which conclusions can be drawn. However, the effect on the lone
researcher, even with a small number of participants, can be overwhelming.
How do we scale this up to include larger numbers of participants? While
eye-tracking analysis programs such as Tobii’s ClearView or Studio
automate some aspects of the data analysis, it does not automate everything
(e.g. pupil dilation calculations are still done “manually”, i.e. the researcher
has to figure out a way of slicing data and creating macros in a program
such as MS Excel). Suggestions on how to manage the data explosion are
given in “Addressing the Challenges”.
1.5 Validity
for example) and this makes it impossible to select long texts. The
challenge for translation process studies lies in the fact that, with the
exception of the localisation industry, translators normally work with larger
chunks of text and, therefore, one could argue that we are not investigating
“real scenarios”. The validity of our research is, therefore, compromised.
Secondly, how participants are translating texts raises concerns about
validity. Translators might spend most of their day working in MS Word,
or MS Excel or in a particular Translation Memory environment. They use
the World-Wide Web for research purposes, but also resort to their
customised glossary and their hardback dictionary. They then volunteer for
a research study and are asked to translate, without using the Web, a
dictionary, or their own glossary, a text type they normally do not work
with in a tool they have never encountered before. It is reasonable to expect
that this will impact on their behaviour and we, yet again, are faced with
the issue of researching non-standard behaviour while assuming it is
standard. Translation process research does not find itself alone in this
dilemma, but it is one that has been largely unacknowledged to date and
which needs some consideration and controls.
A third challenge concerns the nature of the texts under study.
Researchers who also translate or teach translation are not unfamiliar with
the fact that the so-called “general” text can sometimes pose even more
difficulties for the translator than the specialised text. At the same time,
few translators earn their living by translating general texts or texts
published in newspapers and, sometimes, the unexpected challenge of the
apparently easy general text presented in a research study can come as a
surprise to the highly specialised translator. Using unfamiliar text types or
domains to investigate processes is valid if we are testing how participants
cope with new or unfamiliar challenges. However, if we want to investigate
what professionals normally do in their usual working environment, then
we need to select texts that will help us answer that question.
In this section we make some proposals for how at least some of the
methodological challenges outlined above can be met.
Eye tracking in translation process research 263
The setting up of a dedicated eye-tracking lab with control for light and
sound is to be recommended. If a “normal” work environment is important
in the research design, the lab should be made to look and feel like a
standard office and not a lab.
The question of increased accuracy (using head-mounted eye
trackers) over ecological validity (using desktop eye trackers) has received
little consideration in translation process research to date. The tacit
agreement, it seems, is that the desktop eye tracker, if combined with other
measurement techniques such as verbal protocols, interviews, keyboard
logging, etc., is accurate enough for the current aims.
Knowing one’s participants prior to commencement of the study is to
be recommended. If the study is concerned with what translators do in their
professional environment, then having information about what type of PC,
operating system, software, versions and keyboards are commonly used is
worthwhile, and efforts should be made to replicate the usual working
scenario. Obviously, this can only be done to a certain extent, as it is not
reasonable to present each participant with an exact replica of their own
working environment. But knowing in advance, for example, that a
participant has only ever used an “azerty” keyboard and is not used to a
“qwerty” keyboard can save the researcher from losing a participant due to
poor data quality caused by having to search for keys on the keyboard.
2.3 Ethics
2.5 Validity
4
http://gams.uni-graz.at/fedora/get/container:tc/bdef:Container/get [Last accessed
15/05/09].
Eye tracking in translation process research 265
reveal all there is to know about how humans translate, it certainly adds a
very rich dimension to the tools and methods we have for investigating this
activity, and the challenges involved in implementing it, while not
insignificant, can be overcome. The more the research community
embraces new methods of investigation, the more mature the research will
become and this will be to the advantage of all who are interested in this
field.
References
Abstract
This paper reports on a study which uses eye tracking in conjunction with
key logging and retrospective protocols to propose a methodology for
integrating data from task execution and post-task protocols. Carried out
under experimental conditions, the study targeted ten professional
translators, performing a direct (English into Portuguese) and an inverse
(Portuguese into English) translation. Both tasks were performed using key
logging and eye tracking. Subsequently, a free protocol (subjects’
verbalization prompted by a replay of their own task) and a guided
protocol (subjects’ verbalization elicited through questions) were conducted
with the aid of an eye tracker synchronized with audio recordings of
subjects’ verbalizations. Structured around three research dimensions, the
study aims at raising methodological questions on, and proposals for,
increasing data reliability and comparability, as well as tapping into
translators’ metacognitive activity as a means to gather insights into the
translation processes of professional translators.
1. Introduction
2. Background
into English the first for the remaining five subjects (S6, S7, S8, S9, S10).
Task performance was recorded using key logging (Translog 2006©) and
eye tracking (Tobii T60© and Tobii Studio 1.5©), the monitor standing 60
to 70 cms away from the subjects‟ eyes. After each task completion,
subjects were asked first to freely recall their performance as they read the
STs and TTs made available to them in the two horizontal windows on the
computer screen. Eye tracking was again used to record gaze movements
across STs and TTs while subjects carried out this recall, their
verbalizations being synchronized with eye movements. Subsequently, the
same equipment and methodology were used to obtain guided protocols
elicited through questions focusing on five language features related to
logical and cohesive relations, nominal group structure and domain
knowledge.
4. Results
This section is divided into three sub-sections, each presenting one of the
three dimensions pursued as described in the introduction in Section 1,
namely (i) the examination of data reliability based on the criterion of
average fixation length values of at least 175 ms in an attempt to verify
whether similar results are obtained for average fixation length provided by
two different fixation filters; (ii) the examination of the impact of
directionality on the time spent and on the number of fixations in the three
phases of the translation process, as well as on the number of fixations in
two major areas of interest (i.e., source-text area, and target-text area)
during free and guided recall; (iii) the analysis of retrospective protocols
related to a particular cohesive device in the direct translation task in an
attempt to tap into levels of metacognitive activity assumed to be at play
during task execution.
1
According to the Tobii Studio 1.5© Tutorial, the Tobii Studio© filter “estimates the
position of the fixations between changes in gaze position by using a median-
estimator and by checking if adjacent fixations (in the time domain) are closer
together than a given threshold (in the spatial domain). The threshold is set under the
fixation filter settings. The value is used when calculating if the gaze point belongs to
a new fixation or not [...]. The threshold value sets the maximum for how far apart
fixations are allowed to be in pixel radius and still belong to the same fixation.” The
ClearView© filter “defines the maximum distance between two points for them to be
considered [as] belonging to the same fixation and the minimum time for which gaze
needs to be within the radius to be considered a fixation”.
A new window on translators’ cognitive activity 273
9000 8626
8000
7000
6000
4855 5040 4856
5000 4707 4888
4435 4452 4281 4331
4000 3896
3620
3211
3000 2628 2716
2372 2192 2346 2088
2124
2000
1000
0
DT IT DT IT DT IT DT IT DT IT IT DT IT DT IT DT IT DT IT DT
S1 S2 S3 S4 S5 S6 S7 S8 S9 S10
Figure 1: Total absolute time (in seconds) spent on the translation tasks
Legend: DT = Direct translation, IT = Inverse translation
Left of the dotted line in Figure 1 are the subjects who carried out a direct
translation task first, right of the dotted line, the subjects who carried out an
inverse translation task first. According to Figure 1, there is an effect of the
first task upon the second, with eight subjects spending more time on the
first task than on the second, regardless of whether it was a direct or an
inverse translation task. Among the subjects who started with the direct
translation task, only S5 spent more time on the second task. Among the
subjects who started with the inverse translation, only S7 spent more time
on the second task (direct translation). A paired samples t test excluding
276 Fabio Alves, Adriana Pagano & Igor da Silva
these two subjects shows that this impact is statistically significant, both if
S10, who presented a very high difference in time allotted to each task, is
included (t = 2,375; df = 7; p = 0,049) and when this subject is excluded
(t = 3,716; df = 6; p = 0,010). Furthermore, Figure 1 also shows that the
subjects tend to allot a similar amount of time (less than 20 % difference) to
either type of translation, regardless of directionality, with the exception of
S3, S8, and especially S10 (27,6 %, 25,0 %, and 43,3 %, respectively).
Taken together, these results suggest that neither the texts themselves nor
directionality of the task account for the variance in total time spent. Task
order seems to be the only variable which has a systematic impact on the
duration of the task of all subjects as it introduces a decrease in time spent.
It should be borne in mind that ST complexity was analogous and
thus did not constitute a variable in our study and that directionality did not
point to substantial differences in performance. Thus the data on the
relative time spent on each phase of the translation process (orientation,
drafting and revision) indicate that the only variable having an impact on
task execution is task order.
Figure 2 shows that, from a within-subject perspective, each subject
exhibits a relatively similar pattern of time distribution for both the direct
and the inverse translation tasks. Subjects S6, S7, S9 and S10 allotted
virtually no time to orientation, and then spent more time revising the task,
a phase which involves resorting to external support (as observed in the
replay of the Tobii Studio 1.5©). The data show that this can be explained
by the strategy chosen for problem solving and the consultation of external
resources for words or collocations while writing the first draft of the
translation.
Figure 2 also reveals that little time is devoted to initial orientation
(less than 2 %, or 1 minute when we look at the absolute values), the major
exceptions being S8 for both tasks and S5 for the inverse translation task
(with an orientation length higher than 3 %, or of approximately 3 minutes
if we look at the absolute values). These are the only situations where the
whole text is fully read before the translation is drafted (see Section 4.3.1).
A new window on translators’ cognitive activity 277
IT
S10
DT
IT
S9
DT
IT
S8
DT
IT
S7
DT
IT
S6
DT Orientation
Drafting
IT
Revision
S5
DT
IT
S4
DT
IT
S3
DT
IT
S2
DT
IT
S1
DT
Figure 2: Relative time allocated to each phase of the translation process per subject and
per task (values not included in order to avoid information overload in the graph)
Legend: DT = Direct translation, IT = Inverse translation
2
According to the Tobii Studio 1.5© Tutorial, “fixation filters are used to group gaze
data into meaningful fixations. Many different algorithms for fixation definition have
been proposed and are available among researchers and practitioners. In part, the
purpose of a study influences which types of fixation filters are most suitable.”
A new window on translators’ cognitive activity 279
S5). In the revision phase, both translation tasks display the same tendency,
though in this case there is a higher fixation count in the TT area. These
patterns for the drafting and revision phases may show that fixation count
per se in ST and TT AOIs may not be a variable accounting for the
different cognitive processes underlying an inverse and a direct translation
task.
Fixation data on ST and TT AOIs can also be compared with
fixations during the retrospective protocols. Tables 11 and 12 provide data
on the fixation count for both free and guided recall.
As shown in Table 11, subjects have different behaviours in terms of
fixation counts on the AOIs of the free recall protocol, although the overall
mean points out to higher fixation counts in the TT AOI. Conversely, Table
12 shows that subjects (except for S10 in the inverse translation task) have
higher fixation counts in the ST area.
Taken together, these data on fixation show that revision as compared with
drafting, and free recall as compared with guided recall, provoke different
processes, especially because they are inherently related to different areas
of the screen. This will be shown in more detail in our analysis of a text
string in the next section.
A new window on translators’ cognitive activity 281
4.3 Third dimension: the study of gaze on a particular text string in three
different tasks
The third dimension focuses on the analysis of one of the five elements
scrutinized in the guided recall protocol, namely a text string in the English
ST, where the conjunctive adjunct “Finally” ends an enumeration. This
string was chosen because the items in the list concluding with “Finally”
were not explicitly signalled by typical enumerating conjunctions and were
distributed across the whole source text. This was assumed to require a
great deal of processing effort on behalf of the subjects. For the guided
protocol, we assumed that during their response to our question on the
enumerating string the subjects‟ eyes would attempt to locate and fixate the
items of the list.
Our analysis of this particular text string is based both on data
retrieved in the orientation, drafting and revision phases of the direct
translation task and on the free and guided retrospective protocols. Just as
in our first and second research questions, the analysis pursued in the third
dimension also raises methodological issues and aims at putting forward a
new way to triangulate key logging, eye tracking and recall protocols.
Figure 3: Heat map overlapping fixation data for the initial orientation phases of S1-S10; circle
indicates the conjunctive adjunct “Finally”
282 Fabio Alves, Adriana Pagano & Igor da Silva
Figure 5: Linear protocols for the conjunctive adjunct “Finally” in the drafting phase of the
direct translation process; indicates a pause interval up to 5 seconds
This linearity in the eye movements can be seen not only in S8‟s orientation
phase, but also in the orientation phases of the other subjects. Figure 5
presents Translog linear representations for the translation of “Finally”,
with subjects predominantly showing a linear pattern in drafting.
The conjunctive adjunct was not a point of concern for any of the
participants, apparently with the exception of S1 and S8, whose data
showed pauses before they actually translated the word. Figures 6 and 7
show the gaze plots for S1 and S8, respectively, as they read the source text
and produced the target text.
The replay of Tobii Studio 1.5© recordings shows that S1 had a pause of
approximately 30 seconds because the subject browsed the Internet to find
a translation for “Finally”. The Tobii gaze plot, however, shows no
regression in the subject‟s gaze. The subject focuses on “Finally” for a
moment, reads the beginning of the clause complex (“Finally, pulmonary
thromboembolism”) and consults Babylon© online for a translation.
As for S8, the Tobii gaze plot shows that the 10-second pause prior
to the translation of “Finally” seems to be connected with the subject
rechecking the previous nominal group (“consequent pulmonary
vasculopathy”) in the ST.
Figure 7: S8‟ gaze plot immediately prior to the translation of “Finally” by “Por fim”
Translog linear protocols and Tobii gaze plots show no revision of the first
and final translation of the conjunctive adjunct “Finally” given by any of
the translators in the drafting phase. The pattern for all subjects is similar to
that shown in Figure 8 for S2.
Figure 8: S2‟s gaze plot onto “Finally” and surroundings in the revision phase
A new window on translators’ cognitive activity 285
In general, the free protocols did not prompt any comments by the subjects,
and no regression in eye movement was found for any subject. Most of the
subjects focused on the TT area while verbalizing. S2‟s gaze in Figure 9
can be considered as a prototypical example of the way free retrospective
protocols unfolded.
Figure 9: S2‟s gaze plot onto “Finally” and surroundings in free protocol
3
We have opted for the standard Translog presentation with the ST being shown above the TT.
One may hypothesize that there could have been an impact on the results, had the texts been
placed in a reverse order or left and right to each other. Nevertheless, we assume that the
Translog upper/lower presentation of ST/TT areas on the screen follows what is standard
practice and does not constitute an unfamiliar visualisation to the subjects.
286 Fabio Alves, Adriana Pagano & Igor da Silva
Subjects could see both source and target texts in their entirety. In the
guided protocol, subjects were asked, among other activities, to find
“Finally” in the ST and then to answer the question “Why is there „Finally‟
in this part of the text?” Table 13 gives an assessment of the answers
provided by the subjects to the question posed, measured as being “non-
satisfactory” when subjects could not provide any adequate explanation;
“partially satisfactory” when subjects showed evidence that they could see
the function of the conjunctive adjunct even if they could not identify the
previous items in the list; and “satisfactory” when subjects managed to
identify the function of “Finally” as a device ending an enumeration and
could somehow find the previous items in the sequence.
Although the free retrospective protocols and the key-logged data did not
point to potential difficulties in the processing of the adjunct, only four out
of ten subjects were able to identify “Finally” as the last element in a list.
The analysis of Table 13 along with eye-tracking data clearly divides
the ten subjects into two groups. The first group, including S1, S2, S3, S4,
S5, and S10, failed to attribute to “Finally” a concluding role in the
enumeration. Neither their gaze nor their verbalizations provided evidence
of metacognitive activity and awareness of the role played by the adjunct.
This group of subjects basically attributed to “Finally” a temporal meaning
or a discourse function role as a device signalling that the text was coming
to an end instead of realizing that there was a series of items making up the
enumeration ended by “Finally” and that this enumeration had been
A new window on translators’ cognitive activity 287
Figure 10: S8‟s random gaze plot while trying to account for the function of “Finally”
At the end of his recall, unprompted by any question, the subject explained
that he immediately opted for “Por fim” („At last‟) rather than
“Finalmente” („Finally‟) in his target text in the drafting phase, the former
being considered by the subject a device in Portuguese which is more
successful in terms of conveying the meaning of conclusion in a line of
argumentation. His reflections seem to refer to the two possible meanings
construed by “Finally” as a concluding device in a series of enumerated
items or a discourse device signalling a final move in a text.
5. Conclusions
data as the choice of filter and the type of data used can yield different
results. As we have sought to show, ClearView, for instance, seems to be
more prone than Tobii to yield data falling below the threshold of 175 ms,
below which subjects have to be excluded. Another relevant conclusion
deriving from dimension 1 suggests that the criterion for minimum average
fixation length with a lower minimal value of 175 ms equally applies to
translation process data and to free and guided retrospective protocols,
which are, therefore, comparable.
Regarding dimension 2, our data shows no significant impact of
directionality on total and relative times spent on direct and inverse
translation tasks. The main impact is task order, as subjects showed a
tendency to allocate more time to the execution of the first task. However,
when the orientation, drafting and revision phases are considered
separately, directionality does have an impact on the process, with an
increase in the amount of revision time carried out by most of the subjects
in the inverse translation task in spite of task order. As for fixations, these
occur predominantly in the source-text area during the drafting and revision
phases both in the direct and inverse translation tasks. In terms of fixation
counts in the AOIs of the free recall protocols, the overall mean points to
higher fixation counts in the TT AOI (though mean in this case is not
representative of the whole sample). As far as the guided protocols are
concerned, higher fixation counts are found in the ST AOI. As stated
before, the data on eye fixations show that revision as compared with
drafting, and free recall as compared with guided recall entail different
processes, especially because they are inherently related to different areas
of the screen. These patterns may suggest that fixation count per se in ST
and TT AOIs may not be a variable accounting for the different cognitive
processes underlying direct and inverse translation tasks.
With respect to dimension 3, it was possible to show how eye-
tracking data and retrospective verbalizations can be used together to tap
into translators‟ metacognitive activity. As far as free retrospective
protocols are concerned, subjects show a tendency to focus on the TT area.
Free protocols do not add information to initial orientation, and subjects
usually start protocols by speaking of strategies and general problems (e.g.,
terminology issues). Subjects usually cease their recalling when they come
to the end of the drafting phase. Their verbalizations primarily focus on
290 Fabio Alves, Adriana Pagano & Igor da Silva
“silence” and “changes” and do not provide any further information for the
conjunctive adjunct “Finally”, which was “easily” translated in the drafting
phase. As for the guided protocols, subjects usually answer questions
hastily without much thought. They show backward and forward eye
movements which seem to concentrate especially on the ST area as a means
of confirmation of the questions asked.
Most subjects clearly failed to provide a satisfactory answer to the
question they were asked and also showed no evidence of being actively
engaged in finding possible answers on the ST or TT which appeared on
the screen. Among the few subjects who reveal higher metacognitive
activity and awareness of the question they were asked, S8 is the only
subject who could actually provide a satisfactory answer to account for the
function of the conjunctive adjunct “Finally” in the rendition he had just
come about with.
On the whole, verbal protocols generated in conjunction with eye-
tracking data prove useful for probing the relation between the allocation of
time and the level of metacognitive activity. In sum, the study shows the
feasibility of synchronizing retrospective verbalizations, gaze movements
and fixations as the subjects‟ gaze moved across the screen from ST to TT,
therefore opening a new window on the investigation of translators‟
metacognitive activity.
References
Abstract
1. Introduction
1
Forståelse og formulering i tolke- og oversættelseshybrider [“Comprehension and
production processes in translation and interpreting hybrids”] funded by the Danish
Research Council for Culture and Communication.
294 Barbara Dragsted, Inge Gorm Hansen & Henrik Selsøe Sørensen
3. Expertise in translation
4.1 Participants
2
Number of years for which participants have studied and practised translation.
3
Language for Special Purposes, including the domains of economics, finance, law,
and science and technology.
298 Barbara Dragsted, Inge Gorm Hansen & Henrik Selsøe Sørensen
Data were collected from five different experimental tasks: a reading task, a
reading-for-the-purpose-of-translation task, a sight translation task, a sight
translation task with the use of SR and a written translation task.
Translations were from Danish into English. In the reading task (Task 1),
the participants were asked to read a short text for the purpose of
comprehension as if they were reading a newspaper article.
In the reading-for-the-purpose-of-translation task (Task 2), the
participants were instructed to read a text, which they would subsequently
be asked to translate.
In the sight translation task (Task 3), the participants were asked to
produce an oral translation of a text. The text was presented to the
participants on the screen and their oral translation was recorded in
Translog Audio (see below).
In the sight translation task with SR (Task 4), the participants were
asked to produce an oral translation of a text using the speech recognition
tool Dragon Naturally Speaking 10.0 (see below). Before producing the
translation, the participants performed the minimum amount of training
required by Dragon Naturally Speaking in order for the SR tool to be able
to recognise and become familiar with each participant‟s voice and accent.
The ST was presented in the top window of the screen and the TT appeared
in the bottom window of the screen as it was produced orally by the
translator. The oral output was recorded in Translog Audio (see below).
In the written translation task (Task 5), the participants were asked to
produce a normal written translation of a text. The ST was presented in the
top window of the screen and the TT was produced in the bottom window.
We logged the participants‟ keystrokes in Translog.
For all five tasks, we tracked the participants‟ eye movements; for an
introduction to eye-tracking during reading, see Clifton et al. (2007),
Radach et al. (2004), Rayner (1998); and for an introduction to eye-
tracking and its use in translation research, see O‟Brien (2006) and
Göpferich (2008). We used a Tobii 1750 remote eye tracker, which runs at
a frame-rate of 50 Hz and uses binocular eye tracking.4 The texts were
4
Tobii 1750, see http://www.tobii.se.
Experts exposed 299
The text produced orally under the sight translation condition (Task C) was
transcribed, and this transcription as well as the SR output and the written
output from Texts D and E were sent to the translators afterwards for post-
editing. The translators were asked to register the time it took them to edit
the text and to bring it up to the standard they would normally aim at if this
had been a task commissioned by a business client.
In addition to collecting quantitative data from eye-tracking,
keystroke logging and SR, we performed retrospective interviews asking
the participants about the experimental situation and their usual working
habits and routines. The interviews served not only as support for the
quantitative data, but also as a data source in its own right for the above
profiles of the three expert translators.
300 Barbara Dragsted, Inge Gorm Hansen & Henrik Selsøe Sørensen
5
For a description of the GWM tool, see http://www.cs.uta.fi/~oleg/gwm.html.
Experts exposed 301
routines, support the quantitative data, which indicates that the translators‟
experimental behaviour was not very different from their normal behaviour
and actual practice.
Figure 1. Task times in Tasks 1–5 (Task 1 = Reading for comprehension, Task 2
= Reading for translation, Task 3 = Sight translation without SR, Task 4 = Sight
translation with SR, Task 5 = Written translation).
As appears from Figure 1, all three translators spent considerably more
time on sight translation with SR than on normal sight translation (without
SR). This is not surprising as sight translation with SR involves the
additional element of monitoring and editing the TT, a process which has
previously been found to be more demanding than the processing of the ST
both for professional translators (Jakobsen and Jensen 2008: 114; Dragsted,
forthcoming) and for students (Sharmin et al. 2008: 46) and sometimes as
time-consuming as the drafting phase (Englund Dimitrova 2005: 87–88).
Still, it is worth noting that the most considerable leap in task time is not
caused by the shift from reading to (sight) translation, but rather from
translation with no written representation of the TT to translation which
includes online monitoring and editing of written TT output. In other
words, time and gaze data indicate that the most demanding complication
of translation, as compared with the reading tasks, is not the actual
drafting/production of the TT, but rather the process of continuously
monitoring, assessing and editing one‟s own output. According to the SR
expert and semi-expert, the cognitive load and the attentional effort
involved in monitoring and revising SR translation output is higher than in
written translation, and revision involves more checks against the ST to
detect output errors.
Experts exposed 303
The task time differences between the oral tasks (Tasks 3 and 4)
indicate the additional time spent on online TT monitoring and editing of
SR output. However, the eye-tracking data showed that all participants also
had a higher number of fixations in the ST area of the screen in Task 4
compared with Task 3 (see Figure 2), indicating that online TT monitoring
and revision triggered regressions to ST words. It also appears from Figure
2 that there was no clear tendency for the TT area to attract more attention
than the ST area as might have been expected. Only one of the translators,
the semi-expert, confirmed this hypothesis.
Figure 2. Fixation counts for Task 3 (without SR) and Task 4 (with SR)
We expected that there would be a time reduction in both of the oral
translation conditions compared with the written translation task. As
expected, all three translators clearly spent less time on normal sight
translation (Task 3) than on the written translation task (Dragsted and
Hansen 2009), whereas one of the translators did not save time under the
SR sight translation condition compared with the written translation
condition (Figure 1). The translator who saved most time was the SR
expert, which is not surprising; it also corresponds with what he stated in
the interview, cf. below. The SR novice had only a marginal time reduction,
and the SR semi-expert produced the written translation faster than the SR
translation. Again, the semi-expert confirmed this result, stating in the
interview that his total time consumption was sometimes higher when he
used an SR system because he often spent more time on end-revision of SR
output. This also appeared to be the case in the experiment reported here,
304 Barbara Dragsted, Inge Gorm Hansen & Henrik Selsøe Sørensen
see Figure 3, where the semi-expert took longer to revise the text in Task 4
than to draft it.
We investigated long pauses (+2.5 sec.) in the two oral tasks (analysed in
the speech analysis and synthesis program Praat6 with an add-on script
developed for pause detection7), which were assumed to be associated with
the planning of, for instance, the next sentence or clause (Butterworth
1980: 156–157, 160), and compared the pauses and planning patterns under
the oral conditions with pauses in the written task (analysed in Translog).
Furthermore, we analysed fixation patterns in relation to spoken output in
6
See www.fon.hum.uva.nl/praat/, visited on June 11, 2009. Praat website with
information and downloads.
7
Developed by Miette Lennes (www.helsinki.fi/~lennes/praat-scripts/, visited on June
11, 2009).
Experts exposed 307
order to determine whether the translators fixated the ST word for which
they were currently producing/speaking a TT equivalent (whether fixations
were in sync with the word being produced), or whether they fixated words
before or after the word being produced (referred to as off sync fixations).
Off sync fixations are taken to indicate planning or monitoring.
The fixation analyses showed a difference between the fixation
pattern of the SR expert on the one hand and the SR novice and semi-
expert on the other in both oral tasks.
The SR novice and semi-expert had a clear preference for in sync fixations
in both sight translation tasks, whereas the SR expert frequently fixated
words other than the one being produced/translated, notably in Task 4
(sight with SR). Contrary to our expectations, the SR expert did not focus
particularly on subsequent ST words, which would be the most obvious
308 Barbara Dragsted, Inge Gorm Hansen & Henrik Selsøe Sørensen
The focus so far has been on the translation process and on how different
oral translation modes differ from one another and from traditional written
translation. However, an essential issue for any professional translator is
whether the quality of the translated output has a sufficiently high standard.
Hence, in the following, we tentatively propose a method by means of
which quality can be tested and compared across different translation
modes. The analysis performed below is not based on genuine ratings of
quality, but merely on frequency of occurrence in Google and on our own
intuitions.
Spoken translation may help a translator create texts that read
naturally in the target language and do this at higher speed than if the
translator had used the keyboard. On the other hand, the produced TT may
be characterised by imprecise terminology because the oral mode
encourages the translator to use words and phrases which spring to mind
instantly (and often are more general), rather than selecting more precise
terms (Chafe and Danielewicz 1987: 86, 88). As a result, the revision phase
310 Barbara Dragsted, Inge Gorm Hansen & Henrik Selsøe Sørensen
8
In some cases, Google may be used for documenting the optimal translation. The
figure in square brackets indicates google hits on July 14, 2009. The missing variant,
“in the international capital markets”, returns 19800 hits. The conclusion in this case
would be that the SR novice‟s solution is less common but nevertheless widely used.
312 Barbara Dragsted, Inge Gorm Hansen & Henrik Selsøe Sørensen
tend to be transferred to the target text” (Toury 1995: 275), i.e. what Toury
refers to as negative transfer. This includes any expression or syntactic
structure introduced into the TL by translating it from the SL in a way that
seems awkward to TL readers. A calque thus implies that a more
appropriate equivalent exists, but that translators subconsciously have
transferred a SL element to the TL because they could not find the optimal
equivalent on the spot.
Table 5. Task 4 (sight with SR): examples of divergent translator decisions
SL SR novice SR semi- SR expert Comment
expert
1 stor finansiel major major major term
virksomhed financial financial financial
company* undertaking company*
2 politiske political political political term
tilkendegivelser statements intentions to indications
this effect*
3 blev reddet was would not be was saved term (was
rescued saved bailed out*)
4 medførte at … frøs led to a as a result … this meant style and term
fast frozen … froze that…froze (caused … to
solid dry up*)
5 markedet for market for liquidity liquidity term (money
likviditet liquid market market market*)
assets
6 Redningspakker rescue rescue bail-out term
packages packages packages*
7 genoprette tilliden til reinforce restore restore style
the trusts in confidence confidence
in* in*
8 Pengeinstitutter financial banks* financial term
institutions institutions
9 og [tilliden] til, at… and the and with the Parataxis (and
(literally: and the trust that … because… further in the
[confidence] that….) benefit that Governments‟
intentions to
…*)
10 indgik en politisk made a entered into entered into style
aftale om political a political a political
agreement agreement agreement
to to* to*
11 medført, at led to resulted in caused […] style and term
[pengemarkederne] frozen […] frozen [..] to freeze (caused … to
var frosset til over dry up*)
Experts exposed 313
6. Conclusions
References
Abstract
1. Introduction
The translation research centre CRITT, founded by Arnt Lykke Jakobsen, has
carried out a number of experiments coupling eye tracking with translation
process research. These experiments have spanned a wide range of linguistic
phenomena and translation aspects, but the issue of syntactic processing
during translation has received little attention. Here, we begin to explore this
issue.
1
We are grateful to Sharon O’Brien for her constructive comments on a previous draft of
this article.
320 Kristian T.H. Jensen, Annette C. Sjørup & Laura Winther Balling
The inspiration for the experiment was the concept of syntactic priming,
which refers to the influence of previously encountered syntactic features on a
construction currently being processed. Syntactic priming is found both within
a language (see e.g. Hartsuiker & Westenberg 2000; Bock 1986) and between
languages (see e.g. Hartsuiker et al. 2004). We investigated whether even
highly proficient users of L2 (English in this case) such as professional Danish
translators would be syntactically primed to transfer L1 syntax to the L2
translation. We chose to study a phenomenon which is purely syntactic and
where transfer is either obligatory or ungrammatical, namely the order of
subject and verb.
Another concern of the current experiment was to investigate whether
eye tracking is a viable measure of syntactic processing in a translation task.
We considered both total gaze time and pupil dilation.
We analysed the eye-tracking data using mixed-effect regression
models to ensure that any differences observed in the descriptive statistics
were verified by inferential statistics. This allowed us to draw tentative
conclusions for a larger population than the present limited sample.
2. Background
(1)
Den store stærke mand køber en hund
The big strong man buys a dog
SUBJECT VERB DIRECT OBJECT
(2)
Omvendt foretrækker den feminine kvinde en kat
Conversely prefers the feminine woman a cat
ADV VERB SUBJECT DIRECT OBJECT
Research shows that there is a strong link between cognitive effort and the
location of the eye’s fixation (Rayner 1998). In reading, the eye remains
fixated on a particular word for as long as the word is being processed (Just &
Carpenter 1980: 330), and Just & Carpenter’s eye-mind assumption posits that
“there is no appreciable lag between what is being fixated and what is being
processed” (1980: 331). Therefore, we assume that observable, quantifiable
gaze data which we can collect with an eye tracker are indicative of
underlying, not directly observable cognitive processes that take place during
a certain task, in this case an L2 translation task. We investigate two eye-
tracking indicators, viz. total gaze time (a) and pupil dilation (b).
3.1 Participants
3.2 Task
The participants were tasked with translating two texts (A and B from Danish
into English. Prior to the experimental texts, they were asked to translate a
similar text (Text C) as a warm-up exercise.2 The presentation order of texts A
and B was randomised. There were no time constraints and no translation aids
of any kind.
The participants were asked to translate the texts to their own
professional satisfaction and were allowed to revise the target texts, which
some participants chose to do while others did not.
3.3 Texts
Texts A and B are constructed texts. Using authentic material was deemed
unfeasible because the data had to contain as many S-V and V-S segments as
possible and because these segments also had to meet other requirements (see
below). Altogether, the experimental texts contained 12 S-V segments and 12
V-S segments. The segments were comparable in terms of number of
characters. All verb phrases (VP) were simple while the noun phrases (NP)
were complex. We chose to use complex NPs in order to maximise the chance
of an effect of constituent reordering. The order of the S-V and V-S segments
was alternated throughout the texts. Also, as some researchers have shown that
reading times are longer at sentence-initial and terminal positions, no S-V and
V-S segments were located at these positions (Rayner & Sereno 1994: 73ff,
Gernsbacher 1990: 7).
Although one text scored 33 (easy) and the other 44 (medium) by the
LIX indicator (Bedre Word 2007) of text difficulty (Björnsson 1983), the
statistical analyses showed no difference between the texts in terms of
processing effort invested by the participants. Segments from both texts could
therefore be analysed together.
2
In text C, there were 7 SV and 2 VS segments which is most likely fewer VS segments
than in Danish in general, given the 40 percent non-subject initial clauses in Dutch,
German, Icelandic, and Swedish; see Bohnacker & Rosén (2007). Since there was no
evidence of a priming effect in the experimental texts (but rather a fatigue effect, see
Figure 2), a priming effect between texts is highly unlikely.
Effects of L1 syntax on L2 translation 325
A Tobii 1750 eye tracker was used to collect eye-tracking data and the key-
logging software Translog was used to present the texts. Due to the eye-
tracking equipment’s average accuracy of 0.5 of a degree (corresponding to an
inaccuracy of up to 1 cm), a relatively large font size was used, viz. Times
New Roman 18. Double line spacing was applied to increase accuracy in the
data. The individual text length was approximately 150 words to ensure
sufficient space for both source and target texts without scrolling.
Areas of Interest (AOIs), which are spatially defined regions of the
monitor, were set up around each S-V and V-S segment using the ClearView
software (the analysis software designed for the Tobii eye tracker). ClearView
also calculated the fixations. The fixation filter settings were set to include
gaze samples that fell within a radius of 40 pixels and within a time window
of at least 100 ms.
Since ClearView records pupil dilation but does not contain any pupil
dilation analysis tools, pupil dilation calculations were done in Microsoft
Excel based on exports of the raw eye-tracking data log files.
326 Kristian T.H. Jensen, Annette C. Sjørup & Laura Winther Balling
We see two possible explanations for our high discard percentage. One
explanation could be the equipment’s age. A discard percentage of 62.5 (data
from ten out of 16 participants had to be excluded) could indicate technical
problems as this figure is somewhat higher than in previous experiments
carried out with the same piece of equipment (e.g. Pavlović & Jensen 2009;
Jensen, in preparation). The second possible explanation could relate to the
participant’s distance from the monitor. Experiments with participants sitting
at a distance of more than 55 to 60 cm generally resulted in lower eye-tracking
data quality based on the average fixation duration criterion previously
mentioned. In future experiments, data quality could probably be improved by
paying more attention to this parameter.
As it is neither possible nor desirable to use head fixation in this type of
experiment, it is extremely difficult to ensure that the participants do not move
further away from the monitor during the experiment. The fact that non-touch
typists look away from the monitor more may also cause drift, which in turn
can lead to poor data quality. This is of course a general issue for all
experiments combining typing and eye tracking and does not offer an
explanation for the very high discard percentage in this specific experiment.
4. Results
The data were analysed using linear mixed-effects regression models (Bates et
al. 2008) in the R environment for statistical computing (R Development Core
Team 2008). These models are mixed in the sense that they include both
random effects and fixed-effect factors and covariates. The random effects are
effects of sampled variables which are not repeatable; here, we include
random intercepts for participant and AOI, which essentially model the
random variation between the participants and the AOIs. Fixed factors and
covariates are variables which are repeatable; in our case, the most important
fixed effect is for the factor Word Order with the levels S-V and V-S. An
Effects of L1 syntax on L2 translation 327
Table 1. Summary of regression model for total gaze time using contrast coding for
the factor Word Order with S-V as the reference level. The model also included
crossed random effects for AOI (s.d. estimated at 0.1810) and Participant (s.d.
estimated at 0.3636). The residual standard error of the model was estimated at
0.4043.
Estimate MCMC HPD HPD p p (t)
mean lower upper (MCMC)
Intercept 12.7103 12.6912 7.0172 17.9804 0.0002 0.0002
Word order: V-S 0.2889 0.2886 0.0769 0.4927 0.0078 0.0217
Position of AOI 0.0099 0.0098 0.0045 0.0154 0.0010 0.0020
Word repetition -0.5120 -0.5054 -0.8389 -0.1748 0.0036 0.0093
Number of characters 0.0569 0.0564 0.0301 0.0837 0.0004 0.0004
Log mean content word -3.5105 -3.4902 -6.4689 -0.4364 0.0252 0.0534
freq (linear)
Log mean content word 0.4499 0.4468 0.0425 0.8594 0.0338 0.0672
freq (quadratic)
The main hypothesis of the experiment was that V-S segments would require
more cognitive effort than S-V segments. Mean total gaze time per character
was 468 ms for the S-V segments and 576 ms for the V-S segments,
representing a 23.1 % increase. The analysis of the total gaze time summarised
in Table 1 clearly confirms that participants looked significantly longer at
segments with V-S order, which has to be reversed for translation into English,
Effects of L1 syntax on L2 translation 329
than at segments with S-V order, for which the Danish word order can be
retained in the English translation. In Table 1, the coefficient for “Word order:
V-S” represents the difference to “Word Order: S-V” which is mapped to the
intercept. The value for “Word order: V-S” is positive, representing longer
gaze time, and the associated MCMC-based p-value is low (p = 0.0078),
indicating a high likelihood that the difference observed in the sample
generalises to the relevant population of translators. The difference between
the word orders is illustrated in Figure 1. This is a partial effects plot showing
the difference between the two word orders while all other significant
variables are held constant at their medians. In other words, Figure 1 shows
the difference between S-V and V-S order, all other things being equal.
Although the difference between S-V and V-S order is the focus of the
experiment, the regression model summarised in Table 1 also includes other
covariates, which both serve the purpose of statistical control of variables
which could otherwise be confounding and provide information about the gaze
behaviour of our participants while they translate. These variables are listed in
Table 1 and illustrated in Figure 2. The variables in the top row have to do
with the context in which each relevant segment occurs: Position is the word
number of the first word of the relevant segment; this has significant
inhibitory effect such that gaze time increased for later parts of each text (p =
0.001). This effect could be caused by fatigue or the fact that more
330 Kristian T.H. Jensen, Annette C. Sjørup & Laura Winther Balling
information has been presented later in the text, which could result in longer
gaze times. The top right panel of Figure 2 shows the effect of word repetition.
A number of words occurred several times in each text; in order to control
this, we included a variable that indexed the mean number of times each word
in the AOIs had been encountered previously in the same text. The plot shows
that word repetition resulted in shorter gaze times, corresponding to a
repetition priming effect for the words in question (p = 0.0036).
Figure 2: Partial effects of position of AOI, mean number of word repetitions, length
in characters, and mean frequency of content words
Effects of L1 syntax on L2 translation 331
The bottom panels of Figure 2 show effects of item characteristics. The effect
of the number of characters is shown in the bottom left panel: as expected,
participants gazed significantly longer at longer AOIs (p = 0.0004). The
bottom right plot shows that AOIs that contained content words of higher
frequency were associated with shorter gaze times, though the effect levelled
off for high frequencies, as indicated by the significant quadratic effect of
frequency. The frequency effect is the most marginal effect in the regression
analysis, a finding which would be surprising for reading but probably less
surprising for translation where many other factors are at play. Mean bigram
frequency had no significant effect.
The experiment showed that translators gazed significantly longer at V-S than
at S-V segments when translating from L1 Danish into L2 English. There was
no significant difference in pupil dilation between the two types of
constructions. These findings contrast with our expectation that the two
indicators would show the same pattern of increased processing effort for V-S
as opposed to S-V segments. In the following, we discuss possible reasons for
this difference between the indicators and the resulting different
interpretations of the measures and results.
One tempting explanation is that total gaze time and pupil dilation are
indicative of different cognitive processes. Several processes occupy the
332 Kristian T.H. Jensen, Annette C. Sjørup & Laura Winther Balling
no reordering is required. However, this does not have to mean that the
processing of V-S segments impose heavier demands on cognitive capacities,
especially for professional translators who might automatise this shift to some
extent. This would explain why the pupil dilation indicator shows no evidence
of increased cognitive effort associated with the V-S segments compared to
the S-V segments. This tentative explanation obviously challenges the general
assumption that increased time consumption reflects an increased amount of
cognitive effort. This challenge to the assumed relation between time
consumption and cognitive effort presupposes that our pupil dilation
measurements are valid and that pupil dilation reflects cognitive effort
invested in syntactic processing but, as discussed above, this is not necessarily
the case.
Assuming that the total gaze time does in fact reflect processing effort,
we may understand the significant difference between processing of S-V and
V-S segments in two ways. One possibility is to understand it in terms of
automatic transfer of L1 syntax to all types of L2 processing, in accordance
with the view that L2 processing is heavily influenced by L1 structure (e.g.
MacWhinney 2005; Frenck-Mestre & Pynte 1997) and with evidence of
syntactic priming between languages (Hartsuiker et al. 2004). However, the
fact that the L1 syntactic structure is present for the translator in the shape of
the source text makes it difficult to draw conclusions, based on this
experiment, about the influence of L1 on L2 processing more generally. A
reading experiment in L2 comparing segments that are parallel to L1 with
those that are not could help clarify this.
A further interesting investigation of the findings of the current
experiment would be a reading experiment in L1 that could establish whether
differences in gaze time between S-V and V-S are limited to translation tasks
like the present.
Another possibility is that the S-V vs. V-S difference is the result of a
more or less conscious translation strategy. Since the subject-verb order of
many Danish clauses may be directly transferred to English, it may be that
such transfer is the initial strategy in all cases. The fact that such transfer is
correct for clauses with unmarked, canonical word order may support this
334 Kristian T.H. Jensen, Annette C. Sjørup & Laura Winther Balling
explanation. On the other hand, the fact that the participants were professional
translators and thus highly skilled users of the L2 may make it unlikely that
they apply such a relatively crude strategy, which is inappropriate in a number
of cases.
Despite the low number of participants, our experiment showed a
significant effect of word order on gaze time. Moreover, the participants were
professional translators for whom the change of word order from L1 to L2
could be assumed to be trivial. The significance of the effect in spite of these
two facts suggests that the reordering of constituents is an important element
of the translation process. By implication, constituents seem to function as
cognitive units in translation (cf. also Dragsted 2004). We would expect this to
generalise beyond the current language pair of L1 Danish and L2 English and
beyond the current relatively small sample.
References
Appendix
Text A
Nogle vil påstå, at mænd og kvinder taler forskellige sprog. Hvis en mand i et
parforhold siger, at han ikke ønsker at tale om sine problemer, føler den
typiske kvinde, at hun bliver udelukket fra hans verden. Hvis kvinden siger til
ham, at hun ikke ønsker at snakke om problemerne, tror de fleste mænd, at
hun ikke har behov for at tale om dem. Disse misforståelser giver anledning til
skænderier. Når store skænderier opstår, reagerer kvinder og mænd også
forskelligt. Ny forskning viser, at mange kvinder græder, når de er vrede. I
disse situationer vælger skræmte mænd at flygte, og de kan føle sig
hjælpeløse. I modsætning til den grædende kvinde bruger den vrede mand
hårde ord. På trods af disse forskellige reaktioner formår de fleste mænd og
kvinder at leve fredeligt sammen, og vi kan jo nok ikke leve uden hinanden.
Text B
PRECISION,
STRATEGIES AND
QUALITY
ASSESSMENT
Translation preferences in legal translation:
lawyers and professional translators compared
Abstract
1. Introduction
important that the legal translator observes the notion of fidelity to the
source text (ST), cf. e.g. Šarčević (1997: 16):
Legal translators have traditionally been bound by the principle of fidelity.
Convinced that the main goal of legal translation is to reproduce the content of
the source text as accurately as possible, both lawyers and linguists agreed that
legal texts had to be translated literally. For the sake of preserving the letter of
the law, the main guideline for legal translation was fidelity to the source text.
Others, e.g. Šarčević (1997), Engberg (2002) and Alcaraz and Hughes
(2002), advocate a different strategy for the translation of legal texts,
namely that of taking both the purpose of the translation and the cultural
expectations of the target-language reader into consideration by producing
“a translation that provides connotative as well as denotative parity with its
original” (Alcaraz and Hughes 2002: 178). Here the focus is on catering for
information needs of the target-text (TT) reader rather than solely on
conveying the explicit content of the ST. This is where explicitations and
implicitations may be relevant, see Section 5.
To our knowledge, these very different „guidelines‟ for legal
translation, equating fidelity with (near) literalness on the one hand and
catering for information needs of the TT reader on the other, have not been
supported by empirical data showing what translators do in real life when
they translate legal texts or what the attitude to these different viewpoints
are among lawyers today.
In a Danish legal context, translations to and from English are, we
believe, mostly produced by professional translators, sometimes in
collaboration with lawyers. But occasionally Danish lawyers make their
own legal translations. Consequently, we are interested in examining
whether professional translators and lawyer translators approach the
question of fidelity and use of explicitation and implicitation in legal
translation similarly or differently. We will do so by analysing translations
of the same text produced by four lawyers and four professional translators.
Legal translation: lawyers and professional translators compared 341
2. Research questions
1
For an outline of relevance theory, see e.g. Blakemore (1995).
342 Dorrit Faber and Mette Hjort-Pedersen
The data on which the present analysis is based have kindly been made
available to us by a former MA student of LSP translation at the
Copenhagen Business School who, as part of her MA thesis, asked four
lawyers and four professional translators to translate parts of an English
pre-marital contract into Danish (Fischer 2008). Her purpose was different
from ours, in that she was not focusing on explicitation and implicitation
but rather on the two groups‟ preferences for source- and target-language
orientation. The identity of the participants is of course not revealed in the
thesis, but their names are known to us. The lawyers are all employed at a
major Copenhagen-based law firm while the professional translators have
worked in this capacity in various contexts for a number of years. Both
groups of translators have Danish as their mother tongue, and both groups
are familiar with Danish legal texts and their characteristic features. In the
2
For risk taking in translation, see Pym (2005).
Legal translation: lawyers and professional translators compared 343
translation process all of them had access to dictionaries and the Internet.
They had as much time as they needed to complete the task. They were
given a translation brief specifying the purpose of the translation as that of
informing a client of the contents of the pre-marital contract in view of her
intended marriage in England.
The data consist of translated texts and a number of prospective and
retrospective questions relating to the informants‟ translation process, e.g.
prospective questions about their initial choice of translation strategy, and
retrospective questions about elements that they found difficult to convey
in the translation. Thus the data are mainly product data but with some
information about how the informants experienced their own understanding
and translation process.
Pre-marital contract
Explicit information: a contract entered into before a marriage.
Implicated (encyclopaedic) information: This agreement is not binding
(under English law).
Legal translation: lawyers and professional translators compared 345
Ægtepagt
Explicit information: a marriage contract
Implicated (encyclopaedic) information: the contract can be made before or
after the wedding, and it is binding (under Danish law).
Faced with the translation of the term „pre-marital contract‟ into Danish the
translator has to make a choice, taking into consideration of course the
purpose of the translation of the text in question. One option available to
the translator is the following:
1. en ægtepagt indgået inden ægteskabet (ikke-bindende efter engelsk ret)
(“a marriage contract entered into before the marriage (not binding under
English law)”). This represents an explicitation of one implied
consequence, i.e., an implicature in relevance-theoretic terms, of the non-
binding nature of the agreement.
Another option might simply be:
2. en ægtepagt (“a marriage contract”). Here the pre-marital aspect is
implicitated and no implicatures are made explicit.
A third option might be:
3. en før-ægteskabelig kontrakt (“a pre-marital contract”). In this case there
is no explicitation of implied legal consequences and no implicitation of the
time when the contract is entered into (i.e., before or after the marriage,
which in a Danish context is legally less relevant).
6. Analysis
One of the prospective questions that was put to the participants before
they translated the text deals with the issue of fidelity and near-literalness;
both lawyers and translators were asked about their attitude to source- and
target-language bias in legal translation. One of the lawyers responded in
the following way:
I would prefer a very close translation for instance if the text is to be used in
court, or if it’s a certified translation. Alternatively, if the text is to be available
in both English and Danish, I would be less concerned with the fidelity or the
closeness of the translation, as it would typically contain a stipulation like “the
English version shall prevail.” [Our translation]
As far as possible I would have the translation be very close to the source text,
as the document is to be used in England and is about English law, but I would
as far as possible adapt the text to the Danish drafting style. [Our translation]
This would seem to indicate that the use of explicitation and implicitation
would be fairly infrequent with our text, but a detailed analysis of the ST
and TTs shows that this is in fact not the case. The text comprises two
clauses, 4.1 and 4.4 of the pre-marital contract. Table 1 summarises the
results of explicitations and implicitations made by the two groups of
translators.
Table 1. Explicitation and implicitation scores (L = lawyer; T = translator;
A = addition; S = specification; G = generalisation; R = reduction)
4.1 L1 L2 L3 L4 T1 T2 T3 T4
Explicitation A: 9 A: 2 A: 7
A: 5 A: 2A: 4 A: 5 A: 7
S: 1 S: 1 S: 2
S:1 S: 2S: 3 S: 1 S: 1
= 10 =3 =9 =6 =4 =7 =6 =8
Implicitation G: 1 R: 4 G: 1
G :2 G: 1R: 1 R: 1 R: 1
R: 3 G: 1 G: 1
=5 =2 =2
Total 4.1 10 + 1 3 + 4 9 + 1 6 + 5 4+1 7+2 6+2 8+1
4.4 L1 L2 L3 L4 T1 T2 T3 T4
Explicitation A: 4 A: 5 A: 2 A: 3 A: 2 A: 3 A: 3 A: 4
S: 2 S: 2 S: 2
=6 =7 =4
Implicitation R: 2 R: 2 R: 2 R: 5 G: 1 R: 1 G: 1 G: 1
Total 4.4 6 +2 7+ 2 4+2 3+5 2+1 3+1 3+1 4+1
Averages Lawyers Translators
Explicitations 12 9.25
Implicitations 5.5 2.5
Total averages 17.5 11.75
(expl. + impl.)
and Lawyer 4 having the highest scores in this category. This relatively
limited sample of data does not allow any clear-cut conclusions, but it
seems that there is a tendency towards a higher degree of explicitation and
implicitation manoeuvres in the lawyer group, who show a total average of
17.5 as against the translators‟ 11.75.
Example 1
Source text Lawyer 1 Back-translation
4.1 Each of the parties has 4.1 Hver af parterne har 4.1 Each of the parties has
acquired his or her separate erhvervet deres respektive acquired their respective
property independently of aktiver uafhængigt af og assets independently of and
and without any uden økonomisk bidrag fra without financial
contribution from the other den anden part. contribution from the other
and wishes to retain his or party.
her separate property,
whether owned now or Parterne (A) ønsker hver The parties wish both of
acquired later, free from især at bevare den fulde them to retain the full
any claim of the other by ejendomsret til deres ownership of their
virtue of the Marriage respektive aktiver (særeje), respective assets (separate
unless otherwise provided uagtet om aktiverne (A) er property), whether or not
for by this deed or in a erhvervet på nuværende the assets have been
supplemental deed tidspunkt eller erhverves acquired now or are
executed by them. senere. acquired later.
Lawyer 4
4.1 […] wishes to retain his 4.1 Parterne afskærer sig 4.1 The parties are
or her separate property, fra med henvisning til estopped from, by virtue of
whether owned now or ægteskabet at gøre krav the marriage, making
gældende mod den anden
acquired later claims on the other party‟s
parts (R) nuværende og
fremtidige ejendom present and future property
becomes
Example 6
Source text Lawyer 1 Back-translation
4.1 Each of the parties has 4.1 Hver af parterne har 4.1 Each of the parties has
acquired his or her separate erhvervet deres respektive acquired their respective
property independently of aktiver uafhængigt af og assets independently of and
and without any uden økonomisk (A) without financial contri-
contribution from the other bidrag fra den anden part. bution from the other party.
Strictly speaking, „contribution‟ might involve other types of contribution
than the mere financial one. Consequently, narrowing „contribution‟ to
„financial contribution‟ is an instance of explicitating one of several
possible implicatures. Lawyer 1 makes this implicature explicit in his TT
by adding økonomisk. We have no way of knowing whether this is a
conscious strategy, but we assume that he has encoded his interpretation
process linguistically in the TT, thereby narrowing down the number of
possible scenarios to one. A similar situation is seen in Example 7.
Example 7
Source text Lawyer 2 Back-translation
4.4. Neither party shall 4.4. Ingen af Parterne kan 4.4. Neither Party may
make any claim during the gøre krav helt eller delvist make any claim, neither
Marriage to the property in (A) på ejerskab af den wholly or in part to
the separate absolute anden parts fuldstændige ownership of the other
beneficial ownership of the særeje under ægteskabet, party‟s completely separate
other and each of them og hver Part frafalder property during the
agrees to release all rights uigenkaldeligt (A) enhver marriage, and each Party
and not to pursue claims in rettighed en Part måtte irrevocably releases any
respect of any rights which erhverve […] rights a Party may acquire
he or she might acquire […]
[…]
In the English ST, „any claim‟ has a very broad meaning. This meaning is
narrowed down in Lawyer 2‟s translation, in that information is added
about the type of claim, i.e. ejerskab („ownership‟), and the ways in which
this ownership can be held, i.e. helt eller delvist („wholly or in part‟).
Similarly, the ST refers to releasing all rights, and in the translation, this
situation is specified as releasing such rights irrevocably. Again, these
additions may reflect the interpretation processes of Lawyer 2.
Similar instances of narrowing of legal conceptual content are not
seen in the translator translations. As we have no process data, it is not
possible to say whether the same types of inference processes were
undertaken by the translators but without leaving any traces in the TTs.
354 Dorrit Faber and Mette Hjort-Pedersen
7. Conclusion
The analysis shows that explicitations and implicitations are used by both
groups. It also shows that the lawyers, generally speaking, undertake these
transfer operations to a higher degree than the translators, and furthermore
that there is a greater preference for implicitations in the form of reductions
in the lawyer group.
We identify three major reasons behind choices of explicitations and
implicitations. The first has to do with drafting style conventions, which are
different in the two legal systems. Adapting to target language conventions
in terms of dividing sentences leads to the use of additions. Members of
both groups use this strategy, which does not involve any risk since the
operation mainly entails the insertion or repetition of relevant lexical
elements to comply with syntactic needs.
However, when it comes to implicitations a different pattern
emerges. Here quite long qualifying phrases as well as near-synonymous
expressions, which characterise the drafting style and explicitness level of
English contracts, are left out in the lawyer translations. This, of course,
involves a risk since information is transformed from being explicit to
being implicit, which presupposes that the TT reader is able to make the
necessary inferences intended by the ST. The translators do not make such
implicitations; they stay quite close to the ST structure, and do not seem to
want to assume the same kind of risk as the lawyers.
Legal translation: lawyers and professional translators compared 355
References
Paul Kußmaul
Abstract
Translating is not always smooth and steady. Arnt Lykke Jakobsen (2005)
observed translation processes via the typing behaviour of translators and found
that the cognitive rhythm, as he calls it, of translators was disrupted by pauses in
a significant way. Pauses occurred when translators were facing a semantic
obstacle (Jakobsen 2005: 113). When trying to overcome such obstacles,
translators, in my experience, are often faced with the question how precise they
have to be when translating a text. I have been involved in the translation of
social surveys for some years now, and there this question seems especially
pertinent, as will become evident from some of my examples below. The
translation of social surveys created a new incentive for me to reflect again upon
the principle of the necessary degree of precision. I shall begin by discussing
some closely related concepts.
There is a notional connection between precision and literalness. A literal
translation, so the common thinking goes, is more precise than a free translation.
1
I am making use of some of the material from my book Verstehen und Übersetzen (2007),
but in the way it is presented here I hope to have added some new ideas.
360 Paul Kußmaul
risks (Klaudy 2009: 108). The more explicit I am, translators seem to think, the
more precise my translation will be and the more sure I can be that there will be
no loss of information. Being more explicit than the source text is a technique,
no doubt, that can increase precision. But translators need to know the
conditions requiring more explicitness and how explicit they should be. From
the point of view of Descriptive Translation Studies, explicitation has been
described as a common phenomenon; Chesterman even calls it a translation law
(1997: 71). From the point of view of translation didactics, which is actually my
own point of departure, mere description is not quite satisfactory. Further
questions should be asked about how comprehension functions psychologically
and how translation functions when one looks at its purpose, as I shall show
below.
The concept of equivalence, which has been discussed a great deal in
Translation Studies and in translator training right from the beginning, has not
been associated explicitly with minimising risks so far. However, the fascination
that the notion held for translators was based on the hope of finding symmetrical
items between two languages, and if this had been possible, there would no
longer have been risks. With the widening of the field of linguistics, pragmatic
factors were included, and thus the notion of equivalence became more and
more diversified (for overviews, see Shuttleworth & Cowie 1997: 49–51;
Neubert 2004: 338; and Koller 2004: 351). From a functional point of view (i.e.
looking at the purpose of the target text, see below) the notion of equivalence
might still be used if the function of the target text is the same as that of the
source text.
This is the case in the translation of social surveys, which I mentioned
above. Equivalence for survey questions means that exactly the same question
must be asked in the target text as in the source text, in order to guarantee
correct comprehension, and prevent respondents from falsifying their answers.2
In survey translation research this is called the ASQ (Ask the Same Question)
Model (Harkness 2003: 35). To achieve this goal a number of steps are taken
before a translation is used in a large and expensive survey. These are called the
TRAPD procedure, where the acronym stands for Translation, Review,
Adjudication, Pretesting, Documentation (Harkness 2003: 38). This large
number of steps forms the organisational basis for equivalence and can indeed
2
I know that this is a very optimistic, if not unrealistic, aim. Deconstructionists would have a
lot to say to this, but such is the situation in survey translating.
362 Paul Kußmaul
3
Jakobsen observed that pauses in the translation process were often followed by “peak
performances”. This kind of behaviour, he says, might be a sign that “having spent
processing time on identifying a subject domain or a communicative intention or on finding
an appropriate style, experts are able to use their specific-subject domain identification or
their understanding of a communicative intention to control the construction of a familiar
scene in an appropriate style” (Jakobsen 2005: 115). This interpretation, I think, has a close
affinity to the criteria Hönig and I suggested for deciding on the degree of precision and can
be seen as a skill of risk reduction.
The necessary degree of precision revisited 363
whose graduates a large number of the political and economic leaders are still
being recruited).
would be much too detailed (i.e. overdifferentiated). Although the translator has
made explicit what was implied in the English sentence, so much additional
information has been provided that the most important information, namely
“expensive”, has almost been blotted out. If, however, the topic of the text had
been “the British educational system”, such a translation might have been
possible.
In this example the item in question is a cultural term, but the principle of
the necessary degree of precision can be used for solving problems occurring
with the translation of any word; indeed, as we showed in the individual
chapters of Strategie der Übersetzung, for problems occurring at all linguistic
levels: style, syntax, cohesion of sentences, and illocution.4 I shall discuss a non-
cultural example below.
With respect to translation errors, the principle of the necessary degree of
precision was occasionally understood in too general a sense. When my students
read Strategie der Übersetzung they sometimes asked: “Can we still use the
terms „correct‟ and „wrong‟ when discussing translations or do we have to
replace what was called „wrong‟ until now by either „underdifferentiated‟ or
„overdifferentiated‟?”
This is a point Hans Hönig and I should have discussed, and which was
taken up by Anthony Pym ten years later (Pym 1992), who proposes
distinguishing between binary and non-binary errors. Binary errors, he states,
can be described within the dimension of right/wrong; that is, we can make a
qualitative judgement. For instance, if sentence (1) had been translated as
(1c) Im Parlament kämpfte er für Chancengleichheit, aber seinen eigenen Sohn schickte
er auf die öffentliche Schule in Winchester (… but his own son he sent to the state
school in Winchester).
this would simply be wrong. As we know, the famous school there is not
“public” in the sense of a state school, but rather private.
Non-binary errors, however, meet with commentaries such as “It is not
exactly wrong, but …” or “Maybe we can improve it a bit and say…”. These are
quantitative judgements and, from this point of view, translations can be judged
by saying, for example, that they are not good enough, are better than another
4
Some years later I underpinned our principle with Grice‟s maxims of quantity and
relevance (Kußmaul 1995: 92f.).
The necessary degree of precision revisited 365
Let us go back to the stage where we have to decide between variants. For “a
large sum of money” quite a number of translations suggest themselves in
German (tentative English translations have been added in brackets):
366 Paul Kußmaul
From the perspective of risk management, we may say that risk can be reduced
when we verbalise the core notion of the scene.5 It is interesting to see that our
translation is less detailed than the source text as to the amount of money
mentioned, but it makes use of the principle of explicitation, thus facilitating the
evocation of a scene.
Regarding another linguistic level to which the principle of the necessary
degree of precision can be applied, namely that of style, we can see that some of
the other translation variants listed earlier differ on this dimension. For instance,
eine beträchtliche Geldsumme is more formal, i.e. implies a more distant author-
reader relationship than the colloquial eine Menge Geld, which is a sign of
greater familiarity. From a functional point of view, translators will have to
imagine the kind of relationship between interviewer and respondent (the
questionnaire scene) in order to make their stylistic decisions. An oral interview,
for instance, will have a different style from a written interview.
To sum up this section, what translators working in a functionalist
framework do with words is determined by the function of the words (here the
topic “borrowing money”) within a passage of a text that suggests a specific
scene (here the support scene). The text is potentially seen as embedded in a
specific situation (here the questionnaire situation) in a specific culture (here
German culture). For instance, if the questionnaire has to be translated for a
culture where borrowing money is improper, one would have to find some other
equivalent material that can be borrowed, such as perhaps food.
Our focus so far has been on the target text. Writing the target text, however,
occurs in close and constant interaction with understanding the source text. As
normal readers, we may occasionally have the feeling that the meaning of a
word or passage we have read is not quite clear to us, but we usually do not
bother too much about it and are sometimes content with a vague understanding.
Translators are in a different situation. They have to express in writing what they
have understood, and remaining vague is usually not what is expected of them. 6
5
In a more recent article on survey research, Fillmore draws attention to the fact that
questions will evoke prototypical notions in the minds of respondents (1999: 192f., 196).
6
It has been observed in eye-tracking studies that source-text comprehension seems to be a
problem, especially for translation students (Jakobsen & Jensen 2008: 112, 119, Sharmin et
al. 2008: 48). Specifically, measuring the eye-key span, i.e. the time between the eye
The necessary degree of precision revisited 369
fixation on a specific source-text item and its translation, can be used as a means of
identifying words that cause problems (Dragsted & Hansen 2008: passim). Seen from a
didactic point of view, cognitive models can perhaps help to make comprehension a more
conscious affair and overcome problems with the translation of words.
370 Paul Kußmaul
one finds the German equivalent Kommune. This, like Gemeinde, is also a
technical term used in city administration, but for the non-specialist, and perhaps
for some older German readers, this word might also suggest the way anti-
bourgeois young people of the 1968 generation lived together and shared
possessions and responsibilities. A dictionary on the Internet (dict.cc) offers
Gemeinwesen. This is certainly not as ambiguous as Kommune, but is still a
technical term and not easy to understand for the normal reader. The Longman
Dictionary of Contemporary English, which being a monolingual dictionary is
always much more useful in cases such as this, offers: “people who live in the
same area, town etc.”, a definition, however, which still remains vague in terms
of the size.
We can complement linguistic with encyclopaedic information. In
Wikipedia we find
of the city, town or village, because they would not feel concerned about a more
distant area. As a translation I would suggest:
(4b) … in der Gegend, in der Sie wohnen … (… in the area where you live …)
This is a paraphrase of the words in the source text – a very common and indeed
often advisable technique in translating. In the ISSP 2008 in the context of a
different survey topic (religion), the notion of “local community” was expressed
by means of “neighbourhood”, and in the review phase we discussed Gegend,
Viertel and Wohngegend as possible translation variants and finally decided on
Wohngegend. This could also be used here, since it refers to a smaller area and
has the implications needed for our purposes. Translators who feel attached to
the term “equivalence” may use the term “equivalent” – I have no objections –
for both translations.
5. Conclusion
Risk management, as we have seen, does not mean risk avoidance, but rather
risk reduction. Risk reduction can be achieved by translating according to the
principle of the necessary degree of precision. The criterion for precision is
provided by the function of the item to be translated. In our examples, the
function of the survey questions in the translation was the same as in the source
text. Still, translating the questions in such a way that respondents can smoothly
and easily understand them, i.e. imagine the right kind of scene, is not an easy
task. It is via the scenes evoked by the words that comprehension and translation
are closely connected. Making use of the notions of prototypicality and scenes,
we may say that sticking to the core of the scene that is evoked by the source
text will help translators find sufficiently precise words for their translations.
References
Barclay, J. R., Bransford, J. D., Franks, J. J., McCarrell, N. S. & Nitsch, K. 1974.
Comprehension and semantic flexibility. Journal of Verbal Learning and Verbal
Behaviour 13. 471–481.
Chesterman, A. 1997. Memes of Translation. The Spread of Ideas in Translation
Theory. Amsterdam/Philadelphia: John Benjamins.
Dragsted, B. & Hansen, I. G. 2008. Comprehension and production in translation: a
pilot study on segmentation and the coordination of reading and writing
processes. In S. Göpferich, A. L. Jakobsen & I. M. Mees (eds). Looking at Eyes.
372 Paul Kußmaul
Riitta Jääskeläinen
Abstract
1. Introduction
2. Overview of definitions
To start with things that happen with texts, Venuti (1998: 240) describes
translation strategies as follows:
Strategies of translation involve the basic tasks of choosing the foreign text to be
translated and developing a method to translate it. Both of these tasks are
determined by various factors: cultural, economic, political. Yet the many
different strategies that have emerged since antiquity can perhaps be divided into
two large categories. A translation project may conform to values currently
dominating the target-language culture [...] [i.e. domesticate]. Alternatively, a
translation project may resist and aim to revise the dominant [target-culture
values, i.e. foreignise].
1999, Hansen 2006, Shih 2006) that translators often base their decisions
on intuitive criteria along the lines “this sounds better/odd/fluent” or “this
doesn‟t sound like Finnish/Danish/etc.”. It is difficult to determine whether
such criteria reflect internalised knowledge or genuine language intuition
(finding answers to this question could be an interesting research project).
In their textbook of translation, Hönig & Kußmaul (1982) also
mention the need for a strategy which would help the translator to find an
optimal solution for each particular translation problem. According to them
(1982: 58), the translator decides on a general approach on the basis of the
ST and the translation brief; the general approach, or strategy, will guide
the subsequent syntactic and lexical decisions. As will be shown below,
there is evidence that Hönig & Kußmaul‟s didactic recommendation is
what professional translators and advanced students of translation tend to
do in practice.
Scientist. In example (1), the student first states that style may remain the
same, the text should perhaps be simplified a little, which would mean, for
instance, omitting some of the medical terms (e.g. nicotineamide). The
TAP excerpts have been translated from Finnish; the numbers in brackets
refer to the length of pauses in seconds and (.) to pauses shorter than one
second.
(1) Advanced student of translation (at the beginning of the process):
uhm the style of this piece is really quite (.) quite similar to that there (1) in the
column in Hesari (1) it needn‟t be worked on very much (7) except made a little
bit simpler perhaps (4) I don‟t know if (.) the readers of the Helsingin Sanomat
are all that int-interested in (.) what is (2) nicot- nico- nicotine amide (2) etcetera
(15) (source: Jääskeläinen 1999)
In example (2), the same student verbalises her decision to leave out
the reference to Experientia, where the research reported in the ST has been
published. By pointing out that “nobody‟s interested in it” she seems to be
referring to the TT readers whose needs she considered while outlining her
global strategy in example (1).
(2) Advanced student of translation (on leaving out the reference to
Experientia):
that reference needn‟t be written anywhere
nobody‟s interested in that or if somebody is then he can look it up himself (11)
(source: Jääskeläinen 1999)
representation of a style, that is, a way of doing this global action in the most
effective way.
CONTEXT
- socio-cultural context: translation policy etc (Venuti 1998)
- details of the commission: who, what, for what purpose (Skopos) and for whom
TARGET TEXTS
4. Conclusion
We now have a rough idea what translation scholars talk about when they
discuss translation strategies. We also have a sketch of a map showing how
the different strategy notions might be related to each other. Obviously,
more work remains to be done. Often „translation strategies‟ are used
Looking for a working definition of „translation strategies‟ 385
loosely, without any definitions, and the sense in which they are used can
only be deduced from the context. Furthermore, the ways in which and the
extent to which the strategy notions overlap is still not clear. And at this
point of my strategy-related musings it seems premature to suggest
standardisation of terminology or the replacement of some “strategies” by
e.g. policies or methods.
The theoretical survey of the different uses and definitions presented
here must be followed by a thorough empirical study to determine the
viability of the categories proposed and to identify what kind of process-
and product-related phenomena would qualify as one kind of strategy or
another. The research material for such a study should consist of process
and product data, including STs and TTs, translation briefs, keyboard logs,
verbalisations and observational data. Some strategic information will
appear in verbal reports, while others may only appear as consistent
patterns in TTs or keyboard logs, such as those created by Translog (e.g.
Jakobsen 1999, 2003).
References
Gyde Hansen
Abstract
1. Introduction
changes and revisions during the process, e.g. Livbjerg/Mees (1999, 2003)
and Hansen (2003, 2006a, 2008). In Hansen (2006a: 106f), this product
evaluation at a micro level is an important parameter.
In this article, evaluation of products refers to both, the evaluation of
final products and the product evaluation at a micro level.
served the purpose perfectly – despite the extremely high number of errors
in the translated brochures. An example:
These two examples represent extreme ends of the scale but, depending on
the communication situation and receivers’ expectations, in between these
two opposite sides there are several degrees of usability and acceptability.
Interestingly, the same website accessed in September 2009 offers a
machine translation into German which does not make much sense. At
present, however, it is this kind of nonsense we expect when we touch
“Übersetzen”.
It is not always easy clearly to draw the borderline between
“linguistic errors” or “language errors” (e.g. wrong choices as to grammar
and vocabulary) and “translation errors” or “functional errors” (i.e. flaws
that may have a negative influence on the reception of the message in the
communicative situation, as stated by, among others, Pym (1992: 279ff.),
Kußmaul (1995: 129ff), Waddington (2001: 314) and Martínez
Melis/Hurtado Albir (2001: 281). The difficulty lies in the fact that
grammatical and lexical errors usually have an immediate impact on the
acceptability in the communication situation because the readers’ attention
is automatically directed towards them. Umbreit (2001: 253), for example,
394 Gyde Hansen
has observed that linguistic errors affect the way a message is received
because readers lose confidence in the text and cannot take it seriously.
the usability of the text. However, a study I carried out (Hansen 1999: 57)
has shown that these evaluations without the source text are not reliable.
Assessments of this type can perhaps be useful as a supplement to other
evaluations. In translation process research, they are problematical, not
only because they are carried out without the source text but also because
the potential receivers often do not have the necessary concepts at their
disposal to explain and justify their observations and assessments.
Obviously, one has to be trained in the evaluation and revision of
(translated) texts (Hansen 2008, 2009).
which draw on translation process data can be difficult because one has to
be trained in reading log files. They should be carried out by the researcher
or a competent colleague, but also in accordance with the criteria the
evaluators have agreed on.
Even if evaluators agree on the quality criteria, they may still have different
ideas about errors and their gravity. One problem in this connection is the
lack of agreement there may be among evaluators on the units or linguistic
entities assessed. This has an impact on a consistent analysis and
interpretation of the data. In the case of unwarranted omissions, for
example, notably in process research under time pressure (Hansen 2006b),
it has to be taken into account that evaluators make their judgements using
different entities. Some count every missing word as one error while others
are more tolerant and work with larger entities. For process research, it is
crucial to discuss this problem with the evaluators and to reach an
agreement.
A fundamental question is whether it is only errors that should be
registered and evaluated or whether holistic evaluations are needed in
which good translation solutions are taken into consideration (see Martínez
Melis/Hurtado Albir 2001). In the latter case, we would need precise
criteria for “good” translations since statements like “fluent”, “partly
fluent” or “easy to read” (e.g. Jääskeläinen 1999: 112) depend on
individual, subjective perception.
There are several approaches to holistic descriptions, e.g.
Gerzymisch-Arbogast (2001: 230ff; 2004: 69ff) who suggests that
network-based parameters of coherence and thematic and isotopic patterns
are included in the catalogue of evaluation criteria. Williams (2001)
proposes the application of argument macrostructures complementary to
other approaches of translation quality assessment based on the theory of
argumentation by Toulmin (1964).
However, for translation process research, the situation is still that
good translation solutions are assessed differently by different evaluators
and that there is a lack of reliable procedures for holistic assessment,
Evaluation of translation products in translation process research 397
One way of dealing with the above-mentioned problems and avoiding the
risk of bias would perhaps be to give the evaluators predefined criteria and
guidelines prior to their assessment. Martínez Melis/Hurtado Albir (2001:
283) mention several kinds of evaluation, “intuitive assessment” without
criteria and “reasoned assessment” with objective criteria.
If evaluators are not given the criteria to be applied, they will use
their own criteria. However, if they receive predefined criteria and
guidelines from the researcher, the results may lack validity because of
potential bias from the researchers’ interests. Waddington (2001)
investigated evaluation processes in which the evaluators not only got
precise criteria in advance but were also trained in how to apply them. One
of the results of his study was that this method seemed to augment
criterion-related validity (2001: 322ff).
A procedure I would like to propose is an interaction between an
individual spontaneous assessment according to the evaluators’ own norms
followed by a description of the criteria applied by them and a discussion
of their results. They have to be able to describe problems encountered
during the evaluation process and to justify their decisions. As a next step,
the individual problems, decisions and results could be compared and an
agreement reached on a set of shared quality criteria which are
subsequently applied systematically to all the translation products (Hansen
2006a: 112). During the dialogue, different theoretical orientations,
attitudes and ideas can be clarified and an agreement can be reached.
The evaluators can also be asked to try to solve the translation task
themselves. This increases their sensitivity to special characteristics of the
source text and potential translation problems (Hansen 1996: 156). In
addition, they become more alert to other translators’ good, acceptable or
398 Gyde Hansen
References
Hansen, G. 2007. Ein Fehler ist ein Fehler, oder …? Der Bewertungsprozess in
der Übersetzungsprozessforschung. In G. Wotjak (ed.). Quo vadis
Translatologie? Ein halbes Jahrhundert universitäre Ausbildung von
Dolmetschern und Übersetzern in Leipzig. Rückschau, Zwischenbericht und
Perspektiven aus der Außensicht. Berlin: Frank & Timme. 115–131.
Hansen, G. 2008. The speck in your brother’s eye – the beam in your own:
quality management in translation and revision. In G. Hansen, A.
Chesterman & H. Gerzymisch-Arbogast (eds). Efforts and Models in
Interpreting and Translation Research. Amsterdam/Philadelphia: John
Benjamins. 255–280.
Hansen, G. 2009. A classification of errors in translation and revision. In
Forstner et al. (eds). 313–326.
Hansen, G. & Hönig, H.G. 2000. Kabine oder Bibliothek? In M. Kadric, K.
Kaindl & F. Pöchhacker (eds). Translationswissenschaft. Tübingen:
Stauffenburg. 319–338.
Hönig, H.G. 1995. Konstruktives Übersetzen. Tübingen: Stauffenburg.
House, J. 1997. Translation Quality Assessment: A Model Revisited. Tübingen:
Narr.
House, J. 2001. Translation quality assessment: linguistic description versus
social evaluation. Meta 46(2): 243–257.
ISO 9000ff. International Organization for Standardization. British Standards
Institute.
Jääskeläinen, R. 1999. Tapping the Process. Joensuu: University of Joensuu.
Jakobsen, A. L. & Schou. L. 1999. Translog documentation. Copenhagen Studies
in Language 24: 149–184.
Jakobsen, A. L. 2003. Effects of think aloud on translation speed, revision, and
segmentation. In Alves (ed.). 69–95.
Koller, W. 1979. Einführung in die Übersetzungswissenschaft. Heidelberg:
Quelle & Meyer.
Krings, H. P. 1986. Was in den Köpfen von Übersetzern vorgeht. Tübingen: Narr.
Künzli, A. 2007. Translation revision. In Y. Gambier, M. Shlesinger & R. Stolze
(eds). Doubts and Directions in Translation Studies. Amsterdam/
Philadelphia: John Benjamins. 115–126.
Kupsch-Losereit, S. 2008. Vom Ausgangstext zum Zieltext. Berlin: Saxa Verlag.
Kußmaul, P. 1995. Training the Translator. Amsterdam: Benjamins.
Lee-Jahnke, H. 2001. Aspects pédagogiques de l’évaluation en traduction. Meta
46(2): 258–271.
Livbjerg, I. & Mees, I. M. 1999. A study of the use of dictionaries in Danish-
English translation. Copenhagen Studies in Language 24: 135–147.
Livbjerg, I. & Mees, I.M. 2003. Patterns of dictionary use. In Alves (ed.). 123–
136.
Magris, M. 2009. Übersetzungsfehler und deren Auswirkungen in der Praxis. In
Forstner et al. (eds). 301–311.
Martínez Melis, N. & Hurtado Albir, A. 2001. Assessment in translation studies:
research needs. Meta 46(2): 272–287.
Evaluation of translation products in translation process research 401
Copenhagen Business School in 2008. His current research interests are related
to the investigation of human translation processes and interactive machine
translation. Prior to his position in Denmark, he was employed at the Institut für
Angewandte Informationsforschung (IAI) in Saarbrücken, Germany, where he
carried out research on machine translation, terminology tools, and the
implementation of natural language processing software, mainly for extending
and maintaining IAI’s Controlled Language Authoring Tool (CLAT). He has
published more than 60 reviewed articles in journals and international conference
proceedings, and is the editor of a reference book on example-based machine
translation.
E-mail: mc.isv@cbs.dk
Andrew Chesterman was born in England but moved to Finland in 1968 and
has been based there ever since, at the University of Helsinki, where he has
mostly taught English and translation theory. His current position is in the
Department of General Linguistics, where he is professor of multilingual
communication. His main research interests are in contrastive analysis,
translation theory, translation norms and ethics, and translator training. He was
CETRA Professor in 1999 (Catholic University of Leuven), and has an honorary
doctorate from the Copenhagen Business School. Main books: On Definiteness
(1991, CUP), Memes of Translation (1997, Benjamins); Contrastive
Functional Analysis (1998, Benjamins); with Emma Wagner: Can Theory
Help Translators? A Dialogue between the Ivory Tower and the Wordface (2002,
St. Jerome Publishing); and with Jenny Williams: The Map. A Beginners’ Guide
to Doing Research in Translation Studies (2002, St. Jerome Publishing).
E-mail: chesterm@mappi.helsinki.fi
Translation; since 2006 she has been Full Professor of Interpreting and
Translation Studies at Stockholm University. She has published widely on Slavic
linguistics, and on translation and interpretation theory, practice and didactics.
Recent works include the monograph Expertise and Explicitation in the
Translation Process, published by Benjamins in 2005, and the co-editing of the
proceedings from the conference “Critical Link 4”. She is a member of the
advisory board of several international journals in Translation Studies.
Email: birgitta.englund@tolk.su.se
Gyde Hansen, professor dr. habil., has taught at the Copenhagen Business
School since 1978 in the disciplines: comparative linguistics, text linguistics,
intercultural communication, semiotics and marketing, translation theory and
practice, translation revision, philosophy of science and empirical research
methods. Since 2004, she has been Vice President of EST (European Society for
Translation Studies). Her research projects include: TRAP (Translation processes,
from 1996-2002), the Copenhagen Retrospection Project (2004), a longitudinal
406 Notes on contributors
Moshe Koppel received B.A. and M.A. degrees in mathematics from Yeshiva
University and a Ph.D. in mathematics from New York University. He was a
post-doctoral Fellow at the Institute for Advanced Study in Princeton and is
currently on the Faculty of Mathematics and Computer Science at Bar-Ilan
University in Israel. His research focuses mainly on machine learning, especially
text-related applications. He has run a successful stock-trading strategy at a
hedge fund and has served as chief technological consultant and chief scientist of
a number of Internet companies. He has also written two books on Jewish law
and serves as an advisor to Israel’s Knesset on constitutional law.
E-mail: moishk@gmail.com
Iørn Korzen is professor of Italian and responsible for Italian Studies at the
Copenhagen Business School (CBS). He holds an MA in Italian Language and
Literature from Copenhagen University and a Danish doctorate on the Italian
article system from CBS. He is the author of more than 100 publications on
Italian and typological linguistics; during the last 10 years his main focus has
been on discourse structure and anaphoric relations from a contrastive Romance
– Germanic perspective, with special reference to Italian and Danish. He is a
member of Società Internazionale di Linguistica e Filologia Italiana, of which
he was President from 2002 to 2004.
E-mail: ik.ikk@cbs.dk
Brenda Malkiel has the distinction of being the first scholar to use Translog in a
study of Hebrew–English translation. She has published three articles about what
Translog can teach us about the translation process: “Directionality and
Translational Difficulty” (Perspectives), “The Effect of Translator Training on
Interference and Difficulty” (Target), and “From Ántonia to My Ántonia:
Tracking Self-corrections with Translog” (Copenhagen Studies in Language).
Recent publications include “Teaching Translation in the Hebrew Language
Program” (Issues in the Acquisition and Teaching of Hebrew), “Translation as a
Decision Process: Evidence from Cognates” (Babel), and the autobiographical
“This Is Love” (All of Our Lives).
E-mail: brendamalkiel@gmail.com
Inger M. Mees read English at Leiden and Edinburgh and obtained her doctorate
and first university appointment at Leiden. In 1985 she took up her post as
Associate Professor at Copenhagen Business School specialising in English
phonetics and pronunciation. Together with Inge Livbjerg, she has published
research on dictionary use in translation employing think-aloud protocols. With
Beverley Collins, she has co-authored Phonetics of English and Dutch (Brill,
1996), The Real Professor Higgins: The Life and Career of Daniel Jones
(Mouton de Gruyter, 1999). Recent work includes Practical Phonetics and
Phonology, (Routledge, 2008), and editing Daniel Jones’s Selected Works, 8 vols,
(Routledge 2003) and Phonetics of English in the Nineteenth Century, 7 vols,
Routledge 2006).
Email: im.isv@cbs.dk
in 1993. Muñoz coordinates the research efforts of the group Expertise and
Environment in Translation (PETRA, Spanish acronym). His main research focus
is on the interface between cognitive science and empirical research of
translation processes. Muñoz is an associate professor of translation at the
University of Las Palmas de Gran Canaria.
E-mail: rmunoz@dfm.ulpgc.es
conference interpreting from the University of Vienna and from the Monterey
Institute of International Studies, and has been working freelance as a conference
and media interpreter since the late 1980s. Following his doctoral research on
simultaneous conference interpreting, he has worked on community-based
interpreting in healthcare and asylum settings as well as on general issues of
interpreting studies as a discipline. He is the author of Introducing Interpreting
Studies (Routledge, 2004) and co-editor of the journal Interpreting.
Email: franz.poechhacker@univie.ac.at