ASLING 2018 - TC40 @ London Professional Technology for Interpreter - - PowerPoint PPT Presentation

asling 2018 tc40 london professional technology for
SMART_READER_LITE
LIVE PREVIEW

ASLING 2018 - TC40 @ London Professional Technology for Interpreter - - PowerPoint PPT Presentation

ASLING 2018 - TC40 @ London Professional Technology for Interpreter Education About Productivity software for translation revision that speeds up the revision process guarantees consistency and objectivity, and works independent of


slide-1
SLIDE 1

ASLING 2018 - TC40 @ London

slide-2
SLIDE 2

Professional Technology for Interpreter Education

slide-3
SLIDE 3
slide-4
SLIDE 4

About

Productivity software for translation revision that Ø speeds up the revision process Ø guarantees consistency and

  • bjectivity, and

Ø works independent

  • f language or domain
slide-5
SLIDE 5

What is translationQ no not?

Computer-Assisted Translation (CAT) Machine Translation (MT)

slide-6
SLIDE 6

We are a CAR, not a CAT (but we can live together as you can see…)

CAR = computer assisted revision

slide-7
SLIDE 7

Revision today…

slide-8
SLIDE 8

Revision today…

Impossible to get

  • verviews and

statistics and evolutions…

slide-9
SLIDE 9

Meet the translationQ pilot audience

96 admin accounts 45 organisations 22 universities 11 countries 803 users

slide-10
SLIDE 10

How did we gather the “pilot” input?

3 surveys for pilot customers 1 student survey 1 Special Interest Group Numerous e-mails, conversations and visits

slide-11
SLIDE 11

Event and follow-up surveys

  • Which are the top 3 features we should add to translationQ as soon as

possible for you?

Need to be able to go back and edit assignments already published (e.g., due date, title of file) Po Possibility to cu customize e the e er error ca categories Automatic segmentation Po Positive feedback to translators Im Import differ eren ent file e types, es, incl cluding g .docx cx, . , .sd sdlxl xliff, . , .sl sliff, , .xl xlf, , Uploading multiple (Word) documents for revision Ability to import users lists (Excel) Link error memory to user group Statistics Possi ssibility to ch change e th the ma maximum score áf áfter revisi sion process cess Add gen ener eric c feed eedback ck Po Possibility to overwrite th the “system” m” score

slide-12
SLIDE 12

Follow-up surveys

  • Which are the top 3 features we should add to translationQ as soon as

possible for you?

Final translated text next to source text and revised translation Adding multiple revisers to a revision project Uploading source text with original formatting (incl. paragraph structure) and possible image Sh Sharing of error / revision mem emories es acr cross ss a course se tea eam

slide-13
SLIDE 13

Follow up survey 1

  • Wh

Which er error

  • r ca

categori ries do do you you us use tod today?

– TAUS DQF – MeLLANGE – Custom set of categories – Grid used by the Italian Translators' Association (AITI) + the grid we used at the University of Trieste – …

  • Co

Conclusion: allow customisation, but take into account data sharing

slide-14
SLIDE 14

Follow up survey 1

  • How d

How do

  • you

you sc scor

  • re

e tr transl slati tion

  • ns tod

today?

– A score from A to F to 4 questions: style, terminology, fluency, general opinion – PIE-scores – I have my own system, and need to be able to change the maximumscore afterwards to continue using my system – I establish how serious errors are, I grade them from -0.25 to -2, I give points from +0.25 to +2 for good solutions, I decide the pass/fail threshold for each source text translation and then try to apply it consistently – Max score per section (Target language, Meaning, Specialized Meaning and Functional Adequacy) – …

  • Co

Conclusion: allow customisation

slide-15
SLIDE 15

Conclusion

  • translationQ is the result of co

co-cre creation between KU Leuven, Televic and all the participants in the academic pilot project

  • Your input and feedback is crucial to help us achieve our goal:

bui buildi ding ng the he be best revision n tool. Th Thank you!

slide-16
SLIDE 16

But there is more…

Ø Big Wrong Data

slide-17
SLIDE 17

With translationQ… Get ins insig ights hts in

Every translator‘s performances and evolutions Common errors Common error categories …

slide-18
SLIDE 18

Reports available in translationQ

  • Insights in a translator’s performance
slide-19
SLIDE 19

Reports available in translationQ

  • Translation insights
slide-20
SLIDE 20

What we learn from the error memories

  • Data from UNI Padova (grazie mille!)

– Number of different errors – Average frequency on an error – # of unique errors vs. # of repeated errors

  • Impressive numbers of errors

– E.g. Hogeschool Zuyd (NL)

  • # 2 revisions

– 1 with 10 submitted translations – 1 with 9 submitted translations

  • 405 different errors defined
slide-21
SLIDE 21

Revision memory insights

The magic of dynamic tables…

  • Frequency of errors

(à efficiency of translationQ)

  • Score distribution

– with real error view – consistency of scoring

  • Scoring per

error category

slide-22
SLIDE 22

What you can learn from Big Wrong Data

Most frequent translation errors

error categories language pairs

Strengths and weaknesses of each translator Averages and evolutions per translation agency Level of difficulty of a source text …

slide-23
SLIDE 23

What we learned from our database

  • Error categories per translator

– View per translator

  • E.g. Speranza

– View per category

  • Ranking of categories
  • Who’s the category queen/king?

– Relation between suggested/accepted/rejected errors (Is this the path to “generic errors”? More research needed!)

slide-24
SLIDE 24

What we learned from the revisors

  • Different revision “habits”

– E.g. Some use “Feedback” instead of “Correction” – Very different scoring systems

  • Some score on 100

– With scores ranging from -20 to +2

  • Some score on 10

– With scores ranging from -0,25 to 0

  • Too many revisors use “None” or “Uncategorised” as error category

– Should we prevent it?

slide-25
SLIDE 25

What we want to research in the near future

  • How reusable are revision memories?

– Differences per language pair – Within same/across domains – Within one/across institution(s) – With multiple revisors (how to do it efficiently?)

  • Relationship between revision memories and

translation memories

– What happens if revision memories are applied to TMs? – How can revision memories help to improve the quality of TMs? – What if CAT tools link their revision module to translationQ? – What if we compare the errors made by CAT and made by humans?

slide-26
SLIDE 26

What we want to research in the near future

  • Relationship between translation and interpreting

– What we apply translationQ to interpreter errors? Interpreting Error Memories?

  • Allow evaluators to bookmark interpreting errors
  • Allow evaluators to categorize interpreting errors

and adding

  • Having overviews and samples of
slide-27
SLIDE 27

Ongoing research

  • Time and effort spent in

– human revision vs. – car

  • Intra-rater reliability
  • Inter-rater reliability

– scores – error memories (do they find the same errors?) – how consistent are they (same errors, same penalties?)

  • Measure the influence of (teaching) activities (and feedback)
slide-28
SLIDE 28

What What is is tr trans anslatio lationQ nQ us used d for?

Ev Evaluating and ranking lar large groups of f tran ansla lators fo for selection and recruitment Translation

  • n student

evaluation

  • n and revision
  • n.

Or Organize and

  • b
  • bjectively scor
  • re

hi high h stakes es trans nslation n exams ms To

  • find and reuse

co common errors to to create te adaptive exercises De Define the difficulty y level

  • f
  • f sou
  • urce texts

ts. 1st

st lan

languag age te text evaluation and revision (n (not tran ansla latio ions, eg

  • eg. ar

arbit itratio ion la law)

To

  • screen and rank

th the quality ty of ex external translators and freelancers an and tran ansla latio ion ag agencie ies Us Use tr translati tion

  • nQ to

tr train stu tudents ts to bec becom pr profes essiona nal re revisors “a “a tool to teach revision in in Revis isio ion n technique hnique clas lass”

slide-29
SLIDE 29

“Reverse” revision: finding good translations

  • Preselected Item Evaluation

– Perfectly possible with this tool – Finding good translations of key sentences, words or full segments – Positive scoring – 100% consistent scoring – Psychometric analysis – Examples in literature, medical, legal texts…

slide-30
SLIDE 30
  • PIE steps

2 9 / 1 8 3 2

A h A human re revisor preselects i items/keyw ywor

  • rds

in in the the sour urce text t or que questio tion 100 % 100 % ob

  • bjective/con
  • nsistent
slide-31
SLIDE 31
  • Example

2 9 / 1 8 3 3

How do I prepare for this procedure [PI 3]? Prior to beginning the study, you will be asked to drink enough water [PI 4] or fluids to fill your

  • bladder. You may start filling your bladder from home providing your travelling distance is within
  • ne hour [PI 5]. It is important that the bladder should feel full (to a point that you would

normally pass urine) but not be uncomfortably full. Do not overfill the bladder [PI 6]. When the bladder feels full, you will be asked to urinate [PI 7] into the special uroflow study

  • container. It is important that you pass urine in as normal a pattern (for you) as possible. If you

think that this urination was not normal (for you) tell the nurse [PI 8] and you may be asked to repeat the study. If you have any mobility problems, or any other special needs, please let us know in advance so that we can make any necessary arrangements to accommodate you. What happens after the procedure? Immediately after the uroflow, an ultrasound probe will be passed over your abdomen. The nurse performing the procedure [PI 9] will lightly press the probe into the lower part of your abdomen and measure the amount of urine left in your bladder. The results of these studies [PI 10] will be discussed with you on your next visit.

slide-32
SLIDE 32
  • Example

2 9 / 1 8 3 4

slide-33
SLIDE 33
  • PIE Steps

2 9 / 1 8 3 5

Ca Calculation

  • n of t
  • f the s

scor

  • res of t
  • f the c

candidates

  • n
  • n t

the p preselected i items

  • 1. Raw score

(number of hits/total number of PI’s)

  • 2. If enough candidates (50+), calculation of
  • the item di

difficul ulty level el (p p value) ue) and

  • the di

discrimi mina nation n inde ndex (d d inde ndex) of the items

slide-34
SLIDE 34
  • PIE steps

2 9 / 1 8 3 6

§ th the d e difficulty l ty level el (p value): adequate p value: between 0.27 and 0.79 § th the d e disc scriminator

  • ry p

y power er (d index) of the items: adequate d index: higher than 0.29

à El Elimi mina nati ting ng “i “invalid” d” items ms

§ Bringing the raw score back to the fin inal al PIE E score

slide-35
SLIDE 35
  • What does the PIE method tell you?

2 9 / 1 8 3 7

NO NOT an an evalua aluatio tion n of the the full ull ans answer quality quality Bu But: a a r reliable ranki king of a

  • f all y

you

  • ur c

candidates

100% consistent Objective Much faster than full text correction

And a a r reliable r ranki king of y g of you

  • ur i

items

slide-36
SLIDE 36

bert.wylin@kuleuven.be d.minta@televic.com b.wylin@televic.com