SlideShare une entreprise Scribd logo
1  sur  19
TAUS OPEN SOURCE MACHINE TRANSLATION SHOWCASE


“Moses, Moses:
Let my people go”

Moses MT engine feasibility
study

By Serge Gladkoff
Do we witness a breakthrough?


1.   There’s a hype. The number of publications and actual deployments is growing. (Twitter posts on MT
     made 50% of all twits on LocWorld Berlin 2010!).

2.   Statistical systems seem to demonstrate real quality progress. Marriage with RBMT is especially
     promising.

3.   Reports claim economic efficiency of some MT deployments, although few reliable ROI data is available.

4.   New factors: a) huge amount of linguistic data available, b) further leaps in processing power, c) cost and
     time pressures.

5.   OpenSource decreases the cost of development and lowers the barrier to entry into new technology.

6.   Key clients have already integrated MT in their workflows and some projects are coming with
     components of MT output.




                                               (C) 2003-2010 LOGRUS LLC
HYPOTHESIS and goals of the experiment


1. TEST HYPOTESIS – Is the Plateau of quality behind? How far is the next plateau
   of quality from Human Translation?
2. Is the MT technology available for LSPs without multimillion funding?
3. Is OpenSource SMT Moses a viable place to start?
4. Can we base our own productivity solutions on Moses in any near future?
5. What are the possible LSP MT-related process implications?
6. What technical issues could arise?
7. What are the methods to calculate ROI?
8. What are post-editing differences in MT scenario?
9. Is my company capable of handling this?




                                 (C) 2003-2010 LOGRUS LLC
Resources


1. ONE (very powerful) computer.
2. ONE (very experienced) Project Manager – MT skeptic!
3. ONE (very experienced), talented programmer.
4. Very limited task definition and scope.
5. NO prior experience with MOSES – starting from zero.
6. Intentionally very limited Moses tweaking – no going deep into Moses code.
7. TAUS DATA Founding Member status + €15,000 to get full access to corpus.
8. Corporate culture of scientific research, automation, innovation.
9. Total number of internal hours for this pilot =~ 360 man-hours.




                                  (C) 2003-2010 LOGRUS LLC
First steps with Moses


1. We downloaded the system, installed and set the minimum standard, default
   configuration.
2. Data was largely taken from TDA corpus, but we have also used our
   translation memories on IT (English - Russian pair).
   The data has to be cleared from garbage to train Moses – we removed
   complex formatting and tags, also simply dropped invalid segments. This has
   been identified as major obstacle and project effort.

   Currently TDA is working on cleaning the data, but at that moment they were
   far from clean (10-20% of all units were irrelevant for Moses training)

        - Removed tags for RTF, HTML, XML
        - Removed/fixed strings with wrong encoding
        - Removed untranslated strings (source = target);
        - Removed strings with no text (numbers, dates) or empty target




                                (C) 2003-2010 LOGRUS LLC
Playing around


    1. We evaluated not only language quality as such, but also the convenience of
       the system installation and setup, difficulty of fine-tuning and use, and system
       stability.
    2. We have tried various Moses distributions, language models and Linux
       distributions.
    3. We have decided that IRST is the optimal language model.
    4. Various language parameters of Moses have been played with, to get the
       optimal result.
 ∙ -weight-d (-d): weight(s) for distortion (reordering components)                  [input-factors]
∙ -weight-e (-e): weight for word deletion                                           0
∙ -weight-file (-wf): file containing labeled weights                                [mapping]
∙ -weight-generation (-g): weight(s) for generation components                       T0
∙ -weight-i (-I): weight for word insertion                                          [ttable-file]
∙ -weight-l (-lm): weight(s) for language models                                     0 0 5 /home/atma/BASES/db/msao_e/lngbase/msao_e-MAIN
∙ -weight-t (-tm): weights for translation model components                          [lmodel-file]
∙ -weight-w (-w): weight for word penalty                                            1 0 3 /home/atma/BASES/db/msao_e/lm/1.gz
∙ -phrase-drop-allowed (-da): if present, allow dropping of source words             [ttable-limit]
∙ -distortion: configurations for each factorized/lexicalized reordering model.      20
                                                                                     0
∙ -distortion-limit (-dl): distortion (reordering) limit in maximum number of words (0
  = monotone, -1 = unlimited)                                                        [weight-d]
∙ -max-phrase-length: maximum phrase length (default 20)                             0.6
                                                                                     [weight-l]
∙ -mbr-size: number of translation candidates considered in MBR decoding (default 200)
∙ -minimum-bayes-risk (-mbr): use miminum Bayes risk to determine best translation   0.5000
∙ -monotone-at-punctuation (-mp): do not reorder over punctuation                    [weight-t]
∙ -n-best-factor: factor to compute the maximum number of contenders (=factor*nbest- 0.2
  size). value 0 means infinity, i.e. no threshold. default is 0                     0.2
                                                                                     0.2
                                                                                     0.2
                                                                 (C) 2003-2010      LOGRUS LLC
                                                                                     0.2
                                                                                     [weight-w]
Final Logrus Moses system configuration


1. Hardware: Quad-Core AMD-64, 16 Gbytes of RAM, 64-bit Linux Ubuntu, 100
   Gbytes of disk space.
2. Moses configuration: the most recent build, IRST language model, GIZA++
   (training component).
3. We have created Web interface to the system.
4. Moses training has to be automated, and was automated, to try the system with
   different SME corpuses
5. Other SW: our proprietary scripts developed to train Moses, its setup and
   remote (Web) client access.




                                 (C) 2003-2010 LOGRUS LLC
Training, Training, Training


Training and data cleaning was lengthy and iterative process, had to be automated.




                                 (C) 2003-2010 LOGRUS LLC
Results

                                                         Edit Distance (average value of Character Edit Rate1)
    Language pair                 SMT           SMT (*)        SMT (**)         RBMT             RBMT           <# 4>    <# 5>
    MT system (developer)
                                 (Bing —          Logrus         Logrus       (ProMT v8)        (SYSTRAN)        (SDL)   (Google)
                                Microsoft)        Moses          Moses
    English – French               0.45              -              -             0.54             0.51            -         -
    English – Italian              0.46              -              -             0.59             0.51            -         -
    English – Portuguese           0.42              -              -             0.58             0.58            -         -
    English – Russian              0.53            0.57           0.61            0.63             0.69            -         -
    English – Spanish           0.35– 0.37           -              -             0.46             0.45          0.40      0.35

                   (*) Logrus Moses after deep fine-tuning, trained on very narrow IT domain.
                   (**) Logrus Moses after deep fine-tuning, trained on much wider IT domain (TAUSDATA)



1. We have obtained Moses translation quality results in standard configuration, without deep fine-tuning.
2. The same after deep fine-tuning.
3. The MT output results from various systems were evaluated on one and the same sample.
4. Translation quality has been evaluated with Edit Distance algorithm (the software tool was developed in Logrus).
5. The “pure” result was better than with untrained PROMT 8 and was close to BING Translator from Microsoft.
6. IN DOMAIN – vs GENERAL, NOT IN-DOMAIN engines


1   rate of the number of changes introduced after post-editing, to the sentence length.




                                                       (C) 2003-2010 LOGRUS LLC
Problems of MT integration into actual process


1. Large paragraphs (>255 characters). Moses fails to build a match table on
   them, so we had to break them down.
2. Punctuation marks are treated as letters. Multi-sentence segments are not
   treated correctly and add noise.
3. Basic Moses configuration only work with lowercase letters. Additional
   processing is required to restore case.
4. RTF, XML, HTML tags are not supported.
5. Training corpus often contains ambiguous, obsolete terms. Extensive
   terminology support is required.
6. Every Moses user has to develop his own interface between Moses and his
   own workflow, CMS, TMs, projects, etc.




                                 (C) 2003-2010 LOGRUS LLC
Moses Features, Advantages, and Disadvantages


1. ADVANTAGES:
   - Free
   - Amazingly high quality of basic system after proper tweaking
   - Minimal setup provides competitive language quality level
   - Practical solutions can be based on Moses
2. DISADAVTAGES:
   - Moses is not a ready tool – it’s a construction set
   - Serious efforts are required to bring it to practical use
3. PROBLEMS:
   - Training corpuses require cleaning
   - TMX is poorly supported by existing software
   - Domains of feasibility are so specific that it is not clear whether the effort
   worth it yet for “generic” LSP




                                    (C) 2003-2010 LOGRUS LLC
Conclusions – “scientific” aspects


1.   Hypothesis is true – it does look like SMT breaks the MT quality barrier known until
     recently. Another plateau is within usable domain in certain situations.

2.   We experienced the benefits of statistics-based method and its limitations.

3.   We obtained information on areas where SMT produces best results.

4.   We have found the configuration setting and training technique – Moses experience that
     we can take further.

5.   We concluded that further research is required and beneficial.

6.   Terminology requires special attention. Large corpus is not always better than smaller
     one. Very serious inconsistencies in terminology in corpus, particularly detrimental for
     SMT engines.

7.   SMT vs RBMT choice criteria refined.




                                        (C) 2003-2010 LOGRUS LLC
Conclusions – technical aspects


1.   TAUS DATA: insufficient number of data categories (“genres”, content types). Extremely
     diverse data are grouped together, requires domain separation.

2.   Too few TMX editors on the market, most of them too expensive. Multiple problems with
     TMX support, language support and/or stability of these editors.

3.   Multiple problems with Import/Export in all these editors. Logrus had to develop a TMX
     editor of its own, and we are planning to provide it for free to cooperating parties in the
     near future. Other useful proprietary tools and techniques have been developed can be
     shared.

4.   No MT – TM – Translation Environment integration yet.

5.   Terminology-Related Problem: Linking to Various Data Corpora
     RBMTs require grammatical attributes, while STAT-based MTs require fine, “genre”-
     related classification.




                                         (C) 2003-2010 LOGRUS LLC
Conclusions – practical aspects

1.   Overall linguistic quality of MT output is still low: Output from MT engines are much closer to each
     other than to final quality human translation. (We are still very far from Turing barrier.)

2.   Mass post-editing is required. Editing with reference to the source is a MUST.

3.   Dedicated terminology support is a must. Complete, 100% terminology post-verification is required.

4.   MT post-editing is different from human editing, and editors must be trained.

5.   With “Difficult" language pairs the cost of post-editing would be 50%-75% of the cost of human
     translation, in best case scenario, after costly preparation. Overall ROI must include internal R&D
     and/or license and deployment costs.

6.   Continuous internal R&D is required to make use of the technology.

7.   ROI as a consequence is very specific for particular project and company; further research into methods
     to evaluate ROI is required.

8.   The domain of positive ROI increases. LSPs should pay attention.

9.   Significant localization process reengineering is required for MT deployment on all stages: corpus, MT
     setup and administration, deployment, vendor selection, post-editing processes.

10. Languages vary – the process would be different for each language (SMT vs RBMT, parameters, quality
    level, etc.). MLVs are at disadvantage – SLVs have opportunity again with more focus into particular
    language.




                                             (C) 2003-2010 LOGRUS LLC
Last Mile Phenomenon


Outputs from MT engines are much closer
to each other than to final quality human
translation.




                 (C) 2003-2010 LOGRUS LLC
Last Mile Phenomenon

2.   Proximity phenomenon: the closer you are to the goal, the more subtle details are coming to
     stand out and still divide you and the prize. “Final 10% to the goal” is the Last Mile you need to
     cover with the largest effort and cost, deployment, preparation, process change and 100%
     editing.




                                        (C) 2003-2010 LOGRUS LLC
HYPOTHESIS and goals of the experiment


1.   TEST HYPOTESIS – Is the Plateau of quality behind? How far is the next plateau of quality from Human
     Translation?
     Yes, but we are about to hit another plateau, within usable domain. Human quality is still very far, the
     last mile gap is WIDE. Current predictions of machines passing Turing Test are 2029.

2.   Is the MT technology available for LSPs without multimillion funding?
     Not multimillion, but continuous, significant effort is required.

3.   Is OpenSource SMT Moses a viable place to start?
     YES and NO. For immediate deployments other commercial products may be more feasible. Moses is
     however very promising. A very good training ground to prepare for MT or evaluate current MT level.

4.   Can we base our own productivity solutions on Moses in any near future?
     Timeframe is several years still. Numerous technical problems are to be solved.

5.   What are the possible LSP MT-related process implications?
     We’ve got some answers to ourselves.

6.   What are the methods to calculate ROI?
     We have developed our method.

7.   What are the post-editing differences in MT scenario?
     Under development.

8.   Is my company capable of handling this?
     YES.


                                             (C) 2003-2010 LOGRUS LLC
(C) 2003-2010 LOGRUS LLC

Contenu connexe

En vedette

En vedette (12)

Multilingual Term Extraction as a Service from Acrolinx, CHAT2013
Multilingual Term Extraction as a Service from Acrolinx, CHAT2013Multilingual Term Extraction as a Service from Acrolinx, CHAT2013
Multilingual Term Extraction as a Service from Acrolinx, CHAT2013
 
TAUS OPEN SOURCE MACHINE TRANSLATION SHOWCASE, Monaco, Diego Bartolome, tauyo...
TAUS OPEN SOURCE MACHINE TRANSLATION SHOWCASE, Monaco, Diego Bartolome, tauyo...TAUS OPEN SOURCE MACHINE TRANSLATION SHOWCASE, Monaco, Diego Bartolome, tauyo...
TAUS OPEN SOURCE MACHINE TRANSLATION SHOWCASE, Monaco, Diego Bartolome, tauyo...
 
TAUS MT SHOWCASE, Is the Translation Industry Ready, Jaap van der Meer, TAUS...
TAUS MT SHOWCASE,  Is the Translation Industry Ready, Jaap van der Meer, TAUS...TAUS MT SHOWCASE,  Is the Translation Industry Ready, Jaap van der Meer, TAUS...
TAUS MT SHOWCASE, Is the Translation Industry Ready, Jaap van der Meer, TAUS...
 
TAUS MT SHOWCASE, The WeMT Program, Olga Beregovaya, Welocalize, 10 October 2...
TAUS MT SHOWCASE, The WeMT Program, Olga Beregovaya, Welocalize, 10 October 2...TAUS MT SHOWCASE, The WeMT Program, Olga Beregovaya, Welocalize, 10 October 2...
TAUS MT SHOWCASE, The WeMT Program, Olga Beregovaya, Welocalize, 10 October 2...
 
TAUS MT SHOWCASE, Moses in the Mix. A Technology Agnostic Approach to a Winni...
TAUS MT SHOWCASE, Moses in the Mix. A Technology Agnostic Approach to a Winni...TAUS MT SHOWCASE, Moses in the Mix. A Technology Agnostic Approach to a Winni...
TAUS MT SHOWCASE, Moses in the Mix. A Technology Agnostic Approach to a Winni...
 
TAUS MT SHOWCASE, The Open Source MT System Moses and Its Use in the Industry...
TAUS MT SHOWCASE, The Open Source MT System Moses and Its Use in the Industry...TAUS MT SHOWCASE, The Open Source MT System Moses and Its Use in the Industry...
TAUS MT SHOWCASE, The Open Source MT System Moses and Its Use in the Industry...
 
TAUS OPEN SOURCE MACHINE TRANSLATION SHOWCASE, Paris, Manuel Herranz, Pangean...
TAUS OPEN SOURCE MACHINE TRANSLATION SHOWCASE, Paris, Manuel Herranz, Pangean...TAUS OPEN SOURCE MACHINE TRANSLATION SHOWCASE, Paris, Manuel Herranz, Pangean...
TAUS OPEN SOURCE MACHINE TRANSLATION SHOWCASE, Paris, Manuel Herranz, Pangean...
 
TAUS MT SHOWCASE, Moses and Other Resources, Rahzeb Choudhury, TAUS, 10 April...
TAUS MT SHOWCASE, Moses and Other Resources, Rahzeb Choudhury, TAUS, 10 April...TAUS MT SHOWCASE, Moses and Other Resources, Rahzeb Choudhury, TAUS, 10 April...
TAUS MT SHOWCASE, Moses and Other Resources, Rahzeb Choudhury, TAUS, 10 April...
 
TAUS OPEN SOURCE MACHINE TRANSLATION SHOWCASE, Beijing, Chengqing Zong, Casia...
TAUS OPEN SOURCE MACHINE TRANSLATION SHOWCASE, Beijing, Chengqing Zong, Casia...TAUS OPEN SOURCE MACHINE TRANSLATION SHOWCASE, Beijing, Chengqing Zong, Casia...
TAUS OPEN SOURCE MACHINE TRANSLATION SHOWCASE, Beijing, Chengqing Zong, Casia...
 
TAUS MT SHOWCASE, Microsoft Translator, Chris Wendt, Microsoft, 10 October 2013
TAUS MT SHOWCASE,  Microsoft Translator, Chris Wendt, Microsoft, 10 October 2013TAUS MT SHOWCASE,  Microsoft Translator, Chris Wendt, Microsoft, 10 October 2013
TAUS MT SHOWCASE, Microsoft Translator, Chris Wendt, Microsoft, 10 October 2013
 
TAUS OPEN SOURCE MACHINE TRANSLATION SHOWCASE, Beijing, Yu Gong, Adobe, 23 Ap...
TAUS OPEN SOURCE MACHINE TRANSLATION SHOWCASE, Beijing, Yu Gong, Adobe, 23 Ap...TAUS OPEN SOURCE MACHINE TRANSLATION SHOWCASE, Beijing, Yu Gong, Adobe, 23 Ap...
TAUS OPEN SOURCE MACHINE TRANSLATION SHOWCASE, Beijing, Yu Gong, Adobe, 23 Ap...
 
TAUS MT SHOWCASE, Google Translator Toolkit, Patcharin Areewong, Google, 10 A...
TAUS MT SHOWCASE, Google Translator Toolkit, Patcharin Areewong, Google, 10 A...TAUS MT SHOWCASE, Google Translator Toolkit, Patcharin Areewong, Google, 10 A...
TAUS MT SHOWCASE, Google Translator Toolkit, Patcharin Areewong, Google, 10 A...
 

Similaire à TAUS OPEN SOURCE MACHINE TRANSLATION SHOWCASE, Monaco, Serge Gladhoff, Logrus, 25 March 2012

เฉลยคำตอบ O
เฉลยคำตอบ Oเฉลยคำตอบ O
เฉลยคำตอบ O
Surapong Jakang
 

Similaire à TAUS OPEN SOURCE MACHINE TRANSLATION SHOWCASE, Monaco, Serge Gladhoff, Logrus, 25 March 2012 (20)

TAUS USER CONFERENCE 2010, Moving on with MT: building Open Source MT with yo...
TAUS USER CONFERENCE 2010, Moving on with MT: building Open Source MT with yo...TAUS USER CONFERENCE 2010, Moving on with MT: building Open Source MT with yo...
TAUS USER CONFERENCE 2010, Moving on with MT: building Open Source MT with yo...
 
BloombergGPT.pdfA Large Language Model for Finance
BloombergGPT.pdfA Large Language Model for FinanceBloombergGPT.pdfA Large Language Model for Finance
BloombergGPT.pdfA Large Language Model for Finance
 
Session on machine translation batu 19 march2016
Session on machine translation batu 19 march2016Session on machine translation batu 19 march2016
Session on machine translation batu 19 march2016
 
Rust
RustRust
Rust
 
Secure Programming Language Cs
Secure Programming Language CsSecure Programming Language Cs
Secure Programming Language Cs
 
Birendra_resume
Birendra_resumeBirendra_resume
Birendra_resume
 
เฉลยคำตอบ O
เฉลยคำตอบ Oเฉลยคำตอบ O
เฉลยคำตอบ O
 
MS SQL Server 2008, Implementation and Maintenance
MS SQL Server 2008, Implementation and MaintenanceMS SQL Server 2008, Implementation and Maintenance
MS SQL Server 2008, Implementation and Maintenance
 
Unveiling FME 2018
Unveiling FME 2018Unveiling FME 2018
Unveiling FME 2018
 
Birendra_resume
Birendra_resumeBirendra_resume
Birendra_resume
 
Chatbot ppt
Chatbot pptChatbot ppt
Chatbot ppt
 
KIRANKUMAR_MV
KIRANKUMAR_MVKIRANKUMAR_MV
KIRANKUMAR_MV
 
Introduction
IntroductionIntroduction
Introduction
 
Speech-Quality-and-MOS Factors, Voice Quality improvement
Speech-Quality-and-MOS Factors, Voice Quality improvementSpeech-Quality-and-MOS Factors, Voice Quality improvement
Speech-Quality-and-MOS Factors, Voice Quality improvement
 
Mp &mc programs
Mp &mc programsMp &mc programs
Mp &mc programs
 
Distributed Computing & MapReduce
Distributed Computing & MapReduceDistributed Computing & MapReduce
Distributed Computing & MapReduce
 
Scaling Engineering with Docker
Scaling Engineering with DockerScaling Engineering with Docker
Scaling Engineering with Docker
 
고급컴파일러구성론_개레_230303.pptx
고급컴파일러구성론_개레_230303.pptx고급컴파일러구성론_개레_230303.pptx
고급컴파일러구성론_개레_230303.pptx
 
A Taxonomy for Program Metamodels in Program Reverse Engineering
A Taxonomy for Program Metamodels in Program Reverse EngineeringA Taxonomy for Program Metamodels in Program Reverse Engineering
A Taxonomy for Program Metamodels in Program Reverse Engineering
 
Opw presentation of open office-lgma
Opw presentation of open office-lgmaOpw presentation of open office-lgma
Opw presentation of open office-lgma
 

Plus de TAUS - The Language Data Network

Plus de TAUS - The Language Data Network (20)

TAUS Global Content Summit Amsterdam 2019 / Beyond MT. A few premature reflec...
TAUS Global Content Summit Amsterdam 2019 / Beyond MT. A few premature reflec...TAUS Global Content Summit Amsterdam 2019 / Beyond MT. A few premature reflec...
TAUS Global Content Summit Amsterdam 2019 / Beyond MT. A few premature reflec...
 
TAUS Global Content Summit Amsterdam 2019 / Measure with DQF, Dace Dzeguze (T...
TAUS Global Content Summit Amsterdam 2019 / Measure with DQF, Dace Dzeguze (T...TAUS Global Content Summit Amsterdam 2019 / Measure with DQF, Dace Dzeguze (T...
TAUS Global Content Summit Amsterdam 2019 / Measure with DQF, Dace Dzeguze (T...
 
TAUS Global Content Summit Amsterdam 2019 / Automatic for the People by Domin...
TAUS Global Content Summit Amsterdam 2019 / Automatic for the People by Domin...TAUS Global Content Summit Amsterdam 2019 / Automatic for the People by Domin...
TAUS Global Content Summit Amsterdam 2019 / Automatic for the People by Domin...
 
TAUS Global Content Summit Amsterdam 2019 / The Quantum Leap: Human Parity, C...
TAUS Global Content Summit Amsterdam 2019 / The Quantum Leap: Human Parity, C...TAUS Global Content Summit Amsterdam 2019 / The Quantum Leap: Human Parity, C...
TAUS Global Content Summit Amsterdam 2019 / The Quantum Leap: Human Parity, C...
 
TAUS Global Content Summit Amsterdam 2019 / Growing Business by Connecting Co...
TAUS Global Content Summit Amsterdam 2019 / Growing Business by Connecting Co...TAUS Global Content Summit Amsterdam 2019 / Growing Business by Connecting Co...
TAUS Global Content Summit Amsterdam 2019 / Growing Business by Connecting Co...
 
Achieving Translation Efficiency and Accuracy for Video Content, Xiao Yuan (P...
Achieving Translation Efficiency and Accuracy for Video Content, Xiao Yuan (P...Achieving Translation Efficiency and Accuracy for Video Content, Xiao Yuan (P...
Achieving Translation Efficiency and Accuracy for Video Content, Xiao Yuan (P...
 
Introduction Innovation Contest Shenzhen by Henri Broekmate (Lionbridge)
Introduction Innovation Contest Shenzhen by Henri Broekmate (Lionbridge)Introduction Innovation Contest Shenzhen by Henri Broekmate (Lionbridge)
Introduction Innovation Contest Shenzhen by Henri Broekmate (Lionbridge)
 
Game Changer for Linguistic Review: Shifting the Paradigm, Klaus Fleischmann...
 Game Changer for Linguistic Review: Shifting the Paradigm, Klaus Fleischmann... Game Changer for Linguistic Review: Shifting the Paradigm, Klaus Fleischmann...
Game Changer for Linguistic Review: Shifting the Paradigm, Klaus Fleischmann...
 
A translation memory P2P trading platform - to make global translation memory...
A translation memory P2P trading platform - to make global translation memory...A translation memory P2P trading platform - to make global translation memory...
A translation memory P2P trading platform - to make global translation memory...
 
Shiyibao — The Most Efficient Translation Feedback System Ever, Guanqing Hao ...
Shiyibao — The Most Efficient Translation Feedback System Ever, Guanqing Hao ...Shiyibao — The Most Efficient Translation Feedback System Ever, Guanqing Hao ...
Shiyibao — The Most Efficient Translation Feedback System Ever, Guanqing Hao ...
 
Stepes – Instant Human Translation Services for the Digital World, Carl Yao (...
Stepes – Instant Human Translation Services for the Digital World, Carl Yao (...Stepes – Instant Human Translation Services for the Digital World, Carl Yao (...
Stepes – Instant Human Translation Services for the Digital World, Carl Yao (...
 
Farmer Lv (TrueTran)
Farmer Lv (TrueTran)Farmer Lv (TrueTran)
Farmer Lv (TrueTran)
 
Smart Translation Resource Management: Semantic Matching, Kirk Zhang (Wiitran...
Smart Translation Resource Management: Semantic Matching, Kirk Zhang (Wiitran...Smart Translation Resource Management: Semantic Matching, Kirk Zhang (Wiitran...
Smart Translation Resource Management: Semantic Matching, Kirk Zhang (Wiitran...
 
The Theory and Practice of Computer Aided Translation Training System, Liu Q...
 The Theory and Practice of Computer Aided Translation Training System, Liu Q... The Theory and Practice of Computer Aided Translation Training System, Liu Q...
The Theory and Practice of Computer Aided Translation Training System, Liu Q...
 
Translation Technology Showcase in Shenzhen
Translation Technology Showcase in ShenzhenTranslation Technology Showcase in Shenzhen
Translation Technology Showcase in Shenzhen
 
How to efficiently use large-scale TMs in translation, Jing Zhang (Tmxmall)
How to efficiently use large-scale TMs in translation, Jing Zhang (Tmxmall)How to efficiently use large-scale TMs in translation, Jing Zhang (Tmxmall)
How to efficiently use large-scale TMs in translation, Jing Zhang (Tmxmall)
 
SDL Trados Studio 2017, Jocelyn He (SDL)
SDL Trados Studio 2017, Jocelyn He (SDL)SDL Trados Studio 2017, Jocelyn He (SDL)
SDL Trados Studio 2017, Jocelyn He (SDL)
 
How we train post-editors - Yongpeng Wei (Lingosail)
How we train post-editors - Yongpeng Wei (Lingosail)How we train post-editors - Yongpeng Wei (Lingosail)
How we train post-editors - Yongpeng Wei (Lingosail)
 
A use-case for getting MT into your company, Kerstin Berns (berns language c...
 A use-case for getting MT into your company, Kerstin Berns (berns language c... A use-case for getting MT into your company, Kerstin Berns (berns language c...
A use-case for getting MT into your company, Kerstin Berns (berns language c...
 
QE integrated in XTM, by Bob Willans (XTM)
QE integrated in XTM, by Bob Willans (XTM)QE integrated in XTM, by Bob Willans (XTM)
QE integrated in XTM, by Bob Willans (XTM)
 

Dernier

Artificial Intelligence: Facts and Myths
Artificial Intelligence: Facts and MythsArtificial Intelligence: Facts and Myths
Artificial Intelligence: Facts and Myths
Joaquim Jorge
 
+971581248768>> SAFE AND ORIGINAL ABORTION PILLS FOR SALE IN DUBAI AND ABUDHA...
+971581248768>> SAFE AND ORIGINAL ABORTION PILLS FOR SALE IN DUBAI AND ABUDHA...+971581248768>> SAFE AND ORIGINAL ABORTION PILLS FOR SALE IN DUBAI AND ABUDHA...
+971581248768>> SAFE AND ORIGINAL ABORTION PILLS FOR SALE IN DUBAI AND ABUDHA...
?#DUbAI#??##{{(☎️+971_581248768%)**%*]'#abortion pills for sale in dubai@
 

Dernier (20)

AWS Community Day CPH - Three problems of Terraform
AWS Community Day CPH - Three problems of TerraformAWS Community Day CPH - Three problems of Terraform
AWS Community Day CPH - Three problems of Terraform
 
Powerful Google developer tools for immediate impact! (2023-24 C)
Powerful Google developer tools for immediate impact! (2023-24 C)Powerful Google developer tools for immediate impact! (2023-24 C)
Powerful Google developer tools for immediate impact! (2023-24 C)
 
Partners Life - Insurer Innovation Award 2024
Partners Life - Insurer Innovation Award 2024Partners Life - Insurer Innovation Award 2024
Partners Life - Insurer Innovation Award 2024
 
Automating Google Workspace (GWS) & more with Apps Script
Automating Google Workspace (GWS) & more with Apps ScriptAutomating Google Workspace (GWS) & more with Apps Script
Automating Google Workspace (GWS) & more with Apps Script
 
2024: Domino Containers - The Next Step. News from the Domino Container commu...
2024: Domino Containers - The Next Step. News from the Domino Container commu...2024: Domino Containers - The Next Step. News from the Domino Container commu...
2024: Domino Containers - The Next Step. News from the Domino Container commu...
 
Boost PC performance: How more available memory can improve productivity
Boost PC performance: How more available memory can improve productivityBoost PC performance: How more available memory can improve productivity
Boost PC performance: How more available memory can improve productivity
 
Understanding Discord NSFW Servers A Guide for Responsible Users.pdf
Understanding Discord NSFW Servers A Guide for Responsible Users.pdfUnderstanding Discord NSFW Servers A Guide for Responsible Users.pdf
Understanding Discord NSFW Servers A Guide for Responsible Users.pdf
 
Tata AIG General Insurance Company - Insurer Innovation Award 2024
Tata AIG General Insurance Company - Insurer Innovation Award 2024Tata AIG General Insurance Company - Insurer Innovation Award 2024
Tata AIG General Insurance Company - Insurer Innovation Award 2024
 
Mastering MySQL Database Architecture: Deep Dive into MySQL Shell and MySQL R...
Mastering MySQL Database Architecture: Deep Dive into MySQL Shell and MySQL R...Mastering MySQL Database Architecture: Deep Dive into MySQL Shell and MySQL R...
Mastering MySQL Database Architecture: Deep Dive into MySQL Shell and MySQL R...
 
Apidays New York 2024 - The Good, the Bad and the Governed by David O'Neill, ...
Apidays New York 2024 - The Good, the Bad and the Governed by David O'Neill, ...Apidays New York 2024 - The Good, the Bad and the Governed by David O'Neill, ...
Apidays New York 2024 - The Good, the Bad and the Governed by David O'Neill, ...
 
Apidays Singapore 2024 - Building Digital Trust in a Digital Economy by Veron...
Apidays Singapore 2024 - Building Digital Trust in a Digital Economy by Veron...Apidays Singapore 2024 - Building Digital Trust in a Digital Economy by Veron...
Apidays Singapore 2024 - Building Digital Trust in a Digital Economy by Veron...
 
Artificial Intelligence: Facts and Myths
Artificial Intelligence: Facts and MythsArtificial Intelligence: Facts and Myths
Artificial Intelligence: Facts and Myths
 
+971581248768>> SAFE AND ORIGINAL ABORTION PILLS FOR SALE IN DUBAI AND ABUDHA...
+971581248768>> SAFE AND ORIGINAL ABORTION PILLS FOR SALE IN DUBAI AND ABUDHA...+971581248768>> SAFE AND ORIGINAL ABORTION PILLS FOR SALE IN DUBAI AND ABUDHA...
+971581248768>> SAFE AND ORIGINAL ABORTION PILLS FOR SALE IN DUBAI AND ABUDHA...
 
Strategize a Smooth Tenant-to-tenant Migration and Copilot Takeoff
Strategize a Smooth Tenant-to-tenant Migration and Copilot TakeoffStrategize a Smooth Tenant-to-tenant Migration and Copilot Takeoff
Strategize a Smooth Tenant-to-tenant Migration and Copilot Takeoff
 
Bajaj Allianz Life Insurance Company - Insurer Innovation Award 2024
Bajaj Allianz Life Insurance Company - Insurer Innovation Award 2024Bajaj Allianz Life Insurance Company - Insurer Innovation Award 2024
Bajaj Allianz Life Insurance Company - Insurer Innovation Award 2024
 
Strategies for Unlocking Knowledge Management in Microsoft 365 in the Copilot...
Strategies for Unlocking Knowledge Management in Microsoft 365 in the Copilot...Strategies for Unlocking Knowledge Management in Microsoft 365 in the Copilot...
Strategies for Unlocking Knowledge Management in Microsoft 365 in the Copilot...
 
Strategies for Landing an Oracle DBA Job as a Fresher
Strategies for Landing an Oracle DBA Job as a FresherStrategies for Landing an Oracle DBA Job as a Fresher
Strategies for Landing an Oracle DBA Job as a Fresher
 
presentation ICT roal in 21st century education
presentation ICT roal in 21st century educationpresentation ICT roal in 21st century education
presentation ICT roal in 21st century education
 
Scaling API-first – The story of a global engineering organization
Scaling API-first – The story of a global engineering organizationScaling API-first – The story of a global engineering organization
Scaling API-first – The story of a global engineering organization
 
From Event to Action: Accelerate Your Decision Making with Real-Time Automation
From Event to Action: Accelerate Your Decision Making with Real-Time AutomationFrom Event to Action: Accelerate Your Decision Making with Real-Time Automation
From Event to Action: Accelerate Your Decision Making with Real-Time Automation
 

TAUS OPEN SOURCE MACHINE TRANSLATION SHOWCASE, Monaco, Serge Gladhoff, Logrus, 25 March 2012

  • 1. TAUS OPEN SOURCE MACHINE TRANSLATION SHOWCASE “Moses, Moses: Let my people go” Moses MT engine feasibility study By Serge Gladkoff
  • 2.
  • 3. Do we witness a breakthrough? 1. There’s a hype. The number of publications and actual deployments is growing. (Twitter posts on MT made 50% of all twits on LocWorld Berlin 2010!). 2. Statistical systems seem to demonstrate real quality progress. Marriage with RBMT is especially promising. 3. Reports claim economic efficiency of some MT deployments, although few reliable ROI data is available. 4. New factors: a) huge amount of linguistic data available, b) further leaps in processing power, c) cost and time pressures. 5. OpenSource decreases the cost of development and lowers the barrier to entry into new technology. 6. Key clients have already integrated MT in their workflows and some projects are coming with components of MT output. (C) 2003-2010 LOGRUS LLC
  • 4. HYPOTHESIS and goals of the experiment 1. TEST HYPOTESIS – Is the Plateau of quality behind? How far is the next plateau of quality from Human Translation? 2. Is the MT technology available for LSPs without multimillion funding? 3. Is OpenSource SMT Moses a viable place to start? 4. Can we base our own productivity solutions on Moses in any near future? 5. What are the possible LSP MT-related process implications? 6. What technical issues could arise? 7. What are the methods to calculate ROI? 8. What are post-editing differences in MT scenario? 9. Is my company capable of handling this? (C) 2003-2010 LOGRUS LLC
  • 5. Resources 1. ONE (very powerful) computer. 2. ONE (very experienced) Project Manager – MT skeptic! 3. ONE (very experienced), talented programmer. 4. Very limited task definition and scope. 5. NO prior experience with MOSES – starting from zero. 6. Intentionally very limited Moses tweaking – no going deep into Moses code. 7. TAUS DATA Founding Member status + €15,000 to get full access to corpus. 8. Corporate culture of scientific research, automation, innovation. 9. Total number of internal hours for this pilot =~ 360 man-hours. (C) 2003-2010 LOGRUS LLC
  • 6. First steps with Moses 1. We downloaded the system, installed and set the minimum standard, default configuration. 2. Data was largely taken from TDA corpus, but we have also used our translation memories on IT (English - Russian pair). The data has to be cleared from garbage to train Moses – we removed complex formatting and tags, also simply dropped invalid segments. This has been identified as major obstacle and project effort. Currently TDA is working on cleaning the data, but at that moment they were far from clean (10-20% of all units were irrelevant for Moses training) - Removed tags for RTF, HTML, XML - Removed/fixed strings with wrong encoding - Removed untranslated strings (source = target); - Removed strings with no text (numbers, dates) or empty target (C) 2003-2010 LOGRUS LLC
  • 7. Playing around 1. We evaluated not only language quality as such, but also the convenience of the system installation and setup, difficulty of fine-tuning and use, and system stability. 2. We have tried various Moses distributions, language models and Linux distributions. 3. We have decided that IRST is the optimal language model. 4. Various language parameters of Moses have been played with, to get the optimal result. ∙ -weight-d (-d): weight(s) for distortion (reordering components) [input-factors] ∙ -weight-e (-e): weight for word deletion 0 ∙ -weight-file (-wf): file containing labeled weights [mapping] ∙ -weight-generation (-g): weight(s) for generation components T0 ∙ -weight-i (-I): weight for word insertion [ttable-file] ∙ -weight-l (-lm): weight(s) for language models 0 0 5 /home/atma/BASES/db/msao_e/lngbase/msao_e-MAIN ∙ -weight-t (-tm): weights for translation model components [lmodel-file] ∙ -weight-w (-w): weight for word penalty 1 0 3 /home/atma/BASES/db/msao_e/lm/1.gz ∙ -phrase-drop-allowed (-da): if present, allow dropping of source words [ttable-limit] ∙ -distortion: configurations for each factorized/lexicalized reordering model. 20 0 ∙ -distortion-limit (-dl): distortion (reordering) limit in maximum number of words (0 = monotone, -1 = unlimited) [weight-d] ∙ -max-phrase-length: maximum phrase length (default 20) 0.6 [weight-l] ∙ -mbr-size: number of translation candidates considered in MBR decoding (default 200) ∙ -minimum-bayes-risk (-mbr): use miminum Bayes risk to determine best translation 0.5000 ∙ -monotone-at-punctuation (-mp): do not reorder over punctuation [weight-t] ∙ -n-best-factor: factor to compute the maximum number of contenders (=factor*nbest- 0.2 size). value 0 means infinity, i.e. no threshold. default is 0 0.2 0.2 0.2 (C) 2003-2010 LOGRUS LLC 0.2 [weight-w]
  • 8. Final Logrus Moses system configuration 1. Hardware: Quad-Core AMD-64, 16 Gbytes of RAM, 64-bit Linux Ubuntu, 100 Gbytes of disk space. 2. Moses configuration: the most recent build, IRST language model, GIZA++ (training component). 3. We have created Web interface to the system. 4. Moses training has to be automated, and was automated, to try the system with different SME corpuses 5. Other SW: our proprietary scripts developed to train Moses, its setup and remote (Web) client access. (C) 2003-2010 LOGRUS LLC
  • 9. Training, Training, Training Training and data cleaning was lengthy and iterative process, had to be automated. (C) 2003-2010 LOGRUS LLC
  • 10. Results Edit Distance (average value of Character Edit Rate1) Language pair SMT SMT (*) SMT (**) RBMT RBMT <# 4> <# 5> MT system (developer) (Bing — Logrus Logrus (ProMT v8) (SYSTRAN) (SDL) (Google) Microsoft) Moses Moses English – French 0.45 - - 0.54 0.51 - - English – Italian 0.46 - - 0.59 0.51 - - English – Portuguese 0.42 - - 0.58 0.58 - - English – Russian 0.53 0.57 0.61 0.63 0.69 - - English – Spanish 0.35– 0.37 - - 0.46 0.45 0.40 0.35 (*) Logrus Moses after deep fine-tuning, trained on very narrow IT domain. (**) Logrus Moses after deep fine-tuning, trained on much wider IT domain (TAUSDATA) 1. We have obtained Moses translation quality results in standard configuration, without deep fine-tuning. 2. The same after deep fine-tuning. 3. The MT output results from various systems were evaluated on one and the same sample. 4. Translation quality has been evaluated with Edit Distance algorithm (the software tool was developed in Logrus). 5. The “pure” result was better than with untrained PROMT 8 and was close to BING Translator from Microsoft. 6. IN DOMAIN – vs GENERAL, NOT IN-DOMAIN engines 1 rate of the number of changes introduced after post-editing, to the sentence length. (C) 2003-2010 LOGRUS LLC
  • 11. Problems of MT integration into actual process 1. Large paragraphs (>255 characters). Moses fails to build a match table on them, so we had to break them down. 2. Punctuation marks are treated as letters. Multi-sentence segments are not treated correctly and add noise. 3. Basic Moses configuration only work with lowercase letters. Additional processing is required to restore case. 4. RTF, XML, HTML tags are not supported. 5. Training corpus often contains ambiguous, obsolete terms. Extensive terminology support is required. 6. Every Moses user has to develop his own interface between Moses and his own workflow, CMS, TMs, projects, etc. (C) 2003-2010 LOGRUS LLC
  • 12. Moses Features, Advantages, and Disadvantages 1. ADVANTAGES: - Free - Amazingly high quality of basic system after proper tweaking - Minimal setup provides competitive language quality level - Practical solutions can be based on Moses 2. DISADAVTAGES: - Moses is not a ready tool – it’s a construction set - Serious efforts are required to bring it to practical use 3. PROBLEMS: - Training corpuses require cleaning - TMX is poorly supported by existing software - Domains of feasibility are so specific that it is not clear whether the effort worth it yet for “generic” LSP (C) 2003-2010 LOGRUS LLC
  • 13. Conclusions – “scientific” aspects 1. Hypothesis is true – it does look like SMT breaks the MT quality barrier known until recently. Another plateau is within usable domain in certain situations. 2. We experienced the benefits of statistics-based method and its limitations. 3. We obtained information on areas where SMT produces best results. 4. We have found the configuration setting and training technique – Moses experience that we can take further. 5. We concluded that further research is required and beneficial. 6. Terminology requires special attention. Large corpus is not always better than smaller one. Very serious inconsistencies in terminology in corpus, particularly detrimental for SMT engines. 7. SMT vs RBMT choice criteria refined. (C) 2003-2010 LOGRUS LLC
  • 14. Conclusions – technical aspects 1. TAUS DATA: insufficient number of data categories (“genres”, content types). Extremely diverse data are grouped together, requires domain separation. 2. Too few TMX editors on the market, most of them too expensive. Multiple problems with TMX support, language support and/or stability of these editors. 3. Multiple problems with Import/Export in all these editors. Logrus had to develop a TMX editor of its own, and we are planning to provide it for free to cooperating parties in the near future. Other useful proprietary tools and techniques have been developed can be shared. 4. No MT – TM – Translation Environment integration yet. 5. Terminology-Related Problem: Linking to Various Data Corpora RBMTs require grammatical attributes, while STAT-based MTs require fine, “genre”- related classification. (C) 2003-2010 LOGRUS LLC
  • 15. Conclusions – practical aspects 1. Overall linguistic quality of MT output is still low: Output from MT engines are much closer to each other than to final quality human translation. (We are still very far from Turing barrier.) 2. Mass post-editing is required. Editing with reference to the source is a MUST. 3. Dedicated terminology support is a must. Complete, 100% terminology post-verification is required. 4. MT post-editing is different from human editing, and editors must be trained. 5. With “Difficult" language pairs the cost of post-editing would be 50%-75% of the cost of human translation, in best case scenario, after costly preparation. Overall ROI must include internal R&D and/or license and deployment costs. 6. Continuous internal R&D is required to make use of the technology. 7. ROI as a consequence is very specific for particular project and company; further research into methods to evaluate ROI is required. 8. The domain of positive ROI increases. LSPs should pay attention. 9. Significant localization process reengineering is required for MT deployment on all stages: corpus, MT setup and administration, deployment, vendor selection, post-editing processes. 10. Languages vary – the process would be different for each language (SMT vs RBMT, parameters, quality level, etc.). MLVs are at disadvantage – SLVs have opportunity again with more focus into particular language. (C) 2003-2010 LOGRUS LLC
  • 16. Last Mile Phenomenon Outputs from MT engines are much closer to each other than to final quality human translation. (C) 2003-2010 LOGRUS LLC
  • 17. Last Mile Phenomenon 2. Proximity phenomenon: the closer you are to the goal, the more subtle details are coming to stand out and still divide you and the prize. “Final 10% to the goal” is the Last Mile you need to cover with the largest effort and cost, deployment, preparation, process change and 100% editing. (C) 2003-2010 LOGRUS LLC
  • 18. HYPOTHESIS and goals of the experiment 1. TEST HYPOTESIS – Is the Plateau of quality behind? How far is the next plateau of quality from Human Translation? Yes, but we are about to hit another plateau, within usable domain. Human quality is still very far, the last mile gap is WIDE. Current predictions of machines passing Turing Test are 2029. 2. Is the MT technology available for LSPs without multimillion funding? Not multimillion, but continuous, significant effort is required. 3. Is OpenSource SMT Moses a viable place to start? YES and NO. For immediate deployments other commercial products may be more feasible. Moses is however very promising. A very good training ground to prepare for MT or evaluate current MT level. 4. Can we base our own productivity solutions on Moses in any near future? Timeframe is several years still. Numerous technical problems are to be solved. 5. What are the possible LSP MT-related process implications? We’ve got some answers to ourselves. 6. What are the methods to calculate ROI? We have developed our method. 7. What are the post-editing differences in MT scenario? Under development. 8. Is my company capable of handling this? YES. (C) 2003-2010 LOGRUS LLC