Machine translation (MT) was developed as one of the hottest research topics in the natural language processing (NLP) literature. One important issue in MT is that how to evaluate the MT system reasonably and tell us whether the translation system makes an improvement or not. The traditional manual judgment methods are expensive time-consuming unrepeatable and sometimes with a low agreement. On the other hand the popular automatic MT evaluation methods have some weaknesses. Firstly they tend to perform well on the language pairs with English as the target language but weak when English is used as the source. Secondly some methods rely on many additional linguistic features to achieve good performance which makes the metric unable to replicate and apply to other language pairs easily. Thirdly some popular metrics utilize incomprehensive factors which result in low performance on some practical tasks. In this thesis to address the existing problems we design novel MT evaluation methods and investigate their performances in different languages. Firstly we design augmented factors to yield highly accurate evaluation. Secondly we design tunable evaluation models ...
Piracy-free
Assured Quality
Secure Transactions
Delivery Options
Please enter pincode to check delivery time.
*COD & Shipping Charges may apply on certain items.