Post Survey Note: Thank you to all those who completed the survey. It’s no longer live, but you can see the final results in the article.
For the last couple of years I’ve been enjoying the TCLoc Masters degree at the University of Strasbourg. It’s been a really interesting time for me helping to fill in a lot of gaps and widen my technical knowledge around localization, and introducing me to the world of Technical Communication in general. This latter part was particularly interesting because half of our business at SDL relates to this; so having spent my time since 2006 working with our localization products it’s been an eye opener in many ways. I have done this in my own time and not as part of my job, but TCLoc does look like a course that’s tailor made for SDL employees!
“The badass is an uncommon man of supreme style. He does what he wants, when he wants, where he wants.” (Urban Dictionary by dougdougdoug). There are in fact many definitions of what a badass is, but I like this part of one of the definitions because it really reflects what this article is about and why it’s needed. No clues so far… but let’s think anonymization!
AdaptiveMT was released with Studio 2017 introducing the ability for users to adapt the SDL Language Cloud machine translation with their own preferred style on the fly. Potentially this is a really powerful feature since it means that over time you should be able to improve the results you see from your SDL Language Cloud machine translation and reduce the amount of post editing you have to do. But in order to be able to release this potential you need to know a few things about getting started. Once you get started you may also wonder what the analysis results are referring to when you see values appearing against the AdaptiveMT rows in your Studio analysis report. So in this article I want to try and walk through the things you need to know from start to finish… quite a long article but I tried to cover the things I see people asking about so I hope it’s useful.
I don’t know if you can recall these games from when you were a kid? I used to spend hours trying to find all the differences between the image on the left and the one on the right. I never once thought how that might become a useful skill in later life… although in some cases it’s a skill I’d rather not have to develop!
You may be wondering where I’m going with this so I’ll explain. Last weekend the SFÖ held a conference in Umeå, Sweden… I wasn’t there, but I did get an email from one of my colleagues asking how you could see what changes had been made in your bilingual files as a result of post-editing Machine Translation. The easy answer of course is to do the post-editing with your track changes switched on, then it’s easy to spot the difference. That is useful, but it’s not going to help with measurement, or give you something useful to be able to discuss with your client. It’s also not going to help if you didn’t work with tracked changes in the first place because you’d need some serious spot the difference skills to evaluate your work!
In the last year or so I’ve had the pleasure of watching Patrick Hartnett use the SDL Openexchange APIs and SDK to develop SDLXLIFF Compare, then Post-Edit Compare, the Studio Timetracker and a productivity tool that combined all of the first three into one and introduced a host of productivity metrics and a mechanism for scoring the quality of a translation using the Multidimensional Quality metrics (MQM) framework. This last application was never released, not because it wasn’t good, but because it keeps on growing!
Then last month I got to attend the TAUS QE Summit in Dublin where we had an idea to present some of the work Patrick had done with his productivity plugin, get involved in the workshop style discussions, and also learn a little about the sort of things users wanted metrics for so we could improve the reporting available out of the box. At the same time TAUS were working on an implementation around their Dynamic Quality Framework (DQF) and were going to share a little during the event about their new DQF dashboard that would also have an API for developers to connect.
The AutoSuggest feature in Studio has been around since the launch of Studio 2009 and based on the questions I see from time to time I think it’s a feature that could use a little explanation on what it’s all about. In simple terms it’s a mechanism for prompting you as you type with suggested target text that is based on the source text of the document you are translating. So sometimes it might be a translation of some or all of the text in the source segment, and sometimes it might be providing an easy way to replicate the source text into the target. This is done by you entering a character via the keyboard and then Studio suggests suitable text that can be applied with a single keystroke. In terms of productivity this is a great feature and given how many other translation tools have copied this in one form or another I think it’s clear it really works too!
AutoSuggest comes from a number of different sources, some out of the box with every version of the product, and some requiring a specific license. The ability to create resources for AutoSuggest is also controlled by license for some things, but not for all. When you purchase Studio, any version at all, you have the ability to use the AutoSuggest resources out of the box from three places: Continue reading
Best practice! This is a phrase I’ve had a love/hate relationship with over the course of my entire career… or maybe it’s just a love to hate! The phrase is something that should perhaps be called “Best Suggestions” and not “Best Practice” because all too often I think it’s used to describe the way someone wants you to work as opposed to anything that represents the views of a majority of users over a long period of time, or anything that takes account the way different people want to work. In fact with new technology how can it be “Best Practice” when it hasn’t been around long enough in the first place? I think for a clearly defined and well established process then “Best Practice” has it’s place… but otherwise it’s often the easy answer to a more complex problem, or just a problem that is considered too hard to address.
It would be very arrogant of me to suggest that I have the solution for measuring the effort that goes into post-editing translations, wherever they originated from, but in particular machine translation. So let’s table that right away because there are many ways to measure, and pay for, post-editing work and I’m not going to suggest a single answer to suit everyone.
But I think I can safely say that finding a way to measure, and pay for post-editing translations in a consistent way that provided good visibility into how many changes had been made, and allowed you to build a cost model you could be happy with, is something many companies and translators are still investigating.
I believe this interesting quote can be found in “Prometheus Bound”, a play by a Greek dramatist called Aeschylus. I haven’t read the play, but I like the quote, and it certainly lends itself to the importance of memory… even when we refer to a Translation Memory rather than your own built in capability. It’s because your Translation Memory is such an important asset to you that you need to regularly maintain it, and also reuse it wherever possible to expand the benefits you get from it.
Updated: 14 January 2015
Today SDL is all about SDL Language Cloud and not BeGlobal, but I hope the article is still as relevant today. There are more ways to look at how you use Machine Translation so if you’re interested take a look at these two more recent articles as well.
The ins and outs of AutoSuggest
Language Cloud… word-counts… best practice?
The title of this post could be quite tricky to translate in many languages because not everyone uses the expression in the same way, and certainly don’t use the same words. I chose this especially because I thought I’d write a little about using Machine Translation in SDL Trados Studio.
I’m not going to talk about properly trained Machine Translation engines such as SDL BeGlobal, which can be configured and improved to provide remarkably good translations in a short period of time for very large numbers of words… so achieving economies of scale that would be unthinkable with human resources alone. Instead, I’m going to talk about how a Translator can make use of the growing number of Machine Translation resources in a way that might make sense for them.