Measuring translation quality is a shared priority
Part of what makes Mozilla projects so unique is the involvement of the community. Our community prides themselves on being experts in the Web, Mozilla, and all of Mozilla’s products. Thus, delivering high quality localizations of Mozilla products to users is not only a priority for the l10n-drivers, but one that is close to the community’s heart.
Currently, Mozilla has no criteria-based framework for evaluating a localization/translation’s accuracy or quality. Determining how to evaluate translation quality is a very difficult task because language itself is very flexible and subjective. Critical to creating a successful framework for evaluating translation quality is including elements of a project’s scope as well as the most objective pieces of language, such as orthography, grammar, and corporate style guide. A translation quality assessment framework would need to be flexible, robust, interoperable, and easy for graders to use. Developing such a framework is difficult, but I believe we’ve experienced a breakthrough!
Background of the MQM standard: Mozilla pilot projects
Translation Studies researchers from Brigham Young University (along with other North American and European organizations) set out to create a translation quality standard that would be flexible enough to accomodate project specifications and the needs of individual langauges. The Multidimensional Quality Metric (MQM) framework was born!
The MQM framework allows an organization to identify their highest priorities in assessing the translation quality of a specific project, or series of projects. The framework’s issue types are determined by the project’s organization. Graders for the project’s translations are recruited and trained on the meaning of these issue types and the process of grading a translation according to the selected issue type framework. Graders spend between 30-60 minutes per day marking the translation errors they find and categorizing them by issue type. The translation is then assigned a “score” according to the combined grades issued by the graders.
Delphine and I collaborated with these researchers to create MQM evaluation sprints for Firefox and Firefox OS localizations. Together, we found language and subject matter experts from within the university and the community to participate in these sprints. There were a total of three sprints organized over the last six months, each evaluating different languages and utilizing a variety of issue types. Participants were involved both virtually and in a physical space at Brigham Young University and were able to collaborate closely using IRC. Each individual grader evaluated between 400 – 4,000 strings in a single week!
Results of those projects
The results from these sprints were very positive. Localization teams were able to receive specific, actionable feedback on where their translations could be improved. This feedback was organized by issue type and called out specific strings as needed correction in one or more of the issue type categories. Some teams, like the French l10n team, were able to act immediately on that feedback and incorporated into their Firefox OS localization.
Community perception of MQM
Surveying the members of the community concerning their experience with the MQM framework returned very positive results. Generally speaking, localizers enjoyed how thorough the framwork’s definition was for the sprints and felt that the process was easy to understand. Many felt that the experience was valuable to their l10n work and were in favor of the l10n-drivers implementing the standard across the Mozilla l10n program.
Plans to implement MQM framework
There are preliminary plans to develop a Mozilla l10n QA tool based on the MQM framework. Ideas have included a gamified system, allowing users to choose to grade by project or by isolated issue type, and a sleak, web-based GUI complete with product screenshots for each string translation being evaluated. Some of the community has expressed interested in being involved in creating such a tool. We’ll be incorporating their feedback and getting their help as we move forward designing this MQM-based tool.
The MQM standard is governed by ASTM International. Mozilla has been invited to participate in the governing technical committee for the MQM standard, and we’re currently evaluating what our involvement in that committee would look like.
If you have any questions concerning the MQM framework at Mozilla, please feel free to email the Mozilla L10n mailing list.
Julen wrote on :
Jeff Beatty wrote on :
Arle Lommel wrote on :
Jeff Beatty wrote on :