open translations database



Hello Gustavo

We are currently creating a Brazilian Portuguese Style Guide. 
 You are more than welcome to have a copy on completion (est 
mid Dec). The completed languages available at present is 
English, Japanese, French, German, Spanish, Italian, Swedish. 
We are also currently working on completing Simplified 
Chinese, Traditional Chinese and Korean.

Not to effect everyone's mail server, I will send you a 
separate mail with a copy of our English Style Guide which 
you can use as a basis to generate the Brazilian Portuguese 
Style Guide. I will also attached our International Spanish 
Style Guide as reference.

We work with approved freelancers who perform linguistic 
quality assurance feedback on our projects.  They are mainly 
responsible for performing QA's on various samples of the 
files and reporting their findings.  The QA feedback report 
is returned to the translators who implements the recommended 
changes.  However, depending on the language and the project 
requirements, the can implement linguistic changes to the 
files if necessary.  The QA is aimed to ensure consistency 
across applications and/or components such as UI, Help and 
documentation.  Attached is a copy of the template which is 
completed during the QA, "SUN_QA_Matrix2.html"

Gustavo, I can have our Brazilian Portuguese freelancer 
perform a QA on GNOME translations and report his/her 
feedback to the Brazilian Translation Team.  Let me know if 
you are interested.

Bye for now
Aoife







> Date: Wed, 01 Nov 2000 21:25:22 -0200
> From: Gustavo Maciel Dias Vieira <gdvieira@zaz.com.br>
> X-Accept-Language: en
> MIME-Version: 1.0
> To: Aoife Dunne - Sunsoft ELC <Aoife.Dunne@ireland.sun.com>
> CC: gnome-i18n@gnome.org
> Subject: Re: open translations database
> Content-Transfer-Encoding: 8bit
> X-MIME-Autoconverted: from quoted-printable to 8bit by 
ireserver.Ireland.Sun.COM id AAA15619
> 
> Aoife Dunne - Sunsoft ELC wrote:
> > 
> > 
> > Style Guides
> > ------------
> > We have some localised versions of a style guidelines.  
These
> > guidelines are used to aid the translators.  For example, 
in
> > France  how the date, time formats should be localised.  
In many
> > countries such data is correct in many formats, however, 
the use
> > of style guides decide on the preferred format for the 
use of
> > consistency.   Our style guides could be used as 
reference and
> > updated to create a GNOME specific style guide for all 
languages.
> > Let me know if you are interested and I will send you a 
copy of
> > our country specific style guides.
> 
> Hello, Aoife
> 
> Do you fine folks at Sun have a brazilian portuguese team?
> I'm starting to write a style guide for brazilian 
translators and would
> love to receive a copy of yours style guide, if available 
for brazilian
> portuguese.
> 
> 
> > 
> > How else can Sun help,
> > 
> > * possible act a the host for the translation memory 
database,
> > populating newly translated products.
> > 
> > * provide linguistic quality assurance feedback and 
implement
> > linguistic changes if necessary checking for grammar, 
spelling,
> > inconsistencies etc.
> 
> How do you plan to provide QA for translations? Is it 
avaiable for
> brazilian portuguese?
> 
> Abraços,
> Gustavo
> 
> -- 
> gdvieira@zaz.com.br
> http://www.dcc.unicamp.br/~ra941486

Aoife Dunne
Program Manager
European Localisation Centre
Sun Microsystems Ireland Ltd
Hamilton House
East Point Business Park
Dublin 3
Ireland
Tel.:  	+353-1-8199-266
Fax:.	+353-1-8199-261
Email:	aoife.dunne@Ireland.Sun.COM

Title: SUN Quality Assurance Matrix

SUN Quality Assurance Document

 

 

 

QA CHECKLIST – To be filled in by LS

 

Product name

 

Platform/Product line/Version

 

Code

 

Language

 

QA phase

 

Date

 

List of files QAed

 

Completed Call-for-Work document received?

 

Any outstanding queries in the sample QAed?

 

Important note: Please note that the Linguistic Specialist is required to fill in all applicable fields and most importantly that a Pass/No Pass valuation is given for each applicable category as well as for the overall assessment. Please also supply a high-level summary assessment at the beginning of you report which should be directed to the Project Manager. All comments apart from specific language examples and corrections must be given in English.

Important note: If the linguist at any stage during the QA cycle deems the translation to be not fit for QA, the QA cycle should be interrupted and the translation together with an initial report returned to the vendor immediately for editing/correcting. Another day should be earmarked for a second QA cycle. It is of greatest important that the vendor gives priority to correcting the translation so that the second QA cycle can take place as soon as possible.

 

SUMMARY REPORT ON THE TRANSLATION

SUMMARY:

 

 

 

 

 

 

 

 

 

 

 

SIGNATURE & DATE: ___________________________________________________________

 

 

 

 

REPORT

Location of error

US text

Translated text

Description of error/issue

Required/suggested correction

Comments

           
           
           
           
           
           
           

Explanatory comments:

‘Location of error’ may include a book name, page number, file name, URL, path followed to SW item a numbering structure, etc. The location must be unique or in the case of a general comment, this must be clearly marked as ‘global’.

The ‘US text’ may be optional but must be included if the translation error is of the type mistranslation, omission, or misunderstanding of original.

The ‘Translated text’ must be quoted exactly as it appears in the document.

The ‘Decription of error/issue’ must be as per the error categories listed above. All descriptions of errors must be given fairly and non-subjectively.

A ‘Required/suggested correction’ must be offered by the Linguistic QA Specialist, except where there are large passages left untranslated which must be provided by the original vendor/translator.

‘Comments’ is an optional column which may be used by the vendor if he strongly disagrees with corrections suggested by the Linguistic QA Specialist.

Important note: Sometimes corrections are of the type that instead of one grammatical construction (e.g. noun construct) another grammatical form (e.g. infinitive) should be used, particularly in (sub)headings or captions. E.g. not ‘Bearbeiten von Bildern’, but rather ‘Bilder bearbeiten’. In such cases, these global changes need be quoted only once (but marked ‘global’) if the corrections are to be implemented by the vendor. The must be quoted for each and every instance if the corrections are to be implemented by an engineer in SUN.

 

 

COMPLIANCE CHECKLIST – DOCUMENTATION/HELP

 

ERROR CATEGORY: Accuracy

Error type

Max. errors allowed

Errors found

Omissions/additions (e.g., text, callouts, titles, index markers, graphics)

   

S/W references incorrect

   

Mistranslations

   

Headers/footers incorrectly translated

   

TOC/Index errors

   

Cross-references/index markers deleted

   

Acceptable level: 0 - 3 error per 1000 words

No. of errors: Rating : Pass  No Pass 

 

ERROR CATEGORY: Terminology

Error type

Max. errors allowed

Errors found

Glossary not adhered to

   

S/W references incorrect

   

Inconsistent terminology

   

Acceptable level: 0 - 3 error per 1000 words

No. of errors: Rating : Pass  No Pass 

 

ERROR CATEGORY: Language

Error type

Max. errors allowed

Errors found

Semantics

   

Spelling

   

Grammar

   

Punctuation

   

Hyphenation

   

Cross-referencing

   

Acceptable level: 0 - 3 error per 1000 words

No. of errors: Rating : Pass  No Pass 

 

ERROR CATEGORY: Style

Error type

Max. errors allowed

Errors found

General style

   

Register/tone

   

Style guide not followed

   

Language variants/slang

   

Acceptable level: 0 - 4 error per 1000 words

No. of errors: Rating : Pass  No Pass 

 

 

ERROR CATEGORY: Country

Error type

Max. errors allowed

Errors found

Country standards

   

Local suitability

   

Company standards

   

Translation guide-lines not followed (if applicable)

   

Acceptable level: 0 - 4 error per 1000 words

No. of errors: Rating : Pass  No Pass 

 

Overall rating: Pass/No Pass

Requires second QA cycle: Yes/No

 

 

COMPLIANCE CHECKLIST – SOFTWARE

ERROR CATEGORY: Accuracy

Error type

Max. errors allowed

Errors found

Omissions/additions

   

Mistranslations

   

Acceptable level: 0 - 3 error per 1000 words

No. of errors: Rating : Pass  No Pass 

 

ERROR CATEGORY: Terminology

Error type

Max. errors allowed

Errors found

Glossary not adhered to

   

Inconsistent terminology

   

Acceptable level: 0 - 3 error per 1000 words

No. of errors: Rating : Passed  Failed 

 

ERROR CATEGORY: Language

Error type

Max. errors allowed

Errors found

Semantics

   

Spelling

   

Grammar

   

Punctuation

   

Abbreviations

   

Acceptable level: 0 - 3 error per 1000 words

No. of errors: Rating : Pass  No Pass 

 

ERROR CATEGORY: Functionality

Error type

Max. errors allowed

Errors found

Core functionality

   

Hotkey/shortcut functionality

   

Platform/hardware compatibility

   

Keyboard layout/code page

   

UI problems (alignment, sizing, positioning, truncation)

   

Acceptable level: 0 - 3 error per 1000 words

No. of errors: Rating : Pass  No Pass 

 

 

ERROR CATEGORY: Country

Error type

Max. errors allowed

Errors found

Country standards

   

Local suitability

   

Company standards

   

Translation guide-lines not followed (if applicable)

   

Acceptable level: 0 - 4 error per 1000 words

No. of errors: Rating : Pass  No Pass 

 

Overall rating: Pass/No Pass

Requires second QA cycle: Yes/No

 

A Pass/No Pass must be given for both the overall rating and for all the different quality categories!

Important note: A No Pass means that the vendor must carry out an internal QA/edit of the entire translation, not just the marked up pages/chapters.

A Pass means that the suggested corrections must be implemented but that the rest of the translation can be left as is.

 

 

Reference

Ratings and ratings parameters

The aim – excellent quality

Documentation, help

Language: Describes functionality very accurately. Information correctly translated, concise and easily accessible. Consistent use of terminology and style. Adhering to glossary, style guide and any guidelines. No grammar, punctuation or spelling errors. All page references and cross-references correct. Complete index. Correct country standards used.

Formatting: All standard fonts and templates used. No paragraph style or character formatting errors. All art referenced correctly. No index, formatting or reference errors. All page references correct. Help is complete with correct alignment and no display errors.

Functionality: Documentation and help fully consistent with software. Correct art pieces inserted throughout with all callout references in documentation accurate. Technically accurate localisation with no misleading statements, showing a thorough understanding of the product. No topic or footnote errors in help.

Software

Language: No language bugs, Completely, accurately localised. Examples well localised. UI terms consistent and appropriate to platform. Adhering to glossary and SUN guidelines. Describes functionality accurately.

Formatting/functionality: Dialog boxes all correctly sized. No duplicate hotkeys. All text items localised, aligned correctly and fully visible on the screen.

 

What to look out for – unsatisfactory quality:

Documentation, help

Language: Product difficult to use due to omissions in translated text, mistranslations due to misunderstanding the source or lack of information. Discrepancies between software and documentation. Information difficult to access. Examples not or not well localised. Writing style poor, terminology inconsistent. Numerous grammar, punctuation or spelling errors. Serious inconsistencies, errors in page references, cross-references or index. Unacceptable errors in country standards.

Formatting: Serious formatting and layout errors. Standard fonts and templates not adhered to. Paragraph style or character formatting errors. Art not referenced correctly. Help does not compile or contains alignment and display errors. Requires a significant amount of rework to comply with market requirements

Functionality: Inconsistencies between documentation, help and software. Jumps/popups not functioning in help. Incorrect art pieces inserted or inaccurate callout references in documentation. Poor understanding of product leading to inadequate technical localisations throughout. Topic or footnote errors in help.

Software

Language: Language bugs, inadequate or incomplete localisation. Examples not or not well localised. Inconsistent UI terms or not suitable for platform used. Not adhering to glossary or SUN guidelines. Functionality not accurately expressed.

Formatting/functionality: Significant number of localisation bugs. It takes longer to report the bugs than to test the product, demonstrating a lack of any QA check before hand over. Duplicate hotkeys. Text items not aligned correctly or truncated on the screen. A significant amount of rework is required to bring the software to an acceptable standard.

 

 

The aim of this document is to outline SUN’s language quality assurance matrix.

Goals

This document provides information on how the language quality of translated products should be assessed. It outlines the criteria used for evaluating quality and describes the recommended reporting procedures.

This document is addressed to the Linguistic QA Specialists. Their core responsibilities are:

Note: Since all Linguistic QA Specialists used by SUN are a) freelance linguists and b) not involved in the translation/localisation process prior to delivery by the vendor they have no role in language support during a translation project, stylistic or terminological guidelines for translation or any scheduling issues. Their role is purely to assess the ‘finished’ and delivered translation and to suggest corrective action.

Prerequisites for a succesful QA cycle

Theses are dealt with in more detail in a separate document. The following are purely some main checkpoints.

 

 

Language criteria to be taken into consideration

The criteria used by SUN to determine quality and errors in printed and online documentation and software are:

Accuracy omissions, additions, cross-references, headers/footers

Terminology glossary adherence, consistency, context, abbreviations

Language grammar, semantics, punctuation, spelling

Style general style, register/tone, language variants/slang

Country country standards, local suitability, company standards

Functional criteria in software – GUI terminology, cross-application/cross-platform terminology, abbreviations, key combinations and sequences, core functionality, hotkey/shortcut functionality, codepage, platform/application compatibility

Formatting criteria in software – alignment, sizing, truncation, non-displaying characters

 

Materials to be used for the QA:

Supplied glossaries

Supplied style guide

List of outstanding queries if any

Complete Call-for-Work document

QA Feedback form supplied

 

Project elements and fundamental QA prerequisites

All material is provided in electronic format. QA takes place on-screen, any comments are in report format. There is no mark-up on paper except in exceptional cases such as marketing and packaging material or artwork. The file formats used for almost all printed and online documentation are HTML, SGML, and Frame (MIF). In the case of marketing material it is important to QA the PostScript files. When checking for consistency across all components, the software takes precedence over all other project elements, i.e. documentation and online help are evaluated against the software and made to conform to it. It is furthermore important to ensure that the translation is in line with other products in the product line. The major product lines in SUN are Java, Solaris and hardware-related products. There may in future be an additional educational product line.

As a rule of thumb, between 10% and 20% of a translation are QAed, or in the case of short documents a minimum of one day is set aside for QA, in which case the entire document may be checked. For longer translations the word count must be known and provided by the PMs on the Product Information Sheet.

The Linguistic QA Specialists will select 10-20% of the translation based on the word count provided. It is important to select random samples of a translation for QA, i.e. not chapters 1 to 3 but rather e.g. 10 pages from chpt. 4, all of chpt. 7, 2 appendices plus the letter M from the index checked against the relevant pages.

When checking online help, it is important to follow this sequentially by proceeding through the TOC (again the above rules for random sampling apply). Links and context-sensitivity should be checked for functionality but not as a method for proceeding through the help system. In the case of software, all of the UI must be checked plus 10-20% of other messages.

For scheduling/pricing purposes, it is assumed that a Linguistic QA Specialist can check 8000 words of printed/online documentation per day and 4000 words of software per day.

QA of printed and online documentation:

The translated document is evaluated against the base product documentation. In the case of context-sensitive online help, the compiled files are used in the evaluation. The QA takes place sequentially and not by proceeding via links or the context-sensitivity function. All printed and online documentation and context sensitive help are evaluated against the software and made to conform to it.

Software QA:

The localised compiled software is evaluated against the US software. The software QA is of critical importance. All of the User Interface as well as 10-20% other messages should be evaluated. Software QA may also involve some software testing to a lesser or higher degree. In such cases test scripts would be of great help to the Linguistic QA Specialists. On the UI a No Error Pass is required.

Other elements – packaging, marketing, legal material. For these the PostScript files or printouts are to be QAed. On all packaging, marketing and legal material a No Error Pass is required.

 



[Date Prev][Date Next]   [Thread Prev][Thread Next]   [Thread Index] [Date Index] [Author Index]