Case Studies
Discover transformative solutions in action. See firsthand how our Pioneers’ commitment to tackling diverse challenges delivers tangible results. Explore our tailored approaches for navigating complexities, ensuring measurable success.
I. Problem
A language service of a large life science company was facing the typical challenges of common translation workflows. It was difficult to merge translation memory and machine translation into a single paradigm. There was a huge in-house effort to export files, select matching translation memories, send and receive translation packages, import files, and update databases. Additionally, each LSP (Language Service Provider) had a preferred platform to exchange files. The department had little visibility into how the translations were actually made, thus lacking the data to improve the process. Its valuable Multilingual Knowledge System was only used for term recognition.
II. Solution
After a successful PoC (Proof of Concept), the company decided to deploy a language factory. A language factory centralizes all automatic steps such as content recycling, machine translation, automatic correction, quality estimation, etc. The secret of excellence in production often lies not so much in the individual machines but in how they work smoothly together. Collecting data at every process step allows for constant optimization of the factory’s performance. The language factory uses three simple, standardized API calls to communicate with the company’s LSPs for handover. Files to be reviewed by the expert-in-the-loop are posted, the status is polled, and the finished files are fetched.
The language factory connects in a similar automatic way with the company’s content management systems. It uses the COTI standard to collect work and also to place the translated files back in the right place.
By analyzing human edits the factory can train its AI and constantly improve its estimations. The linguistic assets collected in the content repository are used to train the machine translation. The Multilingual Knowledge System identifies domains and topics. This information is used to ensure that the largest chunks of the most relevant content are recycled. Sudden domain switches trigger QA warnings and lower QE scores.
With every project, the factory collects more data, which is nicely visualized in a dashboard. This way, the factory can not only be easily monitored, but certain parameters can be controlled to optimize its operation. Finally, the cost-time-quality triangle can be smartly adjusted to meet business needs.
III. Experiences, Benefits, and Metrics
Months after deployment the language factory has already processed millions of words into 36 languages supported by three LSPs. Besides already delivering significant cost savings of around 28% it allows the department to focus on more value-generating tasks than before. Its language experts can now enforce source text quality, prepare and train MT models, manage multilingual knowledge, define and adapt post-editing criteria, monitor the solution, and analyze process data.
Perhaps most importantly, though, is the constant collection of high-quality multilingual data. These linguistic assets are used for other applications, solving NLP tasks, and training LLMs. The vision is that the department delivers the data and knowledge to support any textual AI initiative of the company. Therefore, it has renamed itself to Language Operations.
I. Problem
What is MVLP and What is a Localization Product?
MLVP or minimum viable localization product is a term that describes the initial stage of a localization product. Apart from human workflow orchestration, MVLP contains no human input apart from data management and curation that is used to train the AI responsible for generating the MVLP. It is a workflow that consists of a raw MT output, followed by a trained AI post-editing process and corresponds to the LangOps manifesto to “Build Language-agnostic”.
Localization product is a concept that can be compared to a software product in many ways. It is helpful to think about it in terms of DevOps practices and how product is manipulated and iterated in sprints until it reaches its final form. Each version of localization product presents different added value that is defined by localization sprint objectives.
II. Solution
Why We Need a Minimum Viable Localization Product?
MVLP serves as the base working product of localization. Having been machine translated and AI post-edited, it is a full placeholder for content that, while not finalized, can be used in in gathering data on user behavior and traffic analysis.
This was also the case where a client of Native Localization agreed to create a workflow that introduced real-time data into their localization decision-making.
A client, who is a software development company in the Fintech sector maintains a product platform, as well as the knowledge base for their customers to be able to use their product to its maximum efficiency. After Native had performed product string localization, it made sense to follow up with the localization of the knowledge base. However, their localization budget was spent for that fiscal year and this portion of content was not included, which was quite substantial as supporting documentation tends to be.
It would not be efficient from a UX perspective to have disparity between content, so the solution required a LangOps based approach, which would dictate that we must “leverage all data and tech” in efforts to make smart localization decisions. An MVLP was created for five main topic articles into 16 languages using DeepL MT engine, accompanied by OpenAI powered AI engine, trained with approved translation memory data and terminology data from previous localization work. The AI engine was further prompted on style, untranslatables, product names, etc. In this case this data was enough to reduce the parity towards human output to minimum. Articles were published in an MVLP form for a month. This provided enough Google Analytics data allowing Native to then execute a Blackbird automation which gathered the Analytics data and then created a report of which articles in which languages generated over 10 000 impressions, over 5000 impressions and over 2000 impressions. Based on the report Native proposed a staggered localization effort with three priority levels, giving our client a chance to invest in localization where it mattered the most.
A surplus solution was later applied to marketing, where MVLP was used in marketing A/B testing with smaller audiences. The previously trained AI model (now supplemented with marketing related data such as ICPs, content pillars, messaging intent, etc.) was used to gauge what ideas resonate more than others in 16 global markets at once. The data was later used for marketing specialists to draw insights from and create campaigns that performed on average 20% better than their previous ones. In addition to that, the insights cost fraction of what localized A/B testing would cost a year before.
III. Experiences, Benefits, and Metrics
Learnings and Benefits
Applying comparatively new LangOps concepts to real time use cases often provide a lot of conclusions and data for further iteration. In this case we learned that in order for MVLP to be as close to human parity, the data used to train the AI matters a lot. Results may warry, but in order to create a sufficient level MVLP, data classification and baseline has to be created to ensure the MVLP is actually useable.
MVLP significantly brings down the cost of A/B testing because the workflow is rather simple. Reduced costs potentially allow designers to be a bit more positively careless in their ideas, inviting more creativity and freedom without worrying about the expenses. Wider A/B testing means better live results.
MVLP invites live data to localization workflows. Current digital landscape relies on applying data as quickly as possible to create impact. MVLP is the first iteration of a localization product that is adjusted and polished with each localization sprint in efforts for the software to become a truly relevant and resonating on the global scale.
I. Problem
Language Service providers and internal language departments often face the challenge that they receive content for translation which is technologically and linguistically unsuitable for translation. In order to improve the quality of the source language, it is important to reach out to upstream processes and win them over as sponsors for a global content delivery process. Also, as stated in the LangOps manifesto, the world is changing from one-way communication paradigm to a conversation or bidirectional flow of information. And of course corporate end customers are demanding AI-driven solutions. LSPs and internal language services will have to cater to these new needs in order to stay relevant.
The challenge has been that traditional TMS and CAT tools are built for experts and not for upstream, non-linguistic stakeholders. Therefore it has always proven difficult to onboard content creators, developers, engineers and the like onto a common platform.
II. Solution
Our LangOps solution combines all the functionality and data access points which corporate end users need in order to interact with “language”. This includes manual and automatic content and translation project creation, like you find in traditional localization portals, of course. But much more than this, it also provides terminology retrieval, management and verification options, machine translation solutions, taxonomies and structure data, systematic translator query management which helps pinpoint content issues, review or quality management features and more. These functionalities are completely customizable to keep the user interface simple and deliver optimal, tailored user experience. That way, onboarding enterprise-wide stakeholders is much easier and faster.
On the back-end, our portal integrates with the traditional TMSs and BMSs, but also authoring tools, content management platforms and proprietary or commercial corporate tools which can consume language data. We make sure all these platforms are kept up to date on the data. By integrating linguistic assets into corporate tools and platforms, we bring their functionality directly to the end users and thus increase the benefits and values customers get out of them.
III. Experiences, Benefits, and Metrics
We believe our platform is a major step towards a true LangOps platform. It gives corporate users exactly the tools and data they require, integrates with all the required upstream and downstream processes and hides the complexities of language technology from those who do not need to be exposed to it directly.
It has made corporate language management much easier to use and spreads the benefit of linguistic assets to a much larger audience in corporate environments. This in turn makes it much easier to obtain budgets and define upstream processes to improve content and communication throughout the entire organization and in all languages.
I. Problem
In today’s globalized world, businesses often encounter significant challenges when it comes to software development and localization. The traditional silos between these two critical processes can lead to inefficiencies, delays, and even errors in the final product. This divide between software development and localization teams has long been a stumbling block for organizations striving for a global reach.
II. Solution
With LangOps—an innovative approach that serves as a natural extension to DevOps, uniting the worlds of software development and localization seamlessly. LangOps empowers organizations to break down the silos between these traditionally often separate domains, fostering collaboration and accelerating the delivery of localized software products.
III. Experiences, Benefits, and Metrics
With LangOps we accomplishes this by integrating localization considerations into the software design and development pipeline from the very beginning. Here’s how it works:
Early Integration: With LangOps, localization isn’t an afterthought; it’s an integral part of the development or even the design process. Designers and developers work alongside localization experts to ensure that internationalization is considered early on. This prevents common localization issues.
Continuous Localization: LangOps encourages continuous integration and continuous localization. As new features and updates are developed, they are simultaneously localized. This ensures that localized versions are always up-to-date and reduces the lag time between development and localization. This can be achieved with our L10n Portal and Services which include nMT and AI to automate steps along the way, leading to a lean and agile end-to-end process.
Automated Workflows: Automation plays a key role in LangOps. Automated testing, quality assurance, and deployment pipelines streamline the localization process, reducing the potential for human error and saving valuable time.
I. Problem
The demand for rapid, precise, and context-aware translations in the language services industry is at an all-time high. Traditional machine translation systems often miss the subtleties of language, requiring extensive post-editing and failing to meet the specific needs of diverse projects and clients. This challenge necessitates a solution that can understand and replicate the nuances of human language, adapt to various styles and tones, and integrate seamlessly into existing translation workflows.
II. Solution
GPT Integration in translate5
translate5’s innovative approach integrates Generative Pre-trained Transformer (GPT) technology as a customizable machine translation engine. This solution enables project managers (PMs) to create bespoke language resources tailored to each project’s unique requirements, leveraging:
Visual Translation Feature
translate5 offers a “What You See Is What You Get” (WYSIWYG) interface, allowing translators to work with the text within the layout for various source file formats, including CMS, Office, InDesign, video subtitling, and Android/iOS apps. This feature ensures translations fit the visual and cultural context of the original document, addressing challenges such as text length and layout compatibility.
Custom Training for GPT
PMs can train GPT with system messages, example data, and terminology, utilizing linguistic resources stored in translate5. This process, similar to onboarding a new translator with a style guide, ensures the AI’s output aligns closely with project expectations.
Collaborative Development and the Open Source Advantage
The successful integration of GPT within translate5 is the result of collaboration between the translate5 team, led by MittagQI, and World Translation. This partnership has facilitated technical development and ensured the solution meets the high standards required by professional translation services. As a third-generation open-source project, translate5 is backed by MittagQI, driving innovation, development, support and maintenance.
III. Experiences, Benefits, and Metrics
Evaluation and Impact
Translating technical documentation for Leica Geosystems from English to German showcased GPT’s capabilities, with its output compared against DeepL. Independent evaluations by experienced translators highlighted GPT’s fluency, idiomatic precision, and alignment with the client’s desired style and tone. Feedback emphasized GPT’s superior handling of style and readability, though noting the need for improvement in translation precision.
This advancement enables PMs to quickly create MT language resources in translate5, customized for each client or project. This transformation requires PMs to possess a deep linguistic understanding, making them prompt engineers who tailor AI output to client expectations, enhancing both efficiency and quality.
Conclusion
The integration of GPT into translate5 marks a significant advancement in translation technology, offering a customizable, efficient, and accurate solution for language service providers. This case study exemplifies the potential of AI and human expertise to meet the translation industry’s evolving demands, setting new benchmarks for quality and innovation. As translate5 continues to explore GPT’s use for various applications, it builds its leadership in leveraging AI to enhance language services.
I. Problem
The demand for rapid, precise, and context-aware translations in the language services industry is at an all-time high. Traditional machine translation systems often miss the subtleties of language, requiring extensive post-editing and failing to meet the specific needs of diverse projects and clients. This challenge necessitates a solution that can understand and replicate the nuances of human language, adapt to various styles and tones, and integrate seamlessly into existing translation workflows.
II. Solution
GPT Integration in translate5
translate5’s innovative approach integrates Generative Pre-trained Transformer (GPT) technology as a customizable machine translation engine. This solution enables project managers (PMs) to create bespoke language resources tailored to each project’s unique requirements, leveraging:
Visual Translation Feature
translate5 offers a “What You See Is What You Get” (WYSIWYG) interface, allowing translators to work with the text within the layout for various source file formats, including CMS, Office, InDesign, video subtitling, and Android/iOS apps. This feature ensures translations fit the visual and cultural context of the original document, addressing challenges such as text length and layout compatibility.
Custom Training for GPT
PMs can train GPT with system messages, example data, and terminology, utilizing linguistic resources stored in translate5. This process, similar to onboarding a new translator with a style guide, ensures the AI’s output aligns closely with project expectations.
Collaborative Development and the Open Source Advantage
The successful integration of GPT within translate5 is the result of collaboration between the translate5 team, led by MittagQI, and World Translation. This partnership has facilitated technical development and ensured the solution meets the high standards required by professional translation services. As a third-generation open-source project, translate5 is backed by MittagQI, driving innovation, development, support and maintenance.
III. Experiences, Benefits, and Metrics
Evaluation and Impact
Translating technical documentation for Leica Geosystems from English to German showcased GPT’s capabilities, with its output compared against DeepL. Independent evaluations by experienced translators highlighted GPT’s fluency, idiomatic precision, and alignment with the client’s desired style and tone. Feedback emphasized GPT’s superior handling of style and readability, though noting the need for improvement in translation precision.
This advancement enables PMs to quickly create MT language resources in translate5, customized for each client or project. This transformation requires PMs to possess a deep linguistic understanding, making them prompt engineers who tailor AI output to client expectations, enhancing both efficiency and quality.
Conclusion
The integration of GPT into translate5 marks a significant advancement in translation technology, offering a customizable, efficient, and accurate solution for language service providers. This case study exemplifies the potential of AI and human expertise to meet the translation industry’s evolving demands, setting new benchmarks for quality and innovation. As translate5 continues to explore GPT’s use for various applications, it builds its leadership in leveraging AI to enhance language services.
I. Problem
Thousands of translation and audio-to-text submissions ran through the localization department with manual processes annually. This was maintained by a significant administrative investment. Also, other departments, such as Editorial and Marketing departments, needed increasing quantities of transcription and translation services.
II. Solution
After a thorough review of available technology, the non-profit decided to design its own component-based language factory, with a set of microservices, tied together with the Blackbird.io workflow orchestrator as its backbone with a recursive microservice architecture.
A folder portal (or “hot folder”) was created on Dropbox in which translation or audio-to-text submissions can be placed. A submission is automatically classified based on file type and file name and automatically assigned to the appropriate semi-automated workflows. The file names are created with a simple-to-use file name builder on a Google sheet and those file names are automatically interpreted through Regex (regular expressions) classifications within Blackbird.io.
As each file submission travels through the workflow, the steps are semi-automatically updated on Slack channels and Trello (Kanban style) through Blackbird.io. Trello has its own automations set up to remove and assign individuals to cards at various steps. These product-based cards are templated in Trello and the workflow automatically copies the correct template based off the filename of the original submission.
Aside from translation submissions, Blackbird also enabled a microservice to be built for audio-to-text by using Transkriptor and OpenAI’s API and its Whisper feature. Through prompting and classifications, this microservice can transcribe audio, add paragraphs, timestamps and speaker diarization.
Aside from a MT-only microservice, another TMS microservice classifies files for TMS into four different domains in order to populate four different Translation Memories and Glossaries.
Other microservices can convert files automatically to use the correct one for each tool and can also be used to convert output files to the desired final file formats.
Once a file submission has triggered a workflow, the workflow is setup to add useful information to the files, such as word counts and it archives file versions to an archive folder automatically.
III. Experiences, Benefits, and Metrics
Manual administrative tasks have reduced by around 40 hours a week.
Various digital tools have been, or are in the process of being, sunsetted and replaced by API enabled equivalents and people outside of the localization department can submit translations and audio-to-text requests with ease.
LangOps staff can easily analyse and adjust each step of the process. 3rd party components of the workflow can easily be switched out or adjusted. Language assets can be used, independent from TMS, for further LangOps workflows and model training.
I. Problem
MakesYouLocal, an e-commerce localisation agency , specialises in helping online businesses thrive in new markets through localised customer service, marketing, and translation solutions. While they were already delivering quality translations, MakesYouLocal recognised an opportunity to enhance their efficiency and scalability to better serve their expanding client base. The goal was to increase productivity, reduce costs, and maintain the high quality their clients expected without compromising on speed or accuracy.
II. Solution
To achieve these objectives, MakesYouLocal partnered with EasyTranslate to implement HumanAI, an advanced technology that merges artificial intelligence with human expertise. This solution was designed to optimise their translation process, making it faster, more cost-effective, and scalable, while preserving the essential human touch that ensures cultural and linguistic accuracy.
Key elements of the solution included:
- AI-Driven Pre-Translation: HumanAI uses sophisticated machine learning algorithms to pre-translate content, significantly reducing the workload for human translators. The AI was trained to align with MakesYouLocal’s specific linguistic preferences and terminologies, ensuring that the initial output was already of high quality.
- Human Oversight and Refinement: After the AI completes the pre-translation, language leads step in to review and perfect the content. This collaboration ensures that the final translations meet their stringent quality standards and reflect the appropriate cultural and brand nuances.
- LangOps Platform: EasyTranslate’s LangOps platform provided an intuitive, collaborative interface where MakesYouLocal’s team could manage, review, and edit translations. This platform enhanced workflow efficiency and enabled seamless communication among team members.
III. Experiences, Benefits, and Metrics
The integration of EasyTranslate’s HumanAI technology into MakesYouLocal’s operations brought substantial improvements, delivering significant benefits in terms of efficiency, cost savings, and translation quality.
1. Increased Efficiency:
– Result: The time required to complete translation projects was dramatically reduced, with projects that previously took 20 hours now being completed in just 2 hours.
-Impact: This tenfold increase in productivity allowed MakesYouLocal to take on more projects, enhance their service offerings, and better meet client demands without adding strain to their resources.
2. Cost Savings:
– Result: Translation costs were reduced by 90%, significantly lowering operational expenses.
– Impact: These savings allowed MakesYouLocal to maintain competitive pricing while increasing profit margins, further strengthening their market position.
3. Maintained High-Quality Standards:
– Result: HumanAI achieved an exceptional accuracy rate of one mistake per 1,000 words, far surpassing traditional benchmarks.
– Impact: This high level of accuracy ensured that MakesYouLocal could continue to deliver consistent, high-quality translations that met their clients’ expectations for brand voice and cultural relevance.
4. Enhanced Scalability:
– Result: With the efficiency gains and reduced costs, MakesYouLocal was able to scale its operations more effectively.
– Impact: They expanded their capacity to handle more projects, allowing them to grow their client base and enhance their competitive edge in the e-commerce localisation market.
Conclusion
The implementation of EasyTranslate’s HumanAI technology was a strategic move that allowed MakesYouLocal to optimise its translation processes, resulting in enhanced efficiency, significant cost savings, and maintained high-quality standards. By leveraging the strengths of both AI and human expertise, MakesYouLocal was able to better serve its clients, expand its business, and solidify its position as a leader in e-commerce localisation.
I. Problem
A large manufacturing company struggled with fragmented technical documentation workflows. Centralizing diverse content formats, translation memory, and machine translation processes into a single data-driven enviroment was challenging. There was a huge in-house effort to reuse siloed technical content in different formats and platforms, send and receive translation packages, import files, and update databases. Each department used different tools for writing, translating, and searching documents, leading to inefficiencies and a lack of organized, centralized data.
II. Solution
Following a successful proof of concept (PoC), the company implemented LOGOSYS, the Logos Multilingual Content Hub. LOGOSYS integrates AI-powered solutions for:
- Content Conversion: Transforms unstructured content into any Dita XML-based CMS format.
- Content Optimization: Uses the myAuthorAssistant app to apply custom terminology, authoring data and writing rules
- Content Generation: Produces coherent and relevant content based on AI patterns and datasets.
- Content Search: Quickly retrieves information using AI-driven search capabilities.
- Translation: Enhances translation quality by combining human edits with Neural Machine Translation (NMT) outputs. This approach refines the accuracy of translations and leverages stored linguistic resources to train both generative AI models and NMT systems. The system continuously monitors and adjusts key parameters, optimizing performance while balancing cost, time, and quality. By adhering to the COTI standard, LOGOSYS ensures efficient collection of work and seamless integration of translated files back into the CMS/NMT translation with human post-editing.
III. Experiences, Benefits, and Metrics
Since implementing LOGOSYS, the hub has efficiently processed millions of words in 30 languages, achieving a 40% reduction in overall costs.
Key Benefits:
- Cost Efficiency: Reduced translation and documentation costs by 40%.
- Increased Productivity: Enabled focus on enhancing source quality, training generative AI models, and managing multilingual knowledge.
- Enhanced Quality: Ensured consistent terminology and writing standards across documents.
- Faster Retrieval: Improved response times for technical support with efficient content search.
- Valuable Data: Built a comprehensive data repository for NLP tasks and custom large language model training, supporting broader AI initiatives.
I. Problem
Many organizations are rushing to adopt new technologies and tools based on what their competitors have announced, without evaluating if these solutions fit their specific needs. This leads to wasted resources, frustrated teams, and inefficient processes. A common scenario involves businesses investing in AI-driven platforms and automation technologies that either don’t integrate with their internal tools or fail to meet the unique requirements of their workflows. As a result, companies end up with expensive solutions that don’t deliver the expected efficiency gains.
This issue is further complicated in the context of multilingual operations, where linguistic quality, cultural adaptation, and functional integration must be considered. Failing to accurately assess these aspects early in the customer journey, especially during discovery calls and pre-sales discussions, results in a misalignment between customer expectations and the solutions provided. The challenge is not only technical but also communicative—how do we ensure that customers understand what they need versus what they think they want based on market trends?
II. Solution
Achieving true efficiency gains starts long before the implementation phase—it begins with communication. A comprehensive, structured assessment of each customer’s specific needs and existing ecosystem is critical. This includes:
1. Deep Dive Discovery Sessions: Engaging with the customer to map out their current tools, workflows, and business goals. By involving both technical and linguistic experts from the very beginning, we can assess not only the technology fit but also the linguistic accuracy required for multilingual applications.
2. Tailored Solution Architecture: Rather than pushing a one-size-fits-all approach, we co-create a solution blueprint that aligns with the customer’s existing systems and anticipates future growth. This involves a thorough evaluation of AI integration, automation capabilities, and linguistic quality management. The aim is to ensure that any technology investment results in streamlined workflows, not additional complexity.
3. Pre-Implementation Simulations: Before finalizing the tech stack, we conduct simulations using the customer’s real data and scenarios. This helps in identifying potential bottlenecks and ensures that the selected tools can seamlessly handle the workload, maintaining the desired levels of quality and accuracy in multilingual content.
4. Transparent Metrics and Feedback Loops: Establishing clear KPIs from the outset ensures that both parties can measure success. Regular feedback loops during pre-sales allow for adjustments in the solution design, making sure that the customer’s expectations are aligned with achievable outcomes.
III. Experiences, Benefits, and Metrics
Customers who underwent this thorough pre-sales and pre-implementation assessment have experienced significant efficiency improvements, often surpassing their initial expectations. For example:
One global financial client, initially fixated on implementing a high-profile Machine translation tool, discovered through our assessment that their internal systems were not optimized for such a solution. By adjusting the strategy to fit their existing tools and focusing on key integrations, they saved 25% in implementation costs and improved translation turnaround times by 40%.
Another client, a multinational pharmaceutical company, faced issues with linguistic inconsistencies in their multilingual documentation. By involving linguistic experts early in the discovery phase, we ensured that the AI tools chosen for the project were tailored to their specific domain terminology and regional language variations. As a result, translation quality improved by 30%, and post-editing efforts were reduced by half.
Overall, customers report greater confidence in their technology investments, as the solutions are clearly aligned with their operational goals. More importantly, by establishing communication as the foundation of the process, we’ve eliminated the guesswork that often leads to project delays and budget overruns.
Conclusion
Efficiency gains in technology adoption, particularly in language operations, don’t happen by accident. They are the result of deep customer understanding, transparent communication, and a tailored approach to solution design. By focusing on these aspects during the pre-sales and pre-implementation phases, companies can avoid costly mistakes, unlock the full potential of their tech investments, and ultimately drive better outcomes. This proactive, communicative approach is not just a best practice—it’s essential for long-term success.
I. Problem
An organisation with over 2500 language professionals, team leads, and managers was responsible for curating and localising millions of data points every month, the content produced was to be used in global products and services serving millions of users on daily basis.
Written documentation was the source of truth for the whole organisation, which included curation guidelines, style guides, language guides and SOPs that everyone followed for their daily operations, but as the products expanded in scale and scope, it became apparent that the existing decentralised structure of documentation was constantly challenging consistency, recency and usability across the organisation.
Products were localised into more than 50 languages, each with their own unique requirements that needed to be considered at all stages of the product’s life cycle. Such requirements were hosted sporadically across documents with various levels of access and ownership, leading to the aforementioned issues of consistency and accuracy.
Due to the distributed nature of the workforce which came from over 70 different countries, English was chosen as the official language for all communication and documentation, so a set of standards and quality measures were needed to always ensure clarity of instructions and purpose, and accessibility for all levels of English proficiency.
The vast number of cultures in the project also meant that guidance and training on cultural sensitivities was also needed for a cohesive and harmonious work environment that encouraged respectful collaboration.
II. Solution
A central platform to host all documentation that adhered to strict privacy and security requirements and provided workers of all levels with the required up-to-date documentation for their products and locales.
The new platform design started by identifying the needs of the different teams, and their existing documentation structure, which provided great insight on possible structures that could be applied to the design of the new system.
The new platform needed to meet the following requirements:
- Multilayered access levels for the different users of the platform.
- A clear and easy to navigate structure that enables users to find the information they need with ease.
- Easy editing and formatting tool to enable documentation creators and users to create and consume appealing and informative documents.
- A robust infrastructure with near 100% uptime due to the highly distributed nature of the organisation.
Due to the products nature and the high level of language integration, it was crucial to accommodate the specific needs of each language team in the requirement design, and how it fits within the overall structure of the knowledge management system. Language guidance was not only for linguists, but for anyone creating the product to use as a reference of what may be needed to accommodate specific language needs.
The platform was built based on the specified criteria, and work was ongoing to sift through the massive volume of documentation (around 7000 unique documents pre-migration) to evaluate the quality of the existing material and avoid any of its shortcomings.
Redundant and obsolete knowledge was identified and isolated, general level instructions were consolidated into general-purpose documentation, style guides and templates were standardised to ensure consistency of source language, language styles and terminology were reviewed, updated and publicised for the whole organisation, training was provided to all stakeholders on the new platform and pilot programs were conducted to ensure stability and proper functionality before the full scale migration.
III. Experiences, Benefits, and Metrics
Before the start of the migration, it was crucial to inform and educate all stakeholders of the upcoming change and addressing any concerns they have about the process. Feedback was collected and acted upon, generating new requirements to be implemented in the new platform.
The involvement and support from stakeholders at all levels ensured that the migration was adopted, rather than forced, which led to the successful and seamless migration of more than 5000 documents (after the culling of obsolete and redundant material), 80% of which were completed in the first week of the three weeks needed to fully migrate.
After the migration was completed, a significant improvement in product quality was noticed (over 20% according to the internal quality scores), which included higher quality localisation due to the integration of language needs throughout the product lifecycle.
The centralisation of knowledge across these many teams enabled the organisation to have consistent and clear messaging that was not possible before, which reflected in the positive reception of the new products that were released after the migration.