Meta’s new AI brings us one step closer to a universal language translator

    Meta has taken another step towards creating a universal language translator.

    The company has an open source AI model that translates more than 200 languages, many of which are not supported by existing systems.

    The research is part of a Meta initiative launched earlier this year.

    Greetings, Humanoids

    Sign up for our newsletter now for a weekly roundup of our favorite AI stories delivered to your inbox.

    “We call this project No language left behindand the AI ​​modeling techniques we’ve used from NLLB help us create high-quality translations on Facebook and Instagram for languages ​​spoken by billions of people around the world,” Meta CEO Mark Zuckerberg said in a Facebook post.

    NLLB focuses on languages ​​with fewer resources, such as Maori or Maltese. Most people in the world speak these languages, but they lack the training data that AI translations typically need.

    Meta’s new model is designed to overcome this challenge.

    To do this, the researchers first interviewed speakers of disadvantaged languages ​​to understand their needs. They then developed a new data mining technique that generates training sentences for: languages ​​with few resources.

    They then trained their model on a mix of the mined data and human-translated data.

    The result is NLLB-200 — a massive multilingual translation system for 202 languages.

    The team assessed the model’s performance against the FLORES-101 dataset, which evaluates translations of low-resource languages.

    “Despite doubling the number of languages, our final model performs 40% better than the prior art on Flores-101,” the study authors wrote:

    SOTA comparison