Unbabel among first AI startups to win millions of GPU training hours on EU supercomputers
The European Union has announced the winners of a âLarge AI Grand Challengeâ it kicked off earlier this year in a bid to accelerate the pace of homegrown innovation by large-scale AI model makers.
Four startups will share âŹ1 million in prize money and â perhaps more importantly â eight million GPU hours to train their models on a couple of the blocâs high performance computing (HPC) supercomputers over the next 12 months. The Commission reckons this will enable them to shrink model training times âfrom years to weeksâ, as its PR puts it.
The winning four startups are â in alphabetical order â : French fintech Lingua Custodia, which does financial document processing using natural language processing (NLP); Belgian startup Textgain, which also uses NLP for text processing but focuses on analysis of unstructured data, such as monitoring social media chatter for hate speech; Latvian startup Tilde, another language specialist thatâs focuses on Balto-Slavic languages â offering machine translation and AI-powered chatbots in the target tongues; and Portugalâs Unbabel, which has historically blended machine translation with the expertise of native human speakers â applying AI for customer service and productivity use-cases for enterprise customers.
The Commission said the AI Challenge received a total of 94 proposals.
Unbabel likely has the highest profile of the four winners. The Y Combinator-backed translation business has been around for the best part of a decade and raised close to $100M over its run, per Crunchbase.
Whether Unbabel needs an extra quarter million euros or even 2 million freebie GPU training hours is up for debate â but even veteran AI startups may feel every little helps given the fast-paced developments in generative AI over the past 1.5 years or so.
At the end of the training period, the EU expects all the winners to release their developed models under an open-source license for non-commercial use or publish their research findings.Â
EU supercomputers to support AI startups
The EU unveiled a plan to expand startup access to the blocâs supercomputing hardware in president Ursula von der Leyenâs state of the union address last fall â saying at the time that it wanted âethical and responsible AI startupâ to be first in line to tap computational support.
The European High Performance Computing Joint Undertaking (aka EuroHPC JU) â to give the blocâs supercomputer initiative its full name â currently has eight operational (nine procured) supercomputers â two of which will be providing the allocation of eight million GPU hours to the four winners: Namely, Finland-based Lumi and Italy-based Leonardo (which are both pre-exascale HPC supercomputers).
A fifth startup â Spain-based Multiverse Computing, which is focusing on trying to improve the energy efficiency and speed of large language models using âquantum-inspired tensor networksâ â just missed out on any prize money but thereâs a consolation: It will be allocated 800,000 computational hours on another of the supercomputers, Spainâs (pre-exascale) MareNostrum 5.
This handful of European startups building large scale AI models wonât be the first to get a taste of what HPC hardware can do. French general purpose AI model maker Mistral was a participant in an early pilot phase of the supercomputing provision last summer, using Leonardo to ârun a few small experimentsâ, as co-founder and CEO Arthur Mensch told TechCrunch back in December â though he said it had not been used for model training at that point.
The EuroHPC JU has also historically provided some capacity to commercial players. However demand for the supercomputers typically far outstrips supply, so the AI startups are essentially getting bumped to the front of the queue.
EU policymakers have also recognized thereâs a need to reconfigure and retool the HPC infrastructure for the generative AI age. Hence why, back in January, the Commission announced a package of âAI innovationâ measures that included proposals for upgrading the supercomputers and building out a support layer to improve accessibility so that AI startups can more easily tap the infrastructure.