French startup FlexAI exits stealth with $30M to ease access to AI compute
A French startup has raised a hefty seed investment to ârearchitect compute infrastructureâ for developers wanting to build and train AI applications more efficiently.
FlexAI, as the company is called, has been operating in stealth since October 2023, but the Paris-based company is formally launching Wednesday with âŹ28.5 million ($30 million) in funding, while teasing its first product: an on-demand cloud service for AI training.
This is a chunky bit of change for a seed round, which normally means real substantial founder pedigree â and that is the case here. FlexAI co-founder and CEO Brijesh Tripathi was previously a senior design engineer at GPU giant and now AI darling Nvidia, before landing in various senior engineering and architecting roles at Apple; Tesla (working directly under Elon Musk); Zoox (before Amazon acquired the autonomous driving startup); and, most recently, Tripathi was VP of Intelâs AI and super compute platform offshoot, AXG.
FlexAI co-founder and CTO Dali Kilani has an impressive CV, too, serving in various technical roles at companies including Nvidia and Zynga, while most recently filling the CTO role at French startup Lifen, which develops digital infrastructure for the healthcare industry.
The seed round was led by Alpha Intelligence Capital (AIC), Elaia Partners and Heartcore Capital, with participation from Frst Capital, Motier Ventures, Partech and InstaDeep CEO Karim Beguir.
The compute conundrum
To grasp what Tripathi and Kilani are attempting with FlexAI, itâs first worth understanding what developers and AI practitioners are up against in terms of accessing âcomputeâ; this refers to the processing power, infrastructure and resources needed to carry out computational tasks such as processing data, running algorithms, and executing machine learning models.
âUsing any infrastructure in the AI space is complex; itâs not for the faint-of-heart, and itâs not for the inexperienced,â Tripathi told TechCrunch. âIt requires you to know too much about how to build infrastructure before you can use it.â
By contrast, the public cloud ecosystem that has evolved these past couple of decades serves as a fine example of how an industry has emerged from developersâ need to build applications without worrying too much about the back end.
âIf you are a small developer and want to write an application, you donât need to know where itâs being run, or what the back end is â you just need to spin up an EC2 (Amazon Elastic Compute cloud) instance and youâre done,â Tripathi said. âYou canât do that with AI compute today.â
In the AI sphere, developers must figure out how many GPUs (graphics processing units) they need to interconnect over what type of network, managed through a software ecosystem that they are entirely responsible for setting up. If a GPU or network fails, or if anything in that chain goes awry, the onus is on the developer to sort it.
âWe want to bring AI compute infrastructure to the same level of simplicity that the general purpose cloud has gotten to â after 20 years, yes, but there is no reason why AI compute canât see the same benefits,â Tripathi said. âWe want to get to a point where running AI workloads doesnât require you to become data centre experts.â
With the current iteration of its product going through its paces with a handful of beta customers, FlexAI will launch its first commercial product later this year. Itâs basically a cloud service that connects developers to âvirtual heterogeneous compute,â meaning that they can run their workloads and deploy AI models across multiple architectures, paying on a usage basis rather than renting GPUs on a dollars-per-hour basis.
GPUs are vital cogs in AI development, serving to train and run large language models (LLMs), for example. Nvidia is one of the preeminent players in the GPU space, and one of the main beneficiaries of the AI revolution sparked by OpenAI and ChatGPT. In the 12 months since OpenAI launched an API for ChatGPT in March 2023, allowing developers to bake ChatGPT functionality into their own apps, Nvidiaâs shares ballooned from around $500 billion to more than $2 trillion.
LLMs are pouring out of the technology industry, with demand for GPUs skyrocketing in tandem. But GPUs are expensive to run, and renting them from a cloud provider for smaller jobs or ad-hoc use-cases doesnât always make sense and can be prohibitively expensive; this is why AWS has been dabbling with time-limited rentals for smaller AI projects. But renting is still renting, which is why FlexAI wants to abstract away the underlying complexities and let customers access AI compute on an as-needed basis.
âMulticloud for AIâ
FlexAIâs starting point is that most developers donât really care for the most part whose GPUs or chips they use, whether itâs Nvidia, AMD, Intel, Graphcore or Cerebras. Their main concern is being able to develop their AI and build applications within their budgetary constraints.
This is where FlexAIâs concept of âuniversal AI computeâ comes in, where FlexAI takes the userâs requirements and allocates it to whatever architecture makes sense for that particular job, taking care of the all the necessary conversions across the different platforms, whether thatâs Intelâs Gaudi infrastructure, AMDâs Rocm or Nvidiaâs CUDA.
âWhat this means is that the developer is only focused on building, training and using models,â Tripathi said. âWe take care of everything underneath. The failures, recovery, reliability, are all managed by us, and you pay for what you use.â
In many ways, FlexAI is setting out to fast-track for AI what has already been happening in the cloud, meaning more than replicating the pay-per-usage model: It means the ability to go âmulticloudâ by leaning on the different benefits of different GPU and chip infrastructures.
For example, FlexAI will channel a customerâs specific workload depending on what their priorities are. If a company has limited budget for training and fine-tuning their AI models, they can set that within the FlexAI platform to get the maximum amount of compute bang for their buck. This might mean going through Intel for cheaper (but slower) compute, but if a developer has a small run that requires the fastest possible output, then it can be channeled through Nvidia instead.
Under the hood, FlexAI is basically an âaggregator of demand,â renting the hardware itself through traditional means and, using its âstrong connectionsâ with the folks at Intel and AMD, secures preferential prices that it spreads across its own customer base. This doesnât necessarily mean side-stepping the kingpin Nvidia, but it possibly does mean that to a large extent â with Intel and AMD fighting for GPU scraps left in Nvidiaâs wake â there is a huge incentive for them to play ball with aggregators such as FlexAI.
âIf I can make it work for customers and bring tens to hundreds of customers onto their infrastructure, they [Intel and AMD] will be very happy,â Tripathi said.
This sits in contrast to similar GPU cloud players in the space such as the well-funded CoreWeave and Lambda Labs, which are focused squarely on Nvidia hardware.
âI want to get AI compute to the point where the current general purpose cloud computing is,â Tripathi noted. âYou canât do multicloud on AI. You have to select specific hardware, number of GPUs, infrastructure, connectivity, and then maintain it yourself. Today, thatâs thatâs the only way to actually get AI compute.â
When asked who the exact launch partners are, Tripathi said that he was unable to name all of them due to a lack of âformal commitmentsâ from some of them.
âIntel is a strong partner, they are definitely providing infrastructure, and AMD is a partner thatâs providing infrastructure,â he said. âBut there is a second layer of partnerships that are happening with Nvidia and a couple of other silicon companies that we are not yet ready to share, but they are all in the mix and MOUs [memorandums of understanding] are being signed right now.â
The Elon effect
Tripathi is more than equipped to deal with the challenges ahead, having worked in some of the worldâs largest tech companies.
âI know enough about GPUs; I used to build GPUs,â Tripathi said of his seven-year stint at Nvidia, ending in 2007 when he jumped ship for Apple as it was launching the first iPhone. âAt Apple, I became focused on solving real customer problems. I was there when Apple started building their first SoCs [system on chips] for phones.â
Tripathi also spent two years at Tesla from 2016 to 2018 as hardware engineering lead, where he ended up working directly under Elon Musk for his last six months after two people above him abruptly left the company.
âAt Tesla, the thing that I learned and Iâm taking into my startup is that there are no constraints other than science and physics,â he said. âHow things are done today is not how it should be or needs to be done. You should go after what the right thing to do is from first principles, and to do that, remove every black box.â
Tripathi was involved in Teslaâs transition to making its own chips, a move that has since been emulated by GM and Hyundai, among other automakers.
âOne of the first things I did at Tesla was to figure out how many microcontrollers there are in a car, and to do that, we literally had to sort through a bunch of those big black boxes with metal shielding and casing around it, to find these really tiny small microcontrollers in there,â Tripathi said. âAnd we ended up putting that on a table, laid it out and said, âElon, there are 50 microcontrollers in a car. And we pay sometimes 1,000 times margins on them because they are shielded and protected in a big metal casing.â And heâs like, âletâs go make our own.â And we did that.â
GPUs as collateral
Looking further into the future, FlexAI has aspirations to build out its own infrastructure, too, including data centers. This, Tripathi said, will be funded by debt financing, building on a recent trend that has seen rivals in the space including CoreWeave and Lambda Labs use Nvidia chips as collateral to secure loans â rather than giving more equity away.
âBankers now know how to use GPUs as collaterals,â Tripathi said. âWhy give away equity? Until we become a real compute provider, our companyâs value is not enough to get us the hundreds of millions of dollars needed to invest in building data centres. If we did only equity, we disappear when the money is gone. But if we actually bank it on GPUs as collateral, they can take the GPUs away and put it in some other data center.â