Databricks spent $10M on new DBRX generative AI model, but it can’t beat GPT-4
If you wanted to raise the profile of your major tech company and had $10 million to spend, how would you spend it? On a Super Bowl ad? An F1 sponsorship?
You could spend it training a generative AI model. While not marketing in the traditional sense, generative models are attention grabbers â and increasingly funnels to vendorsâ bread-and-butter products and services.
See Databricksâ DBRX, a new generative AI model announced today akin to OpenAIâs GPT series and Googleâs Gemini. Available on GitHub and the AI dev platform Hugging Face for research as well as for commercial use, base (DBRX Base) and fine-tuned (DBRX Instruct) versions of DBRX can be run and tuned on public, custom or otherwise proprietary data.
âDBRX was trained to be useful and provide information on a wide variety of topics,â Naveen Rao, VP of generative AI at Databricks, told TechCrunch in an interview. âDBRX has been optimized and tuned for English language usage, but is capable of conversing and translating into a wide variety of languages, such as French, Spanish and German.â
Databricks describes DBRX as âopen sourceâ in a similar vein as âopen sourceâ models like Metaâs Llama 2 and AI startup Mistralâs models. (Itâs the subject of robust debate as to whether these models truly meet the definition of open source.)
Databricks says that it spent roughly $10 million and eight months training DBRX, which it claims (quoting from a press release) âoutperform[s] all existing open source models on standard benchmarks.â
But â and hereâs the marketing rub â itâs exceptionally hard to use DBRX unless youâre a Databricks customer.
Thatâs because, in order to run DBRX in the standard configuration, you need a server or PC with at least four Nvidia H100 GPUs. A single H100 costs thousands of dollars â quite possibly more. That might be chump change to the average enterprise, but for many developers and solopreneurs, itâs well beyond reach.
And thereâs fine print to boot. Databricks says that companies with more than 700 million active users will face âcertain restrictionsâ comparable to Metaâs for Llama 2, and that all users will have to agree to terms ensuring that they use DBRX âresponsibly.â (Databricks hadnât volunteered those termsâ specifics as of publication time.)
Databricks presents its Mosaic AI Foundation Model product as the managed solution to these roadblocks, which in addition to running DBRX and other models provides a training stack for fine-tuning DBRX on custom data. Customers can privately host DBRX using Databricksâ Model Serving offering, Rao suggested, or they can work with Databricks to deploy DBRX on the hardware of their choosing.
Rao added:
Weâre focused on making the Databricks platform the best choice for customized model building, so ultimately the benefit to Databricks is more users on our platform. DBRX is a demonstration of our best-in-class pre-training and tuning platform, which customers can use to build their own models from scratch. Itâs an easy way for customers to get started with the Databricks Mosaic AI generative AI tools. And DBRX is highly capable out-of-the-box and can be tuned for excellent performance on specific tasks at better economics than large, closed models.
Databricks claims DBRX runs up to 2x faster than Llama 2, in part thanks to its mixture of experts (MoE) architecture. MoE â which DBRX shares in common with Llama 2, Mistralâs newer models, and Googleâs recently announced Gemini 1.5 Pro â basically breaks down data processing tasks into multiple subtasks and then delegates these subtasks to smaller, specialized âexpertâ models.
Most MoE models have eight experts. DBRX has 16, which Databricks says improves quality.
Quality is relative, however.
While Databricks claims that DBRX outperforms Llama 2 and Mistralâs models on certain language understanding, programming, math and logic benchmarks, DBRX falls short of arguably the leading generative AI model, OpenAIâs GPT-4, in most areas outside of niche use cases like database programming language generation.
Rao admits that DBRX has other limitations as well, namely that it â like all other generative AI models â can fall victim to âhallucinatingâ answers to queries despite Databricksâ work in safety testing and red teaming. Because the model was simply trained to associate words or phrases with certain concepts, if those associations arenât totally accurate, its responses wonât always accurate.
Also, DBRX is not multimodal, unlike some more recent flagship generative AI models including Gemini. (It can only process and generate text, not images.) And we donât know exactly what sources of data were used to train it; Rao would only reveal that no Databricks customer data was used in training DBRX.
âWe trained DBRX on a large set of data from a diverse range of sources,â he added. âWe used open data sets that the community knows, loves and uses every day.â
I asked Rao if any of the DBRX training data sets were copyrighted or licensed, or show obvious signs of biases (e.g. racial biases), but he didnât answer directly, saying only, âWeâve been careful about the data used, and conducted red teaming exercises to improve the modelâs weaknesses.â Generative AI models have a tendency to regurgitate training data, an major concern for commercial users of models trained on unlicensed, copyrighted or very clearly biased data. In the worst-case scenario, a user could end up on the ethical and legal hooks for unwittingly incorporating IP-infringing or biased work from a model into their projects.
Some companies training and releasing generative AI models offer policies covering the legal fees arising from possible infringement. Databricks doesnât at present â Rao says that the companyâs âexploring scenariosâ under which it might.
Given this and the other aspects in which DBRX misses the mark, the model seems like a tough sell to anyone but current or would-be Databricks customers. Databricksâ rivals in generative AI, including OpenAI, offer equally if not more compelling technologies at very competitive pricing. And plenty of generative AI models come closer to the commonly understood definition of open source than DBRX.
Rao promises that Databricks will continue to refine DBRX and release new versions as the companyâs Mosaic Labs R&D team â the team behind DBRX â investigates new generative AI avenues.
âDBRX is pushing the open source model space forward and challenging future models to be built even more efficiently,â he said. âWeâll be releasing variants as we apply techniques to improve output quality in terms of reliability, safety and bias ⊠We see the open model as a platform on which our customers can build custom capabilities with our tools.â
Judging by where DBRX now stands relative to its peers, itâs an exceptionally long road ahead.