While everyone is competing to replace Nvidia, Groq appears to be getting close.
Groq, a leader in fast AI inference, has achieved a staggering $2.8 billion valuation following a successful $640 million Series D funding round. This milestone not only underscores Groq’s rapid ascent in the tech world but also highlights the growing importance of AI chip technology in advancing artificial intelligence systems.
The funding round was led by funds and accounts managed by BlackRock Private Equity Partners, with participation from both existing and new investors. Notable additions include Neuberger Berman, Type One Ventures, and strategic investors such as Cisco Investments, Global Brain’s KDDI Open Innovation Fund III, and Samsung Catalyst Fund.
Samir Menon, Managing Director at BlackRock Private Equity Partners, expressed enthusiasm about the investment, stating, “The market for AI compute is meaningful and Groq’s vertically integrated solution is well positioned to meet this opportunity. We look forward to supporting Groq as they scale to meet demand and accelerate their innovation further.”
Thank you to our partners at @BlackRock Private Equity Partners for their support and leadership through our Series D funding. We’re excited to continue delivering instant AI inference for the world to use to innovate. pic.twitter.com/HSu94ZgNDj
— Groq Inc (@GroqInc) August 5, 2024
Groq specializes in developing cutting-edge semiconductors and software designed to achieve optimal AI performance. The company’s unique, vertically integrated AI inference platform has generated skyrocketing demand from developers seeking exceptional speed. This technology addresses a critical challenge in the AI industry: the growing bottleneck in computing power, especially as AI applications continue to proliferate across various sectors.
Groq’s LPU™ AI inference technology is designed with a software-first approach to meet the unique characteristics and needs of AI. This innovative architecture has given Groq an advantage in bringing new models to developers quickly and at instant speed. The recent investment will enable the company to accelerate the development of the next two generations of LPU, further cementing its position as a leader in AI inference technology. Morgan Stanley & Co. LLC served as the exclusive Placement Agent to Groq on this landmark transaction, further underscoring the significance of this funding round in the AI industry. Marco Chisari, Head of Samsung Semiconductor Innovation Center and EVP of Samsung Electronics, highlighted Groq’s technological prowess: “We are highly impressed by Groq’s disruptive compute architecture and their software-first approach. Groq’s record-breaking speed and near-instant Generative AI inference performance leads the market.”
Jonathan Ross, CEO and Founder of Groq, emphasized the crucial role of inference compute in AI development: “You can’t power AI without inference compute. We intend to make the resources available so that anyone can create cutting-edge AI products, not just the largest tech companies.” Ross also revealed ambitious plans for the newly secured funds, stating, “This funding will enable us to deploy more than 100,000 additional LPUs into GroqCloud. Training AI models is solved, now it’s time to deploy these models so the world can use them.”
The company’s success has attracted top talent to its leadership team. Stuart Pann, formerly a senior executive from HP and Intel, has joined Groq as Chief Operating Officer. Pann expressed his excitement about joining the company at this pivotal moment: “We have the technology, the talent, and the market position to rapidly scale our capacity and deliver inference deployment economics for developers as well as for Groq.”
Additionally, Groq has gained world-class expertise with its newest technical advisor, Yann LeCun, VP & Chief AI Scientist at Meta who said, “The Groq chip really goes for the jugular,”.
“The Groq chip really goes for the jugular,” says Meta chief scientist Yann LeCun, Ross’s former computer science professor at NYU who recently joined Groq as a technical advisor. @Forbes @richardjnievahttps://t.co/USYY1P3gWI
— Groq Inc (@GroqInc) August 5, 2024
Groq’s impact on the developer community has been remarkable, with over 360,000 developers now building on GroqCloud™. These developers are creating AI applications using openly-available models such as Llama 3.1 from Meta, Whisper Large V3 from OpenAI, Gemma from Google, and Mixtral from Mistral. The company plans to use the new funding to scale the capacity of its tokens-as-a-service (TaaS) offering and add new models and features to GroqCloud.
Mark Zuckerberg, CEO and Founder of Meta, recently acknowledged Groq’s contributions in his letter entitled “Open Source AI Is the Path Forward,” stating, “Innovators like Groq have built low-latency, low-cost inference serving for all the new models.”
To meet the surging demand from developers and enterprises, Groq has announced plans to deploy over 108,000 LPUs manufactured by GlobalFoundries by the end of Q1 2025. This will be the largest AI inference compute deployment of any non-hyperscaler, positioning Groq as a key player in the AI infrastructure landscape.
The company is also expanding its global reach through partnerships with enterprises like Aramco Digital and Earth Wind & Power to build out AI compute centers worldwide. Tareq Amin, Chief Executive Officer of Aramco Digital, highlighted the significance of their collaboration: “Aramco Digital is partnering with Groq to build one of the largest AI Inference-as-a-Service compute infrastructure in the MENA region. Our close collaboration with Groq is transformational for both domestic and global AI demand.”
Groq’s advances are met with scepticism nonetheless. While praising Groq’s strategy as “novel,” a venture capitalist who chose not to participate in the company’s Series D financing expressed doubts about the intellectual property’s long-term defensibility. Furthermore, when scaled, several critics doubt Groq’s chips’ affordability.
Groq has also powered the development of Hero, the first daily assistant designed to simplify life. Hero allows users to manage everything in one place, eliminating the need to switch between multiple apps. Tasks that used to require manual input can now be handled with simple voice commands, and instead of lengthy text exchanges, coordination happens instantly. Hero was created to help busy individuals, couples, and parents streamline their lives. After extensive discussions with users, it became clear that a new approach was needed—one that does away with checking multiple apps, multi-step processes, and constant back-and-forth communication.
Check out this new app powered by Groq! 🤩 https://t.co/7YXL3OvUH9
— Groq Inc (@GroqInc) August 5, 2024
Jonathan Ross had planned to raise $300M which was going to allow them to deploy 108,000 LPUs into production by the end of Q1 2025. But now they are also expanding their cloud and core engineering teams and are hiring.
What are we doing with this capital? Originally we intended to raise $300M which was going to allow us to deploy 108,000 LPUs into production by end of Q1 2025. We raised 2x that, so we're also expanding our cloud and core engineering teams.
— Jonathan Ross (@JonathanRoss321) August 5, 2024
We're hiring! https://t.co/rnk6PGzfXy
While we feel they are the new Nvidia, Ross says, “It’s sort of like we’re Rookie of the Year,”. “We’re nowhere near Nvidia yet. So all eyes are on us. And it’s like, what are you going to do next?”