Follow Datanami:
October 26, 2023

Groq Unveils New Low Latency LPU System at SC23, Paving the Way for Future AI Applications

MOUNTAIN VIEW, Calif.Oct. 26, 2023 — Groq, an artificial intelligence (AI) solutions company, announced today that it will have a booth and multiple talks at the premier industry conference for high performance compute, SC23, from November 12-17 in Denver, CO. Groq and their team will be showcasing a demo of the world’s best low latency performance for Large Language Models (LLMs) running on a Language Processing Unit system, its next-gen AI accelerator. Subject matter experts from Groq will be presenting four sessions during the conference on a range of HPC, AI, and research-related topics.

Groq will be showcasing a demo of the world’s best low latency performance for LLMs running on a Language Processing Unit system, its next-gen AI accelerator.

Jim Miller, VP of Engineering at Groq, and former engineering leader at Qualcomm, Broadcom, and Intel, shared, “The scale and performance of systems used for AI today is enormous, and will get larger if built with legacy technology. At Groq we are setting a new standard with our LPU-based systems that improve performance, power, and scale when serving a large customer base. This is thanks to the hard work and innovative ideas of our dedicated team of engineers at Groq who are committed to solving truly novel problems.”

The LPU accelerator is the Groq response to the next level of processing power required by enterprise-scale AI applications. With a clear market need for a purpose-built and software-driven processor, the Groq LPU accelerator will power LLMs for the exploding GenAI market.

Yaniv Shemesh, Head of Cloud & HPC Software Engineering at Groq, said, “Groq’s groundbreaking speed, in the form of tokens-as-a-service, was a major milestone for my organization and the company. Running your own hardware and building a large scale HPC can be hard, but Groq’s token-as-a-service ease of use and consumption-based model are very attractive to customers. Our performance is beyond fast and is opening new possibilities and innovative customer use-cases previously unimaginable given existing market solutions limitations.”

To date, the company has showcased record-breaking performance of the open source foundational LLM, Llama-2 70B by Meta AI, now running generated language at over 280 tokens per second per user. Groq also recently deployed Falcon, a powerful language model available for both research and commercial use that’s currently at the top of the Hugging Face Leaderboard for pre-trained open source LLMs, and Code Llama, one of the newest LLMs from Meta AI helping users generate code.

Attend the show and you can meet Groq and interact with its technology in the following forums:

  • Visit the booth, #1681, to see a live demo of Groq running LLMs on its LPU system,
  • Schedule a time to talk with Groq specialists in their VIP Lounge by reaching out to [email protected].

About Groq

Groq is an AI solutions company and the inventor of the Language Processing Unit accelerator that is purpose-built and software-driven to power Large Language Models (LLMs) for the exploding AI market. For more information, visit

Source: Groq