Groq Chat
Start new thread
Chris Messina

1yr ago

Groq® - Hyperfast LLM running on custom built GPUs

An LPU Inference Engine, with LPU standing for Language Processing Unit™, is a new type of end-to-end processing unit system that provides the fastest inference at ~500 tokens/second.
J T

1yr ago

Groq Chat - World's fastest Large Language Model (LLM)

This alpha demo lets you experience ultra-low latency performance using the foundational LLM, Llama 2 70B (created by Meta AI), running on the Groq LPU™ Inference Engine.