Groq is Fast AI Inference

The LPU™ Inference Engine by Groq is a hardware and software platform that delivers exceptional compute speed, quality, and energy efficiency. Groq provides cloud and on-prem solutions at scale for AI applications.

Introduction

What is Groq?

Groq is a fast AI inference platform that provides instant intelligence for openly-available models like Llama 3.1. It allows developers to move seamlessly from other providers like OpenAI by changing just three lines of code.

Features of Groq
  • Fast AI inference for openly-available models
  • OpenAI endpoint compatibility
  • Instant intelligence with Groq Speed
  • Powers leading openly-available AI models like Llama, Mixtral, Gemma, and Whisper
  • Backed by Yann LeCun, VP & Chief AI Scientist at Meta, as a technical advisor
How to Use Groq
  • Get a free API key and start building
  • Move from OpenAI to Groq by changing three lines of code
  • Set the base URL and choose your model to run
  • Explore Groq benchmarks to see the instant speed
Pricing of Groq
  • Get a free API key and start building
  • Explore pricing plans for enterprise access and GroqCloud™ platform
  • Learn more about GroqRack™ cluster and its pricing
Helpful Tips for Using Groq
  • Explore Groq benchmarks to see the instant speed
  • Learn more about Groq's OpenAI endpoint compatibility
  • Get a free API key and start building
Frequently Asked Questions about Groq
  • What is Groq?
    • Groq is a fast AI inference platform that provides instant intelligence for openly-available models.
  • How do I use Groq?
    • Get a free API key and start building, move from OpenAI to Groq by changing three lines of code, and explore Groq benchmarks.
  • What are the features of Groq?
    • Fast AI inference, OpenAI endpoint compatibility, instant intelligence, and powers leading openly-available AI models.

Information

Newsletter

Join the Community

Subscribe to our newsletter for the latest news and updates