Ultra-fast LLM inference β LPU-powered, OpenAI-compatible API for Llama, Mixtral, Gemma
Connect via PincerAPI
Use our proxy to call Groq with your PincerAPI key. No separate signup needed.
# Get instructions for Groq
curl -H "Authorization: Bearer YOUR_PINCER_KEY" \
https://pincerapi.com/api/v1/apis/groq/instructions
# Call through PincerAPI proxy
curl -H "Authorization: Bearer YOUR_PINCER_KEY" \
"https://pincerapi.com/api/v1/connect/groq/your/endpoint/here"Direct Setup
Endpoints
/chat/completionsChat completion (OpenAI-compatible)
π‘ Fastest inference available. Supports tool_use.
/modelsList available models
/audio/transcriptionsTranscribe audio with Whisper
π‘ Same as OpenAI Whisper API format
Related APIs in ai-llm
OpenAI
GPT-4o, o1, DALLΒ·E, Whisper, TTS, embeddings β the leading AI API platform
Anthropic
Claude AI models β Claude 3.5 Sonnet, Claude 3 Opus/Haiku for chat and analysis
Google AI (Gemini)
Gemini models β multimodal AI for text, images, audio, video, and code
Cohere
Enterprise AI β Command for chat, Embed for search, Rerank for relevance
Mistral AI
Open-weight AI models β Mistral Large, Small, and specialized coding models
Together AI
Open-source model hosting β Llama, Mixtral, SDXL with OpenAI-compatible API