Together AI via VerticalAPI
Together AI's open-weights catalog (Llama, Qwen, DeepSeek, Mixtral) via VerticalAPI's OpenAI-compatible endpoint. BYOK, zero markup, fine-tuning friendly.
Together AI models routed by VerticalAPI
Pass the model ID below as model in any OpenAI-compatible request. New Together AI models are typically supported within 24h of release.
| Model ID | Name | Context | Pricing (provider) |
|---|---|---|---|
meta-llama/Llama-3.3-70B-Instruct-Turbo |
Llama 3.3 70B Turbo | 128K | $0.88 per 1M tok |
Qwen/Qwen2.5-72B-Instruct-Turbo |
Qwen2.5 72B Turbo | 32K | $1.20 per 1M tok |
deepseek-ai/DeepSeek-V3 |
DeepSeek V3 | 64K | $1.25 per 1M tok |
mistralai/Mixtral-8x22B-Instruct-v0.1 |
Mixtral 8x22B | 64K | $1.20 per 1M tok |
Pricing reflects Together AI's rates — you pay Together AI directly. VerticalAPI adds zero markup on tokens.
5-line Together AI call via VerticalAPI
Drop-in replacement for the OpenAI SDK. Works with the OpenAI Python client, Node, Go, curl — anything that speaks HTTP.
from openai import OpenAI client = OpenAI( base_url="https://api.verticalapi.com/v1", api_key="vapi_...", default_headers={"X-Provider-Key": "..."} ) response = client.chat.completions.create( model="meta-llama/Llama-3.3-70B-Instruct-Turbo", # Together AI messages=[{"role": "user", "content": "Hello"}] ) print(response.choices[0].message.content)
Four reasons developers route Together AI through us
Zero token markup
You pay Together AI directly with your own key. VerticalAPI's revenue is the gateway subscription, not a tax on your tokens.
One key, every provider
Together AI alongside OpenAI, Anthropic, Gemini and 12 more — same OpenAI-compatible endpoint, same SDK, switchable per-request.
Latency & cost monitoring
Per-request token counts, p50/p95 latency and cost dashboards out of the box. Compare Together AI to other providers on identical prompts.
Observability built in
Every Together AI call gets a trace ID, replayable payload and audit log entry. Wire to Datadog or Sentry via OpenTelemetry.
Where Together AI shines
Common questions about Together AI on VerticalAPI
Are fine-tuned models supported?
Yes. Pass your Together fine-tuned model ID as the model field — VerticalAPI proxies it transparently.
Why use Together via VerticalAPI?
Same OpenAI-compatible endpoint as your other providers, one key, switchable. Mix DeepSeek for reasoning, Llama for general, Qwen for non-English — without 3 SDKs.
All supported LLM providers
Same endpoint, same SDK — just change the model and the BYOK header.
Ship on Together AI in 60 seconds
Free tier — bring your own Together AI key, zero markup, OpenAI-compatible endpoint.
Get your VerticalAPI key →