Google Vertex AI via VerticalAPI

Vertex-hosted Gemini, Claude and Llama via VerticalAPI's OpenAI-compatible endpoint. BYOK with a GCP service account, zero markup, region-pinned inference.

Endpoint: https://api.verticalapi.com/v1/chat/completions  ·  BYOK header: Configure SA JSON in dashboard

Google Vertex AI models routed by VerticalAPI

Pass the model ID below as model in any OpenAI-compatible request. New Google Vertex AI models are typically supported within 24h of release.

Model IDNameContextPricing (provider)
gemini-2.5-pro Gemini 2.5 Pro (Vertex) 2M Vertex pricing
claude-sonnet-4-5@vertex Claude Sonnet 4.5 (Vertex) 200K Anthropic-on-Vertex pricing
llama-3.3-70b@vertex Llama 3.3 70B (Vertex) 128K Vertex Llama pricing

Pricing reflects Google Vertex AI's rates — you pay Google Vertex AI directly. VerticalAPI adds zero markup on tokens.

5-line Google Vertex AI call via VerticalAPI

Drop-in replacement for the OpenAI SDK. Works with the OpenAI Python client, Node, Go, curl — anything that speaks HTTP.

vertex-ai_quickstart.py Python
from openai import OpenAI

client = OpenAI(
    base_url="https://api.verticalapi.com/v1",
    api_key="vapi_...",
    default_headers={"X-Provider-Key": "GCP service-ac..."}
)

response = client.chat.completions.create(
    model="gemini-2.5-pro",  # Google Vertex AI
    messages=[{"role": "user", "content": "Hello"}]
)
print(response.choices[0].message.content)

Four reasons developers route Google Vertex AI through us

Zero token markup

You pay Google Vertex AI directly with your own key. VerticalAPI's revenue is the gateway subscription, not a tax on your tokens.

One key, every provider

Google Vertex AI alongside OpenAI, Anthropic, Gemini and 12 more — same OpenAI-compatible endpoint, same SDK, switchable per-request.

Latency & cost monitoring

Per-request token counts, p50/p95 latency and cost dashboards out of the box. Compare Google Vertex AI to other providers on identical prompts.

Observability built in

Every Google Vertex AI call gets a trace ID, replayable payload and audit log entry. Wire to Datadog or Sentry via OpenTelemetry.

Where Google Vertex AI shines

GCP-native compliance data residency consolidated GCP billing Anthropic-on-GCP

Common questions about Google Vertex AI on VerticalAPI

Can I use Claude on Vertex?

Yes. If your GCP project has access to Anthropic's Vertex deployments, pass the @vertex-suffixed model ID and VerticalAPI routes via Vertex's anthropic.* endpoint.

How are credentials handled?

Upload the service-account JSON in the dashboard (encrypted at rest). VerticalAPI mints a short-lived token per request — the SA JSON is never returned by API.

All supported LLM providers

Same endpoint, same SDK — just change the model and the BYOK header.