Data residency controls let you manage where your data is processed and stored. Two independent settings govern this:
inference_geo API parameter or as a workspace default.The inference_geo parameter controls where model inference runs for a specific API request. Add it to any POST /v1/messages call.
| Value | Description |
|---|---|
"global" | Default. Inference may run in any available geography for optimal performance and availability. |
"us" | Inference runs only in US-based infrastructure. |
curl https://api.anthropic.com/v1/messages \
--header "x-api-key: $ANTHROPIC_API_KEY" \
--header "anthropic-version: 2023-06-01" \
--header "content-type: application/json" \
--data '{
"model": "claude-opus-4-6",
"max_tokens": 1024,
"inference_geo": "us",
"messages": [{
"role": "user",
"content": "Summarize the key points of this document."
}]
}'The response usage object includes an inference_geo field indicating where inference ran:
{
"usage": {
"input_tokens": 25,
"output_tokens": 150,
"inference_geo": "us"
}
}The inference_geo parameter is supported on Claude Opus 4.6 and all subsequent models. Older models released before Opus 4.6 do not support the parameter. Requests with inference_geo on legacy models return a 400 error.
The inference_geo parameter is only available on the Claude API (1P). On third-party platforms (AWS Bedrock, Google Vertex AI), the inference region is determined by the endpoint URL or inference profile, so inference_geo is not applicable. The inference_geo parameter is also not available via the OpenAI SDK compatibility endpoint.
Workspace settings also support restricting which inference geos are available:
allowed_inference_geos: Restricts which geos a workspace can use. If a request specifies an inference_geo not in this list, the API returns an error.default_inference_geo: Sets the fallback geo when inference_geo is omitted from a request. Individual requests can override this by setting inference_geo explicitly.These settings can be configured through the Console or the Admin API under the data_residency field.
Workspace geo is set when you create a workspace and can't be changed afterwards. Currently, "us" is the only available workspace geo.
To set workspace geo, create a new workspace in the Console:
Data residency pricing varies by model generation:
inference_geo: "us") is priced at 1.1x the standard rate across all token pricing categories (input tokens, output tokens, cache writes, and cache reads).inference_geo: "global" or omitted): Standard pricing applies.inference_geo settings.This pricing applies to the Claude API (1P) only. Third-party platforms (AWS Bedrock, Google Vertex AI, Microsoft Foundry) have their own regional pricing. See the pricing page for details.
If you use Priority Tier, the 1.1x multiplier for US-only inference also affects how tokens are counted against your Priority Tier capacity. Each token consumed with inference_geo: "us" draws down 1.1 tokens from your committed TPM, consistent with how other pricing multipliers (prompt caching, long context) affect burndown rates.
The inference_geo parameter is supported on the Batch API. Each request in a batch can specify its own inference_geo value.
If your organization previously opted out of global routing to keep inference in the US, your workspace has been automatically configured with allowed_inference_geos: ["us"] and default_inference_geo: "us". No code changes are required. Your existing data residency requirements continue to be enforced through the new geo controls.
The legacy opt-out was an organization-level setting that restricted all requests to US-based infrastructure. The new data residency controls replace this with two mechanisms:
inference_geo parameter lets you specify "us" or "global" on each API call, giving you request-level flexibility.default_inference_geo and allowed_inference_geos settings in the Console let you enforce geo policies across all keys in a workspace.Your workspace was migrated automatically:
| Legacy setting | New equivalent |
|---|---|
| Global routing opt-out (US only) | allowed_inference_geos: ["us"], default_inference_geo: "us" |
All API requests using keys from your workspace continue to run on US-based infrastructure. No action is needed to maintain your current behavior.
If your data residency requirements have changed and you want to take advantage of global routing for better performance and availability, update your workspace's inference geo settings to include "global" in the allowed geos and set default_inference_geo to "global". See Workspace-level restrictions for details.
Legacy models are unaffected by this migration. For current pricing on newer models, see Pricing.
"us" and "global" are available at launch. Additional regions will be added over time."us" is currently available. Workspace geo can't be changed after workspace creation.View data residency pricing details.
Learn about workspace configuration.
Track usage and costs by data residency.
Was this page helpful?