Inference Services
AI inference APIs and services available to Australian users: latency from Australia, data residency guarantees, and models offered.
Total Services
8
tracked servicesAU-Hosted
4
with AU data residency↑ upAvg AU Latency
15 ms
AU-hosted services→ stableOffshore
4
US-routed services→ stableService Directory
Showing 8 of 8
AWS Bedrock (Sydney)
AU HostedOffshore Owned
Amazon Web Services
AU latency: 12ms
3 models available
Azure OpenAI (Australia East)
AU HostedOffshore Owned
Microsoft Azure
AU latency: 15ms
3 models available
Google Vertex AI (Sydney)
AU HostedOffshore Owned
Google Cloud
AU latency: 14ms
3 models available
OpenAI API (US-hosted)
Offshore Owned
OpenAI
AU latency: 185ms
4 models available
Anthropic API (US-hosted)
Offshore Owned
Anthropic
AU latency: 190ms
3 models available
Oracle AI Services (Sydney)
AU HostedOffshore Owned
Oracle Cloud
AU latency: 18ms
3 models available
Groq Cloud (US-hosted)
Offshore Owned
Groq
AU latency: 195ms
2 models available
Together AI (US-hosted)
Offshore Owned
Together AI
AU latency: 200ms
3 models available
Latency from Australia (ms) — AU hosted vs offshore
Latency Impact
AU-hosted services average ~14 ms round-trip from Australian clients. Offshore US services average ~185 ms — a 13x latency penalty.
For real-time applications (voice AI, code completion), latency >50 ms degrades user experience significantly.
Latency comparison
AWS Bedrock (Sydney) 12 ms
Azure OpenAI (Australia East) 15 ms
Google Vertex AI (Sydney) 14 ms
OpenAI API (US-hosted) 185 ms
Anthropic API (US-hosted) 190 ms
Oracle AI Services (Sydney) 18 ms
Groq Cloud (US-hosted) 195 ms
Together AI (US-hosted) 200 ms