Sovereign AI in Australia: Cloud vs. On-Premise vs. Hybrid — How to Decide
Australia just got its first Cisco Secure AI Factory. OpenAI partnered with Commonwealth Bank to train 1.2 million Australians on AI. The sovereign AI moment is here — and mid-market businesses need a clear decision framework for where their AI actually runs.

Data Engineer. Azure 6x Microsoft Certified. Monash University.
On 23 February 2026, Cisco and SharonAI Holdings launched Australia's first Cisco Secure AI Factory in partnership with NVIDIA — running on 256 NVIDIA Blackwell B200 GPUs hosted in NextDC's Australian data centres. The press release described it as "positioning Australia to process AI workloads at the highest security standards."
That same week, OpenAI announced a partnership with Commonwealth Bank, Coles, and Wesfarmers to deliver AI training to 1.2 million Australian workers and small businesses — the first-ever OpenAI partnership of this kind in the Southern Hemisphere.
Australia's AI infrastructure moment has arrived. And with it comes a question that Australian mid-market businesses can no longer defer: where should your AI actually run?
The answer matters — legally, operationally, and commercially. Getting it wrong creates Privacy Act exposure, APRA regulatory risk, and potentially irreversible technical debt. Getting it right creates a defensible, compliant AI architecture you can build on for years.
This article gives you the decision framework.
Why "Where AI Runs" Actually Matters
For most software, the question of where it runs is an infrastructure detail. For AI systems processing Australian personal information, it is a compliance and strategic question.
Privacy Act implications. Under the Australian Privacy Act 1988, when personal information is transferred to an overseas recipient, the Australian entity that originally collected that information remains accountable for how the overseas recipient handles it. If your AI system sends customer data to an API endpoint hosted in the United States or Europe, you have made an offshore data transfer — and you need to have managed that under Privacy Principle 8.
Most businesses using offshore AI APIs (OpenAI, Anthropic, Google Gemini, Cohere) have not completed a formal cross-border data transfer assessment. Many are in technical breach.
APRA requirements for regulated entities. APRA's CPS 234 and the associated guidance on model risk (CPG 220) require APRA-regulated entities — banks, insurers, superannuation funds — to maintain information security standards that extend to third-party providers. Sending sensitive financial data to offshore AI providers without a structured vendor assessment and contractual protections is an APRA risk event.
Competitive and strategic positioning. Some Australian industries — defence contracting, government services, critical infrastructure — have non-negotiable data localisation requirements. For businesses serving these sectors, offshore AI processing is not an option regardless of Privacy Act compliance posture.
The sovereignty conversation. "Sovereign AI" has become the term used to describe AI infrastructure that operates under Australian jurisdiction, subject to Australian law, with Australian data remaining in Australia. The Cisco Secure AI Factory is explicitly marketed as sovereign AI infrastructure. It is not the only option — but its launch signals that demand for this capability is real enough to justify significant infrastructure investment.
The Three Models: What They Are and When to Use Each
Model 1: Cloud AI (Public APIs)
What it is: Your application sends data to a third-party AI provider's API — OpenAI's GPT-4o, Anthropic's Claude, Google's Gemini, AWS Bedrock, Azure OpenAI — which processes it and returns a result. The data travels to the provider's infrastructure, which may be hosted offshore.
The privacy reality: Most major providers offer Australian-region endpoints. Microsoft Azure Australia East hosts Azure OpenAI. AWS Sydney hosts Amazon Bedrock models. Google Cloud has an Australian region. When you use these region-specific endpoints, your data is processed and stored within Australian jurisdiction.
The issue: many businesses use the default API endpoints — which route to US or European infrastructure — without realising they have made an offshore data transfer choice. Defaulting to api.openai.com routes through OpenAI's US infrastructure.
When cloud AI (Australian region) is appropriate:
- Non-sensitive workloads — content generation, summarisation of public information, internal productivity tools
- Prototyping and proof-of-concept work
- Use cases where the data being processed does not include personal information
- Businesses with under $3M annual revenue (below the Privacy Act APP entity threshold for most obligations)
- Workloads where Australian-region endpoints are available and have been explicitly configured
When cloud AI is NOT appropriate:
- Processing sensitive personal information (health records, financial data, personal identifiers) without a cross-border transfer assessment
- APRA-regulated entities processing customer data without formal vendor assessment
- Defence, intelligence, or government-adjacent workloads
- Any use case where data residency is contractually required by clients
Cost profile: Lowest upfront cost. Pay-per-use. No infrastructure to manage. Best for variable or unpredictable workloads.
Model 2: Private Cloud / Sovereign Cloud AI
What it is: AI infrastructure deployed in Australian data centres, operated by an Australian provider, subject to Australian law. The Cisco Secure AI Factory hosted in NextDC is an example. So is AWS Sydney, Azure Australia East, and Google Cloud Sydney when used with appropriate data residency configuration.
The key distinction: Australian-region cloud is not the same as public API endpoints. When you deploy AI models to AWS Sydney or Azure Australia East — rather than calling the provider's default API — you have much stronger data sovereignty controls. Data does not leave Australian jurisdiction. The infrastructure is subject to Australian law.
When private/sovereign cloud AI is appropriate:
- Financial services companies processing customer financial data
- Healthcare organisations processing clinical or personal health information
- Professional services firms handling confidential client data
- Businesses with contractual data residency requirements
- Mid-market businesses wanting the operational simplicity of cloud with the data sovereignty of on-premise
The options in Australia in 2026:
- AWS Sydney (ap-southeast-2): Amazon Bedrock hosts Claude, Titan, and Llama models with full Australian data residency. SageMaker for custom model deployment.
- Azure Australia East: Azure OpenAI Service hosts GPT-4o with Australian data residency. Fully managed, enterprise SLAs.
- Google Cloud Sydney (australia-southeast1): Vertex AI with Gemini models and Australian data residency.
- NextDC / SharonAI Secure AI Factory: GPU compute specifically positioned for sovereign AI workloads — relevant for organisations that need to run their own models rather than use provider-managed APIs.
Cost profile: Higher than default public APIs. Pricing varies by provider and model. Appropriate for production systems with defined, recurring workloads.
Model 3: On-Premise AI
What it is: AI models deployed on infrastructure physically located in your premises or a co-location facility you control. No data leaves your network.
The practical reality in 2026: On-premise AI is substantially more viable than it was two years ago. Open-weight models — Meta's Llama series, Mistral, Microsoft Phi — can be run on commodity GPU hardware and deliver performance close to hosted API models for specific tasks. The NVIDIA AI Factory model (of which Australia now has an example) provides enterprise-grade on-premise AI infrastructure.
When on-premise AI is appropriate:
- Extremely sensitive data that cannot leave your network under any circumstances — top-tier security classifications, patient records under specific contractual requirements, high-value intellectual property
- Highly regulated workloads where vendor audit rights and contractual protections available in cloud are insufficient
- Organisations with existing data centre infrastructure and IT capability to operate AI systems
- Defence and government contractors with specific network isolation requirements
- High-volume, predictable workloads where on-premise economics outperform pay-per-use cloud
When on-premise AI is NOT appropriate (for most mid-market businesses):
- Organisations without dedicated IT infrastructure capability
- Variable or unpredictable AI workloads (cloud cost model is better)
- Organisations that need model updates and improvements without internal ML team effort
- Most businesses under $100M revenue, for whom the capital and operational cost of on-premise AI infrastructure is disproportionate to the risk they are managing
Cost profile: High upfront capital (GPU hardware + hosting). Lower per-inference cost at scale. Significant operational overhead. Appropriate for large, consistent workloads with stringent data requirements.
The Decision Framework: Four Questions
To choose the right model for your use case, answer four questions:
Question 1: What data will the AI system process?
Map the data sensitivity:
- Low sensitivity: Internal documents, public information, non-personal data → Cloud AI (Australian region) is appropriate
- Medium sensitivity: Employee data, business-confidential information → Cloud AI (Australian region) with vendor assessment
- High sensitivity: Customer personal information, financial records, health data → Sovereign/private cloud minimum; on-premise for most sensitive
- Classified/restricted: Government security classifications, defence data → On-premise or accredited sovereign infrastructure only
Question 2: What regulatory framework applies?
- Privacy Act APP entity (>$3M revenue): Must manage cross-border transfers under APP 8; Australian-region endpoints required for personal data
- APRA-regulated: CPS 234 vendor assessment required; CPG 220 model risk governance
- AHPRA/My Health Record Act: Strict requirements on health data handling; on-premise or certified sovereign cloud
- Defence Industry Security Program: Network isolation requirements; on-premise typically required
- Standard commercial: Privacy Act compliance sufficient; sovereign cloud appropriate
Question 3: What are your contractual commitments to clients?
Review your client contracts and master service agreements. Many mid-market B2B businesses have data handling clauses that impose data residency requirements they may not have consciously agreed to when drafting. If a client contract says "data must remain in Australia" — that commitment flows through to your AI systems.
Question 4: What is your operational capability?
Be honest about what your IT team can manage:
- Small IT team or no dedicated IT: Managed cloud AI services (AWS Bedrock, Azure OpenAI) with Australian region — you get data sovereignty with managed infrastructure
- Medium IT team, some infrastructure capability: Sovereign cloud with self-managed components — deploy models to AWS/Azure but manage configuration and monitoring internally
- Large IT team, data centre capability: On-premise is viable — but validate the business case, because managed cloud has improved significantly
What This Means for Your AI Roadmap
For most Australian mid-market businesses — the $20M to $500M revenue segment — the right answer in 2026 is sovereign cloud with Australian-region configuration.
This means:
- AI models deployed to AWS Sydney, Azure Australia East, or Google Cloud Sydney
- Explicit data residency configuration — not default API endpoints
- Cross-border transfer assessment completed for any remaining offshore AI services
- Vendor assessments under your APRA or Privacy Act obligations
- Australian MSA terms with AI providers that cover data protection obligations
This model gives you:
- Full compliance with Australian Privacy Act and APRA requirements
- Operational simplicity — managed infrastructure, no GPU hardware to run
- Access to the latest AI models without internal ML engineering
- Data sovereignty that satisfies most client contractual requirements
- A credible answer when your board or clients ask "where does our data go?"
On-premise AI is appropriate for a narrower set of Australian businesses — those with genuine classified data requirements, very large and predictable workloads, or specific contractual isolation requirements. The Cisco Secure AI Factory is a signal that this market is growing, but it is not the right starting point for most mid-market businesses.
The Cisco Secure AI Factory: What It Actually Signals
Australia's first Cisco Secure AI Factory — launched in NextDC facilities in February 2026, running on 256 NVIDIA Blackwell B200 GPUs — is not primarily a product announcement. It is a market signal.
It signals that demand for sovereign AI infrastructure in Australia has reached a scale where significant capital investment is justified. It signals that the largest technology infrastructure providers consider Australian data sovereignty concerns legitimate enough to address with dedicated hardware. And it signals that the gap between "we'd like to keep our data in Australia" and "we can actually do it with enterprise-grade AI" has closed.
For Australian businesses, this changes the "where does our AI run?" conversation from theoretical to practical. The infrastructure exists. The question is whether your AI architecture takes advantage of it.
Practical Next Steps
If you have not already done so:
1. Audit your current AI tool usage (this week) List every AI tool your business uses. For each: what data does it process? Where is it hosted? What is the data residency configuration? This audit typically surfaces two to three unrecognised offshore data transfer risks.
2. Reconfigure to Australian-region endpoints (next 30 days) For tools with Australian-region options (Azure OpenAI, AWS Bedrock, Google Vertex AI), reconfigure explicitly to Australian endpoints. This is an infrastructure change, not a rebuild.
3. Complete cross-border transfer assessments for remaining offshore tools For AI tools without Australian-region options — and there are still some — complete a formal cross-border transfer assessment under Privacy Principle 8. Document your assessment and maintain it.
4. Integrate sovereign AI into your AI policy Your organisation's AI governance policy should specify where AI systems can and cannot run, with explicit data residency requirements by data sensitivity level. This is the policy layer that makes individual tool decisions consistent.
*Akira Data designs AI architectures for Australian mid-market businesses with data sovereignty built in from the start — cloud, sovereign cloud, or hybrid, depending on your regulatory and operational requirements. Our [AI Readiness Sprint](/services#readiness) includes a data residency assessment and AI architecture recommendation as core deliverables.*
*This article is general information and does not constitute legal advice. Consult your legal advisers for guidance specific to your organisation.*
Share this article