North Highland Reference Architecture

Provider-Portable AI

Our point of view on enterprise AI architecture—and the working implementation that proves it.

1 Reference Pattern
5 Provider Adapters
1 Working Implementation
Zero Vendor Lock-in

Our Point of View

North Highland believes enterprises must own their AI abstraction layer—not rent it from a cloud vendor.

1

Separation of Concerns

Your application logic should never know which foundation model provider is executing a request. AI consumption and AI provision are separate architectural concerns.

2

Enterprise-Owned Gateway

The abstraction layer must be yours—deployed where you choose, governed by your policies, observable through your tools. Not a managed service you can't inspect.

3

Governance by Default

Security, privacy, audit trails, and cost controls must be architectural constraints, not afterthoughts. Every AI request flows through policy enforcement.

4

Prove It Works

A reference architecture is theory. A working implementation is proof. We built RegRiskIQ on these principles to demonstrate they're not just possible—they're practical.

The Strategic Challenge

You face a fundamental tension in AI adoption for regulatory compliance.

Immediate Value Delivery

You need to leverage cutting-edge AI capabilities for compliance automation, risk assessment, and regulatory intelligence today. Waiting means falling behind competitors and increasing regulatory exposure.

🔒

Strategic Flexibility

You require the freedom to choose the best provider for each workload, switch providers as pricing and capabilities evolve, and avoid lock-in that constrains future technology decisions.

🛡

Governance Requirements

Your organization demands consistent security, privacy, and audit controls regardless of which AI provider processes requests. Compliance cannot be an afterthought.

💰

Cost Optimization

You want to route workloads to the most cost-effective provider without sacrificing quality. Different tasks demand different models, and your architecture should support intelligent routing.

THE REFERENCE PATTERN

The Model Gateway Architecture

Applications never communicate directly with foundation model providers. All AI interactions flow through a Model Gateway the enterprise owns and controls.

Applications
Compliance Automation
Risk Assessment
Regulatory Monitor
RAG Copilot
Model Gateway
Unified API Contract • Intent-Based Routing • Policy Engine • Prompt Registry • Safety Controls • Observability
Providers
Amazon Bedrock
Google Vertex AI
Azure AI Foundry
OpenAI Direct
Private / Ollama

Why This Pattern Works

The Model Gateway acts as an anti-corruption layer between your business logic and external AI providers. This separation delivers three strategic advantages:

  • Provider Independence: Switch providers through configuration changes, not application rewrites
  • Centralized Governance: Apply consistent security, privacy, and audit controls across all AI interactions
  • Optimized Economics: Route each workload to the most cost-effective option automatically
Without Gateway With Gateway
Provider-specific code in apps One API, any provider
Scattered governance Centralized controls
Expensive provider switching Configuration-based routing
Manual cost optimization Intelligent auto-routing
Fragmented observability Unified tracing and logs
💡

The Real Challenge

Building an abstraction layer is straightforward—most enterprises already have one. Governing it is where organizations struggle: consistent security policies, audit trails, compliance controls, and vendor management across all providers. That's what this architecture solves.

FROM THEORY TO PRACTICE

RegRiskIQ: The Implementation

We didn't just design this architecture—we built it. RegRiskIQ is our working implementation of the Provider-Portable AI pattern, deployed in production for regulatory compliance workloads.

5 Production Provider Adapters
63 AI Governance Controls
14 Governance Domains

WHAT WE'VE BUILT

RegRiskIQ Capabilities

Our implementation delivers enterprise-grade AI governance through these integrated components.

🔌

Intent-Based Routing

Your applications specify what they need (regulatory analysis, risk scoring, document extraction) rather than which model to use. The gateway selects the optimal provider based on cost, latency, quality, and policy requirements.

📑

Prompt Registry

Prompts become versioned, deployable artifacts stored in a central registry. Update prompts without changing application code. Test new versions before production rollout. Maintain audit trails of prompt changes.

🛡

Policy Engine

Enforce tenant isolation, data residency requirements, guardrails, and rate limits at the gateway level. Policies apply consistently across all AI interactions regardless of provider.

🔎

Provider Adapters

Each foundation model provider integrates through a dedicated adapter that normalizes request formats, response structures, error codes, and authentication patterns. Adding new providers requires only a new adapter.

📊

Observability

OpenTelemetry instrumentation provides end-to-end visibility across gateway, adapters, and providers. Track token usage, costs, latencies, and error rates per tenant, per provider, and per use case.

📚

RAG Independence

Your retrieval pipeline operates independently from model providers. Switch inference providers without re-indexing document stores or modifying retrieval logic. Your knowledge base stays portable.

Routing Strategies

Strategy Optimizes For Use Case
Cost Minimize spend while meeting quality thresholds High-volume, non-critical workloads
Performance Minimize latency for interactive experiences Real-time compliance Q&A
Quality Maximize output quality for critical decisions Regulatory filing review
Hybrid Balance all factors dynamically Default for most workloads

AI Governance Program

Fully codified AI governance playbook with 63 controls across 14 domains. Controls implemented as code with complete regulatory framework mapping to ISO 42001, EU AI Act, and NIST AI RMF.

63
Total Controls
14
Governance Domains
4
Code Formats
4
Regulatory Frameworks

AI Governance Program Foundation

5 domains | 19 controls | Enterprise-wide governance structure

GO Governance & Accountability 5 controls
RO Regulatory Oversight 5 controls
TP Third-Party Management 2 controls
CO Communications 2 controls
AA Assessment & Assurance 5 controls

AI Use Case/System Lifecycle

9 domains | 44 controls | Per-system governance controls

RM Risk Management 6
LC Lifecycle Mgmt 5
SE Security 6
RS Responsible AI 5
GA Generative AI 6
PR Privacy 4
OM Operations 5
IM Incident Mgmt 3
PL Project Lifecycle Management 4

Regulatory Framework Coverage

ISO 42001:2023
AI Management System - Full Mapping
EU AI Act
Articles 9-26 - Full Mapping
NIST AI RMF v1.0
GOVERN, MAP, MEASURE, MANAGE
ISO 27001 / SOC 2
Security Controls - Integrated

Architectural Value Proposition

What the provider-portable architecture enables for your organization.

Unified Single API for All Providers
Isolated Provider-Specific Code Contained in Adapters
Flexible Route by Cost, Performance, or Quality
Observable OpenTelemetry Tracing Across All Requests

Strategic Advantages

Freedom of Choice

Evaluate and adopt new providers without application changes. Your business logic stays stable while AI capabilities evolve.

Optimized Economics

Route each workload to the most cost-effective option. Use premium models where quality matters, economical models where speed is sufficient.

Consistent Governance

Apply uniform security, privacy, and audit controls across all AI interactions. Meet regulatory requirements once, regardless of provider.

Future-Proofing

Architectural readiness for emerging models and providers. When the next breakthrough arrives, you adopt it through configuration.

The Hyperscalers Agree

AWS, Azure, and Google each publish reference architectures for this exact pattern—because they're competing to be YOUR abstraction layer.

AWS Reference Architecture

"Multi-Provider Generative AI Gateway" — Official AWS guidance for routing to Azure, OpenAI, and other providers through an AWS-hosted LiteLLM gateway on ECS/EKS.

AWS Solutions Library

Azure API Management

"AI Gateway" with native Bedrock support — Microsoft's answer: use Azure APIM to govern AWS Bedrock and non-Microsoft AI providers from your Azure control plane.

Microsoft Learn

Google Vertex AI

Model Garden with multi-provider serving — Google's unified platform supporting Anthropic Claude, Meta Llama, and partner models alongside Gemini.

Google Cloud
💡

The Strategic Takeaway

Each hyperscaler wants to be your gateway to all the others. Our architecture gives you this pattern without the platform lock-in—your gateway runs where YOU choose, not where your cloud vendor prefers.

Honest Considerations

Provider portability is real. But it requires intentional design and ongoing investment.

What We Navigate

Provider Differences Are Real

Tool calling semantics, JSON output reliability, token limits, streaming formats, and content safety features vary across providers. Our adapter layer handles this complexity so your applications stay clean.

Gateway Adds Latency

Every abstraction has overhead. The gateway layer adds processing time for routing, policy evaluation, and request normalization. For latency-sensitive workloads, this impact must be measured and optimized for your specific use cases.

Testing Requires Investment

Proving portability demands evaluation harnesses, golden datasets, and quality metrics across providers. We build this infrastructure as a first-class capability.

How We Mitigate

Adapter Test Harness

Each adapter includes a compatibility test suite that validates behavior against provider-specific edge cases. New provider integrations pass this harness before production.

Performance Budgets

Gateway components are designed with latency budgets in mind. We instrument each stage (policy evaluation, prompt resolution, routing) with OpenTelemetry tracing to identify and address bottlenecks. Specific targets are established during implementation based on measured baselines.

Continuous Evaluation

Automated quality regression tests run against all providers weekly. You get scorecards showing "can I switch to provider X" based on real data.

Ready to Move Forward?

Your path to provider-portable AI compliance starts with a structured engagement.

1
Architecture Review
with Technical Teams
2
Pilot Deployment
Scope Definition
3
Provider & Policy
Configuration
4
Production
Deployment