Image 1 of 1
🔷 LLaMA – Sovereign Transition Model
Pricing to be confirmed
LLaMA – Flexible and Customisable AI Capability
LLaMA is deployed as a flexible, open model within our platform, supporting a wide range of general AI workloads alongside the ability to customise behaviour for specific domains and applications.
As an open model, LLaMA enables organisations to adapt and fine-tune AI capability to meet their own requirements, making it well suited to environments where control, transparency and adaptability are important. It provides a strong foundation for conversational systems, content generation and embedded AI functionality across tools and workflows.
LLaMA supports:
General AI Workloads – conversation, summarisation and content generation
Customisation – fine-tuning for domain-specific applications
Integration – embedding AI into existing systems, tools and workflows
Adaptable Behaviour – modifying outputs and responses based on use case requirements
By combining flexibility with structured deployment, LLaMA enables organisations to develop AI systems that can evolve over time while remaining aligned with operational and governance requirements.
Deployment Model
LLaMA is initially provided through a hosted environment to support development, testing and integration.
This allows organisations to rapidly experiment with use cases, refine models and embed AI capability into existing systems without the need for upfront infrastructure deployment.
As deployment requirements mature, environments can transition to UK-based infrastructure, including dedicated client machines and clustered systems. This enables full control over model behaviour, performance and outputs within the jurisdiction, supporting sovereign operation.
Location and Availability
LLaMA is currently delivered through hosted infrastructure outside the UK, providing immediate access for development and early-stage deployment.
UK-based deployment is aligned to infrastructure expansion, with environments provisioned as additional SN40L capacity is introduced. This supports a structured transition from development access to fully sovereign operation within a controlled UK environment.
Pricing to be confirmed
LLaMA – Flexible and Customisable AI Capability
LLaMA is deployed as a flexible, open model within our platform, supporting a wide range of general AI workloads alongside the ability to customise behaviour for specific domains and applications.
As an open model, LLaMA enables organisations to adapt and fine-tune AI capability to meet their own requirements, making it well suited to environments where control, transparency and adaptability are important. It provides a strong foundation for conversational systems, content generation and embedded AI functionality across tools and workflows.
LLaMA supports:
General AI Workloads – conversation, summarisation and content generation
Customisation – fine-tuning for domain-specific applications
Integration – embedding AI into existing systems, tools and workflows
Adaptable Behaviour – modifying outputs and responses based on use case requirements
By combining flexibility with structured deployment, LLaMA enables organisations to develop AI systems that can evolve over time while remaining aligned with operational and governance requirements.
Deployment Model
LLaMA is initially provided through a hosted environment to support development, testing and integration.
This allows organisations to rapidly experiment with use cases, refine models and embed AI capability into existing systems without the need for upfront infrastructure deployment.
As deployment requirements mature, environments can transition to UK-based infrastructure, including dedicated client machines and clustered systems. This enables full control over model behaviour, performance and outputs within the jurisdiction, supporting sovereign operation.
Location and Availability
LLaMA is currently delivered through hosted infrastructure outside the UK, providing immediate access for development and early-stage deployment.
UK-based deployment is aligned to infrastructure expansion, with environments provisioned as additional SN40L capacity is introduced. This supports a structured transition from development access to fully sovereign operation within a controlled UK environment.