custom background image

L4 GPU instances


Deploy your inference and graphics workloads with NVIDIA L4 GPUs

Optimal performance for AI inference, video processing, and 3D rendering — in a reliable and transparent European cloud.

Why choose NVIDIA L4 GPUs?

Cost-effective

Excellent price/performance ratio for deploying AI inference and video processing at scale.

Optimised

Description: Ideal for inference of compact models (up to 7B) and multimedia workloads.

Sustainable

A reduced carbon footprint and controlled costs.

Sovereign

Available in our Public Cloud, ensuring flexibility, transparency, and European compliance.

Optimized for your inference & graphics workloads

Large-scale AI inference

Deploy your vision, NLP or classification models on cost-effective GPUs.

Video processing and streaming

Enjoy hardware AV1, HEVC and H.264 encoding/decoding, ideal for transcoding, object detection or live streaming.

Lightweight generative creation and AI

Generate visuals, avatars or multimedia content using the Ada Lovelace graphics power, while maintaining a minimal energy footprint.

SPECIFICATIONS

Technical specifications

GPU

1, or 4 GPUs per instance

GPU memory

24 GB per GPU

High-performance storage

NVMe storage on most instances

Public & private network

Up to 25 Gbps included

Automation

Management via your customer space, API, OVHcloud CLI, ...

Secure and Private

Certifications: ISO27001, SOC, Health data.

Maximise your ROI with a flexible GPU infrastructure

Transparent pricing

Benefit from the most economical GPU in the range, perfect for AI inference and video.

AI & graphics versatility

Leverage a single GPU for your GenAI workloads, 3D rendering, and AI inference — unique flexibility for your production pipelines.

Sovereignty and compliance

Your data is hosted on a certified European cloud, ensuring security, transparency, and compliance with regulations (GDPR, ISO, HDS).

Total accessibility

L4 GPUs available without commitment constraints, to accelerate your large-scale AI and multimedia projects.

How to choose your GPU for inference?

Lightweight inference

L4: cost-effective and energy-efficient.

Large models

A100 / H100 / H200 according to your bandwidth and performance needs.

Your questions answered

What SLA does OVHcloud guarantee for a GPU instance?

The SLA guarantees 99.99% monthly availability on GPU instances. For further information, please refer to the Terms and Conditions.

Which hypervisor is used for instance virtualisation?

Just like other instances, GPU instances are virtualised by the KVM hypervisor in the Linux kernel.

What is PCI Passthrough?

Cards with GPUs are served via the physical server’s PCI bus. PCI Passthrough is a hypervisor feature that allows you to dedicate hardware to a virtual machine by giving direct access to the PCI bus, without going through virtualisation.

Can I resize a Cloud GPU instance?

Yes, Cloud GPU instances can be upgraded to a higher model after a reboot. However, they cannot be downgraded to a lower model.

Do GPU instances have anti-DDoS protection?

Yes, our anti-DDoS protection is included with all OVHcloud solutions at no extra cost.

Can I switch to hourly billing from an instance that is billed monthly?

If you have monthly billing set up, you cannot switch to hourly billing. Before you launch an instance, please take care to select the billing method that is best suited to your project.

What is a Cloud GPU?

A Cloud GPU is a cloud computing service that provides graphic processing units (GPUs) for tasks that require high computing power. Examples of these tasks are graphic rendering, machine learning, data analysis, and scientific simulations. Unlike on-premises GPUs, which require a significant investment in hardware, cloud GPUs are more flexible and easier to scale. Users can access high-performance computing resources on demand, and only pay for what they use.

What is an H100 and A100 server?

Servers that are equipped with NVIDIA H100 and A100 GPUs are purpose-built to offer exceptional performance in HPC, AI, and data analytics.

What is NGC?

NVIDIA GPU Cloud (NGC) is a cloud computing platform offered by NVIDIA. It provides a comprehensive selection of software that is optimised for GPU acceleration in AI, ML, and HPC applications. NGC simplifies and speeds up the deployment of AI and scientific computing applications. It does this by providing containers, pre-trained models, SDKs, and other tools that are optimised to leverage NVIDIA GPUs.

Why use a Cloud GPU?

There are several advantages to using a Cloud GPU, especially for companies, researchers, and developers in demanding areas such as artificial intelligence (AI), graphics rendering, machine learning (ML), and high-performance computing (HPC).