L4 GPU instances
Deploy your inference and graphics workloads with NVIDIA L4 GPUs
Optimal performance for AI inference, video processing, and 3D rendering in a reliable and transparent European cloud.
Why choose NVIDIA L4 GPUs?
Cost-effective
Excellent performance-to-price ratio, ideal for AI inference and processing vast amounts of video data.
Optimised
Perfectly suited for running compact models (up to 7B) and multimedia workloads efficiently.
Sustainable
A lower carbon footprint and controlled costs.
Sovereign
Available in our Public Cloud, ensuring flexibility, transparency, and compliance with European standards.
Optimised for your inference and graphics workloads
Large-scale AI inference
Deploy your vision, NLP, or classification models on cost-effective GPUs.
Video processing and streaming
Utilise hardware AV1, HEVC, and H.264 encoding/decoding features for tasks like transcoding, object detection, or live streaming.
Lightweight creation and generative AI
Generate visuals, avatars, or multimedia content using Ada Lovelace graphics power, while keeping energy consumption low.
Specifications
Technical specifications
GPU
1–4 GPUs per instance
GPU memory
24 GB per GPU
High-performance storage
NVMe storage on most instances
Public and private network
Up to 25 Gbps included
Automation
Management via your Control Panel, API, OVHcloud CLI, etc.
Secure and private
ISO27001, SOC certifications, health data hosting.
Maximise your ROI with flexible GPU infrastructure
Pricing transparency
Benefit from the most cost-effective GPU in the range, ideal for AI inference and video.
AI and graphical versatility
Leverage a single GPU for your generative AI, 3D rendering, and AI inference workloads. This provides you with unmatched flexibility for your production pipelines.
Sovereignty and compliance
Your data is hosted in a secure, transparent, and compliant European cloud, certified to meet regulations (GDPR, ISO, HDS).
Complete accessibility
L4 GPUs available to provide the performance needed to accelerate large-scale AI and multimedia workloads — no commitment required
How do I choose my GPU for inference?
Lightweight inference
L4: Budget-friendly and energy-efficient.
Large models
A100/H100/H200 according to your bandwidth and performance needs.
Configure your GPU instances
Your questions answered
What service level agreement is guaranteed by OVHcloud on a GPU instance?
The SLA guarantees 99.99% monthly availability on GPU instances. For further information, please refer to the General Terms of Service.
Which hypervisor is used for instance virtualisation?
Just like other instances, GPU instances are virtualised by the KVM hypervisor in the Linux kernel.
What is PCI Passthrough?
Cards with GPUs are served via the physical server’s PCI bus. PCI Passthrough is a hypervisor feature that allows you to dedicate hardware to a virtual machine by giving direct access to the PCI bus, without going through virtualisation.
Can I resize a Cloud GPU instance?
Yes, Cloud GPU instances can be upgraded to a higher model after a reboot. However, they cannot be downgraded to a lower model.
Do GPU instances have anti-DDoS protection?
Yes, our anti-DDoS protection is included with all OVHcloud solutions at no extra cost.
Can I switch to hourly billing from an instance that is billed monthly?
If you have monthly billing set up, you cannot switch to hourly billing. Before you launch an instance, please take care to select the billing method that is best suited to your project.
What is a Cloud GPU?
A Cloud GPU is a cloud computing service that provides graphic processing units (GPUs) for tasks that require high computing power. Examples of these are graphic rendering, machine learning, data analysis, and scientific simulations. Unlike on-premises GPUs, which require a significant investment in hardware, cloud GPUs are more flexible and easier to scale. Users can access high-performance computing resources on demand, and only pay for what they use.
What is an H100 and A100 server?
Servers that are equipped with NVIDIA H100 and A100 GPUs are purpose-built to offer exceptional performance in HPC, AI, and data analytics.
What is NGC?
NVIDIA GPU Cloud (NGC) is a cloud computing platform offered by NVIDIA. It provides a comprehensive selection of software that is optimised for GPU acceleration in artificial intelligence (AI), machine learning (ML), and high-performance computing (HPC). NGC simplifies and speeds up the deployment of AI and scientific computing applications. It does this by providing containers, pre-trained models, SDKs, and other tools that are optimised to leverage NVIDIA GPUs.
Why use a Cloud GPU?
There are several advantages to using a Cloud GPU, especially for companies, R&D, and development teams in demanding fields, such as artificial intelligence (AI), graphics rendering, machine learning (ML), and high-performance computing (HPC).