Skip to content
Cloud

GPU Cloud

Virtual NVIDIA GPUs for AI, machine learning, and rendering — from Tesla T4 to H200, granularly configurable, 100% data sovereign from Munich.

NVIDIA
T4 · A10 · A100 · H200
vGPU
Flexible GPU Shares
CUDA
cuDNN · TensorRT

Configure GPU Server

Select your GPU model and configure all resources individually.

Configuration

GPU Model

NVIDIA vGPU · dedicated VRAM

vGPU Profile

Dedicated VRAM

GB VRAM
€ 0,00

Dedicated vCores

Intel® Xeon® · dedicated

4 vCores
€ 66,00
4326496128
€ 16,50 / vCore

RAM

ECC · exclusively assigned

16 GB
€ 71,20
16 GB128 GB256 GB384 GB512 GB
€ 4,45 / GB

NVMe Storage

3N Redundancy · Ceph

100 GB
€ 35,00
100 GB4 TB8 TB12 TB16 TB
€ 0,17 / GB

Service Level Agreement

Availability & Support

excl. VAT
Monthly net
€ 0,00

GPU Pricing

All vGPU profiles at a glance — combinable with freely configurable compute resources.

GPU / Profile VRAM FP32 Price/Month
Tesla T4 — T4-4Q 4 GB GDDR6 2,0 TFLOPS € 69,00
Tesla T4 — T4-8Q 8 GB GDDR6 4,1 TFLOPS € 129,00
Tesla T4 — T4-16Q 16 GB GDDR6 8,1 TFLOPS € 249,00
NVIDIA A10 — A10-4Q 4 GB GDDR6 5,2 TFLOPS € 99,00
NVIDIA A10 — A10-8Q 8 GB GDDR6 10,4 TFLOPS € 189,00
NVIDIA A10 — A10-12Q 12 GB GDDR6 15,6 TFLOPS € 279,00
NVIDIA A10 — A10-24Q 24 GB GDDR6 31,2 TFLOPS € 549,00
NVIDIA A100 — 1g.10gb 10 GB HBM2e 2,8 TFLOPS € 179,00
NVIDIA A100 — 2g.20gb 20 GB HBM2e 5,6 TFLOPS € 349,00
NVIDIA A100 — 3g.40gb 40 GB HBM2e 8,4 TFLOPS € 529,00
NVIDIA A100 — 7g.80gb 80 GB HBM2e 19,5 TFLOPS € 999,00
NVIDIA H200 — 1g.20gb 20 GB HBM3e 9,6 TFLOPS € 349,00
NVIDIA H200 — 2g.40gb 40 GB HBM3e 19,1 TFLOPS € 679,00
NVIDIA H200 — 3g.70gb 70 GB HBM3e 28,7 TFLOPS € 999,00
NVIDIA H200 — 7g.141gb 141 GB HBM3e 67,0 TFLOPS € 1.899,00

All prices excl. VAT. vCores, RAM, and storage are configured separately. Multi-GPU available on request.

GPU Power

Dedicated GPU Power — Full Control

Book exactly the GPU power you need — from small vGPU slices for inference to the full GPU for training. All resources are dedicated, no shared instances.

NVIDIA Tesla T4, A10, A100 & H200
Granular vGPU profiles (4–141 GB VRAM)
CUDA, cuDNN & TensorRT Support
Dedicated vCores & DDR5 ECC RAM
No vendor lock-in (Open Source: KVM)
No US Cloud Act — owner-operated German GmbH
Operated in ISO 27001 certified data centers
GPU Cloud Infrastructure

Additional Services

Flexible GPU Scaling

Start with a single vGPU and scale to multi-GPU setups as needed. No long-term hardware commitment.

CUDA Ecosystem

Full support for CUDA, cuDNN, and TensorRT. Compatible with PyTorch, TensorFlow, JAX, and all major ML frameworks.

Data Sovereignty

Your training data and models remain on German infrastructure. No US Cloud Act — owner-operated GmbH.

Direct Connect

Connect GPU instances seamlessly with bare metal servers and colocation hardware — ideal for hybrid AI pipelines.

INGATE Premium Support

Support via email and phone, free 24x7 emergency hotline, personal point of contact, and highly qualified on-site staff.

Container-Ready

Pre-configured NVIDIA container images with CUDA toolkit. Docker & Kubernetes-ready for seamless CI/CD pipelines.

Technical Highlights

State-of-the-art infrastructure in our data centers for your business-critical applications.

Redundant Power Supply

Dual-path A/B power supply down to the rack. Dedicated transformers, UPS, and backup generators.

High-Efficiency Cooling

PUE < 1.20 through free cooling and cold aisle containment. Optimized for high-density up to 20 kW per rack.

Fire Protection

VESDA early detection and damage-free gas extinguishing system.

High-Speed Backbone

Redundant high-performance backbone with multiple 100Gbit/s links. Direct peering at DE-CIX and MuCon-X for lowest latencies.

Physical Security

Security level SK4. Biometric access control and comprehensive video surveillance.

Sustainability

Carbon-neutral operations with 100% green energy. Certified green electricity and waste heat recovery.

Certified Data Centers

Our primary data center EMC Home of Data in Munich holds the following certifications. All additional data centers are at least ISO 27001 certified and powered by 100% renewable energy. Select locations additionally hold SOC 1, SOC 2, and PCI-DSS certifications.

ISO 27001
Information Security
ISO 9001
Quality Management
ISO 50001
Energy Management
DIN EN 50600
DC Availability
CSR 26001
Corporate Responsibility
TÜV Süd
100% Green Energy

Frequently Asked Questions

Answers to the most important questions about GPU Cloud.

What is the difference between Cloud GPU and GPU Server?
Cloud GPU offers virtual GPU instances (vGPU) that can be flexibly scaled — ideal for variable workloads. GPU Servers are dedicated physical servers with exclusively assigned GPUs — optimal for continuous training with maximum performance.
Which GPU models are available?
We offer four NVIDIA GPU classes: Tesla T4 (16 GB GDDR6) as a cost-efficient entry-level GPU, A10 (24 GB GDDR6) as an all-rounder, A100 (80 GB HBM2e) for demanding AI workloads, and H200 (141 GB HBM3e) for maximum AI performance. Each GPU can be divided into different profiles — from small slices for inference to the full GPU for training.
What is the difference between T4, A10, and A100?
The Tesla T4 (Turing, 8.1 TFLOPS FP32) is ideal for cost-efficient inference, VDI, and light ML workloads. The A10 (Ampere, 31.2 TFLOPS FP32) is an all-rounder for ML training, 3D rendering, and virtual desktops. The A100 (Ampere, 80 GB HBM2e, 312 TFLOPS FP16 Tensor) offers MIG isolation for demanding AI workloads. The H200 (Hopper, 141 GB HBM3e, 989 TFLOPS FP16 Tensor) delivers maximum performance for LLM training and large foundation models.
What does vGPU mean?
vGPU (Virtual GPU) allows a physical GPU to be divided into multiple virtual instances. Each vGPU instance receives dedicated GPU resources and VRAM. This way, you can book exactly the GPU power you need — without having to rent an entire GPU.
Which frameworks are supported?
Full support for CUDA, cuDNN, and TensorRT. Compatible with all major ML frameworks such as PyTorch, TensorFlow, JAX, and ONNX Runtime. We provide pre-configured container images.
Can I combine Cloud GPU with Bare Metal?
Yes, via Direct Connect you can seamlessly connect cloud GPU instances with your bare metal servers and colocation hardware — ideal for hybrid AI pipelines.
How is billing handled?
All resources are billed monthly. You configure vGPU, vCores, RAM, and storage individually and pay only for what you use. No minimum contract term.
What are egress costs?
Egress costs are fees that cloud providers charge for outgoing data traffic — i.e., for data leaving the data center toward the internet or other networks. Every API response, every download, every video stream, and every backup replication generates egress traffic. Most hyperscalers charge these costs per GB, which can quickly become a significant and hard-to-calculate cost factor. Especially with data-intensive applications, these fees add up rapidly.
Does INGATE charge egress costs?
No. At INGATE, outgoing data traffic is already included in the price — no per-GB fees, no hidden surcharges. This makes your costs fully predictable and transparent. Especially with data-intensive applications like CDN, streaming, large APIs, or backup replication, this results in a massive cost advantage over the major hyperscalers.
How high are egress costs at hyperscalers?
AWS charges approximately $0.09/GB (first 10 TB), Azure approximately $0.087/GB, and Google Cloud approximately $0.12/GB. A company transferring 10 TB per month pays approximately $900-1,200/month for outgoing traffic alone. At 100 TB, it is already $8,000-9,000+/month. These costs are difficult to predict as they depend on user behavior, API call volumes, and traffic patterns — complicating TCO calculations and frequently leading to unexpectedly high bills ("bill shock"). At INGATE, traffic is included, making total costs fully calculable from day one.

Technology Partners & Memberships

Dell PartnerDirect
Equinix
EMC Home of Data
Juniper Networks
LiveConfig
Microsoft Cloud Solution Provider
Microsoft SPLA Partner
RIPE NCC Member