LLM finetuning & inference
Llama, Mistral, Qwen, DeepSeek finetuning with LoRA / QLoRA / full FT on H100. Or self-hosted inference with vLLM / TGI / Ollama for production model serving.
Offshore NVIDIA GPU servers in Netherlands from $279/mo. RTX 4090, RTX 5090 and H100 SXM5 cards passed through KVM with full root. CUDA 12 + cuDNN preinstalled, PyTorch / ComfyUI / Ollama image presets ready to ssh into. Crypto-only checkout, no KYC, no email — just an account token.
The Netherlands gives you the lowest-latency GPU compute in our network thanks to AMS-IX peering. If you serve AI models to European users (inference under 10ms to Western Europe) or pull large datasets from European model registries, this is the location to pick. AMS-IX bandwidth is also the cheapest per-TB on our network.
All plans include CUDA 12 + cuDNN preinstalled, NVMe SSD, DDR5 RAM, full root access, SSH + JupyterLab and unlimited bandwidth.
| Plan | GPU | VRAM | CPU | RAM | NVMe | Bandwidth | Price | |
|---|---|---|---|---|---|---|---|---|
| NL-S | 1× NVIDIA RTX 4090 | 24 GB GDDR6X | 12 vCPU | 64 GB DDR5 | 1 TB NVMe | Unlimited | $279/mo | Order |
| NL-M Popular | 1× NVIDIA RTX 5090 | 32 GB GDDR7 | 16 vCPU | 96 GB DDR5 | 1.5 TB NVMe | Unlimited | $449/mo | Order |
| NL-L | 1× NVIDIA H100 SXM5 | 80 GB HBM3 | 24 vCPU | 192 GB DDR5 | 2 TB NVMe | Unlimited | $1799/mo | Order |
| NL-XL | 2× NVIDIA H100 SXM5 | 160 GB HBM3 | 32 vCPU | 384 GB DDR5 | 4 TB NVMe | Unlimited | $3399/mo | Order |
GPU servers shine on workloads that scale with VRAM and tensor cores — LLM finetuning and inference, diffusion image generation, AI video, and high-throughput model serving.
Llama, Mistral, Qwen, DeepSeek finetuning with LoRA / QLoRA / full FT on H100. Or self-hosted inference with vLLM / TGI / Ollama for production model serving.
Stable Diffusion, FLUX.1, SDXL with ComfyUI or Forge. Train your own LoRA, batch-generate at scale, or self-host an inference endpoint.
OpenSora, CogVideoX, Wan-2.1, AnimateDiff. Video generation needs serious VRAM — start at RTX 5090 (32 GB) or H100 (80 GB).
Deploy fine-tuned models behind your own API. Predictable costs, no per-token fees, no data leaving your jurisdiction. JupyterLab + FastAPI included.
RTX 4090 (24 GB), RTX 5090 (32 GB), H100 SXM5 (80 GB), 2× H100 (160 GB).
Up to 4 TB NVMe SSD, paired with DDR5 RAM for fast dataset I/O.
From paid order to nvidia-smi output in under 60 seconds.
Full root SSH, plus pre-bound JupyterLab on port 8888 with token auth.
The Netherlands offers the best network peering in Europe through AMS-IX. Sub-10ms latency to most of Western Europe, excellent bandwidth capacity, and a historically tolerant hosting environment.
The Netherlands is an EU member so GDPR applies. However, ServPrivacy does not collect personal data, so GDPR compliance is built into our zero-data architecture. No personal data means no GDPR risk.
Extremely fast. AMS-IX peering delivers sub-10ms latency to Western Europe and excellent global connectivity. Our Dutch servers come with up to 10 Gbps uplink and unlimited bandwidth.
Pay in BTC, XMR, ETH, USDT or 10 other chains. SSH + JupyterLab on a real NVIDIA GPU in Netherlands in under 60 seconds.