NVIDIA B200 with Full-Stack TEE
NVIDIA B200 Blackwell GPU

Next-Gen AI with B200 Blackwell

Deploy NVIDIA B200 Blackwell GPUs with Intel TDX + NVIDIA Confidential Computing. 192GB HBM3e memory, 8 TB/s bandwidth, up to 15x faster inference than H100.

192GB
HBM3e Memory
8 TB/s
Bandwidth
15x Faster
Inference vs H100
Full TEE
Intel + NVIDIA
Starting at $7.99/GPU/hr on-demand or $5.63/GPU/hr with 6-month commitment (29% savings)

Revolutionary Blackwell Performance

15x faster inference with Full-Stack TEE on next-gen architecture

GPU TEE Bento
15x

Faster Inference

vs H100 on large LLMs

192GB

HBM3e Memory (2.4x vs H100)
8 TB/s Bandwidth

2.5x

Faster training
vs H100

Blackwell

Next-gen GPU architecture

Full-Stack TEE

Intel TDX + NVIDIA Confidential Computing on cutting-edge Blackwell GPUs

Dual attestation
Hardware root of trust
8 TB/s

Memory Bandwidth

2.4x faster than H100 for memory-intensive workloads

Available Now

US-East & US-West

1-8 GPUs

Enterprise Clusters

Configure Now

B200 vs H200

Official benchmarks from MLPerf v5.0 and NVIDIA Technical Labs

LLM Inference Performance (Llama 2 70B)

Tokens per second (single GPU) - higher is better

H10021.8K tok/s
29%
H20033.0K tok/s
44%
B20075.0K tok/s
100%

H100: 21.8K tokens/sec (baseline) | H200: 33K tokens/sec (1.51x) | B200: 75K+ tokens/sec (3.4x)

Source: MLPerf Inference v5.0 (2024), NVIDIA Technical Blogs

All benchmarks include Full-Stack TEE protection with <5% overhead

B200 Pricing

Next-gen performance at competitive prices. All prices include Full-Stack TEE protection.

Single GPU

1 GPU per instance

$7.99/GPU/hr

On-demand availability

192GB HBM3e memory
US-West region
Full-Stack TEE included
Dual attestation reports
SAVE 29%
Reserved

6-month commitment

$5.63/GPU/hr

8-GPU cluster pricing

1.54TB total memory (8x 192GB)
US-East region
NVLink 5 interconnect
Priority support included

Available B200 Configurations

1x B200

Available Now
Memory192GB HBM3e
Bandwidth8 TB/s
TEE ProtectionFull-Stack TEE
RegionUS-West
Next-generation Blackwell architecture for the most demanding AI workloads
Configure & Deploy

8x B200

Available Now
Memory1.54TB HBM3e
Bandwidth64 TB/s
TEE ProtectionFull-Stack TEE
RegionUS-East
Next-generation Blackwell architecture for the most demanding AI workloads
Configure & Deploy
Security

Full-Stack TEE Architecture

Complete hardware protection from CPU to GPU. Intel TDX + NVIDIA Confidential Computing working together.

Full VM Isolation

Intel TDX protects CPU, memory, and VM from host access. Complete isolation.

GPU Memory Encryption

NVIDIA CC encrypts all GPU memory. Model weights and data stay secure.

Dual Attestation

Cryptographic proof from Intel + NVIDIA. Independently verifiable.

End-to-End Protection

Data encrypted in transit (TLS), at rest (AES-256), and during processing (TEE).

Multi-Region Deployment

Deploy in US-West and India. Same Full-Stack TEE protection everywhere.

Compliance Ready

GDPR, HIPAA, SOC 2 compliant. Hardware-backed security guarantees.

What You Can Build with GPU TEE

Real-world applications running on Phala Cloud with complete Intel TDX + NVIDIA Confidential Computing protection

Private Enterprise AI

Train and deploy models on sensitive healthcare, financial, or legal data with complete hardware protection. Your data never leaves the TEE.

User-Owned AI Agents

Build autonomous AI agents that securely manage cryptographic keys and digital assets. Powers platforms like Eliza and Virtuals Game Agents.

ZK Proof Generation

Accelerate zkVM and zkRollup proof generation with GPU TEE. SP1 zkVM runs with <5% TEE overhead—verified with dual attestation.

FHE/MPC Acceleration

Use GPU TEE as 2FA for FHE and MPC systems. Secure key generation, computation integrity, and attestation in one platform. Powers Fairblock and Mind Network.

Multi-Proof Systems

Combine ZK proofs with TEE attestation for double security. Hedge against cryptographic bugs while maintaining verifiability.

Regulatory Compliance

Meet GDPR, HIPAA, and SOC 2 requirements with hardware-backed privacy guarantees. Full audit trail with Intel and NVIDIA attestation.

Three Ways to Deploy GPU TEE

Choose the deployment model that fits your needs—from full control to instant deployment

CVM + GPU: Maximum Flexibility

Deploy your own Docker containers with SSH access to TEE-protected GPUs. Perfect for developers who need complete control.

  • Deploy custom Docker containers with full SSH access
  • Fine-tune models on private data with complete hardware protection
  • Intel TDX + NVIDIA Confidential Computing protection
  • Dual attestation reports (Intel + NVIDIA) for verification
Deploy CVM Now