Fine-Tuned Models: Private Customization

Fine-tune foundation models on proprietary data inside TEEs. Better accuracy, zero data leakage. Keep your training data, gradients, and custom weights encrypted with hardware-enforced privacy.

LoRA & PEFT
Multi-GPU training
Sealed checkpoints
Training attestations
Fast deployment
Gradient privacy
Why It Matters

Why Private Fine-Tuning Matters

Custom performance demands private corp data; Phala lets you use it safely.

Data security

Training data contains business secrets

Traditional cloud infrastructure exposes sensitive information to operators and administrators.

More Information
Confidential computing

Fine-tuned weights encode proprietary knowledge

Hardware-enforced isolation prevents unauthorized access while maintaining computational efficiency.

More Information
Zero-trust architecture

Model gradients can leak training examples

End-to-end encryption protects data in transit, at rest, and critically during computation.

More Information
Attestation

Vendors should never see your data or weights

Cryptographic verification ensures code integrity and proves execution in genuine TEE hardware.

More Information

How It Works

Unsloth

End-to-end confidential fine-tuning with hardware attestation and encrypted artifacts.

๐Ÿ”
Remote Attestation
TEE Verified & Keys Unsealed
๐Ÿ“ฆ
Load Base Model
Llama / Mistral / Qwen
๐Ÿ—‚๏ธ
Encrypted Dataset
streamed privately
๐Ÿ”ฅFine-Tuning LoopUnsloth
Apply QLoRA
Train with Unsloth
2ร— Faster ยท 70% Less VRAM
Optional DPO/GRPO
Safety Checks
PII, Toxicity, Bias
๐Ÿงพ
Export Encrypted LoRA
+ Attestation Report
๐Ÿš€
Deploy on Phala TEE
OpenAI / HF endpoint

Fine-Tuning LLaMA 3 with Unsloth on Phala Cloud

7-Step Tutorial: Confidential fine-tuning with hardware attestation and encrypted artifacts

1

Environment Setup

Install Unsloth and Hugging Face libraries with GPU support

2

Loading Chat Dataset Securely

Mount and load encrypted fine-tuning dataset in conversational format

3

Loading LLaMA 3 with Unsloth

Load base model with 4-bit quantization and memory optimization

4

Applying LoRA Adapters

Add Low-Rank Adapters to attention and feed-forward layers

5

Fine-Tuning with TRL

Supervised fine-tuning using HuggingFace TRL SFTTrainer

6

Merging LoRA into FP16 Weights

Merge LoRA adapters into base model for deployment

7

Saving and Uploading Model

Push merged model to Hugging Face Hub for inference

# Install Unsloth and Hugging Face libraries
pip install unsloth transformers accelerate trl datasets

# (Optional) Ensure PyTorch 2.1 with CUDA 12.1 is installed for H200 GPU
pip install torch==2.1.0

# Verify that the GPU is accessible
python -c "import torch; print(torch.cuda.is_available(), torch.cuda.get_device_properties(0).name)"

Industry-Leading Enterprise Compliance

Meeting the highest compliance requirements for your business

AICPA SOC 2ISO 27001CCPAGDPR

Frequently Asked Questions

Everything you need to know about Private Fine-Tuning

FINE-TUNING PROCESS & PERFORMANCE

PRIVACY, SECURITY & COMPLIANCE

DEPLOYMENT & MODEL OWNERSHIP

Start Private Fine-Tuning Today

Customize LLMs on your proprietary data with hardware-enforced confidentiality and zero-knowledge guarantees.

Deploy on Phala
  • LoRA/PEFT support
  • Multi-GPU training
  • Sealed checkpoints
  • Training attestations
  • 24/7 technical support