For Data Scientists

VM-Like GPU Access. No YAML. No DevOps.

A real GPU environment you can SSH into, run Docker, and break dependencies without asking anyone. Spin up in seconds, not tickets.

< 30s
From Click to SSH
Full Root
It's Your Machine
Docker
Build & Run

All the Power of a GPU Server. None of the Ops.

Everything you need to train models — accessed the way you already work.

Access Your Way

SSH, VSCode Remote, or Jupyter — your choice. Connect how you're used to.

Docker Just Works

Build images, run containers, mount volumes. Your workflow, unchanged.

VM-Like Environment

Full root access. Persistent storage. A GPU environment you actually control.

Break Things Freely

Isolated environment. One-click reset when dependencies go wrong.

Serverless GPU Queue

Like Slurm, but instant. No time slots. No waiting.

Pay Only When Running

Stop the instance, stop the bill. Scale down to zero.

Credit-Based Quota. No More Time Slots.

Stop booking 8-hour blocks when you only need 2. Use exactly what you need, when you need it.

Traditional Time-Based Quota

Mon
Booked
Queue
Taken
Tue
Taken
Queue
Wed
8hr block — used only 2hr
Book 8-hour blocks minimum
Wait in queue for your slot
Unused time is wasted

Credit-Based Quota

Team Balance$2,450.00
Used: $550Budget: $3,000
H100 Training
-$12.40
Stopped — no charge
$0.00
Use for 10 minutes or 10 hours
Start instantly, no queue
Stop anytime, keep your credits

Serverless GPU: Slurm for the Modern Era

The queue system you know, rebuilt for instant provisioning. Submit jobs, get GPUs, no waiting.

Traditional Slurm

$ sbatch train.sh
Submitted batch job 847291
$ squeue -u $USER
JOBID   STATE    TIME   NODELIST
847291  PENDING  0:00   (Resources)
# ... 45 minutes later ...
$ squeue -u $USER
847291  RUNNING  2:31   gpu-node-03
Wait 30-60 min for GPU allocation
Poll squeue to check status
Logs only after job completes

Float16 Serverless GPU

train.py
H100 80GB
Job #1847
Running • 00:02:31
GPU provisioned in 8 seconds
Streaming logs...
Epoch 1/10 • Loss: 2.341
Epoch 2/10 • Loss: 1.892
█ Training...
GPU ready in seconds, not hours
Real-time status in dashboard
Stream logs as job runs

MIG: One GPU, Seven Hardware-Isolated Instances

Multi-Instance GPU splits a single H100 into up to 7 fully isolated instances at the hardware level. Each instance has dedicated compute, memory, and cache — no noisy neighbors.

1x H100 80GB → 7 Isolated Instances
Hardware Isolation
1g
1g
1g
1g
1g
1g
1g
Dedicated Memory
Dedicated Compute
Dedicated Cache
1g.10gb
10GB VRAM
Small experiments
2g.20gb
20GB VRAM
Development
3g.40gbPopular
40GB VRAM
Fine-tuning LLMs
7g.80gb
80GB VRAM
Full GPU power
More GPU Access for Your Team

One H100 can now serve 7 team members simultaneously. Seniors get full power when needed, juniors get dedicated instances to learn and experiment — no more waiting for GPU availability.

Give Your Team GPU Access Without the DevOps Burden

Deploy Float16 on your cluster. Your data scientists get VM-like environments they can SSH into — you get a platform that manages itself.