Learn how to deploy GPU instances, serverless endpoints, and build AI applications with Float16 Cloud.
Get up and running in 5 minutes
Complete API documentation
Access the Float16 dashboard
Learn the basics of Float16 Cloud platform
Learn what Float16 Cloud is and how it can help you deploy AI workloads
Get up and running with Float16 Cloud in under 5 minutes
Learn how to sign in to Float16 Cloud
Organize your resources with workspaces
Navigate the Float16 console and understand key features
In-depth guides for using Float16 Cloud features
Create and manage dedicated GPU instances on Float16 Cloud
Run AI workloads on-demand without managing infrastructure
Deploy vLLM models instantly with pre-configured settings
Manage persistent storage for your GPU instances
Use pre-configured templates to deploy common AI workloads
AI and LLM services for inference and deployment
Explore Float16's AI capabilities including LLM deployment, vLLM Playground, and OCR
Deploy and serve large language models on Float16 Cloud
Interactive environment for testing deployed vLLM models
Extract text from Thai and English documents using vision-language OCR
Enable LLMs to interact with external tools and functions
Generate responses in specific formats using JSON Schema, Regex, or Choice constraints
Complete API documentation for Float16 Cloud
Credits, payments, and billing management