Skip to content

Introduction

Qualia is a cloud platform for fine-tuning Vision-Language-Action (VLA) models. VLAs enable you to build spatial agents that can reason about their surroundings, plan, and execute skill-based tasks.

Upload your dataset or use existing libraries from HuggingFace to fine-tune models like SmolVLA, Pi0, ACT, and GR00T on cloud GPUs.

These docs cover the complete workflow for training and deploying VLA models:

  1. Data Collection - Record demonstrations using LeRobot and prepare your dataset
  2. Fine-tuning - Train your model on Qualia using the dashboard or SDK
  3. Inference - Deploy your fine-tuned model locally on your robot
  1. Sign up at app.qualiastudios.dev
  2. Create a project and start a fine-tuning job from the dashboard
  3. To use the SDK or API, create an API key in Settings in the dashboard