Skip to main content

Concepts

This section covers the core concepts you need to understand when working with PaletteAI. Whether you are a platform engineer setting up infrastructure or a data scientist deploying workloads, these concepts explain how PaletteAI organizes resources and manages AI/ML deployments.

Organization & Access

PaletteAI uses a hierarchical structure to organize teams and control access to resources.

ConceptDescription
Tenants and ProjectsOrganizational hierarchy for multi-tenancy, GPU quotas, and team access control
SettingsExternal integrations (Palette API credentials) and namespace-scoped configuration
Roles and PermissionsRBAC roles automatically created for Tenants and Projects

Infrastructure

These resources define where and how your AI/ML applications run.

ConceptDescription
ComputeDiscovers available machines for cluster provisioning
Compute ConfigDefault settings for cluster deployment (networking, SSH, node configurations)
Compute PoolsKubernetes clusters where applications run (dedicated or shared modes)

Applications

These resources define what gets deployed to your infrastructure.

ConceptDescription
Profile BundlesReusable packages combining Palette Cluster Profiles and PaletteAI Workload Profiles
App DeploymentsDeploy AI/ML applications to Compute Pools

Deployment Flow

A typical deployment flow involves these concepts working together:

  1. Platform setup - A platform engineer creates a Tenant with Settings for Palette integration
  2. Project creation - Teams get Projects with GPU quotas and role-based access
  3. Infrastructure provisioning - Compute discovers available machines; Compute Pools provision Kubernetes clusters
  4. Application packaging - Profile Bundles package infrastructure and application configurations
  5. Deployment - Data scientists create App Deployments using Profile Bundles on Compute Pools

For a deeper look at the system architecture, refer to our Architecture guide.