API Reference
Base URL:https://api.gocloudera.com/api
All endpoints require authentication via Bearer token (JWT) or API key (X-API-Key header). All responses return JSON.
Authentication
Login
Refresh Tokens
List Available Tenants (Public)
GPU Instances
List Instances
| Param | Type | Description |
|---|---|---|
state | string | Filter by state: running, stopped, terminated |
instance_type | string | Filter by instance type (e.g., p3.2xlarge) |
gpu_type | string | Filter by GPU type (e.g., V100, A100) |
Get Instance Details
Get Instance Metrics
| Param | Type | Default | Description |
|---|---|---|---|
hours | int | 24 | Lookback period in hours |
limit | int | 1000 | Max data points |
Get Idle Instances
Start / Stop Instance
Costs
Get Cost Data
| Param | Type | Default | Description |
|---|---|---|---|
days | int | 30 | Lookback period |
service | string | - | Filter by service |
instance_type | string | - | Filter by instance type |
start_date | date | - | Start date (YYYY-MM-DD) |
end_date | date | - | End date (YYYY-MM-DD) |
Get Cost Summary
Get Cost Trends
Get Budget Status
Get Costs by Tag
| Param | Type | Required | Description |
|---|---|---|---|
tag_key | string | Yes | Tag key to group by (e.g., environment, team) |
AI Spend (LLM Cost Tracking)
Get AI Spend
| Param | Type | Description |
|---|---|---|
spend_type | string | inference, training, fine_tuning, embedding |
provider | string | openai, anthropic, aws_bedrock, azure_openai |
model_name | string | Filter by model (e.g., gpt-4, claude-3-opus) |
workload_id | int | Filter by AI workload |
project_id | string | Filter by project |
Get AI Spend Summary
Get Unit Economics
Get Spend by Dimension
| Param | Type | Description |
|---|---|---|
dimension | string | project_id, team_id, cost_center, business_unit |
Get Budget Burn Rate
Alerts
List Alerts
| Param | Type | Description |
|---|---|---|
status | string | active, resolved, ignored |
alert_type | string | Filter by type |
instance_id | string | Filter by instance |
limit | int | Max results (default 50) |
Get Alert Summary
Resolve / Ignore Alert
Acknowledge Alert (Stops Escalation)
Alert Rules
List Rules
| Param | Type | Description |
|---|---|---|
metric | string | Filter by metric type |
enabled | boolean | Filter by enabled state |
severity | string | Filter by severity |
page | int | Page number |
limit | int | Items per page |
Create Rule
gpu_utilization, cpu_utilization, memory_utilization, daily_cost, hourly_cost, temperature, error_rate
Supported Operators: gt, lt, gte, lte, eq, not_eq
Scope Options:
all— monitors all instancestagged— monitors instances matching a tag filterspecific_instance— monitors specific instance IDs
Get Rule Trigger History
Update / Delete Rule
Enforcement Policies
List Policies
Get Policy Templates
Create from Template
Simulate Policy (Dry Run)
Create Policy
Toggle Policy
Notification Channels
List Channels
Create Channel
Test Channel
Maintenance Windows
Create Window
Get Active Windows
Exports
Export any data as CSV or PDF.Data Sync (Agent Endpoint)
Sync Data from Agent
Error Responses
All errors follow this format:400— Bad request (missing/invalid parameters)401— Unauthorized (invalid or missing token)403— Forbidden (insufficient role)404— Resource not found429— Rate limited500— Internal server error
Rate Limits
API rate limits are configurable per tenant (default: 1000 requests/minute). Rate limit headers are included in responses:X-RateLimit-Limit— Max requests per windowX-RateLimit-Remaining— Remaining requestsX-RateLimit-Reset— Window reset timestamp