FoundationsBuild budget and forecastModule 1 of 4
Guides
March 6, 2026
By Andrew Day

Cloud and AI Budget Health Check

Score your current budget process and identify the next 30 days of improvements. A practical maturity scorecard for teams that want confidence, not perfection.

Share this post

Send it to someone managing cloud or AI spend.

LinkedInX

Use this when you have a budget but are not sure whether it is good enough — or when a board question, an invoice surprise, or a new planning cycle makes you want to check.

The fast answer: score your budget process on four dimensions, then pick the next 30 days of improvements from the gaps. Aim for confidence, not perfection.

What "budget health" actually means

A healthy budget is not one that is accurate to the dollar. It is one that gives you enough confidence to make decisions.

Specifically, a healthy budget should let you answer three questions on any given day:

  1. Is today's spend in line with what we planned?
  2. Where is month-end pace heading?
  3. If something moves unexpectedly, can we explain it and correct it quickly?

If you can answer all three questions without checking more than one or two places, your budget is healthy. If answering any one of those requires a 20-minute spreadsheet exercise or a conversation across three teams, there is room to improve.

The goal of this health check is not to identify everything that could theoretically be better. It is to identify the one or two gaps that are causing the most friction right now, and fix those first.

Budget maturity scorecard

Score each dimension from 1 (weak) to 5 (strong). Be honest. The goal is to find the biggest gap, not to feel good about where you are.

1. Budget foundation

Score What it looks like
1 No written budget. Spend is reviewed after the fact.
2 Single number budget. No breakdown by category or provider.
3 Budget by category (inference, compute, storage, networking). Reviewed monthly.
4 Budget by category and provider. Updated when assumptions change.
5 Budget by category, provider, and feature. Tied to product and growth assumptions.

What each level feels like in practice:

A score of 1 or 2 means the team is constantly surprised. Invoices arrive and the numbers are higher or lower than expected with no clear explanation. Someone manually reconciles the bill after the fact but has no way to predict what next month will look like.

A score of 3 means the team has a framework but limited precision. They know inference is the biggest cost center, and they have a rough number in mind for the month, but they cannot explain a 15 percent variance without digging.

A score of 4 or 5 means the team can explain cost changes quickly. If OpenAI spend jumped, they can point to a specific feature or model tier. If the forecast moves, they understand why.

Why teams get stuck between 2 and 3: The jump from a single budget number to a category-level budget requires categorizing historical spend. Most teams avoid this because it feels like a large upfront investment. In practice, it takes about four hours the first time and becomes much easier to maintain afterward. The payoff — in reduced investigation time and better forecast confidence — usually exceeds that investment within the first month.


2. Forecasting quality

Score What it looks like
1 No forecast. Month-end is a surprise.
2 Last month plus a guess. No model.
3 Simple baseline + trend. Updated weekly.
4 Base, growth, and stress cases. Updated as the month progresses.
5 Forecast by workflow with confidence ranges. Compared to actual and refined monthly.

What each level feels like in practice:

A score of 1 or 2 means every month-end invoice is treated as new information. The team cannot tell leadership what to expect in advance, and any significant variance requires a reactive explanation.

A score of 3 means the team can usually give a reasonable month-end estimate, but it is based on momentum rather than structure. When something changes mid-month, the forecast takes a while to catch up.

A score of 4 or 5 means the team treats the forecast as a living document. It is updated weekly based on actual pace, and there is a stress case that the team can point to when a conversation with finance or leadership goes to worst-case scenarios.

Why teams get stuck at 3: Moving from a trend-based estimate to a structured three-case forecast feels like more work than it is. The base, growth, and stress cases can usually be built in two hours from historical spend and a few growth assumptions. The discipline of updating them weekly takes about 15 minutes. Most teams that try it find that the stress case alone is worth the effort — it prevents last-minute surprises from escalating into difficult leadership conversations.


3. Review process

Score What it looks like
1 No regular review. Ad hoc when someone asks.
2 Monthly only. Explain after the fact.
3 Weekly review with pacing and top deltas.
4 Weekly review with actions, owners, and escalation path.
5 Weekly review plus monthly reconciliation. Decisions logged. Forecast accuracy tracked.

What each level feels like in practice:

A score of 1 or 2 means cost visibility is reactive. The team looks at cost when something goes wrong, not as a regular practice. This means problems compound for weeks before they are noticed.

A score of 3 means the team has a cadence but limited accountability. They meet weekly and look at the numbers together, but often leave without clear owners for follow-up.

A score of 4 or 5 means the review produces decisions. Every meeting ends with named owners and due dates. The decision log is the starting point for the following week's meeting.

Why teams get stuck at 3: The jump from reviewing to deciding is a cultural one, not a process one. A review that produces no actions is still useful as awareness. Adding the discipline of "this meeting ends with one to three named owners" takes one sentence added to the standing meeting invite. The discipline compounds quickly — within a month, the team builds a backlog of completed cost improvements that would otherwise never have been prioritized.


4. Tooling and visibility

Score What it looks like
1 Provider invoices and dashboards only. No unified view.
2 Spreadsheet or manual aggregation. Updated weekly.
3 Daily spend visibility. Single view across providers.
4 Daily spend, forecast vs budget, category breakdowns. Alerts for anomalies.
5 Full stack: daily visibility, forecast, alerts, category analysis, feature attribution. Shared review surface.

What each level feels like in practice:

A score of 1 or 2 means the team spends significant time preparing for any cost conversation. Pulling together the multi-provider view requires opening multiple dashboards, exporting CSVs, and doing manual aggregation. The view is always slightly stale by the time it is ready.

A score of 3 means someone can answer "what did we spend yesterday?" in under 30 seconds. This single capability changes how the team relates to cost — it shifts from a monthly accounting exercise to a daily operating signal.

A score of 4 or 5 means anomalies are surfaced automatically before someone notices them manually. Forecast vs budget is visible without calculation. The shared surface means any team member can check cost status without needing a designated "cost person."

Why teams get stuck at 2: The manual spreadsheet is good enough until it is not. Teams tolerate it until either a large invoice surprise or a new hire who asks "do we have a real dashboard for this?" motivates the upgrade. The investment in daily visibility is usually smaller than teams expect.


What your total score means

Total (out of 20) Maturity Next focus
4–8 Early Get daily visibility and a simple budget by category
9–12 Developing Add forecasting and weekly review
13–16 Solid Strengthen attribution, stress cases, and action ownership
17–20 Mature Refine accuracy, automate more, extend to feature-level

A worked example: running the scorecard

Here is what running this scorecard looks like for a real team.

Context: A 15-person startup that has been using OpenAI and AWS for eight months. Monthly AI and cloud spend is around $11,000. The team has started noticing that invoices vary more than expected and a board member recently asked for a forecast.

Scoring:

  • Budget foundation: 2 — They have a rough monthly number in mind but no category breakdown. OpenAI, AWS EC2, and S3 are all lumped into one line.
  • Forecasting quality: 1 — Month-end is essentially a surprise. The last two months closed 20 percent and 12 percent above what the team expected.
  • Review process: 2 — They look at the bills monthly when invoices arrive, but no one owns the review and nothing actionable comes out of it.
  • Tooling and visibility: 2 — OpenAI dashboard for API spend, AWS cost explorer for cloud. No unified view. Preparing for a cost conversation takes 30 minutes.

Total: 7 out of 20 — Early maturity.

30-day focus: The engineering lead and a senior engineer spend one afternoon building a category-level budget (inference, compute, storage, networking). They identify that OpenAI inference is 62 percent of total spend — significantly higher than they assumed. They set up a shared dashboard that pulls both providers into one view. They schedule a recurring 20-minute weekly slot to review it together.

By the end of the month, they have answered the board question with a three-case forecast, and the engineering lead can check yesterday's spend in under a minute each morning.

New scores at end of month:

  • Budget foundation: 3 — Category breakdown exists and is reviewed weekly
  • Forecasting quality: 2 — Still last-month-plus-a-guess but the guess is now informed by category data
  • Review process: 3 — Weekly review running with pacing and top deltas
  • Tooling and visibility: 3 — Daily spend visible in one place

Total: 11 out of 20 — Developing.

One month of focused work moved the team from 7 to 11. The next month's focus would be adding structured forecasting and action ownership to the weekly review.


30-day improvement plan

Pick one improvement per dimension where you scored lowest. Do not try to fix everything at once — compounding small improvements is more reliable than a large-scale transformation.

If foundation is weak (1–2)

  • Week 1: Write a budget by category (inference, compute, storage, networking).
  • Week 2: Add provider breakdown for top 3 providers.
  • Week 3: Tie budget to last month's actual plus a growth assumption.
  • Week 4: Share the budget with at least one stakeholder and get feedback.

If forecasting is weak (1–2)

  • Week 1: Create a simple forecast (last month + trend). Accept that it will be imprecise.
  • Week 2: Add a growth case and stress case. They do not need to be complex.
  • Week 3: Update forecast weekly as the month progresses and compare to actual daily pace.
  • Week 4: At month-end, compare forecast to actual and record the biggest source of variance.

If review process is weak (1–2)

  • Week 1: Schedule a recurring 20-minute weekly review with the engineering owner.
  • Week 2: Add a named agenda: pacing, top deltas, any anomalies, one action.
  • Week 3: Log at least one decision per review, even if the decision is "no action needed."
  • Week 4: Add monthly reconciliation to the calendar.

If tooling is weak (1–2)

  • Week 1: Get daily spend visibility for your largest provider.
  • Week 2: Add a second provider or category to the view.
  • Week 3: Set one anomaly or forecast alert.
  • Week 4: Create a shared dashboard or report accessible to both engineering and product.

Common gaps — and why they happen

Gap Why it happens Fix
Budget exists but is never updated Created once for a planning exercise, nobody owns updating it Tie budget to a monthly review. Assign one person as owner.
Forecast is a single number No one has taken the time to model scenarios Add base, growth, and stress cases. Show a range rather than false precision.
Review has no actions Meeting is structured as an update, not a decision End every review with 1–3 named actions. Put them in the shared log.
Visibility is provider-by-provider Each provider has its own dashboard, unified view was never built Use a cross-provider view for total cost and category breakdown.
No one owns cost Cost was everyone's responsibility, which means no one's Assign a named engineering or platform owner for the weekly review.
Category breakdown feels unnecessary Teams assume they know where cost goes Run the breakdown once — most teams discover that one category is larger than expected.

How StackSpend helps

StackSpend supports budget maturity by providing:

  • cross-provider daily visibility
  • category-based analysis
  • daily forecast vs budget tracking
  • anomaly alerts
  • a shared surface for weekly and monthly review

See cloud + AI cost monitoring.

What to do next

Share this post

Send it to someone managing cloud or AI spend.

LinkedInX

Know where your cloud and AI spend stands — every day.

Connect providers in minutes. Get 90 days of visibility and start receiving daily cost updates before the invoice lands.

14-day free trial. No credit card required. Plans from $19/month.