AI & Data Architecture

Your CS Team Isn't Ready for AI — Your Data Architecture Is the Problem

The NRR Partners Team 6 min read

Every CS leader I talk to wants AI. Automated risk scoring. Account synthesis. Predictive churn. The pitch writes itself and the tools exist. Claude, Gemini, GPT — they're all capable of doing the analysis.

So why isn't it working yet?

Because AI for Customer Success is a data problem before it's a technology problem. You can have Claude Enterprise sitting in your org right now and it won't help if the data feeding it is incomplete, siloed, stale, or structurally incoherent.

Before you buy another tool or hire an AI consultant, audit your data architecture against these five layers. If any one of them is broken, AI will produce confident-sounding garbage.

Layer 1: Identity Resolution

Can you answer this question instantly: "For customer X, show me every data point we have across every system"?

If the answer involves cross-referencing Salesforce Account IDs with your CS platform's Company IDs with your support tool's Organization IDs with your product's tenant IDs — you have an identity resolution problem.

What this looks like in practice: A CSM asks "what's going on with Acme Corp?" and has to check Salesforce for the revenue picture, Catalyst or Planhat for health scores, Zendesk for recent tickets, and the product dashboard for usage. Each system has its own version of "Acme Corp" and they don't always agree on spelling, hierarchy, or which contacts belong where.

What AI needs: A single, canonical company record that every system maps to. In Planhat, this is the Company object, and every other model (Licenses, End Users, Conversations, Assets) links back to it. If you're using external IDs to map Salesforce Accounts to Planhat Companies, those mappings need to be 100% complete. One unmapped account is one account invisible to every automation you build.

The fix: Run an audit. Pull your Company list from Planhat. Pull your Account list from Salesforce. Pull your organization list from your support tool. Match them. Find the gaps. Fix them. This is unglamorous work and it's the most important thing you'll do.

Layer 2: Behavioral Signal Capture

AI can only analyze behavior it can see. If product usage data isn't flowing into your CS platform, the intelligence layer has no behavioral signal to work with. It's trying to predict churn based on financials and CSM gut feel alone.

The signals that matter:

SignalSourceWhy It Matters
Login frequencyProduct telemetryDeclining logins = declining engagement
Feature adoptionProduct telemetryAre they using what they're paying for?
Support ticketsZendesk/IntercomVolume, severity, and sentiment trends
NPS/CSATSurvey toolStated satisfaction (lagging but useful)
Stakeholder engagementCRM activityAre they taking your meetings?
Contract/License dataBilling or CRMUpcoming renewals, seat count changes
Email/chat sentimentConversational AIAre interactions trending negative?

In Planhat specifically: Usage data can come in via Segment integration, Planhat's tracking script, API pushes, or data warehouse sync (Snowflake/BigQuery). Planhat stores this as time-series Metrics on the Company or End User model. These Metrics power Calculated Metrics — the engine behind any meaningful health score.

If your product team isn't sending usage events to your CS platform, that pipeline is job number one. Everything downstream depends on it.

Layer 3: Temporal Consistency

Health scoring and AI risk assessment need time-series data, not just point-in-time snapshots. The question isn't "what's this customer's usage today?" It's "how has this customer's usage changed over the last 30, 60, 90 days?"

The common failure mode: A health score built on current-state fields. Usage above threshold? Green. Below? Red. That tells you nothing about trajectory. A customer at 90% utilization (green) who was at 120% utilization 60 days ago is declining. A customer at 50% utilization (yellow) who was at 20% utilization 60 days ago is ramping.

What AI needs: Consistent, granular time-series data with enough history to detect patterns. In Planhat, this means:

  • Custom Metrics ingested daily or weekly, not monthly
  • Calculated Metric templates configured for trailing windows (7-day, 30-day, 90-day averages)
  • The "Deviations from Normal" ML template enabled — this uses Planhat's built-in machine learning to detect when a metric deviates significantly from a customer's historical baseline, without you having to define the threshold manually

If your data is point-in-time only, or it updates monthly, or it has gaps, the AI will hallucinate trends that don't exist. Fix the cadence before you build the model.

Layer 4: Relationship Graph

Who are the humans involved in each account, what roles do they play, and how engaged are they?

Most CS platforms store contacts as a flat list. Planhat's End User model is more nuanced — each End User has a Relevance Score (0-100, calculated automatically via ML based on activity level and interactions), plus you can track role, engagement cadence, and sentiment at the individual level.

Why this matters for AI: The single biggest predictor of churn that most health scores miss is stakeholder change. When a champion leaves, the account is at risk regardless of what the usage numbers say. When a new VP joins a customer org and has no relationship with your team, that's a window of vulnerability.

AI can detect these signals if the data exists:

  • End User records updated when contacts join or leave
  • Activity data (emails, meetings, calls) logged against specific contacts
  • Sentiment scored at the End User level via conversational AI

If your CSMs are logging activities against the Company but not against specific contacts, the relationship graph is invisible and the AI can't assess champion risk.

Layer 5: Feedback Loop Infrastructure

The data architecture isn't done when you build it. It needs a feedback loop: are the AI outputs actually correct? When the model says an account is at risk, does it actually churn? When it says an account is healthy, does it actually renew?

In Planhat, this means:

  • Track every prediction against the actual outcome
  • Use License Status changes (active → churned) as ground truth labels
  • Review Calculated Metric accuracy quarterly: are the leading indicators actually leading?
  • Retrain your scoring weights based on results

Planhat's Health Score supports multiple dimensions with configurable weights. Those weights should change over time as you learn which signals are predictive for your specific customer base.

The Practical Audit

Before you spend a dollar on AI for CS, answer these questions:

  1. Can you map every customer to a single canonical record across all systems?
  2. Is product usage data flowing into your CS platform automatically, daily?
  3. Do you have at least 90 days of time-series data for key behavioral metrics?
  4. Are contacts tracked with engagement data at the individual level, not just the company level?
  5. Do you have a mechanism to measure prediction accuracy after the fact?

If you answered "no" to any of these, fix the data before you build the AI. It will save you months of rework and ensure that when you do turn on the intelligence layer, it actually works.

The companies getting real value from AI in CS aren't the ones with the fanciest models. They're the ones who did the data plumbing first.

The NRR Partners team helps B2B SaaS companies build predictive health scoring, AI-powered CS automation, and revenue intelligence. Previously at Planhat. Based in New York, Paris, and Dubai.

The CS Ops Playbook

Five chapters. One system. Read them in order or jump to what matters most.