Working draft for Thu 2026-04-30 demo. Two parts: "Walk Us Through Your Process" answer + questions for the panel.
Built AI Enrichment flows on OpenAI per the brief. Active toggle green, scheduled, triggered manually — History Logs stayed empty. No error message. Spent ~45 minutes cycling through models, context fields, scheduled vs. trigger-based, before trying Azure. gpt-5.3-chat worked the first try.
That experience IS Signal 2 ("not sure if this is a data quality issue or if we set something up wrong") — but lived as the candidate. With Claude as a sounding board, no production data on the line, and time to brute-force my way through configurations. The customer with $185K expansion pending doesn't have that runway.
The 6 linked support PDFs are 2–5 years old — including the History Logs article, the one explicitly recommended for verifying AI enrichment output. References to features that aren't in the dropdown (Perplexity Sonar). The model lineup in the docs doesn't match the UI.
So I learned by clicking. Some of it was figure-out-able with patience (the AI Enrichment step config panel, Trigger Settings). Some of it was hostile — the Scheduled Flow row showing "Deleted" status while the Active toggle stayed green, or field dropdowns offering 100+ options with 7 different "Account Name" variants and no contextual hints.
Same opacity Signal 6 describes ("we're spending a lot of time explaining classifications we can't fully explain ourselves") — visible upstream, in the docs themselves.
Custom fields didn't appear in list view pickers until I hard-reloaded. The AI Enrichment configuration panel had loading delays. Field dropdowns rendered slowly with the long option lists.
Each is minor in isolation. Stacked with the cognitive overload of 100+ option dropdowns, every small lag is one more chance to fat-finger an option I didn't mean to pick and not realize for 20 minutes.
I used Claude throughout as a thinking partner. The dynamic that worked: I brought directionality, Claude brought comprehensiveness and pushback.
Three concrete examples from this prep:
What I deliberately didn't outsource: which observations matter, what the recommendation should be, what to scope in vs. defer. Claude is good at more — more options, more risk angles, more comparisons. I'm still the one deciding what's load-bearing.
The failure mode I tried to avoid: letting AI generate the answer and then editing it. That produces something polished but not mine, and in a panel environment, you can hear it.
I expected to find AI Enrichment under "Complete Clean" (data quality) or near "Internal Match" (entity classification). Instead it lives inside "Assignment Flows" — a category I associate with routing and territory work.
A RevOps user thinking "I need to clean up the industry classifications on my accounts" might not look in Assignment Flows for this. Was this a deliberate IA choice, or has the product evolved past its original menu structure?
The Integrations panel surfaces Anthropic alongside OpenAI, Perplexity, and Google. I added an API key, then went to configure an AI Enrichment flow. The provider dropdown only accepts OpenAI and Azure auth. So the Integrations panel advertises Anthropic, but the AI Enrichment auth handling doesn't actually accept it.
That same shape — UI advertising more than the product actually delivers — showed up a few other places this week:
My actual question is the pattern, not Anthropic specifically: does the gap between what the UI advertises and what the product actually delivers map to what Signal 2's customer is reacting to ("not sure if this is a data quality issue or if we set things up wrong") — and is that a connection the team is actively working through?
If Anthropic specifically is on the roadmap, Claude has a strong enterprise/compliance posture (relevant to Signals 1, 3, 5, 8), Bedrock could unlock AWS-shop customers, and it tends to perform well on structured classification — what has the trade-off looked like internally?
Q&A handling for AI usage questions:
If asked organically about AI use (likely from any of the three panelists, especially given the brief's emphasis), the directionality-vs-comprehensiveness framing transfers as a one-liner: "I bring the directionality, Claude brings comprehensiveness and pushback. The decisions are still mine."
If pressed on specifics ("show me an example"), have one ready that's small enough to land in 30s:
- The risk-surfacing on the feedback loop is the cleanest — I had the "this is a customer asset" insight, Claude listed five risks I hadn't considered, I kept four (cold start, bias, GDPR, messaging) and discarded one as out-of-scope. That's the dynamic in miniature.
If asked something pointed like "how do we know this isn't just Claude's output?" — answer with what was deliberately NOT outsourced: which observations matter, what the recommendation is, what to defer. Plus the sandbox time was real and unprompted — Claude wasn't with me when I was clicking through field dropdowns at 11pm Tuesday.