✗
Can't place themselves on the AI stack. If they don't know whether they're infrastructure, horizontal, vertical, or agentic, they don't have a thesis — they have a product.
✗
Per-seat pricing with no roadmap to outcome pricing. Signals they're pricing against IT budgets, not services budgets. The TAM ceiling is 10× lower than it needs to be for the valuation.
✗
No measurable ROI metric for customers. If the customer can't explain what it saved them in hours or dollars, churn is a question of when, not if.
✗
100% API-dependent, no fine-tuning, no proprietary pipeline. A prompt wrapper. If OpenAI adds one feature, the company evaporates. This is the "Jasper AI" failure mode.
✗
Demo works; production is unverified. Ask to run your own data through the system live. If they resist, that's the answer. Devin benchmarked at 14% real-world task completion despite a viral demo.
✗
Gross margin below 50% after inference. Structurally dangerous. Compute costs will compress further at scale before they improve. Model provider pricing power over COGS is the hidden risk.
✗
No domain expert on the founding team for a vertical play. AI engineers alone can't identify the right problem to solve in legal, medical, or finance. The founders should have sat in the customer's chair.
✗
TAM = the entire industry. "Legal AI is a $500B market" is not a TAM. The TAM is their reachable segment at their current pricing model. Inflated TAM = they haven't done the bottoms-up.
✗
Customer data used for training without explicit opt-in. Regulatory liability and enterprise deal-killer. Enterprise customers at the major model providers are contractually firewalled — this should be standard.
✗
Open-ended "general agent" pitch with no production deployments. Narrow agents work now. General agents do not. If they can't name 5 specific task types their agent handles reliably, it's still a demo.
✗
NRR below 80%. Churn is outpacing expansion. In an AI product that's supposed to get stickier as it learns customer workflows, declining NRR is a signal the learning isn't happening.
✗
"AI" used 20+ times in the deck with no technical depth. Buzzword density inversely correlates with substance. Ask the technical question they haven't prepared for: "Show me your eval framework."