Success Criteria Extraction
Use Claude to extract and generate measurable success criteria from discovery call transcripts, deal context, and product capabilities
Instructions
Success Criteria Extraction
Given a discovery call transcript, deal context, and product capabilities, use the Claude API to extract prospect-stated goals and generate specific, measurable success criteria. The output is a structured set of criteria the prospect can agree to, each with a metric, target, timeline, and measurement method.
API Call
POST https://api.anthropic.com/v1/messages
Headers:
x-api-key: {ANTHROPIC_API_KEY}
anthropic-version: 2023-06-01
content-type: application/json
Body:
{
"model": "claude-sonnet-4-20250514",
"max_tokens": 3000,
"temperature": 0.2,
"messages": [{
"role": "user",
"content": "You are a success criteria analyst for a B2B sales process. Analyze the following inputs and generate specific, measurable success criteria that the prospect can agree to.\n\nDISCOVERY TRANSCRIPT:\n{transcript}\n\nDEAL CONTEXT:\n- Company: {company_name}\n- Industry: {industry}\n- Headcount: {headcount}\n- Current tools: {current_tools}\n- Deal size: {deal_size}\n- Champion: {champion_name}, {champion_title}\n\nPRODUCT CAPABILITIES:\n{product_capabilities}\n\nHISTORICAL SUCCESS RATES (similar deals):\n{historical_success_json}\n\nExtract and generate success criteria. For each:\n1. The prospect's stated goal (use their exact words when possible)\n2. A specific metric to measure it\n3. A quantified target (use the prospect's number if stated; otherwise derive a conservative estimate from their pain description)\n4. Timeline to achieve it\n5. Who measures it (prospect-side stakeholder)\n6. Baseline value (current state, if mentioned or estimable)\n7. Achievability score (0-100): how likely this criteria is to be met based on historical data and product capabilities. Score 80+ = safe to commit. Score 50-79 = achievable with effort. Score <50 = risky, flag for expectation management.\n8. Category: efficiency, revenue, cost_savings, quality, time_to_value, adoption\n\nReturn JSON:\n{\n \"success_criteria\": [\n {\n \"prospect_goal\": \"\",\n \"metric\": \"\",\n \"target\": \"\",\n \"timeline\": \"\",\n \"measured_by\": \"\",\n \"baseline\": \"\",\n \"achievability_score\": 0,\n \"category\": \"\",\n \"source_quote\": \"\",\n \"measurement_method\": \"\"\n }\n ],\n \"missing_info\": [\"list of things the prospect didn't mention that would strengthen the criteria\"],\n \"risk_flags\": [\"any criteria that seem unrealistic or unmeasurable\"],\n \"recommended_follow_up_questions\": [\"questions to ask in the next call to refine criteria\"]\n}"
}]
}
Input Requirements
transcript: Discovery or follow-up call transcript from Fireflies or Gong (usefireflies-transcriptionorgong-call-recordingfundamental)company_name,industry,headcount,current_tools,deal_size,champion_name,champion_title: From Attio deal recordproduct_capabilities: A JSON summary of what your product can measurably deliverhistorical_success_json: Aggregated data from past deals showing which criteria were achieved and at what rates (query from Attio)
Output
JSON with structured success criteria. Each criterion includes an achievability score that guides whether to commit to it in the mutual success plan.
Post-Processing
- Store each criterion in Attio as a custom attribute on the deal record
- Fire PostHog event
success_criteria_extractedwith properties:deal_id,criteria_count,avg_achievability_score,categories - Flag any criterion with
achievability_score < 50for founder review before including in the mutual success plan
Guardrails
- Never generate success criteria the product cannot measure. Every criterion must have a
measurement_methodthat maps to an observable data point. - Never inflate achievability scores. When in doubt, score lower — it is better to under-promise.
- If the transcript contains fewer than 2 identifiable goals, return
missing_infonoting that additional discovery is needed before criteria can be defined. - Cap the number of criteria at 5. If more goals are identified, rank by achievability and prospect emphasis, and include the top 5.
Alternatives
| Tool | Method | Notes | |------|--------|-------| | Claude (Anthropic) | Messages API | Best structured extraction quality | | GPT-4 (OpenAI) | Chat Completions API | Alternative LLM | | Gemini (Google) | Generative AI API | Alternative LLM | | Gong AI | Deal insights | Built-in extraction, less customizable | | Manual | Spreadsheet template | Fallback for enterprise deals requiring human judgment |