January 19

Tags

A simple framework for deciding where AI fits in your marketing

Marketers do not need more inspiration about what AI can do. They need structure for deciding what AI should do in their specific context.

Most AI marketing content focuses on possibilities: impressive demonstrations, ambitious use cases, transformational potential. This creates excitement but rarely creates clarity.

What is missing is a practical, reusable framework that helps marketing teams evaluate specific AI opportunities against their actual constraints, capabilities, and strategic priorities.

This blog is part of a practical guide to making sense of AI, automation and agentic marketing as one connected change, rather than three separate problems. Its purpose is to provide that decision framework – one that works whether you are evaluating your first AI pilot or your fifteenth, whether you are choosing between vendor options or deciding whether to adopt at all.

Why structure matters more than enthusiasm

AI adoption without decision structure follows a predictable pattern:

Month 1: Excitement. Vendor demonstrations. Budget allocated. Pilot approved.

Month 3: Confusion. Outputs are inconsistent. Integration is harder than expected. Value is unclear.

Month 6: Disappointment. Pilot is quietly discontinued. Team concludes “AI does not work for us.”

The failure is rarely technical. The failure is that no one systematically evaluated whether the conditions for success existed before committing resources.

Example: A B2B software company piloted AI for content generation. After 4 months and £40,000, the pilot was abandoned.

Post-mortem revealed:

  • No clear definition of what “success” meant beyond “using AI”
  • Brand voice was undefined, making consistency impossible to evaluate
  • Content workflow had no review process for AI outputs
  • No one was assigned to own the system or optimise its performance
  • Data quality problems (incomplete customer research, outdated personas) meant AI had poor inputs

Every problem was foreseeable. None were identified before launch because no structured evaluation occurred.

What would have worked: Systematically assessing readiness across key dimensions before committing. The framework below provides that systematic assessment.

The AI Decision Framework: Four core questions

Before adopting AI, automation, or agentic systems for any marketing use case, evaluate using four foundational questions. Each question surfaces conditions that determine success or failure.

Question 1: Does this improve decision quality, or just decision speed?

Why this matters: Speed without quality erodes value. Many AI failures come from optimizing for the wrong dimension.

How to evaluate:

A. Identify the core decision being made

Not “what does this tool do?” but “what choice does this help us make better?”

  • Poor clarity: “We want AI for email marketing”
  • Better clarity: “We need to decide which subject lines will drive engagement for specific audience segments”

B. Determine whether speed or quality is the constraint

  • If quality is sufficient but speed is limiting → AI can add value
  • If quality is insufficient → AI may amplify poor quality faster
  • If both are problems → fix quality first, then add speed

C. Test the improvement hypothesis

Ask: “Will AI make this decision materially better, or just faster?”

Decision speed is valuable when:

  • Opportunity windows are short (paid search bidding, real-time personalisation)
  • Volume is high and manual processing is impractical (hundreds of A/B tests)
  • Delay creates competitive disadvantage (responding to market shifts)

Decision quality is valuable when:

  • Current decisions are suboptimal due to limited analysis capacity
  • Patterns exist in data that humans cannot detect at scale
  • Consistency across many decisions improves outcomes

Neither speed nor quality improvements justify adoption when:

  • Current decisions are already good enough
  • AI introduces new risks that outweigh gains
  • Organizational readiness is insufficient

Example evaluation:

Use case: AI to generate social media post variations

Core decision: Which messages will resonate with specific audience segments?

Current constraint: Team can create 3-5 variations manually; limited testing means suboptimal message-market fit.

Improvement hypothesis: AI can generate 20+ variations, enabling more extensive testing and better message optimisation.

Evaluation: Quality improvements through testing volume justify adoption. Speed is secondary benefit.

Decision: Proceed, but focus success metrics on message performance improvement, not just production speed.

Use case: AI to generate quarterly earnings announcement blog post

Core decision: How to communicate financial results in brand voice with appropriate context and stakeholder sensitivity.

Current constraint: No meaningful constraint. Current process produces high-quality, well-received content.

Improvement hypothesis: AI could produce drafts faster.

Evaluation: Speed improvement is marginal (quarterly task). Quality risk is significant (earnings communication requires judgment, context, sensitivity).

Decision: Do not proceed. Human-led process is already effective for this low-frequency, high-stakes communication.

Question 2: Is scale genuinely essential, or just appealing?

Why this matters: “We could do this at greater scale” is seductive but often misleading. Scale without corresponding value is waste, not progress.

How to evaluate:

A. Distinguish between volume that creates value and volume that creates noise

  • Value-creating scale: Personalisation that improves customer experience and conversion
  • Noise-creating scale: Generic content published across more channels without strategic purpose

B. Assess whether current volume is actually a constraint

Ask: “Are we limited by inability to produce volume, or by unclear strategy about what volume serves?”

If strategy is unclear, adding volume amplifies confusion.

C. Calculate whether scale economics justify complexity

More volume through AI introduces:

  • System setup and maintenance costs
  • Monitoring and quality assurance overhead
  • Explanation and governance workload

Does value from additional volume exceed these costs?

Example evaluation:

Use case: AI to create localized landing pages for 200 cities

Scale proposition: Instead of 10 core landing pages, create 200 city-specific variants with local terminology, landmarks, and context.

Evaluation questions:

  • Do we have traffic volume across 200 cities to justify this? (Check: Most traffic comes from 12 cities)
  • Will localization actually improve conversion? (A/B test: Localized pages convert 8% better)
  • Can we maintain quality at 200x scale? (Check: AI-generated content needs 20% human editing time)
  • Is this better than improving our 12 core pages? (Economic analysis: Improving 12 pages has higher ROI)

Decision: Do not proceed with 200-city scale. Instead, create high-quality localized pages for top 12 cities manually, use AI to accelerate research and initial drafting.

Use case: AI to automate A/B testing subject lines at scale

Scale proposition: Test 50+ subject line variations per campaign instead of current 3-5.

Evaluation questions:

  • Do we have email volume to support meaningful testing? (Check: Yes, 100k+ sends per campaign)
  • Will more variations improve performance? (Historical data: Expanding from 2 to 5 tests improved performance 15%)
  • Can we maintain quality at this scale? (Check: All variations generated from approved templates, reviewed before sending)
  • Does learning compound? (Check: Insights from testing inform future campaigns)

Decision: Proceed. Scale is justified by volume, performance gains, and compounding learning.

Question 3: What happens when it goes wrong, and can we recover?

Why this matters: All systems fail eventually. The question is whether failure is catastrophic or manageable.

How to evaluate:

A. Identify plausible failure modes

Not “what could possibly go wrong?” but “what is likely to go wrong given what we know about our systems, data, and organization?”

Common failure modes:

  • AI generates inappropriate content (brand voice, tone, sensitivity)
  • Automation executes based on outdated data or broken integrations
  • Agentic system optimises for wrong goal or games metrics
  • System behaviour drifts over time without visibility
  • Edge cases overwhelm exception handling capacity

B. Assess damage potential

For each likely failure mode, evaluate:

  • Customer impact: Does failure affect customer experience, trust, or satisfaction?
  • Brand impact: Does failure damage reputation or market position?
  • Financial impact: Does failure create direct costs or opportunity costs?
  • Regulatory impact: Does failure create compliance or legal exposure?
  • Recovery time: How quickly can we detect, stop, and fix the problem?

C. Design for graceful degradation

Can the system fail safely, with human failsafes preventing catastrophic outcomes?

Risk assessment matrix:

Failure ImpactRecovery SpeedRisk LevelMitigation Required
HighSlowUnacceptableDo not proceed without major mitigation
HighFastManageableRequire monitoring and quick escalation
LowSlowManageableAcceptable with periodic review
LowFastMinimalProceed with standard oversight

Example evaluation:

Use case: AI-powered customer service chatbot

Likely failure modes:

  • Misunderstands complex customer issues
  • Provides incorrect information about products or policies
  • Responds inappropriately to frustrated customers
  • Fails to escalate when human intervention needed

Impact assessment:

  • Customer impact: High (poor experience, unresolved issues, potential churn)
  • Brand impact: Moderate to high (depending on response quality)
  • Financial impact: Moderate (support costs, potential refunds, churn)
  • Regulatory impact: Low to moderate (depending on industry)
  • Recovery speed: Fast for individual cases (escalate immediately), slow to fix systematic issues (requires retraining)

Risk level: Manageable with strong mitigation

Required mitigation:

  • Human escalation for any negative sentiment detection
  • Strict boundaries on what chatbot can claim or promise
  • 100% logging and review queue for quality assurance
  • Named owner monitoring daily
  • Easy override for customers (“speak to human” always available)

Decision: Proceed with comprehensive guardrails, or start with even narrower use case (FAQs only, complex issues immediately escalate).

Use case: AI to generate investor relations communications

Likely failure modes:

  • Inaccurate financial information
  • Inappropriate tone or positioning
  • Failure to account for market sensitivity or regulatory requirements
  • Missing context that management would include

Impact assessment:

  • Customer impact: N/A (investors, not customers)
  • Brand impact: Catastrophic if errors occur
  • Financial impact: Potentially massive (stock price, lawsuits)
  • Regulatory impact: Extreme (SEC violations, legal liability)
  • Recovery speed: Very slow (corrections to public markets are highly visible)

Risk level: Unacceptable

Decision: Do not proceed. Risk-reward ratio does not justify AI autonomy for this use case. Consider AI for research and draft support only, with complete human authorship and review.

Question 4: Who owns the outcome, and is accountability clear?

Why this matters: Ambiguous accountability kills adoption faster than technical failure. When responsibility is unclear, trust evaporates, and systems are abandoned regardless of performance.

How to evaluate:

A. Name the specific individual accountable

Not “the marketing team” or “the AI system.” A named person who can be asked “why did this happen?” and “what are you doing about it?”

B. Verify decision authority matches accountability

The person accountable must have authority to:

  • Modify system parameters
  • Pause or stop the system
  • Escalate issues
  • Make trade-off decisions

If accountability exists without authority, the structure is broken.

C. Define what “ownership” means operationally

Ownership includes:

  • Daily monitoring and anomaly investigation
  • Regular performance review and adjustment
  • Stakeholder communication and explanation
  • Escalation response and problem resolution
  • Strategic alignment and goal adjustment

D. Assess owner capability and capacity

Can the owner actually fulfil these responsibilities given their other commitments and capabilities?

Accountability assessment:

Ownership ElementDefined?Resourced?Action Required
Named individual accountable☐ Yes ☐ No☐ Yes ☐ No
Decision authority established☐ Yes ☐ No☐ Yes ☐ No
Monitoring process exists☐ Yes ☐ No☐ Yes ☐ No
Review rhythm defined☐ Yes ☐ No☐ Yes ☐ No
Escalation path clear☐ Yes ☐ No☐ Yes ☐ No
Owner has capacity☐ Yes ☐ No☐ Yes ☐ No

If any element is “No,” accountability is insufficient for proceeding.

Example evaluation:

Use case: Agentic lead scoring and routing system

Named owner: Sarah Chen, Marketing Operations Manager

Authority verification:

  • Can Sarah modify scoring rules? → Yes, within defined parameters
  • Can Sarah pause the system? → Yes, immediately
  • Can Sarah escalate issues? → Yes, to CMO
  • Can Sarah make trade-off decisions? → Yes, within budget and strategic constraints

Operational ownership:

  • Daily monitoring: Sarah reviews dashboard, 15 minutes daily
  • Weekly review: Sarah conducts performance analysis, 1 hour weekly
  • Monthly reporting: Sarah reports to leadership
  • Escalation response: Sarah investigates and responds within 4 hours

Capacity assessment:

  • Current workload: 75% utilized
  • Additional time required: ~3 hours weekly
  • Conclusion: Capacity exists

Accountability verdict: Clear and adequate

Decision: Accountability structure supports proceeding.

Use case: AI content generation across multiple product lines

Named owner: “The content team” (no individual named)

Authority verification:

  • Can “the team” modify parameters? → Unclear who decides
  • Can “the team” pause the system? → Unclear who has authority
  • Can “the team” escalate issues? → No defined escalation path

Accountability verdict: Insufficient

Decision: Do not proceed until named owner is established with clear authority and capacity.

Integrating the framework: The full evaluation sequence

The four questions work together to provide comprehensive assessment. Use this sequence:

Step 1: Start with Question 4 (Ownership)

If ownership is unclear, stop. Resolve this before evaluating anything else.

No other evaluation matters if no one can be held accountable.

Step 2: Evaluate Question 3 (Failure modes)

If risks are unacceptable or unmitigable, stop.

No amount of opportunity justifies catastrophic risk.

Step 3: Assess Question 1 (Quality vs Speed)

If neither quality nor speed improves meaningfully, stop.

Adoption without clear value is waste.

Step 4: Analyse Question 2 (Scale necessity)

If scale does not create proportional value, reconsider scope.

Smaller, focused application may be more appropriate than broad deployment.

Step 5: Make the decision

Only proceed if all four questions have satisfactory answers:

  • ✓ Ownership is clear and adequately resourced
  • ✓ Risks are acceptable or well-mitigated
  • ✓ Quality or speed improves meaningfully
  • ✓ Scale (if required) creates corresponding value

The framework in practice: Four scenarios

Scenario 1: Email campaign optimisation

Use case: AI-powered send-time optimisation for email campaigns

Question 4 – Ownership:

  • Owner: Email Marketing Manager (named: James Liu)
  • Authority: Can adjust parameters, pause campaigns, escalate issues
  • Capacity: Adequate (monitoring fits existing workflow)
  • Verdict: ✓ Clear ownership

Question 3 – Failure modes:

  • Likely failures: Sub-optimal send times for some segments, edge cases (time zones, holidays)
  • Impact: Low (suboptimal engagement, not brand damage)
  • Recovery: Fast (next campaign corrects)
  • Mitigation: Monitor performance, manual override available
  • Verdict: ✓ Manageable risk

Question 1 – Quality vs Speed:

  • Current: Manual send time selection based on historical averages
  • AI improvement: Individual-level optimisation based on engagement patterns
  • Value: Estimated 5-10% improvement in open rates
  • Verdict: ✓ Quality improvement justified

Question 2 – Scale:

  • Current volume: 200k+ emails monthly
  • Scale benefit: Individual optimisation impossible manually
  • Scale economics: Setup cost justified by volume
  • Verdict: ✓ Scale necessary and valuable

Final decision: ✓ Proceed with pilot

Scenario 2: Automated blog content generation

Use case: AI to generate weekly blog posts with minimal human involvement

Question 4 – Ownership:

  • Owner: “Content team” (no specific individual)
  • Authority: Unclear who can pause or modify
  • Verdict: ✗ Ownership insufficient – STOP HERE

Decision: Do not proceed until specific owner is named with clear authority

(If ownership were resolved, would continue through remaining questions)

Scenario 3: AI-powered pricing optimisation

Use case: Agentic system to dynamically adjust promotional offer pricing

Question 4 – Ownership:

  • Owner: Revenue Operations Director (named: Maria Santos)
  • Authority: Adequate for operations, requires CMO approval for strategic changes
  • Verdict: ✓ Ownership clear

Question 3 – Failure modes:

  • Likely failures: Price wars, margin erosion, brand positioning damage, gaming by customers
  • Impact: HIGH (brand, revenue, competitive position)
  • Recovery: Moderate (requires market repositioning)
  • Current mitigation: Insufficient
  • Verdict: ✗ Risk unacceptable without major mitigation – STOP HERE

Decision: Do not proceed with autonomous pricing. Consider constrained version: AI recommends, humans approve, or AI operates within very tight boundaries (max discount 15%, change frequency limits, segment restrictions)

Scenario 4: Customer journey orchestration

Use case: Agent to coordinate multi-channel nurture sequences

Question 4 – Ownership:

  • Owner: Marketing Automation Manager (named: David Park)
  • Authority: Full operational control, escalates strategic issues
  • Capacity: Will require additional 4 hours weekly (currently has capacity)
  • Verdict: ✓ Ownership adequate

Question 3 – Failure modes:

  • Likely failures: Over-messaging, poor channel selection, context misses
  • Impact: Moderate (customer experience, unsubscribes)
  • Recovery: Fast (individual, immediate pause)
  • Mitigation: Frequency caps, sentiment monitoring, easy opt-out, escalation triggers
  • Verdict: ✓ Risk manageable with guardrails

Question 1 – Quality vs Speed:

  • Current: Manual coordination across 3 channels, limited personalisation
  • AI improvement: Contextual adaptation based on behaviour across all channels
  • Value: Estimated 20-30% improvement in progression to sales-qualified
  • Verdict: ✓ Quality improvement significant

Question 2 – Scale:

  • Current volume: 50,000 leads in active nurture
  • Scale benefit: Manual coordination impossible at this volume
  • Alternative: Rigid automation cannot adapt to context
  • Verdict: ✓ Scale justifies complexity

Final decision: ✓ Proceed with pilot, strong emphasis on guardrails and monitoring

Beyond adoption: Review and refinement framework

AI decisions are not “set and forget.” Successful deployments include ongoing evaluation and adjustment.

Establish three review rhythms:

Daily (or high-frequency) monitoring:

  • Performance dashboard review (5-15 minutes)
  • Anomaly investigation when triggers fire
  • Escalation queue management
  • Purpose: Catch problems fast

Weekly strategic review:

  • Pattern analysis and emerging issues
  • Constraint and parameter adjustment discussions
  • Learning synthesis (what worked, what did not)
  • Purpose: Optimise continuously

Monthly strategic alignment:

  • Goal relevance (are we optimizing for what still matters?)
  • Organizational changes (structure, strategy, market shifts)
  • Scaling or scope adjustment decisions
  • Purpose: Maintain strategic fit

Key questions for ongoing review:

Performance questions:

  • Are we achieving the quality/speed improvements that justified adoption?
  • Are there unexpected negative consequences we need to address?
  • Is the system getting better or worse over time?

Organizational questions:

  • Does ownership remain clear and adequate?
  • Is monitoring burden sustainable?
  • Are stakeholders maintaining confidence?

Strategic questions:

  • Do business priorities still align with what the system optimises?
  • Should we expand scope, maintain current state, or reduce deployment?
  • What have we learned that should inform other AI decisions?

When to stop, pause, or pivot:

Stop when:

  • Promised value is not materializing after reasonable pilot period
  • Risks have proven larger than anticipated
  • Maintenance burden exceeds execution benefit
  • Organizational confidence has eroded
  • Strategic priorities have shifted making the use case irrelevant

Pause when:

  • Temporary issues need resolution (data quality, integration problems)
  • Organizational capacity is temporarily constrained
  • Market conditions make current approach inappropriate

Pivot when:

  • Core use case has issues, but adjacent applications show promise
  • Scope needs to narrow or expand based on learning
  • Different ownership or governance structure would improve outcomes

Stopping is not failure. It is good management when evidence shows a use case is not working.

Creating organisational decision discipline

Individual AI decisions matter. Decision discipline across the organization matters more.

Three practices that create systematic decision quality:

1. Require completed framework assessment before any AI investment

Make the four-question evaluation mandatory, documented, and reviewed before proceeding.

This prevents enthusiasm-driven adoption and ensures every use case has been rigorously evaluated.

2. Maintain a decision log

Document:

  • Use cases evaluated (approved and rejected)
  • Assessment rationale
  • Outcomes over time
  • Lessons learned

This creates institutional memory and prevents repeating mistakes.

3. Conduct quarterly portfolio review

Evaluate all active AI/automation/agentic deployments as a portfolio:

  • Which are delivering value?
  • Which should be stopped?
  • Which should be expanded?
  • What patterns are emerging?

This prevents accumulation of zombie projects that consume resources without delivering value.

What to do next

If evaluating a specific AI use case:

  1. Complete the four-question assessment systematically. Write down answers. If any core question reveals problems, address those before proceeding.
  2. If all four questions have satisfactory answers, design a small pilot with clear success criteria, timeline, and review rhythm.
  3. If any question reveals gaps, stop and resolve foundational issues (establish ownership, mitigate risks, clarify value proposition, rightsize scope).

If managing multiple AI initiatives:

  1. Apply the framework to all current deployments. How many would pass rigorous evaluation today?
  2. Create a portfolio view showing which initiatives are delivering value vs. consuming resources without clear returns.
  3. Establish decision discipline requiring framework assessment for all future AI investments.

The final post in this series examines why clarity has become the real competitive advantage in AI-enabled marketing – and why the organisations that succeed are not those with the most tools, but those with the clearest thinking about where and how to use them.

Next in the series: Why clarity is the real competitive advantage in AI marketing – As access to AI becomes universal, advantage shifts to judgment.


Discover more from jam partnership

Subscribe to get the latest posts sent to your email.