servicesaboutinsights
Back to all notes

#4 — Prompting

November 1, 20233 min read

#4 — Prompting

Why it matters: Effective prompt engineering is becoming table stakes for AI-powered startups. The right approach can dramatically improve your product's performance while reducing costs.

The Essentials

Start with prompts for rapid prototyping before investing in fine-tuning or custom models. Most founders overlook basic techniques that could deliver 80% of the value at 20% of the effort.

The bottom line: Master these five core strategies to outperform competitors who are wasting resources on unnecessary complexity.

Core Techniques That Actually Work

1. Fundamental Techniques Every Founder Should Know

N-shot prompting delivers outsized ROI:

  • Include 5+ representative examples to guide your model
  • Ensure examples match your production data distribution
  • Demonstrate tool usage patterns when applicable

Chain-of-Thought (CoT) reduces hallucinations:

  • Force your model to show its work before concluding
  • Specify exact reasoning steps to improve accuracy
  • Example implementation: "First extract key decisions from the transcript, then verify against meeting details, finally synthesize into key points"

ReAct framework combines reasoning with action:

  • Explicitly instruct your model to reason, plan, then execute
  • Design clean interfaces between your LLM and external tools
  • Integration quality here often separates winning products

2. Structure Drives Performance

The insight: Structure isn't just about aesthetics—it's about performance. Well-structured inputs dramatically improve output quality.

Model preferences matter:

  • Claude responds best to XML
  • GPT models prefer Markdown and JSON
  • Adapt your approach to your chosen model family

Pro tip: Prefilling response templates guarantees your outputs start exactly as needed.

3. Focus Beats Complexity

The trap: Most founders create bloated, catch-all prompts that underperform.

The solution: Break complex tasks into focused micro-prompts:

  • Instead of one massive meeting summarizer
  • Create separate prompts for extraction, verification, and synthesis

Why it works: Single-purpose prompts are easier to optimize, test, and maintain.

4. Context Optimization Is Your Edge

Challenge assumptions about what context you actually need.

Ruthlessly eliminate:

  • Redundant instructions
  • Self-contradictory language
  • Poor formatting that wastes tokens

Structure context to highlight relationships between components.

5. Evaluation Is Non-Negotiable

Before engineering: Establish reliable evaluation metrics.

The minimum viable process:

  • Manually label ~100 evaluation examples
  • Create initial prompt and benchmark performance
  • Iterate based on concrete metrics, not intuition
  • Test on held-out examples before deployment

What Actually Doesn't Matter

Skip the pleasantries. "Please" and "thank you" don't improve output quality.

Forget threats and tips. Modern models don't need to be cajoled or warned.

The takeaway: Focus your engineering time on techniques that move metrics, not on prompt superstitions that waste your runway.

More than just words

Don't fumble in the dark. Your ICPs have the words. We find them.

Strategic messaging isn't marketing fluff—it's the difference between burning cash on ads that don't convert and building a growth engine that scales.