The Token Economy

Every prompt has a price. Understanding token economics is the difference between a proof-of-concept that impresses and a production system that survives.

How Pricing Works

LLM providers charge per token, with a critical distinction: input tokens (your prompt) and output tokens (the response) are priced differently. Output tokens typically cost three to five times more than input tokens. This asymmetry has direct implications for prompt design.

A verbose system prompt that produces concise outputs may be cheaper than a minimal prompt that triggers lengthy responses. Prompt engineering is not just about quality — it is about optimizing the cost-quality-speed triangle.

The Cost-Quality-Speed Triangle

Every prompt decision involves trade-offs between three dimensions:

  • Cost: Total token expenditure across input and output
  • Quality: Accuracy, completeness, and usefulness of the response
  • Speed: Latency from request to completed response

You cannot maximize all three simultaneously. A longer, more detailed prompt improves quality but increases cost and latency. A smaller model reduces cost and latency but may sacrifice quality. Knowing which dimension to prioritize for a given use case is a core prompt engineering skill.

Token Budgeting

Professional prompt engineers think in token budgets. A customer-facing chatbot might have a budget of 2,000 input tokens and 500 output tokens per interaction. A document analysis pipeline might allocate 100,000 input tokens but only 1,000 output tokens.

Setting token budgets before writing prompts forces you to make deliberate choices about what context to include, how detailed your instructions should be, and how much output to request.

The Hidden Costs

Token costs are only part of the picture. Consider:

  • Retry costs when prompts produce unusable output
  • Human review costs when output quality is inconsistent
  • Latency costs when slow responses degrade user experience
  • Opportunity costs when a cheaper model would have sufficed

The most expensive prompt is one that does not work. Investing tokens in clear, well-structured instructions almost always reduces total cost by reducing failures and retries.