Guide 9 min read

AI Tokens Explained: What Real Estate Agents Actually Pay For

RW
Ryan Wanner

AI Systems Instructor • Real Estate Technologist

Every time you use AI, you're spending tokens. Whether you realize it or not, tokens are the currency behind every prompt, every response, and every bill. Here's exactly what they are -- and what they cost in real estate terms.

What Is a Token? (The 30-Second Version)

A token is a chunk of text that an AI model reads and produces. Not a word. Not a character. A chunk. On average, one token equals roughly 4 characters, or about three-quarters of a word.

"Listing" is 1 token. "Real estate agent" is 3 tokens. "3-bedroom ranch with updated kitchen" is about 7 tokens. You can see exactly how any text splits into tokens using OpenAI's free tokenizer tool.

Why does this matter? Because tokens are how AI models measure everything. Your prompt has a token count. The AI's response has a token count. The model's memory -- how much it can "hold" in a single conversation -- is measured in tokens. And if you're using the API, your bill is calculated in tokens.

Think of tokens as the square footage of an AI conversation. Your prompt is the rooms you're furnishing. The AI's response is the rooms it builds for you. The context window is the total floor plan. Everything fits inside that footprint -- or it doesn't.

Tokens in Practice: Real Estate Examples

Let's make this concrete with tasks you actually do.

Listing description: A typical prompt (property details + instructions + your Context Card) runs about 400-600 input tokens. The AI's response (a 200-word listing description) is about 250-350 output tokens. Total: roughly 700-950 tokens. At GPT-4o API rates, that's about $0.005-$0.01. Less than a penny.

Buyer follow-up email: A prompt with lead details and your email style is about 300-500 input tokens. The drafted email is 200-400 output tokens. Total: 500-900 tokens. Cost: under $0.01.

Neighborhood market analysis: If you paste in MLS data, recent sales, and neighborhood stats, you might send 3,000-5,000 input tokens. The AI's analysis might be 1,500-2,500 output tokens. Total: 4,500-7,500 tokens. Cost: about $0.03-$0.05.

Inspection report summary: A full inspection report pasted into Claude could be 15,000-25,000 input tokens. The summary might be 1,000-2,000 output tokens. Total: 16,000-27,000 tokens. Cost: about $0.06-$0.10. That's a dime to summarize a 30-page report.

The pattern is clear: individual tasks cost almost nothing in tokens. Even if you process 100 listing descriptions per month via the API, you're looking at about $1 total. Tokens only become a meaningful cost factor at high volume or with very long documents.

Context Windows: Your AI's Memory Limit

The context window is the total number of tokens an AI model can process in a single conversation -- your prompts, the AI's responses, and everything in between. When you hit the limit, the AI starts "forgetting" the earliest parts of your conversation.

Here's where the major models stand:

  • GPT-4o: 128,000 tokens (~96,000 words)
  • Claude (Anthropic): 200,000 tokens (~150,000 words)
  • Gemini 1.5 Pro: 1,000,000 tokens (~750,000 words)

For real estate, context window size matters most when you're working with long documents. A 50-page purchase agreement is roughly 15,000-20,000 tokens. A full CMA packet might be 30,000 tokens. An inspection report runs 15,000-25,000 tokens. All of these fit comfortably inside any modern model's context window.

Where agents run into trouble: long, multi-turn conversations. If you've been going back and forth with ChatGPT for 30 messages about a complex deal, and the responses start feeling disconnected from what you discussed earlier, you've likely filled the context window. The solution isn't a bigger model -- it's a new conversation with a fresh Context Card that includes the key details from your previous chat.

Token Costs: Subscription vs API Pricing

OptionPriceToken LimitBest For
ChatGPT Plus$20/month flatUnlimited messages (rate-limited)Individual agents, daily use
Claude Pro$20/month flatUnlimited messages (rate-limited)Long documents, detailed instructions
OpenAI API (GPT-4o)~$2.50/1M input, ~$10/1M output128K per requestAutomations, bulk processing
Anthropic API (Claude Sonnet 4)~$3/1M input, ~$15/1M output200K per requestComplex workflows, large documents
OpenAI API (GPT-4o-mini)~$0.15/1M input, ~$0.60/1M output128K per requestHigh-volume, simple tasks

Subscription plans charge flat monthly rates. API pricing is per-token, pay-as-you-go. Prices as of February 2026.

When Tokens Matter (and When They Don't)

Here's the honest breakdown: if you're using ChatGPT Plus or Claude Pro at $20/month, tokens don't affect your bill at all. You've paid the flat fee. Use it as much as you want. The only limit is rate throttling during peak times, and for most agents, you'll never notice it.

ChatGPT Plus has 10 million subscribers paying that flat $20. The vast majority never think about tokens. And that's fine. If you're using the app for conversation-based work -- writing, research, analysis, brainstorming -- tokens are invisible to you. Stay on the subscription. Don't overthink it.

Tokens start mattering in three scenarios:

1. API integrations. If you're building automations or your brokerage is running AI through code, you're paying per token. Understanding token costs helps you estimate monthly expenses and choose the right model. GPT-4o-mini at $0.15 per million input tokens is 16x cheaper than GPT-4o -- and for simple tasks like drafting short emails, the output quality is nearly identical.

2. Very long documents. Pasting a 100-page document into the AI uses a lot of tokens from your context window. If you're working with large files regularly, a model with a bigger context window (Claude's 200K vs GPT-4o's 128K) gives you more room. This isn't about cost -- it's about capacity.

3. Context window overflow. When your conversation gets so long that the AI starts losing track of what you said earlier, that's a token problem. The fix: start fresh with a new conversation and include the essential context upfront in your Context Card.

The OODA Loop for Token Management

If you're using the API or managing AI costs for a team, the OODA Loop gives you a framework for optimizing token spend.

Observe: Track your actual token usage. OpenAI's dashboard shows usage by day and model. Anthropic provides the same. Look at how many tokens each workflow consumes. You might find that one automation accounts for 80% of your bill.

Orient: Map token usage to business value. That lead-response automation using 500K tokens per month at $6 might be generating $4,000 in additional commissions. That's a 66,000% return. But that research tool pulling 2M tokens per month at $25 for data nobody reads? That's waste.

Decide: Choose the right model for each task. You don't need GPT-4o for every job. Simple tasks (email subject lines, quick summaries, data extraction) work fine on GPT-4o-mini at a fraction of the cost. Reserve the powerful models for complex analysis, long-document processing, and tasks where quality directly impacts revenue.

Act: Implement model routing. Send simple requests to the cheap model, complex requests to the powerful model. Trim unnecessary context from prompts -- don't paste your entire Context Card when you only need one section. Set usage alerts so you're not surprised by a spike. Review monthly and adjust.

For individual agents on the $20 subscription: you can skip all of this. Tokens are abstracted away. Focus on getting better at prompting instead -- that's where the real value lives.

Sources

  1. OpenAI Tokenizer Tool -- Visualize How Text Splits Into Tokens
  2. OpenAI Pricing -- GPT-4o, GPT-4o-mini, and API Token Costs
  3. Anthropic Pricing -- Claude Sonnet, Opus, and API Token Costs
  4. Master of Code -- ChatGPT Statistics: 10M Plus Subscribers, Revenue Data
  5. Anthropic Documentation -- Claude Model Specs (200K Context Window)
  6. NAR 2025 Technology Survey -- AI Tool Adoption Among Realtors
  7. OpenAI / Spendflo -- ChatGPT Pricing Plans and Features Breakdown

Frequently Asked Questions

What is a token in AI?
A token is a chunk of text that AI models process. One token equals roughly 4 characters or three-quarters of a word. 'Real estate agent' is 3 tokens. AI models measure everything in tokens: your prompt's length, the response length, the model's memory (context window), and API pricing. You can visualize token splitting with OpenAI's free tokenizer tool at platform.openai.com/tokenizer.
How much does one AI token cost?
It depends on the model and whether you use a subscription or API. ChatGPT Plus ($20/month) and Claude Pro ($20/month) offer unlimited use -- tokens don't affect your bill. On the API, GPT-4o costs about $2.50 per million input tokens and $10 per million output tokens. A listing description costs roughly $0.01 in API tokens. GPT-4o-mini is about 16x cheaper at $0.15 per million input tokens.
How many tokens does a listing description use?
A typical listing description workflow uses about 700-950 total tokens. Your prompt (property details, instructions, style preferences) runs 400-600 input tokens. The AI's response (a 200-word description) is about 250-350 output tokens. At GPT-4o API rates, that costs under a penny. On a ChatGPT Plus or Claude Pro subscription, it costs nothing extra beyond the $20/month.
What is a context window in AI?
The context window is the maximum number of tokens an AI can process in one conversation -- including your prompts, the AI's responses, and all conversation history. GPT-4o has a 128K token window (~96,000 words). Claude has 200K tokens (~150,000 words). When you exceed the context window, the AI loses track of earlier conversation. The fix: start a new conversation with a Context Card that includes the essential details.
Should I worry about tokens if I use ChatGPT Plus?
No. ChatGPT Plus is a flat $20/month with no per-token charges. The same applies to Claude Pro. Tokens only matter for billing if you're using the API (pay-per-token). The only token-related issue for subscription users is the context window -- if your conversation gets very long, the AI may lose track of earlier messages. Starting a new conversation with a Context Card solves this.
Why does Claude have a bigger context window than ChatGPT?
Claude's 200K token context window vs GPT-4o's 128K reflects different engineering priorities by Anthropic and OpenAI. For real estate agents, the practical difference is that Claude can hold about 50% more text in a single conversation. This matters most when working with long documents like inspection reports, purchase agreements, or CMA packets. Both windows are large enough for the vast majority of real estate tasks.

Related Terms

Keep Reading

Related Articles

Stop Guessing With AI

Join The Architect workshop and learn the frameworks that turn AI from a toy into a business tool.