Using AI Without Disclosure Can Cost You Your License
Three separate authorities have drawn lines in the sand. Cross any one of them and you're facing discipline, fines, or criminal charges.
California AB 723 made it a criminal misdemeanor to use digitally altered images in property advertising without disclosure. Effective January 1, 2026. Not a suggestion. Not a best practice. A criminal offense under the California Real Estate Law.
NAR's AI guidelines explicitly warn against using AI to draft contracts, modify legal documents, or generate content that constitutes unauthorized practice of law. Agents aren't lawyers. AI doesn't change that.
HUD's 2024 guidance confirms the Fair Housing Act applies fully to AI systems. If your AI ad targeting tool excludes protected classes, you're liable. If your AI tenant screening tool produces disparate impact, you're liable. Even if the AI made the decision, not you. Even if a third-party vendor built the tool.
This isn't theoretical risk. The National Fair Housing Alliance reported 32,321 complaints in 2024, with 54.6% involving disability discrimination. AI tools that screen tenants, target ads, or generate listing descriptions are now under the same microscope as every other business practice in real estate.
The OODA Loop applies here: Observe the legal landscape. Orient around the specific rules that affect your practice. Decide on a disclosure protocol. Act before enforcement catches up with you. Because enforcement is coming.
California AB 723 — The Landmark Law
AB 723 is the first state law specifically targeting AI-altered images in real estate. Every other state is watching it. If you list anywhere, understand this law. Your state is likely next.
What It Requires
Any real estate broker, salesperson, or person acting on their behalf who uses a digitally altered image in property advertising must:
1. Include a clear disclosure that the image has been digitally altered.
2. Provide access to the original, unaltered image(s) — either included directly in the listing or accessible via a publicly available link, URL, or QR code.
That's it. Two requirements. Disclose and show the original. Not complicated. But the penalty for ignoring them is severe.
Who It Applies To
Licensed real estate brokers and salespersons. Developers. Marketing staff acting on behalf of licensees. Third-party photographers and staging companies working under contract with a licensee. If you're in the chain of creating or posting altered listing photos, this law applies to you.
What Counts as "Digitally Altered"
Bornstein Law's analysis breaks this down clearly:
Requires disclosure: Adding or removing furniture, fixtures, appliances, or flooring. Changing paint colors, landscaping, or building exteriors. Altering views, neighboring structures, or street features visible from the property. Virtual staging. AI-generated items that don't physically exist in the property. Decluttering tools that remove real objects. Renovation preview tools that show hypothetical remodels.
Exempt (no disclosure needed): Lighting or exposure corrections. Color or white-balance adjustments. Image sharpening or straightening. Cropping and resizing. HDR blending of bracketed exposures. Basically: making the photo look better without changing what's in the room.
The Penalty
AB 723 adds Section 10140.8 to the California Business and Professions Code. It sits inside the California Real Estate Law, where willful violations are prosecutable as a misdemeanor. Beyond criminal exposure, violations create grounds for license discipline by the Department of Real Estate (DRE). That means suspension or revocation.
Put differently: a $5 virtual staging image without a disclosure line could trigger criminal charges and a license review. The risk-reward math doesn't work in any scenario.
What CRMLS and MLSs Are Doing
California Regional MLS (CRMLS) and other MLSs are updating listing input forms to include disclosure fields for digitally altered images. Expect a checkbox or text field on your listing input form asking whether photos have been digitally altered. Some MLSs are adding automated flags for images detected as AI-generated. If your MLS hasn't updated yet, don't wait. Add the disclosure manually.
State-by-State AI Disclosure Requirements
| State | Law / Regulation | Effective | RE-Specific? | Key Requirement |
|---|---|---|---|---|
| California | AB 723 (digitally altered images) | Jan 1, 2026 | Yes | Disclose altered listing images + provide originals. Misdemeanor for willful violation. |
| California | AB 2655 (AI watermarking) | Jan 1, 2026 | No (general) | Large AI platforms must watermark synthetic content. Affects image generation tools agents use. |
| New York | AI avatar disclosure bills | 2025-2026 | Partial | Disclose AI-generated human figures in advertising. Affects listing videos using AI avatars. |
| Colorado | SB 24-205 (AI Consumer Protection) | Jun 30, 2026 | No (general) | High-risk AI systems must disclose AI use to consumers. Covers AI used in housing decisions. |
| Utah | SB 0226 (AI Amendments) | May 7, 2025 | Partial | Regulated occupations must prominently disclose AI use. RE agents are regulated occupations. |
| Texas | TRAIGA (Responsible AI Governance) | Jan 1, 2026 | No (general) | Prohibits certain AI uses with biometric data. Impacts AI tenant screening and facial recognition. |
| Illinois | AI content labeling bills | 2025-2026 | No (general) | AI-generated content in advertising must be clearly labeled. Applies to listing marketing. |
| All States (Federal) | Fair Housing Act + HUD AI Guidance | Active now | Yes | AI in tenant screening, ad targeting, and housing decisions must comply with FHA. Providers liable for third-party AI. |
Tracked via Orrick's U.S. AI Law Tracker. Laws evolving rapidly — verify current status before relying on this table.
NAR Guidelines — The 3 Traps to Avoid
NAR's official AI guidance identifies three specific traps that agents fall into. Each one can end your career.
Trap 1: AI-Generated Contracts
Do NOT use ChatGPT, Claude, or any AI tool to draft, modify, or customize real estate contracts. Period.
NAR's guidance is explicit: generating or substantially modifying contract language constitutes unauthorized practice of law unless you are a licensed attorney. It doesn't matter how good the AI output looks. It doesn't matter if you "just used it for the first draft." If a dispute arises and opposing counsel discovers the contract was AI-generated by an unlicensed person, you're facing UPL charges on top of whatever the dispute is about.
Use AI for everything around the contract: listing descriptions, marketing emails, client communication, market analysis. But the contract itself? That's your attorney's job. Or use the state-approved forms your association provides. Don't get creative.
Trap 2: AI Ad Targeting That Excludes Protected Classes
Facebook's AI ad targeting settled a $115 million lawsuit with HUD for excluding protected classes from housing ads. Meta changed its system. But the liability didn't disappear. It shifted to you.
If you use any AI-powered advertising platform to target housing ads, you are responsible for ensuring that targeting does not exclude people based on race, color, religion, sex, national origin, familial status, or disability. That includes lookalike audiences. That includes AI-optimized targeting that "learns" to exclude demographics based on engagement patterns.
HUD's guidance is clear: housing providers are liable for Fair Housing violations even when using third-party AI tools. The vendor built the tool. But you chose to use it. HUD's 2024 statement leaves zero ambiguity.
Practical safeguard: use broad geographic targeting for housing ads. Avoid interest-based or demographic-based exclusions. Review your ad platform's AI targeting settings quarterly. Document your compliance process.
Trap 3: Not Disclosing AI-Generated Content
WAV Group published disclosure templates that have become the industry standard. Their recommended language: "This content was generated with the help of AI and reviewed for accuracy."
The disclosure applies to more than just images. If AI wrote your listing description, disclose it. If AI generated your market report, disclose it. If an AI chatbot is responding to buyer inquiries on your website, make sure visitors know they're talking to a bot, not a person.
Utah's SB 0226 already requires this for regulated occupations — and real estate agents are regulated occupations. Colorado's law (effective June 2026) will require it for high-risk AI interactions, which includes housing decisions.
Don't wait for your state to mandate it. Add disclosure now. It costs nothing, builds trust, and eliminates future compliance risk. That's the 5 Essentials principle: systematize the simple stuff so you never have to think about it again.
Your AI Disclosure Checklist
Use this checklist for every listing. Print it. Tape it to your monitor. Make it part of your listing launch workflow.
Virtual Staging Disclosure
Every virtually staged image includes a "Virtually Staged" watermark. Original, unaltered photos are included in the listing or accessible via link/QR code. Listing description states which photos are staged.
Listing Description Disclosure
If AI wrote or assisted with the listing description, add: "This description was prepared with the assistance of AI and reviewed for accuracy by [Your Name]." Review every AI-generated description for factual accuracy before publishing.
Email and Marketing Disclosure
AI-generated emails, newsletters, and marketing materials include: "This content was generated with the help of AI." WAV Group's template language works. Put it in your email footer and forget about it.
Contract Protection
NEVER use AI to draft, modify, or customize contracts. Use state-approved forms only. AI can help you understand a clause. AI cannot write the clause. Document this policy in your team handbook.
Ad Targeting Compliance
Housing ads use broad geographic targeting only. No demographic or interest-based exclusions. Review AI-optimized targeting settings quarterly. Document your targeting parameters for every campaign. If your platform uses "smart" audience optimization, verify it doesn't exclude protected classes.
Chatbot and AI Assistant Disclosure
Website chatbots clearly identify themselves as AI, not human agents. Lead capture forms powered by AI include disclosure. Any automated response system that could be mistaken for a human conversation is labeled.
Photo and Media Audit
Review all listing photos for AI alterations before publishing. Maintain an archive of original photos for every listing. Consider tools like FairSentry (99.97% accuracy) for automated Fair Housing compliance monitoring across your listings and marketing materials.