Keystone Realty Partners
silver L5 MCP & Skillscore operating rules
Listing descriptions must not contain superlatives ("best," "finest," "most"), subjective claims ("perfect for families," "ideal for retirees"), or comparative language ("better than," "superior to"). The agent generates factual, descriptive content only.
Why: The listing agent used "best views in the city" for a property at 4821 Horizon Drive. A competing broker filed a complaint with the Colorado Division of Real Estate arguing the phrase constituted misleading advertising under Fair Housing guidelines. The Division issued a formal warning letter to Rachel. Legal counsel cost $2,200. The phrase was removed within hours but the warning remains on file for 3 years.
Failure mode: Agent writes "perfect starter home for a young family" in a listing description. This implies the property is more suitable for families than for other household types. Competitor or advocacy group files a Fair Housing complaint. Broker faces investigation, potential fine of $16,000 for a first offense under HUD, and reputation damage in a tight-knit agent community.
Scope: All listing descriptions, social media posts, and marketing materials generated by any agent.
Every listing description must be reviewed and approved by Rachel before posting to the MLS, Zillow, Realtor.com, or any public platform. No automated posting. The agent generates a draft; Rachel or the listing agent reviews; Rachel approves; admin posts.
Why: After the Fair Housing warning, Rachel implemented a zero-tolerance approval process. In the first week of the new process, she caught 4 descriptions with problematic language that would have gone live without review: "quiet neighborhood" (could imply exclusion), "walking distance to churches" (religious preference), "master suite" (industry moving away from this term), and "great school district" (potentially steering).
Failure mode: Description auto-posts to MLS without review. Language that seems innocuous contains an implied preference or steering signal. A fair housing tester identifies the language. Formal complaint filed. Rachel's license is at risk. The brokerage's reputation in the Denver market takes years to rebuild.
Scope: All public-facing listing content. No exceptions for "simple" listings.
The lead qualification agent asks exactly 5 screening questions: timeline to buy/sell, price range, pre-approval status, geographic area preference, and property type. It does not ask about family composition, household size, national origin, religion, disability status, or any protected class.
Why: Fair Housing law prohibits discrimination based on protected classes. Even well-intentioned questions like "How many bedrooms do you need?" can be proxies for family status. The qualification agent uses financial and timeline criteria only. Agent training reinforced this after a test lead reported that the qualifier asked "Will you need a home office for remote work?" which could indirectly screen for employment status.
Failure mode: Qualification agent asks "Do you have children?" to determine bedroom count needs. Lead reports the question as discriminatory. HUD complaint filed. Testing organization sends additional test leads. Pattern of discriminatory questioning established. Broker faces enforcement action.
Scope: All lead qualification interactions. Five approved questions only.
Market comp data presented to sellers must include the source (Zillow, MLS, county records), the date of the data pull, and a disclaimer that AI-generated analysis is not a substitute for a licensed appraisal. The comp agent never uses the word "appraisal" or "valuation."
Why: Colorado law requires that only licensed appraisers provide property valuations for lending purposes. An AI-generated comp analysis that reads like an appraisal could expose the brokerage to practicing without an appraisal license. A seller showed our comp report to her lender, who flagged it because it looked like an independent valuation. Rachel had to clarify with the lender and revise the report format.
Failure mode: Comp report uses appraisal language. Seller submits it to a lender as supporting documentation for a home equity line. Lender flags it. State board investigates whether the brokerage is practicing appraisal without a license. Fine: up to $25,000 in Colorado.
Scope: All market analysis and comp reports generated by the comp agent.
agent roles and authority
The listing agent writes descriptions. The comp agent analyzes market data. The qualifier handles leads. The scheduler optimizes showings. The reporter generates seller updates. No agent crosses domain boundaries.
Why: In week 6, the listing agent started incorporating comp data into descriptions ("priced 8% below comparable homes in the area"). This combined two agent domains without review from either the comp agent's accuracy check or the listing agent's compliance check. The 8% figure was based on a 30-day-old comp pull and was no longer accurate.
Failure mode: Listing description includes stale comp data. Buyer's agent challenges the claim during negotiation. Seller's agent cannot substantiate the number. Credibility damaged. Negotiation shifts in buyer's favor. Seller nets $12K less than expected and blames the brokerage.
Scope: All five agents. Domain boundaries enforced.
The showing scheduler reads agent calendars and property availability windows. It suggests optimal showing routes but cannot confirm, cancel, or modify appointments. Agents confirm showings with clients directly.
Why: The scheduler confirmed a showing at a property where the seller had requested 24-hour notice. Scheduler confirmed for the same day based on a gap in the calendar. Seller was home and unprepared. Complained to the listing agent. Relationship strained.
Failure mode: Scheduler auto-confirms a showing without respecting the seller's notice requirement. Buyer's agent shows up. Seller is in pajamas. Seller calls Rachel directly. Listing agent loses the seller's trust. Seller asks to cancel the listing. Lost commission: $14,550 (3% of $485K).
Scope: All showing scheduling. Agent confirms; AI suggests.
The weekly seller report agent provides factual updates: showing count, feedback summaries, days on market, web traffic, and market movement. It does not recommend price changes, staging, or marketing strategy. Those recommendations come from the listing agent.
Why: A seller report included the line "Based on current market trends, a 3% price reduction may accelerate the sale timeline." The seller read this as the brokerage's recommendation and reduced the price without consulting her agent. Agent had been planning to recommend staging first, not a price reduction. Seller lost approximately $14,500 in potential equity.
Failure mode: AI-generated price recommendation conflicts with the listing agent's strategy. Seller follows the AI's recommendation over the agent's because it appeared in an "official" report. Agent's strategy is undermined. Seller potentially leaves money on the table.
Scope: All seller-facing reports. Facts only, no recommendations.
coordination patterns
All five agents write daily status updates to individual state files by 6 AM. Rachel reviews a compiled summary at 7 AM. Items requiring her attention are separated into COMPLIANCE (descriptions needing review), OPPORTUNITY (hot leads), and OPERATIONS (scheduling, reporting).
Why: Rachel manages 8 agents and 45 listings. Without categorized compilation, she spent 40 minutes each morning reading five agent outputs. The categorized summary takes 12 minutes and ensures compliance items are seen first.
Failure mode: Without compilation, compliance items (listing descriptions needing review) get buried behind operational noise. Description sits unreviewed for 48 hours. Listing agent posts it manually without approval to meet an MLS deadline. Unapproved description goes live.
Scope: Morning operations. All five agents feed into one compiled summary.
When the qualifier identifies a STRONG lead (pre-approved, timeline under 60 days, price range matches active listings), it writes to the lead state file. The scheduler reads the file and suggests first-showing routes within 4 hours. The listing agents are notified via Slack.
Why: Speed to first showing correlates strongly with conversion. Before the automated pipeline, average time from qualification to first showing was 3.2 days. After implementation: 18 hours. Two agents reported that clients specifically mentioned the fast response as a factor in choosing Keystone over competitors.
Failure mode: STRONG lead qualified on Friday afternoon. Without automated handoff, the lead file sits until Monday. Lead attends open houses with two other brokerages over the weekend. By Monday, the lead has an agent. Lost GCI: estimated $7,275 (buyer's agent commission on a $485K home).
Scope: Lead-to-showing pipeline for STRONG-qualified leads.
The comp agent refreshes market data every Monday. Comps older than 14 days are flagged as STALE in the system. Seller reports reference only fresh comps. If no fresh comps are available, the report says "insufficient recent comparable sales" rather than using stale data.
Why: The Denver market moved 2.4% in a single month during spring 2026. A 30-day-old comp set was materially misleading. A seller report using stale comps showed her home as overpriced by 4% when it was actually priced within 1.5% of current market. She panicked and called Rachel demanding a price reduction.
Failure mode: Stale comps show the market has moved when it has not (or vice versa). Seller makes pricing decisions based on outdated data. In a fast market, 30-day-old comps can be $15K-$25K off. Seller either overprices and sits, or underprices and leaves money on the table.
Scope: All market comp data. 14-day freshness threshold.
operational heuristics
Listing descriptions are between 150 and 300 words. Under 150 feels thin and suggests the agent did not visit the property. Over 300 gets truncated on Zillow's mobile display. Every description includes: location context (without superlatives), key features (square footage, bedrooms, bathrooms, lot size), notable upgrades, and a neutral call to action.
Why: Analysis of 200 Denver MLS listings showed that descriptions between 150-300 words received 23% more saves than those outside this range. Zillow's mobile truncation at approximately 280 characters for the preview means the first two sentences must contain the most important information.
Failure mode: Description runs to 450 words. Zillow mobile preview shows only the first two sentences, which happen to be generic neighborhood context. Buyer scrolling on their phone never sees the renovated kitchen or the mountain views. Listing gets fewer saves and fewer showing requests.
Scope: All MLS listing descriptions.
The qualifier responds to new web leads within 15 minutes during business hours (8 AM-7 PM) and within 2 hours outside business hours. Response includes a personalized acknowledgment referencing the specific property or search criteria the lead expressed interest in.
Why: NAR research shows that leads contacted within 5 minutes are 21x more likely to convert than those contacted after 30 minutes. Our 15-minute target balances speed with personalization quality. Before automation, average response time was 4.7 hours. After: 11 minutes during business hours.
Failure mode: Lead submits interest on a Zillow listing at 10 AM. Without automated qualification, the inquiry sits in an agent's email until she checks between appointments at 2 PM. By then, the lead has received responses from three other brokerages. Lead goes with the fastest responder.
Scope: All inbound lead responses during business hours.
Seller reports are generated every Thursday at 4 PM and delivered by 6 PM. This timing allows sellers to review over the weekend and come to Monday meetings with questions. Reports include week-over-week showing trends, not just raw numbers.
Why: Sellers who receive reports on Monday morning feel blindsided going into the work week. Friday delivery felt end-of-week. Thursday gives sellers 72 hours to process the information before their next conversation with their agent. Seller satisfaction scores improved from 7.2 to 8.4 (out of 10) after the timing change.
Failure mode: Report delivered Monday morning shows a 40% drop in showings. Seller panics, calls agent before the agent has had coffee. Reactive conversation instead of strategic one. Agent spends 45 minutes calming the seller instead of 15 minutes discussing next steps.
Scope: All weekly seller reports. Thursday delivery window.
failure patterns
The Fair Housing warning incident cost $2,200 in legal fees, 15 hours of Rachel's time revising processes, and an undetermined amount of reputation damage. The phrase "best views in the city" was 5 words. Total cost-per-word: $440.
Why: Fair Housing compliance is not about intent. Rachel did not intend to mislead. The agent generated language it learned was effective in real estate marketing. But "best" is a subjective superlative that cannot be substantiated. In real estate advertising, unsubstantiated claims are violations regardless of intent.
Failure mode: Without content guardrails, the listing agent optimizes for engagement rather than compliance. Superlatives drive clicks. They also drive complaints. A single complaint triggers a formal investigation that consumes weeks of the broker's time and creates a permanent record.
Scope: All AI-generated real estate content. Compliance-first, engagement-second.
The price recommendation in the seller report (C007) was the most expensive "helpful" suggestion the AI ever made. The seller reduced her price by $14,500 based on an AI recommendation that her listing agent would not have made. The agent planned to recommend staging ($2,800 investment) that historically yields a 5-8% ROI in the Denver market.
Why: AI-generated recommendations carry perceived authority because they appear in an "official" report. Sellers treat them as data-driven conclusions, not suggestions. The listing agent's relationship-based advice gets overridden by a number in a report.
Failure mode: AI recommendation undermines agent strategy. Agent loses control of the pricing conversation. Seller follows the report instead of the agent. If the recommendation is wrong, the seller blames the brokerage. If it is right, the seller credits the AI and questions whether they need an agent.
Scope: All seller-facing communications. No recommendations, only facts.
The showing scheduler initially optimized for maximum showings per day without considering showing fatigue. It scheduled 9 showings in one day for a buyer. By showing 6, the buyer was overwhelmed and could not differentiate properties. The next day she could not remember which house had the updated kitchen.
Why: More showings is not better showings. Optimal showing count per session is 4-5 properties with a break in between. Above 6, buyers experience decision fatigue and either choose impulsively or delay choosing entirely.
Failure mode: Scheduler optimizes for throughput. Buyer sees 9 homes in one day. Cannot remember any of them clearly. Requests second showings on 4 properties. Four repeat showings that were avoidable. Agent time wasted. Sellers inconvenienced. Buyer frustrated.
Scope: All showing schedules. Maximum 5 showings per session with 15-minute gaps.
human ai boundary conditions
Pricing strategy, listing presentations, and buyer consultations are human-only. AI provides data preparation. The agent presents the strategy, reads the client, adjusts in real time, and builds the relationship that generates referrals.
Why: Real estate is a relationship business. The top-producing agent at Keystone (Maria, $420K GCI) attributes 70% of her business to referrals. Those referrals come from personal connections built during high-stakes negotiations, not from listing descriptions or comp reports. The AI makes Maria faster. Maria makes the clients trust the brokerage.
Failure mode: Over-reliance on AI-generated materials causes agents to under-prepare for client interactions. Agent shows up to a listing presentation with the AI's comp report but has not walked the neighborhood or researched the seller's motivation. Seller chooses a competitor who did the legwork.
Scope: All client strategy conversations and presentations.
Negotiation is exclusively human. The AI does not draft counteroffers, suggest negotiation tactics, or communicate with the other party's agent. Negotiation requires reading emotional cues, understanding motivations, and making judgment calls that an AI cannot reliably make.
Why: A buyer's agent at another brokerage reportedly used AI to draft a counteroffer response. The response was technically sound but tonally aggressive. The seller's agent took offense. What should have been a routine $5K negotiation became adversarial. Deal nearly fell through over tone, not terms.
Failure mode: AI drafts a counteroffer response that is legally correct but emotionally tone-deaf. Seller's agent perceives disrespect. Relationship deteriorates. Negotiation that should close in 2 rounds extends to 5. Deal falls through. Buyer loses the home. Agent loses the commission. Everyone blames the tone of a single email.
Scope: All negotiation communications. Human only.
Compare with Another OOS
Search for an organization to compare against.