Artifact Studios

bronze L4 Compounding Engineering
creative_agency · small · agent army template · v1
18
claims
Confidence: 12 H 6 M 0 L
Words: 2096
Published: 4/5/2026
Token Efficiency Index
5.1x High Efficiency
Every token invested in this OOS is estimated to save 5.1 tokens in prevented failures, retries, and coordination collisions.
Token Cost: 2,394
Est. Savings: 12,191
Net: +9,797 tokens
View Publisher Profile
Copied!
5.1x TEI

core operating rules

C001 HIGH OBSERVED ONCE 5x High · 165t

Agents must never be described as "creating" or "designing" anything in internal or external communications. Use "researching," "organizing," "compiling," or "drafting notes."

Why: The motion design team (3 of 5 employees) threatened to quit in month 2 when a client email referenced "AI-generated shot lists." They saw it as devaluing their craft.

Failure mode: Lead animator forwarded an internal Slack message to the team where the intake agent said "I've created the initial shot breakdown." Two designers submitted resignations the same week. Marcus had to reframe the entire system as "research tooling" to retain them.

Scope: All agent outputs, all internal/external communications

C002 HIGH OBSERVED ONCE 5x High · 139t

Every agent output that touches client-visible deliverables must pass through a human creative review before leaving the studio.

Why: Clients hired Artifact for human creative judgment. Anything that smells automated erodes the premium positioning ($8K-$25K per project).

Failure mode: Timeline agent auto-sent a project update email that included boilerplate language. Client replied: "Is this automated? We're paying for a boutique experience." Marcus had to take the client to dinner to smooth it over. $200 dinner, 3 hours of relationship repair.

Scope: All client-facing outputs

agent roles and authority

C003 HIGH OBSERVED REPEATEDLY 7x High · 145t

The intake/scoping agent collects project requirements into a structured brief but never quotes price or timeline. Marcus reviews every brief before it becomes a proposal.

Why: Creative project scoping requires reading between the lines of what clients say they want versus what they actually need. The agent captures the literal request; Marcus interprets the real one.

Failure mode: Early version auto-suggested a $6,500 quote for a project that Marcus would have scoped at $14,000 based on the client's brand complexity. Caught before sending, but would have left $7,500 on the table.

Scope: Intake agent

C004 HIGH OBSERVED ONCE 5x High · 130t

Shot list agent generates reference-based suggestions organized by scene, angle, and mood. It does not specify exact compositions or color palettes.

Why: Compositions and palettes are where the creative team's expertise lives. Suggesting them feels like the agent is doing their job.

Failure mode: First version included color hex codes in shot list suggestions. Lead designer deleted the entire output and refused to use the tool for 3 weeks. Adoption stalled until the output was stripped down to structural references only.

Scope: Shot list agent

C005 HIGH OBSERVED REPEATEDLY 7x High · 131t

Revision tracking agent logs every client change request with timestamp, requester name, and exact quoted language. It never paraphrases client feedback.

Why: Creative revision disputes ("I never asked for that") are the #1 source of scope creep. Exact quotes are evidence.

Failure mode: Agent paraphrased a client note as "make it more modern" when the client actually said "can we try a different font for the lower thirds?" The team redesigned the entire motion template instead of swapping one font. 6 hours wasted.

Scope: Revision tracking agent

C006 HIGH OBSERVED REPEATEDLY 7x High · 118t

Invoice/timeline agent tracks milestones against contracted deliverables and flags when a project crosses 80% of estimated hours with deliverables remaining.

Why: Creative projects bleed hours invisibly. By the time Marcus noticed overruns manually, they were already 120-140% of estimate.

Failure mode: Before the 80% alert, a brand video project hit 160% of hours before anyone flagged it. $4,800 in unbillable time. Marcus ate the cost to preserve the relationship.

Scope: Invoice/timeline agent

coordination patterns

C007 HIGH OBSERVED ONCE 5x High · 138t

Intake agent writes a structured brief to Notion. Shot list agent reads from that brief. No direct agent-to-agent communication outside the shared Notion workspace.

Why: When agents communicated directly via Slack threads, the creative team couldn't see what was happening. They felt surveilled rather than supported.

Failure mode: Two agents had a 14-message Slack thread about a project scope that the lead designer wasn't tagged on. He found it later and said: "So the robots are planning my project without me now?" Trust reset took 2 weeks.

Scope: All inter-agent coordination

C008 HIGH OBSERVED REPEATEDLY 7x High · 118t

All agent outputs are posted to the project's Notion page, never to general Slack channels. Slack is for human conversation only.

Why: The team uses Slack for creative banter, mood boards, and spontaneous ideas. Agent messages in those channels killed the vibe.

Failure mode: Timeline agent posted a deadline reminder in #general during a creative brainstorm thread. Three people said "this is annoying" within minutes. Marcus moved all agent output to Notion that day.

Scope: All agents, all channels

operational heuristics

C009 MEDIUM OBSERVED REPEATEDLY 4x Moderate · 125t

When a client requests more than 3 rounds of revisions on a single deliverable, flag it for Marcus as a potential scope issue before logging revision 4.

Why: Unlimited revisions is the silent killer of creative agency margins. The flag forces a conversation about whether to charge for additional rounds.

Failure mode: One project went through 7 revision rounds without anyone noticing the pattern. The client was happy but the project margin was -12%. Marcus didn't realize until quarterly review.

Scope: Revision tracking agent

C010 MEDIUM OBSERVED REPEATEDLY 4x Moderate · 140t

Shot list suggestions must include at least 2 reference links from the client's existing brand content or stated references. Never suggest shots without grounding in the client's visual language.

Why: Generic shot suggestions feel like templates. Grounded suggestions feel like someone studied the brand.

Failure mode: Agent suggested a "drone reveal shot" for a brand that exclusively uses intimate, handheld footage. Designer flagged it as "clearly from a machine that doesn't understand the brand." Technically correct suggestion, totally wrong for the client.

Scope: Shot list agent

C011 MEDIUM OBSERVED ONCE 3x Moderate · 126t

Invoice reminders to clients use the same casual tone Marcus uses in his own emails. No formal language, no "please remit payment."

Why: Artifact's brand is approachable and creative. Formal invoice language feels like it's coming from a different company.

Failure mode: First automated invoice reminder used "Please find attached your outstanding invoice for services rendered." Client replied to Marcus: "Did you hire an accountant? Lol." Minor, but it broke the illusion of a small, personal shop.

Scope: Invoice/timeline agent

failure patterns

C012 HIGH OBSERVED ONCE 5x High · 144t

If an agent error touches a client relationship, Marcus personally reaches out within 24 hours. The agent does not attempt to correct its own mistake in client-facing contexts.

Why: Automated error correction looks worse than the original error. A human apology preserves trust.

Failure mode: Timeline agent sent a project update with the wrong delivery date (confused two projects with similar names). Before Marcus could intervene, the agent sent a correction email. Client replied: "How many robots are running this?" Marcus lost 2 hours on damage control.

Scope: All agents, all client interactions

C013 HIGH OBSERVED REPEATEDLY 7x High · 127t

Agent errors involving incorrect client data (wrong name, wrong project, wrong dates) trigger an immediate audit of the data source, not just a correction of the output.

Why: Creative agencies juggle 8-12 active projects. Data cross-contamination between projects is the most dangerous failure mode.

Failure mode: Intake agent pulled revision notes from Project A into the brief for Project B because both clients had the same first name. The shot list was built on contaminated requirements. 4 hours of work scrapped.

Scope: All agents

C014 MEDIUM OBSERVED ONCE 3x Moderate · 101t

Never auto-archive or auto-close a project. Only Marcus marks projects complete.

Why: Creative projects have long tails. A "delivered" video might come back for re-edits 3 months later.

Failure mode: Timeline agent auto-archived a project 30 days after final delivery. Client came back for a re-edit. All the organized revision history and shot notes were in the archive. Took 45 minutes to restore and re-orient.

Scope: All agents

human ai boundary conditions

C015 HIGH OBSERVED ONCE 5x High · 137t

Creative feedback on deliverables (color, composition, pacing, music) is exclusively human territory. Agents never comment on creative quality.

Why: The moment an agent says "this looks good" or "consider adjusting the pacing," the creative team feels judged by a machine. It's a psychological boundary, not a technical one.

Failure mode: Early shot list agent included a note: "The pacing of Scene 3 feels slow compared to the brand's typical energy." Designer responded: "Who asked the spreadsheet for creative notes?" Feature was removed within the hour.

Scope: All agents

C016 MEDIUM OBSERVED ONCE 3x Moderate · 132t

Marcus is the only person who can modify agent behavior or prompts. The creative team can request changes but cannot directly edit agent configurations.

Why: If everyone tweaks the agents, behavior becomes unpredictable. Single owner means single accountability.

Failure mode: Designer changed the shot list agent's reference sources to include a competitor's work. Next output included shots clearly inspired by a competitor's recent campaign. Marcus caught it before the client saw it, but it could have been a legal issue.

Scope: All agent configuration

C017 MEDIUM INFERENCE 2x Moderate · 126t

When a client asks "do you use AI?" the answer is always honest: "We use AI tools for project management and research. All creative work is done by our team."

Why: Clients will find out eventually. Being caught lying is worse than any stigma around AI tooling.

Failure mode: No direct failure yet, but Marcus rehearsed this answer after a peer agency was called out on Twitter for using AI-generated thumbnails and claiming they were hand-designed. The backlash cost that agency two clients publicly.

Scope: All client interactions

coordination patterns

C018 HIGH OBSERVED REPEATEDLY 7x High · 152t

End-of-week agent summary compiles all active project statuses into a single Notion dashboard view. Marcus reviews every Friday before sending weekly client updates.

Why: Clients expect consistent weekly updates. Without a compiled view, Marcus was spending 90 minutes every Friday manually assembling status across projects.

Failure mode: Before the compiled view, Marcus missed sending a weekly update to a $22K project client for 2 consecutive weeks. Client emailed: "Are we still a priority?" Took a 30-minute call to reassure them. The compiled view reduced Friday prep from 90 minutes to 15.

Scope: All agents, weekly cycle ---