Imagine you are two years into running a small ecommerce or DTC shop. Ads are your lifeline. You pay a healthy monthly retainer to an agency that “handles everything” for Meta, TikTok, and Google. You are not thrilled about the cost, but you are scared of touching anything that might crater sales.
One night, scrolling through r/AiForSmallBusiness, you see a post from a founder in the same position. They cancelled their ad agency, spun up an AI-driven creative workflow for product images and short-form video, and… nothing broke. Click-through rates stayed roughly flat. ROAS stayed roughly flat. The only big difference was the line on their P&L where the retainer used to sit.
Their lingering question was sharp:
“Was my agency just using AI behind my back?”
That is the uncomfortable reality many small operators are waking up to. Agencies are not magical. They have talented people, but they also lean heavily on the same AI tools you are testing yourself. In many cases, you are paying a premium for coordination, process, and risk management—not access to creativity you could never generate.
This article is not an “agencies are dead” rant. It is a practical guide for owners and operators who want to:
- Cut fixed creative costs.
- Keep or improve performance.
- Stay in control of brand and offers.
You will see how to design an AI-first, in-house creative stack and where an orchestration layer like Get BOB fits so campaigns remain testable, trackable, and sane.
What agencies actually do for small ecommerce clients (and what AI already covers)
When you feel frustrated with an agency, it is easy to think in all-or-nothing terms: either they own everything or you fire them and go fully DIY. A clearer way is to break their remit into layers.
A typical small-agency scope for ecommerce/DTC includes:
- Strategy and positioning – audiences, angles, offers, channels, budget mix.
- Offer design – bundles, discounts, guarantees, landing page concepts.
- Creative production – product photos, UGC-style clips, scripts, hooks, thumbnails.
- Variations and testing – multiple angles, formats, and audiences.
- Reporting – weekly or monthly summaries, “what worked, what did not”.
- Coordination – managing freelancers, ad platforms, and your internal team.
Today, AI tools can realistically cover parts of this list:
- Creative production: generate and remix product images, backgrounds, and lifestyle scenes; draft scripts and captions; cut raw footage into short-form ads.
- Variations and testing: spin up multiple hooks, CTAs, and visual variations quickly.
- Basic reporting: summarize performance, highlight winners/losers, suggest next tests.
Where humans still matter a lot:
- Strategy and positioning: understanding your margins, lifetime value, and brand constraints.
- Offer design: deciding when to push discounts vs. bundles vs. value adds.
- Interpretation: separating “platform noise” from real shifts in customer behavior.
- Guardrails: deciding which stories you will not tell, even if they might convert.
Instead of agency vs no agency, a better question is:
“Which slices of this work can we internalize with AI, and where do we still want human experts in the loop?”
For many 1–20 person ecommerce and DTC teams, the answer looks like this:
- Internalize: creative production, variations, day-to-day testing, and first-pass reporting.
- Keep human help: strategy, large-budget decisions, big creative bets, and brand guardrails.
Designing an AI-first in-house creative stack
You do not need a giant martech budget to replace a big chunk of agency work. You need a small, well-defined stack and consistent workflows.
Think in four building blocks:
1. Image engine
A tool (or combination of tools) that can:
- Generate product-on-white shots and on-brand backgrounds.
- Create lifestyle scenes that reflect your real customers.
- Resize and adapt assets for different placements (feed, Stories, Reels, TikTok).
2. Short-form video engine
A way to:
- Turn raw product clips or simple B-roll into TikTok/Reels-style ads.
- Add captions, hooks, and simple motion graphics.
- Quickly generate multiple cuts of the same concept.
3. Copy and concept engine
A workflow for:
- Generating hooks, scroll-stoppers, and CTAs.
- Turning product benefits into scripts and captions.
- Translating winning angles across channels (Meta, TikTok, email).
4. Asset library and naming conventions
This is where many DIY stacks fall apart. Without structure, you end up with 200 files called “final_final_ad2.mp4”.
At minimum, define:
- A central library (Drive, DAM, or similar).
- A naming standard that encodes: campaign, audience, angle, format, and version.
- A simple tagging scheme for “status” (draft, in test, winner, paused).
Then, instead of thinking “I need 50 random AI ads,” think in terms of angles.
Pick 2–3 core angles per product line, for example:
- Price saver – “Stop overpaying for X; here is the affordable alternative.”
- Time saver – “What used to take 30 minutes now takes 3.”
- Story / transformation – “From frustrated to delighted in one week.”
Use AI to riff inside those lanes. For each angle, generate:
- 3–5 image concepts.
- 2–3 short video variations.
- 5–10 copy variants (hooks + primary text + CTA).
A small team can now produce a full creative matrix for a campaign in an afternoon instead of a week—without losing the thread of what they are actually testing.
Guardrails for authenticity: avoiding the “AI-perfect but uncanny” trap
One of the biggest fears founders voice is: “If I lean on AI this hard, will my brand start to feel fake?”
The Reddit founder whose store survived the agency breakup put it bluntly: as long as ROAS stays healthy, does it matter if an image is AI-generated or real?
It does—if your creative drifts into a universe your customers do not recognize. You avoid that by putting simple authenticity guardrails around every batch of AI assets.
Before you greenlight a new set of ads, ask:
- Can a real customer see themselves in this scenario or image?
- Does any visual or copy element contradict how the product actually works?
- Does the tone match how you (or your best reps) talk to customers in email and DMs?
- Would you feel comfortable if a loyal customer knew this was AI-assisted?
Operationally, that means:
- Mixing AI assets with real customer photos, UGC, and founder clips.
- Avoiding over-polished visuals that promise impossible outcomes.
- Scheduling regular “manual sampling” days where someone senior reviews a slice of active ads and landing pages, even if most of the work is automated.
Authenticity is not about avoiding AI. It is about ensuring AI amplifies real stories instead of inventing a fantasy version of your brand.
Making the numbers work: agency vs AI-first economics
For this shift to be worth it, the economics have to make sense. Here is one simple way to compare scenarios.
Assume:
- Agency model: $4,000 monthly retainer + 10% of ad spend.
- Ad spend: $20,000 per month.
- Total monthly agency cost = $4,000 + (0.10 × $20,000) = $6,000.
In return, you get, say, 40 fresh ad variants per month across platforms.
Effective cost per tested variant ≈ $6,000 ÷ 40 = $150.
Now consider an AI-first in-house stack:
- AI tools (image, video, copy): $400–$600 per month in subscriptions.
- Internal operator time: 15 hours per month at an internal cost of $60/hour = $900.
- Optional fractional strategist: $1,000 per month for a focused strategy session and periodic reviews.
Total monthly “stack” cost ≈ $2,300–$2,500.
If you can produce 80 variants per month with this setup, your effective cost per tested variant drops to roughly:
$2,400 ÷ 80 = $30 per variant.
This is a rough model, but it illustrates the leverage. You are not trying to drive your agency cost to zero on day one. You are trying to:
- Reduce fixed creative costs per tested idea.
- Keep or improve blended ROAS.
- Increase your visibility into what is being tested and why.
The safest way to run this comparison is not to flip a switch, but to design a controlled test.
For example:
- Weeks 1–4: Let your AI-first stack own 30–50% of spend on one or two key products, while the agency continues to run the rest.
- Tag campaigns clearly so you can compare “Agency-led” vs “AI-first in-house” performance.
- Measure ROAS, CTR, and cost per incremental purchase for each side.
If the AI-first slice holds performance within, say, 10% of your baseline while materially cutting fixed costs, you have strong evidence you can safely renegotiate scope or further internalize creative.
A quick reassurance: you do not have to fire your agency overnight
This is an important mindset point, especially if your agency relationship is long-standing or politically sensitive.
You can treat this as a scope experiment, not a breakup.
For example:
- Phase 1 – Test lane: Ask the agency to focus on higher-level strategy, larger campaigns, or new market launches, while you internalize day-to-day creative refreshes for evergreen products using your AI stack.
- Phase 2 – Performance review: After 4–8 weeks, compare results. Are your AI-led evergreen campaigns holding or beating the agency’s evergreen performance? How much internal time did it actually take?
- Phase 3 – Rescope, not rip-and-replace: If the numbers look good, renegotiate the agency remit around strategy, production of large signature campaigns, or creative direction—while your in-house AI stack handles volume and iteration.
This framing makes the change less threatening. It also aligns incentives: you still value the agency’s expertise, but you are no longer paying premium rates for tasks AI can help you run internally.
Where Get BOB fits: orchestration, logging, and human checkpoints
Once you lean into AI-generated creative, a new risk shows up: chaos.
Without an orchestration layer, you can end up with:
- Dozens of tools and exports with no single source of truth.
- Little visibility into which AI-generated angle actually drove results.
- Ads going live without anyone catching a risky claim or off-brand tone.
Get BOB is designed to be the orchestration layer that keeps your AI-first stack manageable. In an AI-driven creative workflow, a BOB can:
- Watch for triggers: new products, promos, inventory changes, or performance shifts in your store and ad accounts.
- Generate briefs and call AI tools: create structured prompts for image and video generation (“5 lifestyle images for Product X, angle = time saver, audience = busy parents”) and send them to your chosen tools.
- Enforce naming conventions: save outputs into your asset library with consistent metadata (campaign, audience, angle, format, version, status).
- Push assets into platforms: upload selected variants to Meta, TikTok, or Google, mapping each asset to the right ad sets and audiences.
- Log performance: pull key metrics (CTR, CPC, ROAS, revenue) back against each variant and angle, so you can actually see which AI-generated ideas work.
- Create human checkpoints: require human approval before:
- New offers or price claims go live.
- Sensitive imagery is used (for example, health, finance, children).
- A large budget shift is made based on early results.
- Summarize and suggest: send a weekly digest like “5 top-performing AI variants, 3 underperformers to pause, 2 new angles worth testing,” in plain language.
In other words, the moat is not any single AI model. It is your orchestration layer—how you connect tools, codify learnings, and keep humans in the loop where it matters.
A 30-day playbook to test this in your own shop
You do not need a quarter-long project plan to see whether an AI-first creative stack can work for you. Here is a pragmatic 30-day sequence.
Week 1: Map your current agency / creative workflow and costs
- List what your agency actually does for you today, step by step, from brief to reporting.
- Quantify monthly costs: retainer, percent of spend, extra fees for creative refreshes.
- Pick one product or campaign with meaningful spend where you are comfortable running a test.
- Define success criteria, for example:
- “Maintain ROAS within ±10% of current baseline.”
- “Cut effective cost per tested variant by 50%.”
- “Improve time from idea to live ad from 10 days to 2–3 days.”
Week 2: Assemble a minimal AI stack and define angles
- Choose your basic tools for images, video, and copy. Keep it simple; you can always swap later.
- With your team (or agency, if you are collaborating), define 2–3 core angles for the chosen product or campaign.
- Have a BOB set up workflows to:
- Generate briefs for each angle.
- Call your AI tools to create assets.
- Save outputs to a shared library with consistent naming and tagging.
Week 3: Launch a controlled test
- Put your AI-first creative into market for 30–50% of the spend on the test product or campaign.
- Keep the rest of the spend on your existing baseline campaigns so you have a clean comparison.
- Ask your BOB to:
- Track results per variant and angle.
- Send you twice-weekly summaries.
- Flag any ads with unusually low performance or potential brand risk for review.
Week 4: Review, decide, and rescope
- Compare key metrics: ROAS, CTR, CPC, cost per purchase, and cost per tested variant.
- Review qualitative fit: Did any AI-generated assets cross your authenticity lines? How much manual clean-up did they require?
- Estimate internal workload: Did this actually consume 15 hours of operator time, or 40? Where did friction show up?
- Make a clear decision for the next 60–90 days:
- Scale up AI-first share of spend, with tighter guardrails.
- Renegotiate agency scope to focus on strategy and flagship campaigns.
- Or, if performance was clearly worse, adjust the stack and try again with a narrower test.
You come out of this month not with a philosophical opinion about AI, but with your own data.
Bringing it together
Founders on Reddit are demonstrating what many small ecommerce and DTC operators quietly suspect: a well-designed, AI-first creative stack can often match agency-level performance at a fraction of the fixed cost.
The real shift is not from “professional” to “DIY.” It is from paying a middleman who also uses AI, to owning an AI-assisted workflow yourself—backed by clear guardrails and an orchestration layer like Get BOB.
If you structure your experiment carefully, keep authenticity front and center, and let a BOB handle the glue work between tools and platforms, you can:
- Test more creative ideas for less money.
- Understand which angles actually move your numbers.
- Decide, with confidence, which parts of your agency relationship you want to keep—and which you are ready to bring home.