Every agency does weekly reviews. Almost none of them close the loop. Here's what typically happens: someone pulls last week's numbers into a deck, identifies that cost-per-lead spiked 28%, recommends pausing two underperforming ad sets, and the meeting ends. Next week, nobody checks whether those ad sets were actually paused. Nobody knows if the pause helped. The "review" was really just a suggestion box with a slide deck attached.
We ran that system for years before admitting it was broken. The reviews weren't bad — the diagnostics were solid, the recommendations were sound. But recommendations without verification are just opinions with data. So we rebuilt the entire process around one principle: every action proposed in a review must be verified in the next review. We call it the Closed-Loop Review, and it changed how we operate across our Singapore and Jakarta delivery pods more than any tool, platform, or hire.
Key Takeaways
- Most weekly reviews are suggestion boxes, not systems. They diagnose and recommend but never verify whether actions were executed or whether they worked.
- The Closed-Loop Review adds a Loop Review section that checks last week's proposed actions — what was done, what happened, and what it means for this week.
- Actions are categorised as Execute, Test, or Monitor — each with a different verification standard. Executes are binary (done/not done). Tests need a hypothesis and success criteria. Monitors need a trigger threshold.
- Archetype-specific diagnosis matters. A CPL spike for a lead gen client triggers a completely different diagnostic tree than a ROAS drop for ecommerce — same metric movement, different implications.
How to build a weekly marketing performance review that closes the loop — from diagnosis to action to verification. The 6-section Closed-Loop Review structure used by Kaliber Group to ensure campaign optimizations actually get executed and verified.
The Problem With Standard Reviews
Standard weekly reviews follow a simple pattern: diagnose, recommend, forget. Look at this week's numbers, compare them to targets, identify what's off, suggest fixes, move on. Next week, start from scratch. The result is a kind of institutional amnesia — the team generates good insights every week but never compounds them because there's no mechanism for follow-through.
The deeper problem is that without verification, you can't learn. If you recommend shifting budget from Campaign A to Campaign B but never check whether it happened or what resulted, you have no feedback loop. You're making decisions in a void, repeating mistakes you've already solved, and solving problems you've already fixed. The review becomes performative — it looks rigorous but produces no durable value.
We also noticed a subtler failure mode: recommendation inflation. When nobody verifies, people recommend more aggressively. Why not suggest ten changes when you know nobody's tracking? This floods the team with action items, most of which get ignored, and the important ones get lost in the noise. A system without accountability rewards volume over judgment.
The 6-Section Closed-Loop Structure
We rebuilt the weekly review into six sections. Each one exists for a specific reason, and the order matters.
Closed-Loop Review — 6 Sections
Section 1 — Executive Pulse: One sentence that captures the week. "Lead volume is on target but CPL increased 18% due to a single campaign overspending on a broad audience." This forces the reviewer to synthesise rather than enumerate. If you can't summarise the week in one sentence, you don't understand it well enough yet.
Section 2 — Scorecard: Five to seven metrics compared against two benchmarks: the target, and the trailing 4-week average. Targets tell you where you want to be. The trailing average tells you what's normal for this account. A CPL of $45 might look alarming against a $30 target, but if the trailing average is $42, you're seeing a 7% increase, not a 50% spike. Context changes the diagnosis entirely.
Section 3 — Loop Review: This is the section that transforms the review from a report into a system. Every proposed action from the previous week gets reviewed: Was it executed? If yes, what was the result? If no, why not? This creates accountability without blame — the goal isn't to catch people out, it's to learn. Did pausing that ad set actually reduce CPL? Did the new audience test deliver the volume we hypothesised? If we don't ask, we don't learn.
Section 4 — Flags and Observations: Not everything that's notable requires action. Some things need watching. Frequency creeping up from 2.1 to 2.4 over three weeks isn't a crisis, but it's worth flagging. A new competitor entering the auction isn't actionable today, but might explain next week's CPM increase. This section prevents the two failure modes of over-reaction (changing everything that moves) and under-attention (ignoring gradual shifts).
Section 5 — Proposed Actions: A maximum of five actions, each categorised. Execute actions are binary — do this specific thing by next review. Test actions include a hypothesis and success criteria — "we believe increasing budget on Campaign B by 20% will maintain CPL below $35 while increasing volume by 15%." Monitor actions include a trigger threshold — "watch frequency on retargeting; if it exceeds 3.0, we'll pause and refresh creative." The categorisation matters because each type has a different verification standard in next week's Loop Review.
Section 6 — Look-Ahead: What's coming that might affect next week's numbers? Monthly budget resets, seasonal events (Lunar New Year, Hari Raya, National Day — critical for our APAC clients), client meetings where strategy might change, creative refreshes landing mid-week. This section exists because performance marketers are notoriously bad at anticipating — we react to last week's data instead of preparing for next week's conditions.
Archetype-Specific Diagnosis
Here's something most review templates get wrong: they treat all accounts the same. A CPL increase on a lead generation account and a ROAS decrease on an ecommerce account might look like the same kind of problem — performance is getting worse. But they trigger completely different diagnostic trees.
When CPL spikes on a lead gen account, the first question isn't "what went wrong" — it's "did lead quality change?" Sometimes CPL goes up because the algorithm is finding higher-intent leads who are more expensive to reach but convert at a higher rate downstream. If you react to the CPL spike by cutting budget, you might be killing your best-performing period. The diagnostic tree for lead gen starts with quality verification, then moves to volume analysis, then budget allocation, then audience saturation.
For ecommerce, a ROAS decline triggers a different sequence: first check whether average order value shifted (a ROAS drop with stable conversion rate might just mean people bought cheaper items), then check if it's a traffic quality issue (new audiences converting at lower rates), then check competitive pressure (CPMs spiked because a competitor entered the auction), then check creative fatigue.
We built these diagnostic trees into the review process so the analysis follows the right path based on the client archetype — lead gen, ecommerce, B2B SaaS, local services. Same data, different questions, different conclusions. This matters especially in Southeast Asia where we see high variance across markets — what works in Singapore rarely translates directly to Indonesia or Malaysia without recalibration.
The Closed Loop in Practice
The Closed-Loop Cycle
The real power of this system shows up after a few weeks. By week three, the Loop Review section becomes the most valuable part of the review — not because it catches mistakes, but because it builds institutional memory. You can trace the lineage of a decision: "We proposed shifting budget to Campaign B in W08. It was executed in W09. CPL improved 12% by W10, confirming the hypothesis." That's a verifiable learning, not an anecdote.
AI accelerates every part of this loop. Our system pulls live data from Google Ads and Meta APIs at the moment of review — no stale dashboards, no manual exports. It runs archetype-specific diagnostic trees automatically, flagging anomalies and proposing actions with the right categorisation. But the Loop Review — the verification step — still requires human judgment. Did this change actually cause the improvement, or was it coincidence? That's a question AI can inform but humans must answer.
The constraint of "maximum five actions" is deliberate. More than five and things don't get done. Fewer than three and you're probably not looking hard enough. The sweet spot is four to five well-specified actions with clear ownership and clear verification criteria. When everything is an action item, nothing is.
Frequently Asked Questions
What should a weekly marketing performance review include?
A strong weekly review needs six components: an Executive Pulse (one-sentence summary), a Scorecard (5-7 metrics vs. targets and trailing averages), a Loop Review (verification of last week's actions), Flags and Observations (things to watch but not act on), Proposed Actions (3-5 categorised next steps), and a Look-Ahead (upcoming events that might affect performance). The critical piece most reviews miss is the Loop Review — without it, you diagnose but never verify.
How do you track whether campaign optimizations actually worked?
Every proposed action gets reviewed in the following week's Loop Review section. Execute actions are verified as done or not done, with results measured. Test actions are evaluated against the hypothesis and success criteria defined when they were proposed. Monitor actions are checked against their trigger thresholds. This creates a feedback loop where you're not just making changes — you're learning which changes produce results.
What is a closed-loop marketing review process?
A closed-loop review is one where every recommendation is tracked through to execution and verification. The cycle runs: Review (diagnose) to Propose (recommend actions) to Approve (team validates) to Execute (changes made) to Verify (check results in next review). The loop closes when verification insights feed back into the next review cycle. Most review processes are open-loop — they recommend but never verify.
How often should you review Google Ads and Meta campaign performance?
Weekly is the right cadence for strategic reviews — it's frequent enough to catch issues before they compound, but not so frequent that you're reacting to noise. Daily pacing checks (automated, quick, spend-focused) complement the weekly review by catching spend anomalies in real time. Monthly reviews serve a different purpose: they're for trend analysis and strategic shifts that aren't visible at the weekly level.
What is the difference between a marketing report and a marketing review?
A report presents data — here's what happened. A review interprets data and drives action — here's what happened, why it happened, what we should do about it, and how we'll verify it worked. Reports are backward-looking documents. Reviews are forward-looking decision-making sessions. The Closed-Loop Review adds a third dimension: it's also learning-oriented, building institutional memory about which actions produce which results.
How do AI tools help with weekly marketing reviews?
AI accelerates three parts of the review process: data collection (pulling live metrics from Google Ads and Meta APIs instead of manual exports), pattern detection (flagging anomalies and running archetype-specific diagnostic trees), and action proposal (suggesting categorised next steps based on the diagnosis). The Loop Review — verifying whether past actions worked and interpreting causation — still requires human judgment. AI makes the review faster and more thorough, but doesn't replace the decision-maker.
What metrics should be in a weekly marketing performance scorecard?
The specific metrics depend on the client archetype. Lead gen: leads, CPL, conversion rate, spend vs. budget, lead quality score. Ecommerce: revenue, ROAS, conversion rate, AOV, spend vs. budget. B2B SaaS: MQLs, cost per MQL, pipeline value, demo bookings. Every scorecard should compare metrics against two benchmarks — the target (where you want to be) and the trailing 4-week average (what's normal for this account). That dual comparison prevents both complacency and false alarms.