The Discovery Phase: Charging You to Learn What You Already Know

The Discovery Phase: Charging You to Learn What You Already Know

You have been running this company for eleven years. You built it from a freelance operation in a spare bedroom to a team of sixty people with a real HR department and a coffee machine that requires a minor in engineering to operate. You know your customers better than they know themselves. You know which campaigns worked and which ones died quietly, which product lines carry the margins and which ones exist for strategic reasons that made sense in 2019. You know everything. Which is why it’s so refreshing when the agency shows up and tells you they need six weeks to discover your business before they can begin work.

The discovery phase is the consulting industry’s most elegant invention: a paid period of time during which external parties learn basic information about your company, ask questions whose answers are already in the documents you gave them, and ultimately produce a report summarizing what your team has known since approximately the third month of the company’s operation. It costs between €15,000 and €80,000, depending on the agency’s day rate and how many post-it notes they use.

The Workshop and Its Rituals

Discovery typically begins with a workshop. The workshop is held offsite when possible, because breakthroughs require a different conference room than the one you use for regular meetings. The agency arrives with a facilitator, a notetaker, a deck of questions they’ve asked every client since 2017 with the logos swapped out, and a large supply of differently colored sticky notes — because insight, as everyone in consulting knows, is color-coded.

The exercises themselves follow a predictable choreography. There is a “hopes and fears” exercise, where stakeholders write their hopes and fears on — you guessed it — sticky notes of different colors, and then place them on a wall. There is a “customer journey mapping” exercise where everyone maps the customer journey they already know onto a large piece of paper, giving it a visual formality that makes it feel like a discovery rather than a transcription. There is an “align on priorities” exercise that reveals priorities everyone present already agreed on before entering the room.

At the end of the workshop, the facilitator takes photos of all the sticky notes. These photos will form the visual evidence of discovery. They will appear in the discovery report under the heading “Key Findings,” even though the findings are, essentially, the things you told them during the introductory call.

The Report That Validates the Investment

Several weeks later, the agency delivers the discovery document. It is substantial — forty to eighty pages, beautifully designed, with your logo on the cover and a table of contents that suggests it contains more than it does. The document is organized around themes that emerged from the workshops, meaning: themes that the facilitation team decided on before the workshops and then found evidence for in the post-it notes.

Section one: Company Overview. This is where the agency demonstrates that they have read your website. Section two: Market Context. This is where they demonstrate access to a market research subscription. Section three: Customer Insights. This is where the personas live — see our previous discussion of Jennifer. Section four: Strategic Tensions. This is the clever part, where the consultants identify genuine contradictions within your business (there are always genuine contradictions within every business) and present them as discoveries rather than as the permanent condition of operating a complex organization.

The document ends with Recommendations, which are either (a) so broad as to be applicable to any company in any sector, or (b) so specific as to be obviously the direction the agency wanted to take from the beginning, with the discovery process serving as elaborate justification. Either way, the Recommendations justify Phase Two, which is the actual work, which is what you thought you were hiring them for in the first place.

Why Companies Keep Paying For It

Here is the thing: the discovery phase often does produce something valuable, just not what it claims to produce. The real value is not the report. The real value is forcing internal alignment — getting people in a room who don’t normally talk, making them articulate things that are understood but never stated, creating a shared document that gives disparate teams a common reference point. That’s genuinely useful, and it’s worth some money.

The problem is the mythology around it. The myth that external parties will discover something about your business that you don’t already know. The myth that the sticky note exercises produce insights rather than document existing knowledge. The myth that the forty-page report is the output of discovery rather than the justification for it. If companies commissioned “internal alignment workshops” instead of “discovery phases,” the same thing would happen at a fraction of the cost and with considerably less pretense.

But “internal alignment workshop” doesn’t have the same ring as “discovery.” Discovery implies something will be found. Something unknown made known. It suggests that before the agency arrived, you were operating in the dark — which is just condescending enough to feel like expertise.

After the Discovery, Before the Work

Between the discovery report and the actual deliverables, there is often a strategy phase. Then a creative brief phase. Then a brief alignment phase. Then, eventually, something resembling work begins. By that point, the market has shifted, the internal champion who hired the agency has moved to a different company, and the original brief — which was actually pretty clear — has been through enough transformation that nobody remembers what they were trying to do in the first place.

You don’t need six weeks and a roomful of sticky notes to understand what you want. You need a brief that’s honest about the problem, a team that’s honest about what they can solve, and enough mutual trust to skip the theater and get to work. The NoBriefs shop exists precisely for the people who believe that creative work should start with clarity, not with a photo of a sticky-note wall. The Fuck The Brief line isn’t anti-process — it’s anti-pretense. There’s a difference, and the difference costs about €40,000.

LinkedIn Thought Leadership: The Art of Saying Nothing at Scale

LinkedIn Thought Leadership: The Art of Saying Nothing at Scale

It is Monday morning, 8:47 a.m., and somewhere in a home office with a ring light, a marketing director is composing a post about failure. Not a real failure — a curated one. A failure that taught a lesson, revealed a strength, and ultimately led to the breakthrough that now justifies the post. The failure has a three-act structure. It ends with gratitude. It will receive 847 likes from people who also failed productively and learned exactly the right things. It is LinkedIn thought leadership at its most refined: maximum emotion, minimum information, zero actual thoughts.

LinkedIn has achieved something remarkable. It has turned professional networking into a genre of inspirational fiction, a content format so thoroughly colonized by performed wisdom and humblebrag vulnerability that reading it has become a meditative exercise in learning to feel nothing. The thought leadership industrial complex is alive and well, and it is absolutely crushing it in terms of engagement metrics.

The Anatomy of a Thought Leadership Post

The classic LinkedIn thought leadership post follows a structure as reliable as a sonnet. It opens with a short sentence. Often a question. Or a bold claim. Then it subverts expectations. Then the pivot — usually signaled by the word “But.” Here comes the lesson. The lesson is universal. Something about resilience, or curiosity, or the importance of asking for help. The lesson was learned through experience, though the experience is described at a level of abstraction that makes it applicable to literally everyone and therefore useful to no one.

The post ends with a call to action. “What do you think?” or “Drop your thoughts below” or “Tag someone who needs to hear this.” The comments fill with people saying “so true,” “this is everything,” and “sharing this with my team.” Nobody shares it with their team. The team is also on LinkedIn, composing their own posts about different lessons learned through similar non-specific experiences.

The images are either selfies taken at conferences (conference selfies have their own sub-genre — the earnest handshake, the panel photo where everyone looks like they’re mid-profound-statement), or black-text-on-white-background quote cards, or photos of notebooks with the first line of the post written in them, because apparently writing the post wasn’t enough and someone needed to also photograph themselves writing the post.

The Thought That Isn’t Leadership

The term “thought leadership” was coined in the 1990s to describe genuine domain expertise — the kind of thinking that moves an industry forward, challenges conventional wisdom, or introduces a framework that changes how people work. It was a useful concept. Then it was discovered by content strategy, and like all useful concepts discovered by content strategy, it was immediately turned into a template.

Real thought leadership is uncomfortable. It requires having an opinion that some people will disagree with. It requires specificity — not “failure leads to growth” but “here is exactly why this specific approach to product pricing fails in enterprise SaaS, and here is the data.” It requires, at minimum, the presence of an actual thought.

What LinkedIn has produced instead is the aesthetics of thought leadership without the substance: the vulnerability without the risk, the opinion without the argument, the expertise without the specifics. It is possible to post on LinkedIn every day for a year and say absolutely nothing of professional value while accumulating thousands of followers who are waiting for you to say something of professional value. The followers keep waiting. The posts keep coming. The engagement keeps climbing. Everyone involved calls this success.

Why Marketers Are the Worst Offenders

Every profession has its LinkedIn voice, but marketing has a special relationship with the platform because marketers understand, better than anyone, how LinkedIn’s algorithm works and what kind of content it rewards. The result is a self-referential loop: marketers post content optimized for LinkedIn engagement about the importance of authentic content. They write about the dangers of vanity metrics while obsessing over their follower count. They discuss the value of genuine connection while A/B testing their post hooks to maximize the click-through on “see more.”

There is a particular brand of LinkedIn post beloved by marketing professionals that goes: “Nobody talks about this, but [extremely common observation in the marketing industry].” Nobody talks about the importance of knowing your audience. Nobody talks about how email marketing still works. Nobody talks about the fact that strategy should precede tactics. These observations have been talked about, relentlessly, in every marketing conference, podcast, and textbook for the past thirty years. They continue to perform extremely well on LinkedIn because the platform has no memory and the audience is always new.

The Exit, If You Want It

There is a version of LinkedIn that is genuinely useful — for job searching, for connecting with specific people you want to work with, for occasionally finding a piece of writing that is actually good because someone decided to write something true and specific instead of broadly inspirational. That LinkedIn exists. It is outnumbered about forty to one by the thought leadership version, but it’s there.

The antidote to thought leadership is simply leadership — having a point of view you’d defend in a meeting, not just in a post. Knowing something well enough to be wrong about it in an interesting way. Writing something that not everyone will like because it actually says something.

If you’ve ever posted something on LinkedIn and then watched the algorithm reward you for it while knowing, in the quiet part of your brain, that you wrote it for the algorithm and not for the reader — you’re already more self-aware than most. That’s a start. Wear the NoBriefs badge honestly: the Spreadsheet Sloth knows that not all productivity is real productivity, and not all posting is real communication. Sometimes it’s just noise wearing a ring light.

A/B Testing Everything Except Your Strategy

A/B Testing Everything Except Your Strategy

You have tested the green button against the blue button. You have tested “Buy Now” against “Get Started” against “Claim Your Spot.” You have tested hero images, subject lines, headline fonts, CTA placement, and whether the word “free” in an email subject line increases or decreases open rates (the answer changes every six months, depending on who you ask). You have optimized every pixel of the customer journey. You have never once questioned whether the destination is worth reaching.

Welcome to the golden age of A/B testing, where we measure everything and understand nothing, where data replaces judgment, and where the appearance of scientific rigor substitutes for actual strategic thinking. It’s a beautiful system, really — endlessly productive, perpetually inconclusive, and completely immune to the uncomfortable question of whether you’re optimizing the right thing at all.

How Testing Became a Religion

The democratization of A/B testing tools was supposed to bring empiricism to marketing. And for a while, it did. Instead of arguing about opinions in a conference room, teams could run experiments. Instead of HiPPO decisions (Highest Paid Person’s Opinion), data would rule. The shift was genuine and valuable. Then, like most genuinely useful things, it was taken approximately forty steps too far.

The problem began when testing stopped being a tool and became an identity. “We’re a data-driven culture” is now the marketing equivalent of “we move fast and break things” — a slogan that sounds rigorous and signals virtue without committing to anything specific. In practice, data-driven culture often means: we run tests on things that are easy to test, measure things that are easy to measure, and declare victory when a metric goes up, even if nobody can explain why or whether it matters.

The result is teams that have run 200 experiments on their homepage and cannot tell you what their brand stands for. Teams that can produce a confidence interval for a subject line variant but have no idea why customers churn after 90 days. Teams that treat statistical significance as a moral category — as if a p-value of 0.04 is not just a number but a verdict from the universe that green buttons are objectively better than blue ones.

The Local Maximum Trap

Here is what nobody puts in the A/B testing case study: optimization without strategic direction leads you, very efficiently, somewhere you don’t want to be. Every incremental test improves what exists. None of them ask whether what exists should continue to exist.

This is the local maximum problem. You can optimize a mediocre product page to be the best possible version of a mediocre product page. Your conversion rate will climb 0.3% per experiment until it can climb no further. You will have squeezed every drop from a lemon that, strategically speaking, you should have replaced with a different fruit two years ago. But the dashboard looks great. The quarterly report is full of winning experiments. Everyone involved gets a mention in the retrospective under “wins.”

The deeper issue is that A/B testing is structurally incapable of questioning its own premises. You can test two versions of a landing page. You cannot test whether the landing page is the right mechanism for what you’re trying to achieve. You can test two subject lines. You cannot test whether email is the right channel. Those questions require judgment, context, and the willingness to consider that your current approach might be fundamentally wrong — qualities that do not produce clean dashboards and cannot be automated.

Data as Alibi

There is a more cynical function that testing serves in large organizations, and it’s worth naming honestly: it provides cover. When a decision goes wrong, “we had the data” is the modern version of “I was just following orders.” It distributes responsibility so thoroughly that nobody is accountable for anything. The test said to do it. The algorithm recommended it. The confidence interval supported it. Who could argue with that?

This is why genuinely bold creative decisions almost never come out of A/B testing. Testing can tell you which version of an existing concept performs marginally better. It cannot tell you to do something completely different. It cannot tell you to run the campaign that makes your legal team nervous and your competitors jealous. It cannot tell you to make the thing that nobody has made before because, by definition, you can’t test something against itself before it exists.

The great brand-building moments in marketing history were not A/B tested into existence. They were made by people with strong points of view who were willing to be wrong in an interesting way rather than right in a boring one. Testing would have optimized those ideas into something safe and forgettable.

What Good Testing Actually Looks Like

None of this means testing is bad. It means testing is a tool, not a philosophy. Used well, it answers tactical questions quickly and cheaply: which headline communicates the value proposition more clearly? Which onboarding flow reduces drop-off? These are real questions with measurable answers, and testing is exactly the right instrument for them.

The problems start when testing substitutes for strategy, when the question “what should we be doing?” is replaced by “which version of what we’re already doing performs better?” The first question is hard, uncomfortable, and requires people in a room with different opinions and no easy answers. The second question produces a dashboard. Guess which one gets more organizational energy.

If you’re the kind of marketer who has sat in the meeting where someone says “let’s test it” as a way of avoiding a decision, you know exactly what this piece is about. And if you want a daily reminder that your job is more than optimizing button colors, the NoBriefs shop has what you need — starting with the Fuck The Brief collection, for when the brief itself is the thing that’s failing the test.

Meet Jennifer: The Marketing Persona Who Has Never Existed

Meet Jennifer: The Marketing Persona Who Has Never Existed

Somewhere in your company’s shared drive, nestled between the Q3 performance report nobody opened and the brand guidelines nobody follows, lives Jennifer. Jennifer is 34. She’s a mid-level marketing manager. She drives a hybrid car, loves yoga on weekends, and is “digitally savvy but values authenticity.” She has two kids named something tasteful like Mia and Lucas. She earns €52,000 a year and her biggest pain point is “finding time for herself.” She is your target audience. She does not exist.

The marketing persona is one of the most elaborate fictions the industry has ever produced — a literary character with the depth of a fortune cookie and the strategic value of a horoscope. Yet companies spend thousands of euros on workshops, consultants, and wall-sized sticky-note sessions to give birth to Jennifer, only to promptly file her away and never consult her again.

How Personas Are Born (and Die on the Same Day)

The persona creation workshop follows a sacred ritual. A facilitator — who charges €1,200 a day and owns at least one pair of quirky glasses — leads a cross-functional team through a series of exercises. What does our customer fear? What does she aspire to? What does she read? The team, composed entirely of people who are not the customer, makes educated guesses. Someone writes “LinkedIn and maybe TikTok?” Someone else says “she probably shops at Zara but wishes she could afford COS.” Everyone nods. Jennifer is born.

By the end of the workshop, Jennifer has a photo (stock, obviously — a smiling woman of ambiguous ethnicity chosen for maximum inclusivity with minimum effort), a name, and an origin story more detailed than most employee onboarding documents. She has a “jobs to be done,” a “day in the life,” and a quote that nobody actually said but feels like she would say, probably something like: “I want brands that understand me.”

Jennifer is then placed in a beautiful PDF, sent to the design team with the instruction to “keep her in mind,” and never spoken of again. The actual campaign is made for the CEO’s cousin, who saw a competitor’s ad and said they wanted something “similar but more premium.”

The Data Problem Nobody Mentions

Here is the uncomfortable truth about most marketing personas: they are not based on data. They are based on vibes, assumptions, and whatever the Head of Sales said loudly in the last meeting. Real customer research — interviews, behavioral data, purchase pattern analysis — is expensive, time-consuming, and inconveniently messy. Personas, by contrast, can be invented in an afternoon and presented with the visual confidence of a TED Talk.

The result is a character who is simultaneously too specific and completely useless. Jennifer is 34, but what about the 28-year-old who actually converts? Jennifer is worried about sustainability, but what if your actual buyer is a 55-year-old procurement manager who just wants the invoice on time? The persona, instead of clarifying the audience, creates a fictional target that the team aims at while the real customers wander in through the side door.

There is a particular cruelty in this process: the more detailed the persona, the more convincing it feels, and the more dangerous it becomes. A one-pager with a stock photo and a salary band has the aesthetic authority of market research. Nobody questions Jennifer. Questioning Jennifer means questioning the workshop, and the workshop cost €4,000.

The Persona Industrial Complex

The persona business is thriving. There are tools, templates, platforms, and entire methodologies dedicated to helping you build better fictional humans. Some of these tools use AI to generate personas from your CRM data, which is genuinely useful, though it does raise the philosophical question of whether Jennifer generated by an algorithm is any less made-up than Jennifer generated by a roomful of people eating catered sandwiches.

The larger problem is structural. The persona is a tool designed for a world where marketing teams need to humanize abstracted data — to give a face to a segment. The intent is noble. The execution is a game of telephone between your actual customers and a stock photo woman who enjoys weekend yoga. Somewhere between the insight and the output, Jennifer stopped being a tool and became a totem.

What would actually help? Talking to real customers. Reading real complaints. Watching real behavior. Conducting interviews where the person on the other side says unexpected things that ruin your assumptions. It’s less photogenic than a persona card, and it won’t look good framed on the office wall, but it has the rare quality of being true.

Jennifer Sends Her Regards

Jennifer doesn’t mind that you’ve forgotten her. She’s used to it. She lives in the slide deck between the market sizing chart and the competitive landscape table, eternally 34, eternally concerned about work-life balance, eternally waiting for a brand that finally gets her.

Meanwhile, your actual customers are out there — complicated, inconsistent, price-sensitive in ways Jennifer isn’t, loyal to brands Jennifer has never heard of. They don’t have a name or a stock photo. They haven’t been workshopped. They’re a mess. They’re the whole point.

If the absurdity of marketing theater is something you feel in your bones every day, you’re in the right place. The NoBriefs shop was built for people who have sat through one too many persona workshops and lived to tell the tale. The KPI Shark knows that the most dangerous metric is the one that makes everyone feel productive without doing anything useful. Come join the club.

Performance Marketing Killed the Creative Star

Performance Marketing Killed the Creative Star

In the beginning, there was the big idea. The campaign that changed how people felt about a brand. The work that lived in culture for years, that people remembered decades later, that made careers and moved markets in ways that couldn’t be cleanly attributed to any single impression or click. The industry was built on the belief that great creative work was worth making even when you couldn’t fully explain why it worked — that the brand-building effect of advertising was real, even if it was diffuse, delayed, and difficult to measure in a quarterly report.

Then came performance marketing, and the big idea got a conversion rate.

The logic was impeccable: digital advertising allowed you to measure what was working with a precision that had never existed before. You could see exactly how many people clicked, converted, bought, and returned. You could A/B test everything. You could allocate budget to what performed and cut what didn’t. The result, in theory, was a more efficient, more accountable, more rational marketing function. The result, in practice, was something more complicated and considerably more damaging to the craft of advertising than the industry has been willing to fully acknowledge.

What Gets Measured Gets Made (And Everything Else Doesn’t)

The practical effect of optimizing marketing spend around measurable performance metrics is that you create a systematic bias against anything whose value is real but diffuse — brand awareness, emotional resonance, cultural relevance, the slow accumulation of positive associations that determines whether a brand is trusted, liked, and chosen over time. These things are hard to measure in a dashboard. Therefore, they are systematically underfunded in organizations where the dashboard is the decision-making tool.

This produces a specific kind of marketing that is technically excellent and strategically hollow: campaigns that drive clicks without building brand equity, content that converts without creating preference, advertising that performs in the short term by accelerating existing intent without doing anything to create future intent. You can run this playbook for several years and look extremely good in the monthly report while quietly dismantling the brand foundations that made the performance possible in the first place.

The grocery industry has a term for this: “eating the seed corn.” You can solve this year’s food problem by consuming next year’s seeds. The harvest numbers look fine right up until there’s no harvest.

The Creative That Performance Marketing Produces

Ask a performance marketing team to brief creative work and watch what happens to the brief. The audience becomes “people who have previously expressed purchase intent in this category.” The message becomes the offer, the discount, the feature comparison. The format becomes whatever the platform has determined converts best this quarter. The success metric becomes cost per acquisition.

None of this is wrong. All of it, applied exclusively, produces advertising that is indistinguishable from every other piece of advertising in the same category — because it’s optimized against the same data, serving the same audience, making the same offer, in the same formats that the algorithm has decided perform best. The result is creative convergence at scale: entire categories of advertising that look and sound identical because they’ve all been optimized toward the same conversion signals.

The creative professionals working in pure performance environments often describe the experience the same way: the work is technically demanding, the feedback loop is fast and clear, and the creative latitude is approximately zero. You are not making advertising; you are making variables for a test. This is a legitimate discipline. It is not what most people went into advertising to do, and the talent pipeline consequences of making it the dominant mode of the industry are becoming visible in the quality of work being produced.

The Evidence That Brand Building Actually Works

The research on this is not ambiguous. The Binet and Field “The Long and the Short of It” analysis of IPA effectiveness data established clearly that the optimal marketing investment split for most categories is approximately 60% brand-building activity and 40% performance/activation activity — with the brand-building investment providing the context and preference that makes the performance activity more efficient.

The brands that have maintained this balance — that have continued to invest in genuinely creative, genuinely brand-building work even when the short-term attribution is murky — have consistently outperformed brands that shifted entirely to performance marketing, particularly over time horizons longer than one quarter. This finding has been replicated across categories, markets, and time periods. It is as close to settled science as marketing research gets.

The industry knows this. The reversion to performance-only thinking happens anyway, because quarterly performance pressure, short CMO tenures, and the genuine difficulty of defending brand investment in a meeting full of people who prefer dashboards with green arrows make the rational long-term choice organizationally difficult to sustain. Knowing the right thing and doing it are different problems.

The creatives who have spent their careers making brand-building work — work that doesn’t convert immediately, that operates on emotion rather than offer, that tries to make people feel something rather than click something — deserve recognition for doing the harder, more important job. The Fuck The Brief collection is for everyone who’s been told their work “can’t be attributed” and knows that’s the whole point.

Good work takes time to prove itself. In the meantime, the NoBriefs Club shop has your back.

0
    Your Cart
    Your cart is emptyReturn to Shop