Ready to reinvent growth? Let's connect.

How does multi touch attribution work?

Multi-Touch Attribution, or MTA, seeks to better understand what truly drove or “caused” an action to happen. How does multi touch attribution work? Usually, these actions are online and commercial in nature. The most typical use case is understanding what drove an order or a transaction. These “part-worths” can then be summed up across many orders or transactions to form a complete picture of how different marketing channels contributed to the sales accrued in a given period.

For example, say a user purchased a subscription to a monthly ride-sharing service. It is easy to know that the user entered from a paid search ad; each ad has a unique referring URL that can be tracked. However, giving all credit to the paid search ad is most likely unfair. The user might have seen a social media campaign days before, or an ad on a baseball game weeks before. Those touches should also receive credit. The difference between a last-touch and multi-touch attribution is usually substantial. Lower-funnel channels usually look less attractive in a multi-touch approach (see Figure 1 below).

Figure 1: A last-touch versus a multi-touch approach can yield very different results. The “MTA effect” shows less attractive results for paid search and affiliate, and better results for DM, online video, and social in this example.

To accurately apportion credit to all contributing channels, it is necessary to estimate the effect that each touch had on the customer who ultimately ordered. This requires, at the most basic level, knowing that a previous interaction occurred. This is the number one problem with multi-touch attribution: gathering the “touches”.

Historical Context of Tracking

In the first decade of the 2000s, tracking users across digital properties was easier. Customers’ browsers held records of where they had been and what they had done. These cookies functioned as digital breadcrumbs, alerting an advertiser that a consumer had seen an ad weeks before. It was possible for advertisers to piece together customer journeys deterministically—that is, knowing concretely when and what someone saw before they made an order. With these data censuses, early MTAs could estimate each touches’ effects simply—typically applying even credit to each touch, with linear decays of impact over time—or using more complex methods.

Privacy Measures Impacting MTA

However, over the past several years, it has become more difficult to build truly deterministic MTAs. Privacy measures—whether cookie deprecation, walling off app data interchanges, or obfuscation of PII  in third-party data sources—have made it harder and harder to say that a pageview or click was the result of an individual customer with any degree of confidence. The tradeoff between privacy and customization will likely be a forever war (see Figure 2)—suggesting an owned, decoupled approach to attribution that doesn’t rely on fads or black boxes.

Figure 2: The push and pull between privacy and personalization will likely be a forever war.

First-Party Data: The Advertiser’s Advantage

Some data sources are still totally knowable. Anything directly controlled by the advertiser—sometimes called first-party data—can be tracked in principle. Examples of these first-party interactions include email, direct mail, landing pages (or any web pages on the company website), owned or cooperative advertising networks, and call centers—whether customer support or acquisition-focused. In these cases, previous touches and interactions with an individual can be reconstructed as an identity graph, tied together with some unique ID.

Dealing with Unknown Identities

In most cases, however, interactions with customers are “underwater”—unknown to everyone but the individual and the publisher that he interacted with. In these cases, all hope is not lost. There are two basic approaches to dealing with resolving unknown identities.

Identity Resolution Technology

The first is using an identity resolution technology provider. These data brokers provide crosswalks between different walled gardens, by either inking proprietary sharing agreements or by using algorithms to match up individuals by using various pieces of their identity to build “fingerprints.” The fidelity of matching can vary from very low to approaching 50%—which is quite good. The trick is, most of these providers do not want to give up their raw data. Instead, they prefer to keep identities in their domain, for obvious reasons. In fact, most identity resolution providers’ first business is activation, not measurement. They provide value by essentially extending retargeting—the continued advertisement to one user over time—to broader parts of the digital domain.

Even so, it’s worth following up with a few vendors to see what they are willing to provide. In some cases, APIs may be provided that return a proprietary ID (or not) to a given third-party interaction. At scale—over millions of interactions—this approach can help build a deterministic identity graph.

Probabilistic Assignment of Interactions

Another approach is probabilistic assignment of interactions. This approach—while less precise than deterministic assignment of interactions—is also not vulnerable to future changes to platform privacy policies. In this framework, each identity (user) is assigned a probability of having seen an advertisement based on time and location.

For example, say we know that in the Memphis DMA on May 1st, there were 1,045 paid social ads served (these types of data are usually available from platforms, or via the digital agency executing the marketing.) The only other piece of data we would need would be the targeting criteria for the campaign (say, 18-34 year olds), and the number of those people in the DMA. From the American Community Survey (ACS), we know that the 2022 1-year estimates for 18-34-year-olds is 367,709 in the Memphis MSA (Metropolitan Statistical Area)—not an exact match for DMA but close enough. In that case, we can say that on May 1st, an 18-34-year-old individual in that area would have a 1,045 / 367,709 (0.28%) chance of having seen that ad.

This approach can work for virtually any media channel, from targeted to broad reach. It obviously is less accurate and contains significant potential error, but in aggregate, it is a powerful tool to create all encompassing media attribution.

Building a Longitudinal Human Record (LHR)

In reality, a mixture of discrete (identity resolved) and aggregated (estimated) impressions will be used in a comprehensive MTA (see Figure 3). Whatever the combination, the aim is to end up with a longitudinal human record (LHR)—the data structure also used for CDPs (customer data platforms.) This data structure in “tall and skinny”—in other words, it will have many rows and few columns. A typical LHR used for multi-touch attribution will have hundreds of millions or billions of rows, and something like 10-20 columns.

Figure 3: A mixture of record-level (deterministic) and aggregated (probabilistic) data sources to build the longitudinal human record (LHR)

The most important columns in the LHR are the date-time of the interaction in question and the ID of the “individual”. The records should be sorted first by ID, and then by date-time. In other words, all interactions associated with an ID x will be grouped together, and then sorted by date-time descending, with the most recent interaction at the top. In the case of an order—the thing you are trying to attribute in MTA—this will generally be at the top of a lot of previous interactions.

Apportioning Value to Interactions

Once the data are constructed, a method is required to apportion value to each of the previous interactions. The simplest method is equal apportionment—and it’s not necessarily a bad place to start. Say an ID associated with an order has 40 logged interactions in the prior 90 days. In this case, each of the 40 interactions would receive 1/40th of the credit for the order.

This simple method can be made more accurate by discounting “probabilistic” interactions by their probabilities. For example, in the case of the 0.28% chance someone in Memphis saw the social ad on May 1st above, that record would be a 0.0028 in both the numerator and the denominator of the attribution equation. It follows that deterministically known records—for example, an email that was clicked—will get a 1.00, which will swamp the probabilistically estimated interaction (or impression.)

Advanced Methods of Parsing Credit

More advanced methods of parsing credit are certainly possible. There are two primary methods: logistic regression at scale and Markov Chain estimation.

Logistic Regression

Logistic regression at scale attempts to estimate each prior interaction’s relative impact on an event (usually a sale) by transforming each side of the regression equation with a logarithm. The advantage of this approach over more traditional linear regression is that because the input (independent) and output (dependent) terms are logged, any combination of independent variables is bounded between 0 and 1 (or 0% and 100%). This is also called modeling the log of the odds of an event (the transaction). It is out of the scope of this paper to get into the specifics of how logistic regression (or logit) models works, but suffice to say that the output will be a term for each potential input expressed as a log of the odds, which can be translated to probability with the equation:

Where t is the linear combination:

The trouble with using logistic regression for MTA is that it requires ones (successful sales) and zeros (unsuccessful sales—or potential customers who never became customers.) Unsuccessful sales can be hard to come by. When an entire customer journey is known, for example in the case of a health insurance carrier attempting to get members to take a discrete, known action, this is a very feasible approach. However, when prospective customers don’t have a real defined endpoint, more creative approaches to defining zeroes are needed.

Markov Chain Estimation

The second more advanced approach to parsing value in Markov Chain estimation. This flips logistic regression on its head, and focuses on the upstream tactic—say, clicking an email—and then estimates each next downstream outcome possible from that point. A good visualization of this approach is the old “The Price is Right” game “Plinko.” In this game, the player places a disc at the top of a board with a set of pegs. Each peg represents a random coin flip—so the path that the disc takes is a series of .50 probability decisions resulting in an outcome. However, in Markov Chains, each probability is not 50%–and the disc can end up in more than two next states in the next step.

Markov Chains have the fundamental property of “memorylessness”—their behavior at a given state is only dependent on where they are, not where they have been. This simplifies MTA analysis, but by implication, does not take into account the “saturation” effect of marketing. In other words, we only care that you received a piece of mail at time T, but not how many pieces you received prior to time T. The task of the analyst is to build a matrix of the probabilities of the interaction a customer will have next at a given interaction point, sometimes called a transition probability matrix. For example, this matrix might look like this for an advertiser deploying mainly digital tactics:

  Next State
  InitialEmailSocialAffiliateOTTSaleNo Sale
Current StateInitial00.400.200.300.1000
Email00.100.050.0500.010.78
Social00.020.010.020.020.020.91
Affiliate00.040.010.030.010.030.88
OTT00.040.010.010.020.020.90
Sale0000010
No Sale0000001

The above matrix must meet two conditions:

  • Pij is the probability of moving from state i (row, current state) to state j (column, future state); it must also be bounded between 0 and 1 (inclusive)
  • The rows (i) have to sum to 1; sale moving to sale is called “absorption” (yes, this is a bit confusing, you can think of it as the journey ending)

From this matrix, it’s possible to calculate the absorption probabilities of each cell—in other words, the probability that at that point, an individual will eventually turn into a sale. To do this, split thetransition matrix into two sub-matrices, Q, the matrix of transitions between non-absorbing states (touchpoints), and R, thematrix of transitions from non-absorbing states to absorbing states (conversion or non-conversion).

Matrix Q:

 InitialEmailSocialAffiliateOTT
Initial00.400.200.300.10
Email00.100.050.050
Social00.020.010.020.02
Affiliate00.040.010.030.01
OTT00.040.010.010.02

Matrix R:

 SaleNo Sale
Initial00
Email0.010.78
Social0.020.91
Affiliate0.030.88
OTT0.020.90

With these two matrices, absorption probabilities are found by calculating the fundamental matrix, F = 1 – Q)-1 where I is the identity matrix and (1 – Q)-1 is the inverse of the matrix I-Q. Finally, multiplying F by R will produce the ultimate probability (absorption) of a sale happening (or not) at any point in the matrix. In the case of the given data, F for “sale” is:

  Next State
  EmailSocialAffiliateOTT
Initial StateEmail0.0111510.0005690.0005870.000176
Social0.0004880.0202350.0004470.000418
Affiliate0.0014010.0003870.0310110.000324
OTT0.0009250.0002560.0002630.020416

These estimates can then be used to allocate value between touches in a chain. For example, say a user is observed going from email to email to social to OTT, and then finally closing. In this case, the values pulled from the above table would be 0.011151; 0.000569; and 0.000418, with the final step—OTT to absorption—pulled from our first matrix as 0.02. To apportion value, these numbers are summed and divided by the total number of steps in the pathway:

(0.011151 + 0.000569 + 0.000418 + 0.02)  =  0.03214

To get to a final contribution vector for each unique journey:

Touch 1 (Email)Touch 2 (Email)Touch 3 (Social)Touch 4 (OTT)
35%2%1%62%

Of course, this has to be scaled over thousands and millions of orders—and it’s a lot more complicated than the simple example shown above.

The final question is time. In the example above, say the first interaction (email) happened 90 days ago, but the last interaction before close (Social) was two days ago. Is it fair to give the Email touch 35% credit and the Social touch 1%? Almost certainly not.

There are many techniques for decaying a touch’s impact on ultimate conversion, but they all have the same basic concept: an interaction’s impact fades over time. In MMM (media mix models) we use the term adstock—the rate at which a touch’s “power” decays after it is shown. The concept is similar in MTAs, but is reversed, because we are focused on the sales, not the stimulus. In other words, we look back from the order, not ahead from the stimulus.

The simplest approach for MTA is straight-line decay with a look-back window. This is just as it sounds: A touch before the start of the look-back window gets 0 credit; one mid-way through the window gets ½ of the full credit; and a touch on the day of the order gets full (1) credit. More advanced logistic decay approaches are certainly possible, but yield limited additional benefits and add significant complexity.

Bringing It All Together: How Does Multi Touch Attribution Work?

Putting all of these elements together—the longitudinal human record (LHR), probabilistic inference of non-deterministic touches, estimation of contribution via either logistic regression or Markov Chain analysis, and decay—will give a marketing analytics team the basic elements of multi-touch attribution.

One final, important note: black box MTAs provided by vendors claiming a secret sauce for identity resolution are compelling because they seem simple, but buyer beware. Proprietary identity graphs are only as good as the federations the company belongs to and the underlying cross-walk code, and there is really no way to validate them other than by seeing if they “make sense.” For this reason, they tend to struggle over the long run as accuracy—and hence utility—are questioned by financial decisionmakers. A purpose-built, open-code, white box approach to MTA, which can also be used to power downstream applications like dashboards and econometric MMM panels—should be the preferred approach for marketing analytics teams.

Download our whitepaper, “Measuring Marketing’s Effectiveness”​

Access our whitepaper for a deep dive into additional imperatives and methods for CMOs and analytics teams driving measurable marketing ROI.

MMM vs MTA 

In the ever-evolving landscape of marketing attribution, businesses are engaged in an arm race to better measure and maximize ROI and growth. Beyond simple last-touch attribution, the two prominent methodologies often compared are Media Mix Modeling (MMM) and Multi-Touch Attribution (MTA). This article delves into the nuances of MMM vs MTA, exploring their differences, advantages, and how they can be effectively utilized to drive marketing success. The good news is that this isn’t necessarily an either / or decision; MTA can be used upstream and in concert with MMM to combine quick-read and long-run views of marketing’s effectiveness. 

Before diving into MMM vs MTA, it’s important to note that the bread and butter of marketing reporting—last touch attribution—is still a powerful and simple tool, and should not be cast aside in favor of more advanced techniques. Last touch attribution credits the final interaction before conversion, generally using a direct database linkage via some kind of unique ID—a telephone number, URL string, or webpage tag. Last touch attribution will always be simpler and faster than multi-channel techniques. MMM and MTA outputs should always be looked at side-by-side with last touch reporting. It is often these comparisons that yield the most interesting insights. 

Understanding Media Mix Modeling (MMM) 

Media Mix Modeling (MMM) is a statistical technique used to estimate the impact of various marketing channels on sales performance. It aggregates historical data in a time series, usually across a geographic or other cross-sectional key, to measure the effectiveness of marketing efforts and to allocate budgets more efficiently. MMM can interpret any kind of stimulus, whether paid or earned, upper- or lower-funnel, or offline vs. online. It can also be used to understand the impacts of non-promotional factors—including price changes, competitive actions, product launches, and distribution channel strategy. 

Figure 1: The basic idea behind MMM; there is an efficient frontier for marketing achieved by optimally mixing channels. This mix is different at different spend levels, but generally the macro curve exhibits diminishing returns to scale (its slope decreases as a company spends more). 

Key Benefits of MMM 

  • Comprehensive View: MMMs provide a broad and complete overview of how different marketing channels interact and contribute to overall sales. This comprehensiveness is beneficial to understanding the combined effects of multiple marketing efforts and avoids over-crediting.  
  • Long-Term Abilities: One of the strengths of MMM is its ability to account for the longer-term impacts of marketing—whether over weeks and months, or years, in the case of marketing’s impact on brand equity. This is particularly helpful when trying to gain an honest accounting of the effectiveness of upper-funnel channels like TV and print, where effects are usually not immediate. This long-term focus also makes MMM less capable of reading “what just happened”—although techniques like rolling analysis windows can help with trending. 
  • Requires “Smaller” Data: MMM data frames are only a few megabytes, usually a few thousand rows and a few hundred columns. This generally makes it possible to store the data on a traditional device—a hard drive or simple cloud-based storage. MMMs can utilize either spend or impressions as its “x” or independent variables, and can use either sales or revenue as an ultimate dependent variable. Even so, MMMs need extensive historical data, usually a minimum of two years, making them more suitable for established brands with significant data accumulation. Brands without much historical data can still be boot-strapped with Bayesian priors. 

Understanding Multi-Touch Attribution (MTA) 

Multi-Touch Attribution (MTA) is a more granular approach that focuses on assignable channels—usually digital, but also including direct mail and interactions with known customers. It attributes conversions to the multiple touchpoints that a consumer interacts with throughout their journey. This method provides insights into the effectiveness of each touchpoint in the conversion path.  

MTA was very popular in the early days of digital marketing, before privacy concerns and platform data hoarding made it harder to resolve identities across channels. In recent years, it has come under fire as first generation deterministic ID resolution approaches failed, but advances in data clean rooms and probabilistic exposure inference are making “second generation” MTA models a very attractive option for inference. 

Key Benefits of MTA 

  • Close to Real-Time Insights: MTA models are capable of using real-time data, allowing marketers to make quick adjustments to their strategies. This is particularly advantageous in fast-paced digital environments. The devil is in the details, however—real-time results are only possible with very robust data pipelines and fast compute environments. 
  • Potentially Unlimited Granularity: Because MTA models are built at the log level, there is potential unlimited detail available about each individual touchpoint, helping marketers understand the specific role each interaction plays in driving conversions. Keep in mind that this detail is dependent upon robust lookup tables and cross-walks, as well as thought through marketing taxonomies. 
  • Consumer Journey Mapping: A side benefit of building the human longitudinal record required for MTA analysis is a 360-degree view of the journey. Exploratory data analysis (EDA) of this data artifact using big data tools can identify influential touchpoints, find breakdowns in e-commerce pipelines, and discover high-value audience segments. Even so, MTA data frames store much more data, at the record level. Log files, sometimes billions or tens of billions of rows, must be processed, demanding a big data compute environment. 

Combining MMM vs MTA for a Holistic Approach

MMM vs MTA is how these methodologies are often perceived, but they can and should be used together. Integrating the macro-level insights from MMM with the micro-level details from MTA can provide a comprehensive understanding of marketing effectiveness. This integrated approach allows businesses to leverage the strengths of both models. 

The record-level data required for MTA analysis can be used upstream of the MMM econometric panel structure, directly feeding it. In this way, the same “single source of truth” can be used for both analyses. Data that is not used in MTA—for example, survey data—can be joined after the raw data has been grouped and aggregated. Our recent white paper on the Go-to-Market Data Lake architecture details this approach. There are three main steps: 

  1. Creation of Longitudinal Human Record (LHR): Tying customers’ journeys together in longitudinal chains can help locate points of friction, profiles audiences, and conduct multi-touch attribution (MTA). 
  1. Creation of Econometric Panel: The LHR then serves as the base query to create an econometric panel for MMM. This panel is a summation of stimulus (x-variables) and response (y-variables) by day or week, across one or more cross-sectional dimensions. 
  1. Data Aggregation and Supplementation: The panel is then supplemented with aggregated data, such as linear television or unresolved digital marketing data, to fill in gaps and ensure a complete dataset. 

Figure 2: Start with record-level data to build the LHR, and feed the ultimate econometric panel to enable MMM. 

Use Cases for MMM and MTA 

It is best to think about the usage of MMM and MTA in the context of planning cycles. It is helpful to think of three marketing planning cycle types: strategic; tactical; and reactive. 

Strategic Planning 

Strategic planning typically happens annually, with more ambitious strategy resets looking out three or even five years. This type of planning typically looks at the total marketing investment envelope (e.g., $50M or $75M per year); the rough mix by funnel position; and any new channels or types of marketing to be tested at scale. MMM—particularly more advanced modeling taking advertising’s impact on brand equity—is the right tool for this exercise. 

MMMs can be extended from measurement to optimization by extrapolating the curves outputted from statistical inference into “what if” scenarios, and then using machine learning to re-mix marketing until an optimal solution is reached. This optimization step can be a helpful input into the strategic planning process. It is important to note, however, that optimizations based on past results are not accurate when predicting huge budget swings. A good rule of thumb is that beyond a 20% increase or decrease in budget, curves become unreliable. 

Tactical Planning 

Tactical planning typically happens quarterly or annually, and looks at the specific channel and audience mix that will drive maximum ROI—sometimes called ROAS in marketing circles (return on advertising spend). Both MMM and MTA can be useful in the tactical planning phase. MMM is good to understand marginal customer acquisition cost (CAC), allowing campaign planners to re-mix channels to maximize effectiveness given a certain budget. MTA can then be mixed in to identify recent changes in channel effectiveness, and to get to granular detail on specific creative types, landing pages, offers, and cadence.  

Reactive Adjustment 

Reactive planning (or adjustment) happens constantly. Marketing dashboards typically start with last-touch results. They become far more powerful when positioned side-by-side with MTA results. Channel managers who are used to seeing last touch CPAs will now see a mutually exclusive, collectively exhaustive view of CPAs, as well, that credits channels with “halo” if they drive more influence up the funnel. MTA is ideally suited for reactive adjustment because it can be built to update in near real time.  

Showing last-touch next to fully attributed credit can help marketers understand true contribution.

Figure 3: A last touch vs. MTA version of channel contribution. The “MTA effect” is the impact of multi-touch attribution on last touch CPAs. Some channels do better, and others look more expensive. 

Final Thoughts on the MMM vs MTA Debate

In the debate of MMM vs MTA, the answer really is “both”. Both methodologies offer unique benefits and address different aspects of marketing effectiveness. The good news is that they can be built together, using one data pipeline. By understanding their strengths and limitations, marketing leaders can leverage MMM for strategic planning and MTA for tactical optimization and reactive adjustments. Combining both approaches—without throwing away last touch attribution—provides a holistic view of marketing performance, ensuring that every marketing dollar is spent effectively.  

Download our whitepaper, “Measuring Marketing’s Effectiveness”​

Access our whitepaper for a deep dive into additional imperatives and methods for CMOs and analytics teams driving measurable marketing ROI.

MaxDiff explained: Unlocking meaningful insights

Have you ever encountered survey results showing all the respondents think everything is important or, worse, nothing is important? This survey output often arises in survey questions where respondents are asked to rate items on a scale, resulting in a chart like this:

Simply asking people what's important results in charts like these which contain no real useful information

Here, every option seems equally important, making it challenging to pinpoint what truly matters to respondents.

Now, imagine you are responsible for selecting the next best feature for your product, a decision that will trigger a multi-million-dollar investment to bring this feature to life. Or you are a CMO trying to decide what features will resonate with consumers as you break into a new market. Would you feel comfortable making a decision based on the above chart?

Probably not.

Unfortunately, this “everything is important, so nothing is” syndrome is a common quantitative research issue, especially with tools like Net Promoter Score (NPS) or Likert Scales, which struggle to measure incremental changes and extract meaningful differentiation between options.

Enter MaxDiff, a methodology designed to solve these exact problems.

What is Maximum Difference Scaling?

Maximum Difference Scaling, also known as MaxDiff, is an advanced choice-based tradeoff methodology used to help strategic leaders uncover better decisions about their GTM approach, products, and more. Instead of asking respondents to rate the importance of items independently, MaxDiff presents them with a set of options and forces them to make tradeoffs, indicating their most and least preferred items. The output offers crisp insights into what truly matters to respondents.

Example of what a respondent would see in a MaxDiff battery

Why Use MaxDiff?

MaxDiff can be advantageous over other research techniques for a few reasons:

  • Forces Choices and Tradeoffs: By requiring respondents to choose between options, MaxDiff helps highlight preferences more sharply.
  • Generates Clearer Insights: This methodology reveals gaps in importance and preferences, making it easier to identify key priorities.
  • Avoids Yeah-Sayers: Respondents can’t simply agree that everything is important or not important; they must make definitive choices.
  • Creates Intuitive Experience for Respondents: The format is straightforward, making it easier for participants to engage meaningfully with your survey.
  • Enables More Efficient and Effective Survey Design: MaxDiff efficiently captures nuanced data without overwhelming respondents with lengthy rating scales.

When to Use MaxDiff

MaxDiff is a versatile tool that can be applied to various use cases to help marketers make more informed decisions:

  • Prioritization of Features: Determine which features are most important to your customers to drive product evolutions.
  • Needs-Based Customer Segmentation: Understand and identify different customer segments based on their preferences (see below example).
  • Message and Branding Testing: Find out which messages and branding efforts resonate most with your audience.
  • Product Feature Testing: Identify the most and least desirable product features.
  • TURF Analysis: Extend MaxDiff’s output of your product’s most important features or criteria to inform the top combination of features for your offering via a TURF (Total Unduplicated Reach and Frequency) analysis (see below example).

How to Use MaxDiff

To get the most out of MaxDiff, here’s the structured approach we follow:

  1. Qualitative Discovery: Begin with qualitative research to understand the broad range of needs, criteria, and options. This step ensures that your list of options is comprehensive. Note: MaxDiff won’t be able to tell you if your list of options is “good,” “bad,” or all-encompassing, so it’s crucial to identify the options in qualitative discovery to inform your next step.
  2. Quantitative Validation: Once you have a robust list, use the MaxDiff tool in your survey platform to validate (or disprove) each option’s importance and satisfaction levels.
  3. Outcome Analysis: The result is a prioritized list of needs, highlighting the most important and least satisfied criteria to address.

Example output using MaxDiff offers more differentiation between options.

MaxDiff is a powerful tool that brings clarity to the often opaque world of market research. By forcing respondents to make tradeoffs, it provides actionable insights that can help to inform product evolutions, better understand customer needs, or optimize messaging and branding. If you’re struggling with the “everything is important, so nothing is” syndrome in your surveys, it’s time to consider MaxDiff for more precise, more meaningful results.

7 steps to optimize channel partner enablement

For most B2B go-to-market (GTM) leaders, 2023 accelerated a trend that had, until recently, only been lurking in the background—the diminishing effectiveness of direct sales and marketing teams. The traditional reliance on tele- and email-based account management, free trials, and automated motions have the unintended consequence of decreasing the openness or receptivity of the buyers they target. Today, buyers are less receptive than ever to answering a call, much less meeting with a person or organization with whom they do not have a pre-existing relationship.

To counter this dynamic, GTM leaders have predictably turned to their channel partners and the teams that support them to provide essential pathways to expand market reach and drive the top-line growth required to achieve targets.

A Comprehensive Guide to Channel Partner Enablement

Effective channel partner enablement is crucial. It enhances relationships and drives strategic outcomes, such as expanding account relationships and increasing the leverage of commercial organizations.

In this guide, we explore the seven steps of channel partner enablement that ensure these relationships are as productive and profitable as possible. From customizing partner programs to leveraging advanced data analytics, these imperatives can transform your partner strategy into a powerful component of success.

  1. Tailor Partner Programs
  2. Craft and Quantify Compelling ‘To-Partner’ Value Propositions
  3. Personalize Content Around Partner-Oriented Use Cases
  4. Leverage Data as a Collaborative Asset and Differentiator
  5. Optimize Channel Partner Coverage
  6. Analyze Partner Performance, Engagement, and Activity
  7. Develop Ongoing Communication and Feedback Loops

1. Tailor Partner Programs

A well-structured partner program is essential for building new and nurturing existing relationships. One sure sign of an effective program is its ability to be tailored to meet each partner’s diverse needs and expectations. While this was once a fairly straightforward exercise, the proliferation of partner roles and capabilities has blurred the lines between historical partner classifications such as ISV, VAR, GSI, MSSP, etc. In the new age of partnerships, agility, supported by clearly defined and simple program structures, best drives alignment and motivates partners to execute their roles effectively.

Best Practices:

  • Understand Partner Profiles: Segment your partners based on their market position, size, and typical customer base. This segmentation will allow you to tailor programs that align with the most prominent business models and partner capabilities.
  • Design Tiered Incentives: Implement a tiered program structure that provides partners with the requisite financial rewards and non-financial benefits for meeting different performance thresholds. This promotes healthy competition and ensures partners of all sizes can benefit and grow.
  • Perform a Program Health Check: Conduct a comprehensive evaluation of your partner program every two to three years to ensure it’s working as designed and effectively driving the desired partner behaviors and outcomes. This health check helps identify areas of improvement and ensures the program remains aligned with market conditions and strategic objectives.

2. Craft and Quantify Compelling ‘To-Partner’ Value Propositions

Creating and clearly defining compelling value propositions for partners that resonate at the organizational and individual levels is crucial for the success of any channel partnership. This means articulating both the higher-level strategic benefits that partners gain from the program (i.e., increased revenue-enhanced product offerings, etc.) and breaking down what the partnership might mean for the individual contributors (i.e., sales incentives, exclusive training, etc.), all while providing the quantifiable data to back up the claims and benefits. This approach helps partners see the tangible gains from the partnership, driving them to engage more fully and align their efforts with the shared goals of the collaboration.

Best Practices:

  • Build ‘Partner-Centric’ Value Propositions: Put yourself in the shoes of the partners you serve to refine plans for helping them achieve their business goals. Stakeholders often nuance these goals, and there is seldom a ‘one size fits all’ solution. Partner research surveys are often helpful in developing ‘Partner-Centric’ messaging frameworks.
  • Think Organizationally and Individually: In addition to outlining organizational benefits, emphasize how the individuals within the partner company can achieve success from the program. This might include access to exclusive training, tools that make their jobs easier, or performance-based financial rewards.
  • Celebrate and Highlight Success: Use case studies and testimonials from existing partners to demonstrate the real-world benefits and ROI of the partnership. Seeing peers’ success is often a powerful motivator for existing partners and has the added benefit of attracting new partners.

3. Personalize Content Around Partner-Oriented Use Cases

Personalization seems simple but can be challenging to execute. When referring to partner enablement, personalization should provide the right type of information and access to the right resources tailored to the partner’s specific needs. When implementing a personalization strategy into your program, focus on partner-oriented use cases that deliver highly relevant assets that directly support the partners’ sales and marketing motions and strategies.

Best Practices:

  • Start with Identification: Kick off the personalization process by conducting thorough analyses to identify the most relevant and widespread use cases for your partners. This often overlooked but critical step is fundamental for prioritizing where to invest precious time and resources for driving personalization at scale. Regularly revisit and refine these use cases based on evolving product, buyer, and partner dynamics to ensure they align with actual partner needs.
  • Support Content Utilization: Creating content that aligns with the most prevalent use cases and supports partner needs based on scenarios (i.e., sales scripts, marketing assets, training materials, battle cards, etc.) is only half the battle. Once built, systems must be in place that cater to each use case. This might involve specialized training sessions, dedicated support personnel for complex issues, or even white-glove marketing support for certain high-priority partners and use cases.

4. Leverage Data as a Collaborative Asset and Differentiator

Over the last decade, the prevalence of data, technology, and supporting analytics has exploded; it seems as if almost everything related to the marketing and sales process is counted, stored, and analyzed. Channel partnerships, therefore, should be more valuable than ever, with mutual clarity into account bases, pipelines, and near-instantaneous sharing of insights—yet most partnerships do not work this way. A lack of mutual trust underscored by concerns over data security and the perceived risk of losing proprietary insights and information often creates strategic deadlock between partner and provider, where each party shares only just enough information as necessary to receive the requisite compensation—in other words, transactional partnerships.

Integrating mutual data into the partner enablement program is crucial for transforming partnerships from “reactionary” or “transactional” to “proactive.” In doing so, organizations can improve their operational efficiency while enhancing the partnership’s overall value and encouraging long-term collaboration and shared success.

Best Practices:

  • Establish Clear Agreements: Draft formal agreements that clearly outline what data will be shared, how it will be used, and what safeguards will be in place to protect it. This should include defining the purpose and acceptable use of data, the methods and frequency of data exchange, and the duration for which data will be shared.
  • Use a Phased or Tiered Approach: Begin with sharing non-sensitive, less critical information (e.g., aggregate/summary data, market trends, or product performance metrics) to build confidence. Only once a foundation of trust has been established should you move towards sharing more sensitive customer data. This should be done slowly and deliberately—often starting with a single, trusted partner (and may be further bisected by product/region)—before a programmatic roll-out.
  • Leverage an Objective Third Party: In most cases, involving a neutral third party can help ease concerns by managing the upkeep, matching, anonymization, and exchange. This could be a trusted consultant or a data-sharing platform that ensures compliance with all agreed terms and maintains the confidentiality and integrity of the data.
  • Enhance Collaborative Technology: Use technology that enhances collaboration between you and your partners. Shared workspaces, real-time communication platforms, and integrated supply chain systems can help streamline operations and improve transparency.

5. Optimize Channel Partner Coverage

Optimizing channel partner coverage ensures that partnerships across the ecosystem are effectively engaged and supported throughout the partner lifecycle and customer sales processes. Though concepts of coverage, sizing and deployment, and territory assignments or compensation alignment are critical levers of success for leaders of direct selling units within an organization, they tend to be less prevalent for channel organizations (this is especially true when related to enablement). This dynamic is mainly due to organizations being more accustomed to the immediacy and attribution of direct results when managing their own sales teams, where outcomes are easily measured, transparent, and more responsive to change than their indirect counterparts. On the other hand, channel partners inherently add incremental layers of complexity, making it harder to prioritize and optimize partner coverage effectively. As a result, many channel organizations fail to incorporate this powerful lever in their overall enablement strategy.

Ensuring that partner account managers and commercial teams are executing consistent processes is crucial for the scalability and effectiveness of partner enablement programs. Regular training and consistent processes are essential, especially in maintaining alignment across various teams.

Coverage optimization starts with confirming the right types of job roles and commercial functions are in place (e.g., partner account managers, partner marketing managers, channel managers, etc.) and subsequently equipping those personnel to execute consistent processes and maintain a regular cadence of enablement activities. By refining coverage strategies and securing buy-in from sales leadership, organizations can often enhance the overall effectiveness of their channel partnerships, leading to increased sales and a stronger market presence.

Best Practices:

  • Sales Leadership Buy-In: Secure buy-in from sales leadership to ensure enablement efforts are supported with the necessary resources and attention.
  • Utilize Partner Segmentation to Inform Coverage: Segment partners based on key criteria such as tier, relationship strength, capability, market potential, strategic value, etc. Use this segmentation to guide the role requirements and ensure resources are effectively deployed.
  • Consistent Globally Extensible Execution: Develop operating procedures, milestones, and engagement standards for partner managers that drive valuable partner interactions. This consistency helps build trust and strategic alignment and has the added benefit of keeping management up-to-date and well-informed.
  • Flexible Engagement Models: Recognize that different partners may prefer different levels of engagement. To accommodate these preferences, offer various interaction models, from self-service portals to personal account management.

6. Analyze Partner Performance, Engagement, and Activity

Regardless of the type of partnership (e.g., tech, channel, strategic, etc.), a clear understanding of partner performance, engagement, and activity is essential to understanding productivity partnerships and the effectiveness of enablement efforts. As the partner technology stack has continued to expand, partner organizations increasingly leverage advanced technology and tools to gain better insight into how their partners behave.

These tools include sophisticated CRM/PRM systems, comprehensive Business Intelligence (BI) applications, specialized engagement tracking software, and cutting-edge predictive analytics platforms that curate detailed insights into how partners interact with online portals and individual pieces of content, and predict future performance trends that allow for improved enablement strategies and tactics. By integrating these powerful analytics tools into their operational framework, partner programs move beyond simple observation, enhancing partner relations and improving future interactions.

Best Practices:

  • Tools Must Support Strategy: Partners expect modern, easy-to-use, and easy-to-navigate tools —these are now largely table stakes in partner enablement (typically inclusive of Portals, PRM/CRM systems, data analysis tools, and marketing automation platforms). However, these tools should be thought of as additive to the overall enablement strategy rather than the answer to all challenges.
  • Assign Ownership to Measuring Engagement: Using analytics tools to monitor how partners interact with provided content and resources often sounds too good to be true, and many organizations struggle to find tangible value despite a plethora of data. Establishing clear ownership can help alleviate this concern and ensure that the data collected does not go unattended.
  • Customize Support Based on Data: Utilize the insights gathered from analytics to provide customized support to partners, helping them overcome specific challenges and capitalize on opportunities. Consider engagement intelligence when developing coverage or aligning incremental resources with partners.

7. Develop Ongoing Communication and Feedback Loops

Open lines of communication and robust feedback mechanisms are vital for building trust and maintaining long-term relationships with partners. These systems help organizations adjust their strategies and operations in response to direct input from their partners.

Best Practices:

  • Voice of Partner Surveys: Implement well-structured “Voice of Partner” surveys that regularly collect comprehensive feedback from partners on various aspects of the program, products, services, and overall experience.
  • Only Track What You Intend to Act On: It is crucial to collect feedback and act on it. Show partners that their input is valued by making visible changes based on their suggestions and concerns.

Embrace a Holistic Approach to Channel Partner Enablement

In our experience, partner organizations often become hyper-focused on one or two dimensions of enablement while overlooking how these elements work together holistically. In such instances, it’s crucial to remember that effective enablement transcends individual initiatives; it’s about developing a well-defined, cohesive ecosystem of collaboration and mutual understanding. For organizations aiming to fully leverage their partner networks, embracing a comprehensive approach to enablement is a strategic necessity and a transformative opportunity. This journey towards refined enablement empowers partners, enhances collaborative efforts, and drives substantial business growth and success.

Download our framework, “Designing a Best of Breed Partner Program”​

​A well-designed partner program can set the stage for success. Download our framework to learn about the eight components of a best practice partner program and three quick-start areas to accelerate channel revenue growth.

How to measure marketing effectiveness for acquisition and member campaigns

Marketing campaigns typically target one of two main audience pools: 1) Prospects, i.e., those a business aims to acquire; or 2) Members, i.e., those a business seeks to retain or engage. To measure marketing effectiveness, each group presents unique challenges.

In the case of acquisition marketing, information about prospects is often limited. Marketers must allocate spending across various channels to cast a broader net to acquire new customers. They can target addressable markets that meet their specific criteria (i.e., individuals aged 65+ for insurance companies that provide Medicare plans or those that live within specified locations for retail organizations). However, complete information about the prospect pool is often unknowable.

By contrast, member marketing benefits from having extensive data on current members. Much of the person-level data utilized is generally stored within an organization’s first-party data. This allows marketers to utilize personalized channels (such as digital outreach and content-specific mail) for more precise targeting and measurement.

In this blog, we will examine and compare these two types of marketing. For each type, we’ll touch on the broader objectives and measurement techniques, the data required for measurement, the typical challenges marketers face when attempting to gain insights from data, and the ways that analysts can offset these challenges to measure marketing effectiveness.

Why It’s Crucial to Measure Marketing Effectiveness

The measure of marketing effectiveness establishes the incrementality of our marketing, providing clarity into what we gained from a campaign, i.e., would not have happened otherwise. This is essential for discussing budget, ROI, and the net benefits marketing provides to the business. Effective measurement leads to educated, data-driven, strategic marketing decisions that account for unknowns and clarify the ideal focus areas.

Whether the marketing is directed at prospects (acquisition) or members, measurement allows marketers to answer big questions such as:

  • “What is the value that marketing is providing?”
  • “How can we increase our outcomes given budget constraints?”
  • “What marketing tactics are the most effective at driving positive outcomes?”

These are complex questions that drive strategic decisions in organizations. The identification of insights that can provide better answers to these questions and make more informed decisions is invaluable.

Acquisition Marketing Campaign Measurement

Acquisition marketing is complicated. Marketers have a wide range of channels to use to engage new prospects. These include traditional channels, such as Direct Mail and TV (both branded and direct response); the wide array of digital channels, such as Display, Paid Search, Organic Search, Online Video, and Social Media, amongst many others; and more industry-specific channels, such as Events, Affiliates, or Out-of-Home (OOH, e.g. billboards or transit ads).

Measurement Techniques

With so many channel choices, finding the right mix of budget resources can be challenging. Marketing Mix Modeling (MMM) is an excellent method for answering mix-related questions. An MMM is an econometric model that helps measure the effectiveness of marketing activities and their impact on sales or other KPIs. Within MMM model development, audience granularity is not required. A set of aggregated data at various segment levels (e.g., time periods, geographies, product categories, and other types of well-defined metadata) can be used to build a valuable model.

While MMM helps provide a broad understanding of how to allocate spend across channels, it is limited in its ability to provide insights at the more granular customer level. Identifying how to attribute “credit” for each sale (or other output KPI) across marketing channels requires some attribution models.

The default, simpler methodology is last-touch attribution (LTA). This methodology attributes 100% of the sale to the last channel a customer engaged with before converting. This often leads to over-attribution of bottom-of-funnel channels in reporting, causing an overinvestment in these down-funnel channels that is more straightforward to measure; and atrophy within less measurable, upper-funnel channels that play an essential role in the acquisition process through their impact on brand health and audience attitudes.

The far more challenging approach is multi-touch attribution (MTA). MTA aims to solve the LTA problem by providing a more comprehensive view of the customer journey. The model apportions the credit of an application across the various channels that prospects engaged with throughout the buying process. A good MTA model can help marketers better understand the value of each touchpoint. With the complicated ecosystem of marketing and sales efforts, better information can significantly improve strategies.

Data Requirements

One key challenge in measuring acquisition marketing is finding signal within the sheer volume of data available. With a huge market and many potential customers, marketers have access to a wealth of information that can inform their strategies. However, the quality and granularity of the data across channels can vary significantly.

For example, in the case of Direct Mail as an acquisition channel, marketers have complete visibility into the “1s” and “0s” – that is, they know exactly who received a piece of mail and who did not. This level of granularity allows for more precise measurement and analysis.

Alternatively, poor ID resolution for upper- to mid-funnel channels, such as Digital channels and TV, is a crucial hurdle to overcome when building an acquisition MTA. Unlike Direct Mail, which allows for precise tracking, these channels often rely on cookies, device IDs, or other imperfect methods of identification. As a result, these channels usually only yield aggregated impression and response counts. As such, it can be difficult to connect individual users across different touchpoints on these platforms if they don’t use them to convert.

Finally, multiple application capture channels generate different levels of data, which can be challenging to reconcile with each other and the funnel activity. For example, a lead captured through a website form may provide different information than one captured through a call center. Integrating these disparate data sources into a cohesive MTA model can be complex and time-consuming.

Strategies for Overcoming Measurement Obstacles

Marketers can implement several strategies to overcome the challenges associated with poor data granularity within marketing channels. One strategy is an aggregated MTA approach, as opposed to the more traditional discrete MTA detailed above.

An aggregated MTA measures the effect of marketing investments on the last-touch attributed marketing channels or known segments over a given time period. This methodology estimates channel- or segment-specific dependent variables using multiple stimulus variables. In many ways, an aggregated MTA is more similar to an MMM in technical modeling methodology than a discrete MTA. In this instance, it measures the multi-channel impacts on siloed/coded response measures.

Model propensities offer another approach to overcoming the measurement challenges of channels without person-level impression data. This is applicable for channels, such as DRTV or Digital channels, where the individuals who convert (the “1s”) are known, but those who do not convert (the “0s”) are not. The aggregate number of impressions for these channels is known for a specific geography and time period. A model can be built to impute impressions within the broader audience pool. Individuals with higher propensities can then be treated as a “high likelihood” impression within a traditional MTA model.

Member Marketing Campaign Measurement

Member marketing can have many objectives, such as ensuring existing customers remain satisfied with their experience or engaging the individual to take certain actions to add value to the organization (e.g., extending their subscription, buying an add-on, etc.). While having a “known” population makes member marketing campaigns easier from a targeting perspective, these campaigns can face challenges in determining the prioritization of the testing agenda.

Measurement Techniques

A Randomized Control Trial (RCT) is the classic approach for testing incrementality. In this methodology, the population is randomly split into “test” and “control” groups, where the former receives the novel treatment to be evaluated, and the latter gets a comparative baseline. This baseline typically does not include marketing and is considered a universal holdout for establishing marketing’s incrementality in general. This approach allows the team to determine what conversions would have been lost if the campaign had not been run.

While RCT is the most straightforward approach conceptually, there are some opportunity costs to consider. In member marketing campaigns, the impact of traditional RCT universal holdouts tends to be “felt” more by the business because the members being withheld from marketing are known. Additionally, in its purest form, an RCT should ask a simple, one-objective question per test, which, depending on the outcome’s response timeline, can frustrate marketers who want to test and learn on big questions at a faster cadence.

Given the desire to preserve the whole addressable population in most member campaigns, MTA is a measurement route that can provide reads on incrementality while allowing the total population to be messaged. As discussed in our acquisition example, member-level data enables us to apply this modeling concept at a per member, per journey level, or a discrete MTA. The outcome is an understanding of incrementality per conversion while enabling aggregations to multiple different audience cuts given the discrete data origins.

In the member marketing world, a discrete MTA is an attractive foil to RCT, given that it mitigates some of the obstacles noted above. When using this approach, a universal holdout is not necessary to achieve incrementality. However, embedding RCT outcomes into the model itself and comparing the model output against RCT tests is a way to make your discrete MTA more mature.

Since discrete MTA models are built at an individual level and member populations can quickly get into the tens of millions, an advanced data science platform with scalable compute is necessary to enable this type of work (e.g., Databricks or Snowflake). Additionally, temporal constraints on the underlying data feeding these models can sometimes create a barrier to scoring and results. For example, if our campaign’s objective is first to have a member schedule an appointment with their doctor and, after that, schedule another appointment to get a specific test, such as a colonoscopy, the lifecycle of that final response is several months. Additional data complications and refresh cadences can delay this process even further. Given that MTAs look backward on conversions to see what drove the action, the campaigns receiving credit in this scenario would not get proper attribution for several months.

Data Requirements

In a member marketing scenario, we have distinct, discrete data on each person we are marketing to. There is also a clear line of sight into each individual’s outcome, i.e., whether they did or did not act. This means we can cleanly get to the “1s” and “0s,” reflecting the true outcome for a member. This foundation serves RCT and discrete MTA methods, allowing for tailored sampling and individual-level modeling.

Additionally, the marketing typically measured in member campaigns is not usually awareness-based outreach (e.g., TV ads). Instead, it consists mainly of direct response marketing. Channels such as email, outbound calling, SMS, and direct mail can all directly route members to the intended outcome. This facilitates member-level tracking of each marketing piece an individual received and intermediary metrics (opens, clicks) for the applicable channels.

Strategies for Overcoming Measurement Obstacles

Given the tradeoffs between the two outlined approaches, the marketing team should align with stakeholders on a clear learning agenda. If faster results are the priority, RCT would be the best solution despite its tradeoffs. To overcome hesitancy in the control population in the case of a universal holdout, the analytics team can provide various options on how sensitive (what level of confidence) and specific (how small of a difference) the test read will be—the more sensitive and specific, the larger the holdout. Often, a middle ground can be found to produce definitive reads at a reasonable population size.

If there is no desire to withhold individuals from marketing and the conversion metric being tracked has a reasonable lifecycle, a discrete MTA would be the favored method. Once it has been built and given consistent streaming inputs and a relatively swift conversion metric, discrete MTAs can migrate into real-time reporting of conversions reported to date.

Techniques & Strategies for Effective Campaign Measurement

It’s crucial to measure the marketing effectiveness of acquisition and member marketing campaigns to make informed, data-driven decisions that optimize budget allocation and drive positive outcomes. While each marketing method presents challenges, various measurement techniques and strategies can be employed to overcome these obstacles.

For acquisition marketing, a combination of MMM and MTA can provide a comprehensive understanding of how to allocate spend across channels and attribute credit for conversions. In the case of member marketing, RCTs and discrete MTAs can help establish the incrementality of campaigns and identify the most effective tactics for engaging existing customers. By leveraging the appropriate measurement techniques, utilizing the available data, and aligning on clear goals, marketers can make strategic decisions that drive business growth and demonstrate the value of their efforts.

Download our whitepaper, “Measuring Marketing’s Effectiveness”​

Access our whitepaper for a deep dive into additional imperatives and methods for CMOs and analytics teams driving measurable marketing ROI.

5 essential jobs in a go-to-market analytics team

In the fast-paced B2B marketing and sales industry, analytics have become indispensable. Successful go-to-market strategies require a thorough understanding of the five essential jobs that analytics teams perform to unlock success. This blog delves into each job, providing insights and best practices for optimal results.

What are the Five Essential Jobs?

1) The What: Reporting

Definition: The What job is about reporting the facts. How many leads are we generating, how quickly are revenue analytics leads moving through the funnel, how much are we spending on marketing, and how is each rep doing versus their goals? This is the most basic, and also the most important, “table stakes” job. Fortunately, it can be mostly automated via business intelligence tools.

Reporting is the lifeblood of sales and marketing executives, ideally offering near-real-time performance insights. Different types of reports provide insights into various parts of the pipeline.

Types of Reports:

  • Activity-Based Reports provide insights into day-to-day actions and engagements that drive a B2B marketing and sales process.
  • Performance-Based Reports look beyond activities and assess the results and outcomes of sales and marketing efforts.
  • Forecasting Reports anticipate future sales movements and market trends.
  • Customer Reports offer insights into customers’ acquisition cost (CAC), their value over time, and the results of targeted marketing efforts.

Reporting Best Practices:

  • Ensure data quality for trusted reports.
  • Start with a simple report and gradually expand.
  • Establish a common data language (taxonomy).
  • Address duplicate data issues for accurate reporting.
  • Employ centralized data storage for effective reporting.

2) The Why: True Analysis

Definition: The Why job is most akin to the Greek root words of analytics—literally “untying a knot.” Analysts answer never-ending ad hoc questions from executives on almost any topic imaginable. This process cannot be automated; analysts need fast access to clean data and sound data science tools to get results.

Marketing and sales analysts are crucial to the success of a commercial team; they perform the vital task of true analysis, combining business and data science skills to unravel insights and drive informed decision-making.

Key Requirements for Analysts:

  • Fast compute and affordable, limitless storage.
  • Reproducible analysis through text-based data science languages.
  • Timely access to relevant data for analysis.

Agile Project Management for Analysis:

  • Use a Kanban approach for task listing and prioritization.
  • Maintain a clear list of analysis tasks, stakeholders, and due dates.

3) The Who: Targeting and Segmentation

Definition: The Who job deals with accounts and customers—segmenting them, targeting them, and serving them the right content. This job provides the strategic input for Account-Based Marketing (ABM).

Effective B2B targeting involves understanding market dynamics, including segment targeting, within-segment targeting, and within-account segmentation. The analytics team must integrate various data sources to create target lists and buyer archetypes.

Data Integration for Targeting:

  • Segment and Within-Segment Targeting: Industry trends, firmographic data, behavioral data.
  • Account Targeting: Master hierarchical table structure for accounts.
  • Within Account Targeting: Categorize key players, map decision hierarchies, and use qualitative and quantitative research.

Actionability Requirements for Segmentation:

  • Define actionability requirements before segmentation.
  • Assign segments back to leads and contacts with reasonable accuracy.

4) The How: Measurement

Definition: The How job is about measuring marketing and sales: How did we get this lead? And what can we do to get more like it? In B2C companies, media mix modeling (MMM) is commonly used to get these answers. This is trickier for B2B companies but just as critical.

B2B marketing measurement poses unique challenges, requiring a nuanced approach. Small-n deals, multiple objectives, chunkier tactics, long-time scales, audience complexity, and sales integration demand a hybrid measurement approach.

Measurement Challenges and Approaches:

  • Separate measurement for different objectives (awareness, demand generation, sales enablement).
  • Hybrid approaches combining econometric, deterministic, test-based, and heuristic-based methods.
  • Leverage management insights to complement data-driven models.

5) What’s Upcoming: Prediction

Definition: The What’s Upcoming job predicts what customers will do or respond to. It is the “action” side of account-based marketing and depends on machine learning techniques (predicting and classifying based on signals).

Machine learning plays a pivotal role in predicting outcomes in B2B go-to-market analytics. However, B2B predictive modeling faces challenges such as sparser signals, fewer observations, and a need for a deeper understanding of individual features.

Challenges and Considerations:

  • Signals are sparser in B2B predictive modeling.
  • Fewer observations require a more challenging modeling approach.
  • Consider outsourcing to ABM platforms but retain a team of predictive data scientists for deeper insights.

Unlocking Success in B2B Go-to-Market Analytics

Mastering the five essential jobs in B2B go-to-market analytics—reporting, true analysis, targeting and segmentation, measurement, and prediction—is key to unlocking success. Staying ahead with innovative approaches and a robust analytics strategy will pave the way for sustained growth as the landscape evolves.

Download our whitepaper, “A Roadmap for Modern B2B Go-to-Market: Part 2 – Operations and Analytics”​

Download this whitepaper to learn more about the processes, technology, and analytics needed to meet revenue goals.

Deliver brand success backed by data science

A Data Science-Backed Approach to Brand Building

Any marketer will tell you that brand-building takes significant time, concerted effort, and major resources. Measuring that work, however, can feel like an altogether separate beast. Quantifying the impact of upper funnel efforts on eventual sales is an increasingly high priority for effective marketing teams—those seeking to avoid the measurement trap and balance out their marketing investments.

While Media Mix Models (MMM) are often the best tool for evaluating marketing effectiveness and avoiding last-touch bias, these advanced tools can struggle with the long-term and slowly compounding nature of brand-building. Inevitably, we reach a key question: “How do we take a multi-stage and data science-backed (read: econometric) approach to modeling a brand?”

Generalized multi-stage model, where brand marketing drives sales directly, and indirectly via brand equity

Figure 1: Generalized multi-stage model

“Brand” as a Collection of Latent Factors

Let’s start with what we know: there is no single “brand equity” KPI that we can observe and track. There are certainly observable indicators of brand strength, things like awareness (be it aided or unaided), NPS, product ratings, repeat customer rates … the list goes on. And, on the other side of the equation, there are observable drivers of brand equity—this is anything that pushes the brand forward (or backward), things like earned media, competitor spend, celebrity partnerships, paid marketing campaigns … again, the list gets long in a hurry.

Once we have collected the relevant data and built a comprehensive list of our drivers and indicators, we can leverage an econometric technique called structural equation modeling to estimate the factor (or factors) in the “middle”—the unobservable thing that is driven by the drivers and indicated by the indicators—a latent factor. With a bit of data science wizardry, we can load these observable metrics against factors and review the strength of those relationships.

Choosing Brand Latent Factors

Again and again, we find that collecting indicators and drivers to load on a single factor (call it “brand equity”) simply doesn’t work. We’ll often generate strange coefficients with low confidence in our modeling outputs when using a single factor. There is no one-size-fits-all approach here, but with modeling iterations, we can start to think about a collection of brand factors that have distinct sets of drivers and indicators. And this is where things get interesting.

When aligning drivers and indicators to factors and selecting how many factors to model, we start a conversation between quantitative data science and qualitative marketing strategy. We collect a comprehensive data inventory, build a multifaceted “story of the brand” using key events and trends over time, and finally let the data tell us which aspects of our story seem to be resonating econometrically. This give and take will help crystallize our mental model and generate a set of latent factors to model on.

Here are some unobservable latent brand factors that might underpin our brand model:

  • Brand Awareness
  • Category Awareness
  • Brand Affinity
  • Brand Salience or Fame
  • Perceived Quality
  • Brand Loyalty
A structural equation model showing how marketing levers drive latent factors, which are in turn indicated by manifest variables

Figure 2: Illustrative map of SEM results for brand latent factors

What Can We Learn from Modeling Brand?

Once we have a model that (a) loads econometrically with high confidence and (b) lines up with our intuition (our “priors”), we can chart these brand factors over time and begin to investigate causal relationships or key moments in a brand’s timeline.

A brand's key attributes moving through time

Figure 3: Illustrative brand journey—not all factors move together

As we incorporate brand factors into a holistic marketing effectiveness model by layering on other components (i.e., modeling sales response, controlling for seasonality and macro factors, etc.), we can start to ask the juicy questions that brought us to this exercise in the first place, things like:

  • Which brand factor(s) align with the sales trend?
  • Which marketing channels drive which brand factors?
  • How much of a time lag is there between movements in brand factors and the resulting sales impact?
  • How much of a halo effect do our brand factors have on the effectiveness of lower funnel channels, like paid search?
  • How do strategic initiatives impact our brand factors, and are we moving the ones that matter?
  • In sum, what should we—as a marketing organization—be prioritizing?

Delivering Brand Success

A brand is a tricky concept to put a number on, but doing so is exceptionally powerful in a multi-stage model of marketing effectiveness. Using structural equation modeling with the right drivers, indicators, and latent factors, we can take a data science-backed approach and incorporate a cohesive brand story into a holistic effectiveness model.

Download our framework, “Measuring the Impact of Brand Marketing on Business Growth”​

Quantifying the direct impact of upper-funnel branding on tangible business growth is a challenge. Many marketers inadvertently fall into the “Brand Measurement Trap,” overestimating direct marketing’s role at the cost of long-term brand-building. This framework introduces an integrated approach, combining time series-based brand health metrics (awareness, consideration, and affinity) with a multi-stage modeling system, giving businesses the tools to quantify how brand marketing drives positive consumer behavior and generates long-lasting growth.

The power of aggregators in healthcare

In the rapidly evolving landscape of digital marketing, the healthcare industry has witnessed a surge in the demand for affiliate marketing, with aggregators playing a central role. Over the past decade, insurers, brokers, and senior care companies have increasingly turned to aggregators, paying a percentage of the product sold or a flat rate for a lead or referral. The rise of digital marketing has propelled these aggregators, who leverage sophisticated online comparison platforms, digital marketing expertise, and superior online customer experiences. 

The insurance aggregator market in North America stands at an impressive $7 billion, according to data from Allied Market Research, showcasing the substantial influence of aggregators in the healthcare industry.

Lead Generation Aggregators vs. Product Comparison Aggregators

Understanding the nuances between lead generation and product comparison aggregators is essential for companies navigating the affiliate marketing landscape. 

Lead Generation Aggregators

Lead generation aggregators employ their own no-brand-name marketing strategies to create demand, capture consumer interest, and subsequently sell leads to end-users such as insurers, brokers, or senior care companies. This can take various forms, including customer lists, leads, and introductions. 

Exclusivity: In addition to determining the depth of lead generation, executives must weigh the importance of exclusivity in their partnerships with aggregators. Many aggregators sell the same leads to multiple partners unless the end user pays a premium for exclusive leads or referrals.


Figure 1: Higher and lower cost and exclusivity in partnerships with aggregators

Product Comparison Aggregators

Product comparison aggregators sell directly to consumers by collecting branded digital marketing from competing companies on a proprietary online platform. This enables consumers to comparison shop in an unbiased manner. These aggregators often partner with trusted retail, pharmacy, and wellness brands to create co-branded marketing, offering consumers a convenient one-stop-shopping experience. 

Aligning Marketing Goals to the Right Affiliate Marketing Options

Leveraging affiliates for acquisition can be a lucrative endeavor if approached thoughtfully and executed with care. A clear understanding of the quality of member affiliate marketing can be crucial. Oftentimes, there are tradeoffs between lower acquisition costs and potentially lower customer lifetime value. 

Leveraging Aggregators for Acquisition

Acquiring members via affiliates is more than a “fire and forget” exercise. To maximize effectiveness, carriers can take concrete steps: 

  1. Capture Comparison Shoppers: Identify and capture consumers who are likely to comparison shop, ensuring a strong online presence on well-known comparison-shopping websites. 
  2. Increase Awareness for a Halo Effect: Elevate product and brand awareness to create a halo effect. Even when shopping around, consumers want to feel confident in their choices. 
  3. Broaden Reach: Aggregators, with their marketing efforts, have additional reach. Target audience segments that companies may not have otherwise captured, particularly in areas or demographics with lower penetration. 

Optimizing Member Retention

Retaining members acquired via aggregators presents unique challenges. To enhance retention, marketing executives should focus on four key factors: good data, early engagement, strong aggregator partnerships, and holistic analyses of aggregator-captured members. 

  • Good Data: Tracking member cohorts based on sales channels and applying predictive indices for likelihood to switch is essential for identifying high-risk members. 
  • Early Engagement: Members captured indirectly need early engagement strategies to foster a connection. Promote downloading the app, selecting a primary care provider, and completing the member profile. 
  • Strong Aggregator Partnerships: Collaborative communication plans between aggregators and insurers can improve retention among converted consumers, avoiding confusion and reinforcing the plan selection. 
  • Holistic Analyses: Regular analyses of aggregator-captured members are necessary to assess goal achievement and optimize aggregator utilization for future strategies. 

Navigating the Future of Healthcare Marketing

In conclusion, aggregators are an integral part of the health insurance landscape, offering companies the choice to actively capitalize on their strengths or passively receive sales. A scientific view of indirect channel tradeoffs, leveraging data to determine the future role of aggregators in a marketing and sales channel strategy, is key. To delve deeper into the intricacies of sales disruptions and recalibrated selling motions prompted by the pandemic, we invite you to download our whitepaper for invaluable insights. 

Download our whitepaper, “Navigating 5 Fundamental Shifts in Healthcare Marketing and Sales Channels”​

For a more in-depth exploration of the the role of aggregators, and four other disruptions, download our whitepaper.

4 situations when marketing effectiveness measurement is especially helpful

Every organization should have marketing effectiveness measurement tools at its disposal. Marketers can’t optimize on last-touch attribution alone, and educating internal stakeholders on how marketing channels work together has long-term benefits.

We’ve outlined four situations when marketing effectiveness tools, such as MMM and MTA, can be particularly helpful.

  1. Finance-led Organizations
  2. Diminishing Demand Environments
  3. Tightening Budget Environments
  4. Challenging Brand Environments

1. Finance-led Organizations

When the focus is on expense control, lean staffing models, and profitable revenue growth, marketing is frequently viewed as a cost center. CFOs often push marketers to employ media channels with the lowest cost per acquisition (typically demand capture channels such as paid search and affiliate), which may have a negative long-term impact. MMM and MTA can help finance-led organizations understand what portion of sales are driven by marketing without the risk of overattributing to paid search and affiliates.

Additionally, aligning on marketing-driven cost per acquisition can support (or start) conversations around the level of investment needed to generate additional sales to hit targets.

If finance sets a cost-per-acquisition limit at $100, Display could support an average of $34K daily spend, while CTV could only support an average of $3K.

Or, looking at it another way, if the budget increased by 15%, how many more sales would convert, and how would the overall cost per acquisition change?

Finance may be persuaded that the additional $3.6M in budget is worthwhile to generate an additional 13,000 sales.

2. Diminishing Demand Environments

When an organization has over-rotated toward lower funnel activities, sales demand decreases over time, and cost-pers increase. Marketing effectiveness measurement can help identify opportunities to increase upper funnel spend (to generate demand) through response curves. Response curves show where the next dollar should be allocated to maximize return and when a media will see less return on additional spend.

Once marketing decides to shift investment up the funnel, MMM and MTA can validate that this was the correct move, help track cost-pers by channel over time, and refresh spend curves at higher (or lower) levels to continue to optimize spend.

3. Tightening Budget Environments

No organization has an unlimited budget, and marketers often default to last-touch attribution to determine where to cut when budgets are reduced. Unfortunately, this is short-sighted; last touch significantly underweights less trackable activities, such as TV, online video, and other awareness marketing.

Looking only at last touch in the above example, one would likely cut TV spend. However, the actual marketing ROI of TV is significantly better than what can be measured with last touch. Marketing effectiveness measurement helps make marketers smarter with more information about the performance of marketing channels.

4. Challenging Brand Environments

When organizations are struggling with brand—such as a PR incident, long-run customer service issues, a popular new category entrant, or lack of brand awareness activities—it can be challenging to stay the course on brand spend. Brand-building activities take a long time to pay off, and organizational stakeholders (finance or the board, for example) may want to see proof that investments are working.

MMM can measure the impact of marketing on brand health and, consequently, brand health’s impact on sales. While moving brand metrics significantly takes a long time, modeling can confirm and quantify the link between brand health and sales.

An equation on “responsiveness” and “stickiness” can quantify how much brand spend is required to move the brand metric. “Responsiveness” measures how the brand metric responds to marketing (i.e., how quickly marketing can move the metric). “Stickiness” measures how long the change marketing drove lasts (i.e., how quickly the metric returns to previous levels).

Combining these two elements quantifies how much spend is required to improve a particular brand metric and how long that change will last. Example: One week of $1M brand spend will yield a cumulative increase of 1.7 percentage points in Awareness over eight weeks.

Looking at historical sales performance during periods when brand metrics were higher helps quantify the sales impact of increasing a brand metric calculated in the formula above.

Together, these can help stakeholders outside of marketing understand the sales impact of these activities over a shorter period and buy marketing time for the flywheel of long-run brand improvement to take effect.

Conclusion

Usually, we are brought in when things aren’t going well, and frequently, the issue is one of the four situations covered above. All organizations could benefit from a more holistic approach to marketing measurement, though.

Download our framework, “Measuring the Impact of Brand Marketing on Business Growth”​

For a comprehensive exploration of brand marketing’s impact and our multi-stage modeling approach, download our detailed framework.​

From state lines to sales lines: Unraveling the power of ZIP Code analytics

The Value of Zip Code-Based Data in Go-to-Market

Marketing and sales teams rely heavily on data to make informed decisions, drive strategies, and target their efforts effectively. In today’s world, zip code-based geographic and demographic data offer several invaluable insights that can supercharge their endeavors. Different regions can exhibit widely varying consumer behaviors, preferences, and needs. What works in one area may fall flat in another due to cultural, economic, or even regulatory differences.

Consider the Medicare insurance market. In this industry, healthcare policy terms, conditions, and coverages change at the local (zip code) and state levels. Not only do sales and marketing teams face restricted distribution and data collection regulations, but plans and markets are highly competitive. For marketing and sales professionals, the capability to segment by location ensures marketing messages on the benefits of plans are accurate and relevant to the demographic.

Optimizing Marketing’s Impact Using Geographic Data

Many of Marketbridge’s clients are in the thick of geographic-based data every day. The granularity and precision offered by such data can be a game-changer in the competitive landscape, ensuring that efforts are data-driven, precise, and, most importantly, effective. For example, a marketing campaign that resonates strongly in an urban setting may not have the same impact in a rural area. Similarly, a product that sells well in a high-income neighborhood may not perform as well in a less affluent one. By breaking down performance data geographically, organizations can tailor their approaches to align with local conditions, enabling them to allocate resources more effectively and improve ROI.

Predicting Customer Changes

Additionally, a geographic lens on performance data can provide valuable insights into emerging market trends and opportunities that are not immediately obvious when looking at aggregate data. For instance, a sudden spike in sales in a specific region could signal an unmet need or untapped market potential, providing a first-mover advantage to companies that act quickly. Alternatively, a decline in a particular area could be an early warning sign of market saturation, increased competition, or shifting consumer sentiments. Understanding these geographic nuances allows businesses to adapt and pivot their strategies proactively, ensuring they stay ahead in meeting customer needs and expectations.

Examples of Zip Code-Based Use Cases for Modern Go-to-Market

  • Targeted Marketing Campaigns: Zip code demographics allow marketers to tailor campaigns, ensuring better audience resonance and higher conversion rates.
  • Resource Allocation: Sales teams use zip code data to pinpoint high-potential areas, optimizing resource distribution.
  • Personalized Customer Experience: Demographic data informs localized marketing strategies, enhancing brand differentiation.
  • Predictive Analysis: Geographic and demographic insights help companies anticipate product or service demand.
  • Product Development: Demographics guide product features, pricing, and branding for potential launch areas.
  • Optimizing Distribution: Zip code data identifies ideal locations for distribution hubs or outlets.
  • Competitive Analysis: Understanding area demographics helps marketers assess competitor footholds and adjust tactics.
  • Risk Management: Businesses gauge regional risks using socio-economic data, influencing decisions like insurance premium settings.
  • Enhanced Customer Segmentation: Zip code data refines customer targeting, from luxury promotions in affluent areas to budget offerings elsewhere.
  • Compliance and Regulation: Zip code demographics help sectors like finance maintain local regulatory compliance.

The Challenges of Geospatial Data Analysis Using Zip Codes

While the significance of geographic data for sales and marketing teams is clear, transforming raw data—whether latitude/longitude, address, zip code, or census block—into actionable insights requires expertise. In the past, businesses leaned heavily on U.S. postal service data for geographic information, often outdated (by over a decade) and plagued with technical inaccuracies. Data associated with zip codes that traverse state boundaries presented additional challenges. U.S. Census Data provides a more accurate alternative data source for mapping, but one-off analysis is time-consuming.

The Roles and Toolsets Needed

Zip code analysis requires the right combination of talent and automated technology. At Marketbridge, “data detectives”—analysts with a deep data science toolkit who explore diverse data sets to understand why events/trends have happened—play a critical role in this endeavor. They use open-source, fully documented code connected to source systems, vendor files, and public geospatial data to standardize analysis and modeling of geographic data changes over time. The solutions these analysts create are document-based with code embedded in workbooks and are easily knittable to reports mapping changes and trends in consumer behavior for sales and marketing teams.

Analysts at Marketbridge use RStudio and the extendable RMarkdown/Quarto framework for this use case. Unlike Jupyter Notebooks, Quarto documents are entirely text-based, making version control easy via Git/Github, and are then convertible into HTML reports, PowerPoint slide decks, or Word documents.

Building an Open-Source Solution for Zip Code Data

Frustrated by the limitations of R mapping when using Census data to overlay demographics on zip codes, I built Choroplethr, an open-source package for the R language, as a solution. Choroplethr seamlessly connects to U.S. Census Data for mapping, providing more accurate geospatial data and automatically handling visual mapping tasks.

Choroplethr is open-source and available to the public; to use it from R, simply type:

install.packages(“choroplethr”)

Choroplethr works very well with a fairly new R package, zctaCrosswalk. This package contains the U.S. Census Bureau’s 2020 ZCTA to County Relationship File, as well as convenience functions to translate between states, counties, and zip code tabulation areas (ZCTAs). You can install the package like this: install.packages(“zctaCrosswalk”).

Mastering Data for Strategic Growth

In today’s data-driven landscape, the nuanced understanding of geographic and demographic data, down to the zip code, is pivotal for marketing and sales strategy. Through tools like open-source software (R studio), professionals can access refined insights that were previously difficult to harness or done via a one-off analysis. These tools not only aid in crafting precise campaigns but also in understanding market dynamics at a granular level. This journey underscores the importance of evolving technology and innovation in making sense of vast data sets.

Download our report, “The State of Marketing Analytics”​

Download our marketing analytics benchmark report to learn more about 1) Direct-from-the-source challenges and priorities heard from marketing analytics leaders, 2) Key insights on how organizations are maintaining, running, and growing their analytic functions, and 3) Go-forward actions for marketing analytics teams to improve processes and advance analytics.

Skip to content