“On TV And Video” is a column exploring opportunities and challenges in advanced TV and video.
Today’s column is written by Jane Clarke, managing director and CEO at the Coalition for Innovative Media Measurement (CIMM).
Attribution, or the measurement and assignation of an outcome following an ad exposure, has revolutionized media planning and buying with the promise of being able to not only assess the impact of spend but optimize it moving forward.
While there are many different approaches to attribution, with each vendor applying a unique solution, the underlying data definitions that serve as the basis for their analysis should be the same.
And currently they are not – a shortcoming that is inhibiting the overall growth and impact of attribution as a measurement discipline. It’s not that the attribution analyses aren’t valuable, it’s that buyers can’t feel confident that there isn’t hidden error confounding their results.
This reality was made clear in an analysis of TV attribution providers just completed for CIMM and the 4A’s Media Measurement Task Force by Sequent Partners and Janus Strategy and Insights.
The research found that ad occurrence and exposure data are highly inconsistent across providers. The reason was not only the differences between the actual data, but primarily the methodology being used to convert that data into final ad occurrence files and exposure data, including weighting, editing and other data processing rules.
The situation is analogous to everyone working from their own “set of facts,” defining reality in different ways. That is why, if you had a variety of providers examine the same campaign, you would in all probability end up with very different conclusions about ROAS.
Data on ad occurrence and frequency varies greatly from attribution provider to attribution provider because currently there are no standards defining the methodology that should govern processing of this data. Television attribution results will become more transparent, consistent and reliable when providers adopt more stringent media measurement standards.
The Media Rating Council has an effort underway to develop outcome-based measurement standards. Ideally, when it comes to how data processing is defined, these standards should address:
- Weighting. Providers need to implement a robust panel weighting scheme that addresses variables common to TV viewing such as DMA, HH size/Presence of Children and Income/Education/Occupation.
- Unification. A standard process for unifying the database for ROI measurement is needed, providing a common base of viewers where there is both an opportunity for exposure and an opportunity for a response, such as website visitation, retail traffic or purchase.
- Exposure Qualification. There needs to be agreement as to what standard exposure criteria is. Should it be one second, three seconds, five seconds, 10 seconds or a one-minute schedule at 300 GRPs?
- Occurrences. There has to be rigorous quality control in the re-creation of as-run schedules and in the evaluation of Reach reporting from exposure data across schedules.
For attribution to fulfill its promise and for buyers and sellers alike to fully embrace it for ROAS analysis, we need a common approach to defining the starting point of occurrence and exposure data.
In an industry where proprietary exclusivity is all, standards are sometimes bristled at.
Certainly, attribution providers can and should go to market with their own unique approaches. But all we need to start on the same page. And to accomplish that, we need standards to govern how occurrence and exposure is being defined and processed from data, to ensure accurate and consistent assessment of ROAS.