Home Data-Driven Thinking Ad Tech Desperately Needs Data Exchange Standards

Ad Tech Desperately Needs Data Exchange Standards

SHARE:

Data-Driven Thinking” is written by members of the media community and contains fresh ideas on the digital revolution in media.

Today’s column is written by Sanjay Agarwal, vice president of engineering at Drawbridge.

In the ad tech ecosystem before OpenRTB, proprietary protocols caused long integration cycles, more code complexity and maintenance and custom logic for parsing requests.

The OpenRTB protocol replaced all proprietary protocols for sending ad requests from supply-side platforms (SSPs) to demand-side platforms (DSPs), resulting in faster integrations and a scaled ecosystem as a whole.

While this solved the SSP-DSP interaction almost a decade ago, there is still unfortunately no standard format for exchanging data between data management platforms (DMPs) and DSPs, posing a major point of friction in connections between DMPs and DSPs. There is a strong need for the standardization of data formats and the transport layer to push data from a DMP to a DSP.

While this problem has been around for a decade or so now, the proliferation of data providers has made standardization all the more pressing. While the DMPs and DSPs certainly stand to benefit from standardization, the real winners at the end of the day would be marketers. They would be able to more seamlessly leverage first- and third-party data, get campaigns set up quicker, more easily receive data back and never have to worry about integration timelines.

Today, the typical timeframe for adding a new integration from DMP to a DSP is anywhere from two to four months. The “standard” options for the transport layer can be via the file transfer protocol, S3 bucket or HTTP, and the data format itself can be proprietary. Product and engineering teams from both sides are involved in the integration, with calls, email threads and a lot of back-and-forth that could easily be avoided with a standardized data format.

The friction of sending and receiving data has resulted in niche DMPs using bigger DMPs like BlueKai to channel the data to media platforms. But usually this third-party onboarding results in a loss of functionality and scale, with a drop-off occurring at every data hop. Some companies also act as intermediaries to funnel the data from data producers to data consumers, which again results in loss of scale because of cookie translation. Data debugging becomes harder when third parties are involved, but unfortunately, due to today’s landscape, it’s required.

Offline transfer of data, such as through S3 or the file transfer protocol, results in less robust data validation. There is no feedback given to the sender on whether the data is received and processed correctly. Plus, if the data is time sensitive, such as first-party data used for retargeting, offline transfer will usually cause further delays, rendering the data useless.

Data formats are fragmented. Typically, audience data is organized by storing segment IDs by cookies, device IDs or email hashes. Whatever data format is chosen, it should support both adding and removing segment IDs to cookies or device IDs.

Another ideal feature would be bulk updates, where multiple user IDs can be modified, and each modification could even be accompanied by an expiration date until which the segment is valid. This could also support consumer choice and respect privacy through a standard opt-out that resonates across all IDs associated to the consumer. I could go on.

Having a standard format to send data across would drastically cut down the time taken for integration between DMPs and DSPs, resulting in a more seamless exchange of data. Plus, new integrations could be turned on with a flip of a switch, and a real-time transport layer could support time-sensitive data.

Just like OpenRTB has resulted in seamless integrations between SSPs and DSPs, the time is ripe to have a standard data format and transfer process for exchanging audience data between DMPs and DSPs. This will result in more platforms participating in the ecosystem, faster integrations and, ultimately, greater efficiencies for marketers.

The question is, who will lead the charge?

Follow Drawbridge (@Drawbridge) and AdExchanger (@adexchanger) on Twitter.

Tagged in:

Must Read

Scales and hands touching the bowls with index fingers from opposite sides. Arguments, evidence and tricks in trial. Concept of judging, trial and justice

The FTC Bars Kochava From Selling Sensitive Data Without Consent

It’s been nearly four years since the Federal Trade Commission first accused Kochava of selling highly sensitive location data. Now, the two have finally reached a settlement.

Comic: CTV Tracking

Upfronts Advertisers Say They Want Outcomes – And Amazon Licks Its Chops

Amazon has packaged a handful of upgrades to its ads measurement solutions, obviously catered to TV and streaming media advertisers.

AdExchanger Senior Editors Anthony Vargas and Alyssa Boyle.

POSSIBLE 2026: AdExchanger's Hot Takes

AdExchanger Senior Editors Alyssa Boyle and Anthony Vargas share their takeaways from three days chatting about agentic AI at POSSIBLE.

Privacy! Commerce! Connected TV! Read all about it. Subscribe to AdExchanger Newsletters

Reddit Reports A 75% Boost In Q1 Ad Revenue As It Reaches For 100 Million Daily US Users

Generative AI search has pushed traffic off a cliff across most of the internet, but not on social platforms. Reddit included.

POSSIBLE 2026: Can AI Help Agencies Finally Break Down Those Silos?

Domenic Venuto, indie agency Horizon Media’s chief product and data officer, sat down with AdExchanger during POSSIBLE at the Fontainebleau in Miami to unpack the role of AI in today’s media and advertising landscape.

Google Touts Its AI Ad Tech Adoption And New AI Max Features

Google announced new features and ad types for AI Max, its AI-based bidding product for search and shopping or sponsored product ads. The company also touted “hundreds of thousands” of advertisers using AI Max.