WTF is log-level data?

This article is a WTF explainer, in which we break down media and marketing’s most confusing terms. More from the series →

Gaining better visibility of what’s happening in the digital advertising supply chain goes hand in hand with knowing the right questions to ask. One question that vendors are being increasingly pressed for by publishers is: “Where’s my log-level data?” Ad buyers are also asking more for customized log-level data from exchanges.

But what exactly is log-level data, why is it necessary, and why is it difficult to get hold of?

Here’s all you need to know.

What is log-level data?
All data that’s relevant to a single impression. It could be geo data, URLs, cookie IDs, time stamps, viewability levels, and, of course, the good stuff: transaction data. It’s the transaction data more publishers are asking exchanges for. Armed with log-level data, a publisher can see exactly what is occurring in its digital ad supply chain. It can see details on what fee each vendor in the chain takes from the amount the marketer bids on the inventory. Log-level data can also show detail like whether an exchange is running multiple bids on inventory on behalf of the same client, when they shouldn’t be, in order to make their match rates look better and also means they could in theory duplicate their take rates.

Why do publishers want access to this more now?
To highlight any skulduggery in their digital ad supply chains. Having access to log-level data, given to them by their exchanges, enables them to look at what’s going on across their supply chain. Sometimes auction dynamics can be tweaked without the buy or sell side always knowing. For instance, when vendors switched to first-price over second-price auctions, publishers and agencies weren’t always informed. “We want to look under the bonnet at the full transaction data across our supply chain,” said Ryan Skeggs, gm of digital sports publisher GiveMeSport. “We want to know what their take rates are,” he added. The Guardian has also been proactive in requesting log-level data.

Why do ad buyers want access to it?
For buyers, it’s equally critical to have full log-level data, for more mixed reasons. On the one hand, it enables them to fine-tune their media planning if they have more specifics. “It helps us derive more insights so we can do things like understand not just attributed media but full paths to conversions or experiments on brand uplift and impacts of different partners, tactics and platforms,” said Matt McIyntre, head of programmatic for Europe, Middle East and Africa at Essence. It also helps them make more accurate planning around the reach and frequency of ads shown on an individual basis, he added. But it’s also useful for keeping an eye on any opaque or murky auction dynamics. For instance, after Index Exchange’s bid caching embarrassment, agencies requested log-file data to track whether it had been affecting them. Likewise, agencies like Essence have done the same to keep an eye on other auction dynamics changes that haven’t been declared — such as sudden shifts to first-price auctions, and use of dynamic floor pricing, according to McIyntre.

How do you get a hold of the data?
The most complete log-level data comes from the exchanges. There are some vendors who are making it available for free by request. Others have told publishers it will cost them a monthly fee, others have offered to give a small (like 1 percent in some cases) percentage of the data, which is useless for publishers. Others have said they’re hampered from sharing it due to contractual obligations with other partners. Publishers are skeptical of excuses though. “Exchanges are very good at scaremongering publishers about the cost of doing this. But publishers must just be persistent, because it’s not that hard,” said Skeggs.

So once you have the log-level data, then what?
It requires a place to store it, and some analysts to distill what the data means. Publishers and advertisers must be very specific about their objectives before they request what log-level data they want. We’re talking terabytes upon terabytes of data, so it’s worth customizing what is needed for specific objectives. Otherwise, it will be a confusing sea of information, which is costly to store and will take 10 times the amount of time to sift through.

More in Media

After an oversaturation of AI-generated content, creators’ authenticity and ‘messiness’ are in high demand

Content creators and brand marketing specialists on how 2026 will be the year creator authenticity becomes even more crucial in the face of rampant AI-generated “slop” flooding social media platforms.

‘The net is tightening’ on AI scraping: Annotated Q&A with Financial Times’ head of global public policy and platform strategy

Matt Rogerson, FT’s director of global public policy and platform strategy, believes 2026 will bring a kind of reset as big tech companies alter their stance on AI licensing to avoid future legal risk. 

Future starts to sharpen its AI search visibility playbook

Future is boosting AI search citations and mentions with a tool called Future Optic, and offering the product to branded content clients.