It was easier than we thought. It took us not much more time than we'd planned. The developers put in a universal tag. We enabled a spreadsheet. We served cookies (at least that's what I heard). The test data came in. We clicked some buttons in the analytics tool and got dashboards. When the site launched, the early word was that we had increased our visitors.
We were happy! We bought ice cream for everyone. Our most senior marketer told me: "You have built a bridge to the future of marketing through analytics."
Two months later, the senior marketer was gone. He had taken a position behind an espresso machine. The developers were still clinking glasses over the successful launch when they got the sack. I hid in a broom closet until the shouting stopped.
Bridge to Nowhere
We had built a bridge to nowhere. We had no way of checking for accuracy, no way of analyzing the data that came back, no way of answering business questions, and no buy-in from anyone that they would need to change anything depending on what the analytics said.
All we had done was "implement the tool." The tool was much like the other tools, and we followed the guidelines so that all the reports would "come out" looking the way they were supposed to. And that was where the bridge to nowhere stopped - and where we were left lingering at the edge, peering down 500 feet into a swirling maelstrom of uncertainty about what was really going on with our sites.
There were signs that things were wrong: mismatched data with our affiliates, for instance; and more mismatches from our ad networks. Then some of the content changed, and the reports just stopped making any sense. I was reminded of that commercial where the fatuous guy and his partner in the candy business are "building a website - by ourselves!"
It didn't take long for the whole metrics structure to become creaky, unreliable, annoying, frightening, and finally useless. The project was scrapped and tears were shed.
It might have been avoided.
I met someone at a conference who told me that most successful companies don't just "implement a tool." What they do first is rely on a repeatable success process; then establish business goals; then build a site that tries to drive those business goals; then use a well-targeted, expertise-driven web analytics plan (using any of a number of tools) to see how close they got to matching those goals.
"Well, we did that - sort of," I said.
"Let me guess," said my new friend. "You did a pretty good job of planning and setting up the tool."
"You planned for day one after launch. But you didn't plan for day 150."
"No," I said. And it was true.
We didn't plan for new questions coming to us from business stakeholders. We didn't expect to have to make "the tool" jump through flaming hoops to extract small nuggets of data that would inform an important business decision. We didn't expect that the site itself would change and require a reestablishment of goals and principles based on the changes that came from editorial. And we didn't bother measuring them at all, because we didn't have a template for how to do that. Nor did we have the in-house resources to spend on a full understanding of exactly how to bend and shape this tool into an insight engine. Worst, the interns they had originally put in place to "manage the tool" once it was launched had now been redeployed to "tweet" for us (and whose misguided efforts landed us in court) - but that is another story.
What we lacked most was a template for change. Because at a certain point, I got the hang of the tool, and was able to answer questions as they came up. But the problem then was that we still had the old-fashioned mud-fights about whose content had to change. Was it the content that fared worst? No, it wasn't. The claim was that it had been improperly measured (it hadn't been, but without third-party expertise, we had no way to claim objectivity). And what ended up getting changed was the amount of links on the home page, by a factor of at least four - because every content curator, absent measurement to tell them whether they were a light-hitting journeyman infielder or an Albert Pujols of web content, insisted on getting their square of real-estate.
And so the chaos continued. At times I thought the website itself would simply come crashing down around us in a simulacrum of sheer frustration. And we continued to pay licensing fees to the tool vendor.
Forget the Numbers
Things are better now.
The new CIO canceled everything to do with web analytics because, after studying what we had done with it, he concluded "none of it was actionable." He won kudos for saving the company a ton of money. No license fees, no analytics experts, no time wasted on tagging, no reconciliation of numbers, none of this "trying to get it right."
I took a walk out on the bridge the other day. It was swaying in the wind, and still only halfway done. On the other shore I thought I saw smoke signals, but after I used my binoculars I could see it was steam from hot tubs where marketers relaxed knowing they had built their analytics around a plan, had imported some senior expertise, and had made the organization respond to the news about what was pulling people to conversion and what was driving them away.
The edge of our bridge was cold and windy. The water below swirled. Looking back, I could see folks were even dismantling what bridge we'd built - figuring on getting something for it in the scrap market.
We sure had saved money by stopping before it got too late!
But those hot tubs looked awfully comfortable across the bay. And the grass around them was greener than on the side I had to go back toward. For a minute, I thought I would try to swim over to the hot tubs. But the waters looked cold and I had to get back.
What sounded better than staying with the bridge to nowhere? That I knew I could get in my car and maybe take the backroad over to the hot tubs.
This column was originally published on Nov. 7., 2011 on ClickZ.
Meet Your Favorite ClickZ Contributors
Many of ClickZ's leading expert contributors will be at ClickZ Live, the new online and digital marketing event kicking off in New York (March 31-April 3). Hear from the likes of: Jeremy Hull, Lisa Raehsler, Andrew Goodman, Bryan Eisenberg, Mathew Sweezey, Aaron Kahlow, Stephanie Miller, Simms Jenkins, Jeanne S. Jennings, Dave Hendricks and more!
Andrew is a digital marketing executive with 20 years' experience servicing the enterprise customer. Currently he is Managing Partner at Efectyv Digital, a digital marketing consulting company, and Managing Partner at Technology Leaders, a web analytics consulting firm he founded in 2002. He combines extensive technical knowledge with a broad strategic understanding of digital marketing and especially digital measurement, plus hands-on creative in the form of the written word, user-experience and traditional design.
His practice is dedicated to building customers' digital marketing success and helping them save money during the process.
He is a writer, a public speaker and a visual artist as well.
His book "Digital is Destroying Everything—and What Comes Next" will be published by Pearson in the Spring of 2014. He writes a regular column about Analytics for ClickZ, the 2013 Online Publisher of the Year. He wrote the groundbreaking "Dawn of Convergence Analytics" report which was featured at the SES show in New York, and the second report in the series will be featured at the same show in San Francisco.
In addition to speaking at SES, he has presented at eMetrics; and his session was voted one of the top ten presentations at the DMA show in Las Vegas. He is speaking again at the DMA in Chicago in the fall of 2013.
In 2004 Andrew co-founded the Digital Analytics Association and is currently a Director Emeritus. He has designed analytics training curricula for business teams and has led seminars on digital marketing subjects.
He was also an Adjunct Professor at The Pratt Institute where he taught Advanced Computer Graphics for 3 years. Andrew is also an award-winning, nationally exhibited painter.
March 19, 2014