Instead of trying to test endless variations of minutiae, look for the big ideas that impact customer experience and buying process.
Is yours the typical company conducting two to five tests a month, struggling to eek out more from your optimization program, and wasting critical resources of people and website traffic? That's the result of not focusing in on the big idea!
How much should you be testing?
A mid-size company can easily handle 30 to 50 tests a month. The reason most companies never get there is because they waste so many cycles on what I call "slice and dice" optimization. Let's consider the following test, which I found "in the wild" - and which I find ironic because it is for a service offering a marketplace of landing page designers.
Depending on how you want to define your variables, these two pages have around a dozen changes. I hope this is not what you want from your landing page designers. You don't need someone to create endless variations of every variable.
Can you identify all of the variables being proposed for testing? (I'll share with you my list next time.)
For now, let's assume that for each of these variables you test just two variations even if more may be warranted. I'll show you what the problem with that approach is. (Note: I'll make use of Google's Website Optimizer "Test Duration Calculator" to estimate the numbers, but you could easily do this by hand or with a calculator or simple spreadsheet.):
I don't know the true stats for this page, but they don't really matter in order to illustrate the challenge.
Let's assume the following:
That means it takes more than 108 days - over three months! - to complete this test. That's a whole-heck-of-a-lot of visitors and a whole-heck-of-a-lot of time consumed to get one test completed.
The way we teach testing, there are probably three variables worth testing (variables that communicate to a visitor) on this page. Let's assume the same two variations for each, though to be frank, one of the variables I would want at least three or four variations if we were doing this test for an actual client. But for the sake of simplicity, let's keep everything the same.
So now we would have:
This test would be over in just under 18 days, a scant two and a half weeks.
Which way seems more efficient?
Should you test for variables that seem to really matter to visitors vs. testing virtually random variations of elements in the hope something gives you a little lift? You may achieve some gains - that's why this practice is so common - but you'll burn out waiting for the results. This is why so many optimization efforts fizzle out over time.
Next time I will share the variables I found and discuss the three elements I would test on this current page.
In order to give you a leg up on identifying the variables on your own, I'll give you a question as a framework. When was the last time you looked at a page and said to yourself, "The layout is horizontal and not vertical, so gosh darn it, I can't buy from this page"? Vertical vs. horizontal layout could matter as a display of information issue if you are trying to change a lot of what is above and below the fold. But that really wasn't the case in this example. It's just a waste of time and effort unless you have no real idea what will move the needle for customers - and in that case, any test is better than nothing. Maybe.
Instead of trying to test endless variations of minutiae, we teach companies to look for the big ideas that impact customer experience and buying process. The smaller variations we can always come back to after the big ideas establish directionality. Can you find the big ideas being tested (or that should be tested) in this example?
Join the Industry's Leading eCommerce & Direct Marketing Experts in Chicago
ClickZ Live Chicago (Nov 3-6) will deliver over 50 sessions across 4 days and 10 individual tracks, including Data-Driven Marketing, Social, Mobile, Display, Search and Email. Check out the full agenda and register by Friday, Oct 3 to take advantage of Early Bird Rates!
Bryan Eisenberg is coauthor of the Wall Street Journal, Amazon, BusinessWeek, and New York Times bestselling books "Call to Action," "Waiting For Your Cat to Bark?," and "Always Be Testing." Bryan is a professional marketing speaker and has keynoted conferences globally such as SES, Shop.org, Direct Marketing Association, MarketingSherpa, Econsultancy, Webcom, SEM Konferansen Norway, the Canadian Marketing Association, and others. In 2010, Bryan was named a winner of the Direct Marketing Educational Foundation's Rising Stars Awards, which recognizes the most talented professionals 40 years of age or younger in the field of direct/interactive marketing. He is also cofounder and chairman emeritus of the Web Analytics Association. Bryan serves as an advisory board member of SES Conference & Expo, the eMetrics Marketing Optimization Summit, and several venture capital backed companies. He works with his coauthor and brother Jeffrey Eisenberg. You can find them at BryanEisenberg.com.
IBM Social Analytics: The Science Behind Social Media Marketing
80% of internet users say they prefer to connect with brands via Facebook. 65% of social media users say they use it to learn more about brands, products and services. Learn about how to find more about customers' attitudes, preferences and buying habits from what they say on social media channels.
An Introduction to Marketing Attribution: Selecting the Right Model for Search, Display & Social Advertising
If you're considering implementing a marketing attribution model to measure and optimize your programs, this paper is a great introduction. It also includes real-life tips from marketers who have successfully implemented attribution in their organizations.
September 23, 2014
September 30, 2014
1:00pm ET/10:00am PT
October 23, 2014
1:00pm ET/10:00am PT