Can empirical approaches outperform human experts in different disciplines, from baseball scouting to digital marketing? Here's why predictive analytics is more than just a numbers game.
I'm just reading "Super Crunchers" by Ian Ayres. It offers an interesting look at how data mining and predictive analytics are becoming more widespread and are increasingly shaping our lives. Ayres cites examples where empirical approaches outperform human experts in their ability to accurately predict likely outcomes.
I particularly liked his story of an econometrician able to predict the expected quality of Bordeaux wine based on a simple regression analysis of weather data. He could predict the expected quality of a particular vintage based on just three variables: the amount of winter rainfall, the average temperature during the growing season, and the amount of rainfall during the harvest. Most interesting is the resistance and even hostility he got to his predictions from the wine establishment. The wine experts of the time were threatened and affronted by the fact that their art and expertise could be reduced to a simple equation.
Ayres provides examples from other industries where data mining and predictive analytical techniques have changed the rules of the game, from baseball scouting to social policy development to medicine. Quite often, there's been resistance to these techniques from established experts in that field. They would not or could not accept that such empirical methods could be better than the expertise they had developed through years of training and experience. However, numerous studies cited by Ayres have shown that predictive analytics outperforms experts in predicting an outcome correctly. That doesn't mean predictive techniques always get it right, just that they get it right more often than the experts.
In the digital marketing field, Ayres uses the example of A/B and multivariate testing. He points out that today's volume of data and technology allows people to run repeated tests and trials to predict which versions of which page element are most likely to be successful in driving the desired outcome. Anyone familiar with multivariate testing technologies knows that the marketing stance regarding them is often that they eliminate the need for subjectivity in the design process. You just come up with some alternative versions and see which one works best. It's the ultimate tool for overcoming bias and subjectivity of the various stakeholders involved in site development. Who needs usability testing, right?
Ayres' background isn't as a statistician or an analyst but as a lawyer. You don't immediately think of lawyers as being masters of the empirical universe. Why would a lawyer be an expert in number crunching? Being a lawyer could be similar to being an analyst, though. Each tries to prove or disprove a hypothesis and looks for the appropriate evidence to support a theory or disprove somebody else's. Thus, it's a fallacy to believe that econometrics and predictive analytics are purely scientific disciplines.
Predictive analytics is often as much about art as it is about science. To build a good model, you must have a good understanding of the way the system you are trying to model works. More often than not at the beginning of the model building process, there's some subjective opinion about the likely factors influencing the thing you're trying to predict. So where do these opinions come from? They usually come from experts in that particular field. We sometimes called this the domain expertise. In the econometrician predicting wine quality example, the econometrician was also a wine buff so he possessed knowledge about the likely factors that could potentially affect a particular vintage's quality. His skill was in quantifying it.
Likewise, some domain expertise is needed to develop good tests. If we look at multivariate testing, technology can help determine which is the best page design to use. If you test four different versions of an element, say a call to action, then you'll get a winner. That winner may be the one you started out with, but it's still the winner. It doesn't mean that it's the best one, it's just the one that is best out of the various options you examined. There may be a much better option out there that you haven't tested. Usability experts can potentially provide better insights into what versions are the best ones to test in the first place and help understand test results.
We need experts to help us build better models. That expertise may come from years of experience or knowledge gained from understanding previous models' effectiveness. In either case, there's room for both science and art.
Join the Industry's Leading eCommerce & Direct Marketing Experts in Chicago
ClickZ Live Chicago (Nov 3-6) will deliver over 50 sessions across 4 days and 10 individual tracks, including Data-Driven Marketing, Social, Mobile, Display, Search and Email. Check out the full agenda and register by Friday, Oct 3 to take advantage of Early Bird Rates!
Neil Mason is SVP, Customer Engagement at iJento. He is responsible for providing iJento clients with the most valuable customer insights and business benefits from iJento's digital and multichannel customer intelligence solutions.
Neil has been at the forefront of marketing analytics for over 25 years. Prior to joining iJento, Neil was Consultancy Director at Foviance, the UK's leading user experience and analytics consultancy, heading up the user experience design, research, and digital analytics practices. For the last 12 years Neil has worked predominantly in digital channels both as a marketer and as a consultant, combining a strong blend of commercial and technical understanding in the application of consumer insight to help major brands improve digital marketing performance. During this time he also served as a Director of the Web Analytics Association (DAA) for two years and currently serves as a Director Emeritus of the DAA. Neil is also a frequent speaker at conferences and events.
Neil's expertise ranges from advanced analytical techniques such as segmentation, predictive analytics, and modelling through to quantitative and qualitative customer research. Neil has a BA in Engineering from Cambridge University and an MBA and a postgraduate diploma in business and economic forecasting.
IBM Social Analytics: The Science Behind Social Media Marketing
80% of internet users say they prefer to connect with brands via Facebook. 65% of social media users say they use it to learn more about brands, products and services. Learn about how to find more about customers' attitudes, preferences and buying habits from what they say on social media channels.
An Introduction to Marketing Attribution: Selecting the Right Model for Search, Display & Social Advertising
If you're considering implementing a marketing attribution model to measure and optimize your programs, this paper is a great introduction. It also includes real-life tips from marketers who have successfully implemented attribution in their organizations.
October 23, 2014
1:00pm ET/10:00am PT