Elena Grewal presented these slides on a/b testing in the real world (offline experiments not online) at the Big Data Innovation Summit on April 9, 2014.
14. Stepping back
14
+ Many companies have offline operations
+ Can optimize these using experiments
!
!
!
Online Experiments:
We run these all the time too.
If you are curious about on our online experimentation see Jan Overgoor’s tech talk
http://nerds.airbnb.com/tech-talks/
16. Before and after won’t work
16
• Often very little data before professional photos are added
• Seasonality and other confounding factors bias results
17. Selection bias often impacts analysis
17
• Listings that opt to get professional photography are not the
same as listings that do not get photography
18. Without an experiment, we don’t know the causal effect
18
This is the same reason we need online experiments
Date
01−01 01−15 02−01 02−15 03−01 03−15
Product Launch
Product Rollback
Launch initiative:
e.g. Offered Free Professional Photography
19. Traditional A/B Testing Online
Great sources:
http://mcfunley.com/design-for-continuous-experimentation
http://www.evanmiller.org/how-not-to-run-an-ab-test.html
Control Treatment
19
25. Local Operations: Market Level Experiment
25
!
+ Smaller “long tail” markets < 100 reviewed listings
Randomize Markets
93 Treatment / 92 Control
Assess impact of operational strategy on market growth
+ Statistically measure the lift due to local ops teams
+ Measuring active listings, hosts, reviewed listings, and
bookings
27. Finding: Local Ops Efforts Have Positive Impact on Growth
27
Active Listings
Control
17% Growth
Local Ops Kickoff
Treatment
31% Growth
28. Case Study: Campos do Jordão, BR
28
+ Market grew 9x
+ Over 90% of the new listings are from new users
+ Low CPA
+ Primary approach is phone sales
+ Other approaches were less successful
+ 862%
+ 7%
Use qualitative research to understand what happened
Active Listing Growth
Treatment
Control
29. Host Education
Improving listings through outreach
29
+ Initially not launched as an experiment and found positive impact
+ Launched as an experiment and found neutral impact
+ Don’t need market level approach here!
!
30. Some takeaways
Use context to improve operations
30
+ Can investigate heterogeneity in treatment effects with higher N
+ Word of caution: can’t just compare those who were reached
by a call or email to the control (selection bias strikes again)
32. Additional Offline vs. Online Considerations
32
+ Opt-in biases
+ You know you are in an experiment (Hawthorne/John Henry effects)
+ Monetary incentives impact external validity, trade-off take-up rate
+ Takes time to adjust to a change
+ Sample size may be limited by ops capacity
+ Stakeholders may be less data-savvy
+ Real people delivering the experiment!
+ Ethical considerations
!
Always partner with customer support.
!
33. Takeaways
+ Controlled experiments are the way to go if you want to make causal inference
+ Use them to optimize operations!
!
but:
+ Level of randomization - what impact do you want to measure?
+ Cannibalization
+ Compare the right groups - no selection bias
+ Break down results to get the most from the analysis
+ Be practical/ethical - you are dealing with real people here
33