So we run this split test on our ecomm site. 2 sidebars.
Sidebar #1 (control) = All graphics, explains 4 reasons why customers should buy from us, takes up more room, more "flashy." Headline "Let Us Hook You Up!"
Sidebar #2 = Typographic, has 5 reasons why customers should buy from us including a flash widget that shows all positive reviews of our site from ResellerRatings.com. Very clean & professional. More compact. Direct Headline "5 Reasons to Choose Us"
I'm testing using Optimizely.
Split test #1 shows sidebar 2 increase sales by a whopping 100%+, with 96.4% confidence.
Im jumping for joy, running around holding my ass cheeks with both hands.
I think to myself "this is too good to be true."
I talk to my sales rep who tells me that there are a handful of transactions that she placed on behalf of customers that called in. I think to myself "fuck, so much for clean data." There were only a couple of transactions tainted, but in the interest of making a clean, data-based decision, I decide to run the test again, this time tracking more variables such as the amount of people that clicked to chat with us (very important in our space because we can give discounts over chat that we can't give elsewhere).
So I run the same test again, this time filtering out our sales rep's ip. For the first 2 days, sidebar 2 is killing it just like first time. Then sidebar 1 unexpectedly catches up and takes the lead. The test has been running for ~2 weeks now and its showing that sidebar 2 is producing 25% less sales than the control, at a 75% confidence. Also, I'm seeing sidebar 2 perform significantly worse for certain key chat buttons. Overall, sidebar 2 is getting 14.9% less chats, however this is only 85% significant. The only metric that is >99% confident is "engagement" (aka clicks on page), for which sidebar 2 is winning by 7%.
Where I'm at now
I'm honestly not sure what to do right now.
Should I keep letting this test run? If so, for how long? Its already been 3 weeks between the 1st 2 tests.
Should I restart this test using a new tracking platform (e.g. visual website optimizer)? I suppose this could act as a tie breaker.
I'm sure I'm not the first person to experience something like this.
Any wisdom is welcome.
Thanks in advance,
-AP
P.S. Here are some tits for your troubles
Sidebar #1 (control) = All graphics, explains 4 reasons why customers should buy from us, takes up more room, more "flashy." Headline "Let Us Hook You Up!"
Sidebar #2 = Typographic, has 5 reasons why customers should buy from us including a flash widget that shows all positive reviews of our site from ResellerRatings.com. Very clean & professional. More compact. Direct Headline "5 Reasons to Choose Us"
I'm testing using Optimizely.
Split test #1 shows sidebar 2 increase sales by a whopping 100%+, with 96.4% confidence.
Im jumping for joy, running around holding my ass cheeks with both hands.
I think to myself "this is too good to be true."
I talk to my sales rep who tells me that there are a handful of transactions that she placed on behalf of customers that called in. I think to myself "fuck, so much for clean data." There were only a couple of transactions tainted, but in the interest of making a clean, data-based decision, I decide to run the test again, this time tracking more variables such as the amount of people that clicked to chat with us (very important in our space because we can give discounts over chat that we can't give elsewhere).
So I run the same test again, this time filtering out our sales rep's ip. For the first 2 days, sidebar 2 is killing it just like first time. Then sidebar 1 unexpectedly catches up and takes the lead. The test has been running for ~2 weeks now and its showing that sidebar 2 is producing 25% less sales than the control, at a 75% confidence. Also, I'm seeing sidebar 2 perform significantly worse for certain key chat buttons. Overall, sidebar 2 is getting 14.9% less chats, however this is only 85% significant. The only metric that is >99% confident is "engagement" (aka clicks on page), for which sidebar 2 is winning by 7%.
Where I'm at now
I'm honestly not sure what to do right now.
Should I keep letting this test run? If so, for how long? Its already been 3 weeks between the 1st 2 tests.
Should I restart this test using a new tracking platform (e.g. visual website optimizer)? I suppose this could act as a tie breaker.
I'm sure I'm not the first person to experience something like this.
Any wisdom is welcome.
Thanks in advance,
-AP
P.S. Here are some tits for your troubles
