Ok-
So I'm testing 14 images, with 4 different ad texts = 56 ad copies
at 56K impressions in my campaign, each ad has only gotten 1K impressions, which is not enough to actually eliminate any individual ads.
However, if I aggregate my numbers by ad text i find that
Text 1 = 9 Clicks / 14k Impression
Text 2 = 6 Click / 14k
Text 3 = 4 Click / 14k
Text 4 = 1 Click / 14k
So given those aggregates, Ad Text 4 loser to Text 1, by a 96-98% confidence interval.
Q1: Is it fair game to slash all 14 ads with Ad Text 4 then?
I'm pretty sure this kind of aggregation wouldn't fly in ultra sensitive studies (i.e. AIDS medication testing), but for more pragmatic and time sensitive issues like marketing, is it fair game?
Q2: By the same logic, can I weed out certain images by aggregating their performance across all 3 texts?
I'm curious to hear what the prevailing wisdom is regarding this.
Finally, In a campaign, I've heard of people starting with a HUGE amount of ad copies in the first round (100-5000), but then work to narrow those down for further testing.
Q3: Do you work to narrow it down to 1 best ad copy or ~10?
My problem has been that with so many ad copies, it gets down to about 10. Even with 10K impressions or so, some copies will easily have 2x the amount of clicks as others, but thats not statistically significant. I'm really tempted to just say "fuck it" and optimize around the winning ad, but my OCD-self says "Don't do that. You know that a decision based on shit data is a shit decision."
OK - I shut up & listen now. thx
So I'm testing 14 images, with 4 different ad texts = 56 ad copies
at 56K impressions in my campaign, each ad has only gotten 1K impressions, which is not enough to actually eliminate any individual ads.
However, if I aggregate my numbers by ad text i find that
Text 1 = 9 Clicks / 14k Impression
Text 2 = 6 Click / 14k
Text 3 = 4 Click / 14k
Text 4 = 1 Click / 14k
So given those aggregates, Ad Text 4 loser to Text 1, by a 96-98% confidence interval.
Q1: Is it fair game to slash all 14 ads with Ad Text 4 then?
I'm pretty sure this kind of aggregation wouldn't fly in ultra sensitive studies (i.e. AIDS medication testing), but for more pragmatic and time sensitive issues like marketing, is it fair game?
Q2: By the same logic, can I weed out certain images by aggregating their performance across all 3 texts?
I'm curious to hear what the prevailing wisdom is regarding this.
Finally, In a campaign, I've heard of people starting with a HUGE amount of ad copies in the first round (100-5000), but then work to narrow those down for further testing.
Q3: Do you work to narrow it down to 1 best ad copy or ~10?
My problem has been that with so many ad copies, it gets down to about 10. Even with 10K impressions or so, some copies will easily have 2x the amount of clicks as others, but thats not statistically significant. I'm really tempted to just say "fuck it" and optimize around the winning ad, but my OCD-self says "Don't do that. You know that a decision based on shit data is a shit decision."
OK - I shut up & listen now. thx