A Mechanism for Mining Top Impacted Variants used for AB Testing

advertisement
International Journal of Application or Innovation in Engineering & Management (IJAIEM)
Web Site: www.ijaiem.org Email: editor@ijaiem.org, editorijaiem@gmail.com
Volume 3, Issue 2, February 2014
ISSN 2319 - 4847
A Mechanism for Mining Top Impacted
Segments in Web Analytics based on the
Variants used for AB Testing
Mubbashir Nazir1, Umar Mir2 and Anuj Yadav3
1,2
M.Tech Scholar Department of Computer Science & Information Technology DIT University Dehradun
Asiistant Professor Department of Computer Science & Information Technology DIT University Dehradun
2
Abstract
In this paper, a new approach for Impact Analysis of AB Testing is presented that aids in mining Top Impacted Behavioral
Segments in Web Analytics, based on the variants used for AB Testing. First we propose a method for calculating Segment
Exposure Ratio (SER), to determine the user exposure ratio for each segment. Then we calculate the Impact score for each
segment based on the experiments and the corresponding versions to which the segments are exposed to, facilitating the discovery
of top impacted segments. Finally we calculate the interestingness score of the different variants of an experiment based on the
impact score of each segment, providing the interesting patterns if any among the different version of an experiment.
Keywords: AB Testing, Web Analytics, Behavioral Segments, Segment Exposure Ratio (SER), Interestingness, Impact
Score, Interesting Patterns.
1. INTRODUCTION
With the advancements in technology and the advent of internet, the domain of business extended to web based initiatives
such as E-commerce. For any E-commerce site the success largely depends upon the sales and conversions rates. The
design and the usability of the E-commerce site and software substantially contribute to these goals. In order to enhance
the design effectiveness of the E-commerce site, users should be exposed to different variants of the site design and each
variant should be tested for effectiveness in terms of metrics such as number of hits, conversion rates, landing pages etc.
In online marketing such approach is referred to as AB Testing. AB Testing (also termed as Split Testing) is the method
of pitting up two variants of a web page against each other to determine which one performs better. It is a method to
validate that any new design or change to an element on a webpage is improving the conversion rates before making that
change in the site code. In this method the two variants of a web page are exposed to randomly selected sample of users,
and the impact of the two variations on key success measures such as conversion, is statistically analyzed
Testing takes the guesswork out of website optimization and enables the data-backed decisions that shift business
conversations from “We Think” to “We Know”. By measuring the impact that changes have on metrics such as
purchases, add to cart etc.it can be ensured that every change produces the positive results.
Fig 1: Example of AB testing using two design versions (A, B) on randomly distributed visitors
In the Proposed approach we use AB Testing in Web Analytics (a tool that provides objective tracking, collection,
measurement, reporting, and analysis of quantitative Internet data to optimize websites and web marketing initiatives [1])
to perform Impact Analysis of segments in order to determine the Top Impacted segments for each of the design variants
used, thereby enhancing the design effectiveness of the site and also aid in increasing the sales and conversion rates.
2. AB TESTING FRAMEWORK
A user is exposed to multiple tests, assigned form a pool of running tests. Each test has multiple versions which are
mutually exclusive i.e. only one version of a test can be assigned to a user. Each test can ran either exclusively or shared.
Volume 3, Issue 2, February 2014
Page 66
International Journal of Application or Innovation in Engineering & Management (IJAIEM)
Web Site: www.ijaiem.org Email: editor@ijaiem.org, editorijaiem@gmail.com
Volume 3, Issue 2, February 2014
ISSN 2319 - 4847
Based on the success measures such as conversion rates one of the winning versions are assigned. If there is no winning
version then one version from running versions is randomly assigned. Exclusiveness is used to disassociate any biasness
induced by the experimentation running.
Creation of experiment involves two steps. One which will create an Experiment Group (or simply experiment) and the
other will create a new version by providing the experiment Id. All experiments and versions are induced as first party
cookies, thus enabling us to track user behavior with each test the user is exposed to. A new version is allotted when,
 No cookie parameter is passed i.e. user is new
 Inactivity exceeds TTL value of current allotted experiment version
 Currently assigned experiment is paused or closed
Fig 2: Scenario depicted AB Testing framework
Fig 3: AB testing result set based on visits, page views, unique clicks, and avgclicks
Volume 3, Issue 2, February 2014
Page 67
International Journal of Application or Innovation in Engineering & Management (IJAIEM)
Web Site: www.ijaiem.org Email: editor@ijaiem.org, editorijaiem@gmail.com
Volume 3, Issue 2, February 2014
ISSN 2319 - 4847
3. THE PROPOSED APPROACH
In order to mine the top impacted segments we identified various measures which on processing help in determining the
top impacted segments based on AB Testing variants. The measures are listed as below:
 Total user exposure for an experiment(TUE)
 User Exposure per segment for an experiment(USE)
 Segment Exposure Ratio (SER)
 Total user exposure for a version (TUV)
 User exposure per segment for a version (USV)
 Segment Impact Score (SIS)
 Interestingness Score (IS)
Fig 4: Flow Diagram for doing Impact Analysis on AB Testing
3.1 Algorithm for calculating Segment Exposure Ratio (SER)
Stage 1: Generate Clickstream
Stage 2: Filter Clickstream by IP and check if segment record is not null
Stage 3: Extract records having unique experiment Id
Stage 4: Calculate total users exposed (TUE) for each unique Experiment as:
Where DV refers to distinct users visiting the site identified by the unique first party cookie and unique experiment Id
Stage 5: Calculate User exposure for each segment (USE) for each unique Experiment as:
Where DSV refers to distinct users visiting the site identified by the unique first party cookie, unique experiment Id and
unique segment name.
Stage 6: Calculate Segment Exposure Ratio (SER) as:
SER gives traffic distribution ratio for each behavioral segment with unique experiments
3.2 Algorithm for calculating Segment Impact Score
Stage 1: Generate Clickstream
Stage 2: Filter Clickstream by IP and segment record is not null
Volume 3, Issue 2, February 2014
Page 68
International Journal of Application or Innovation in Engineering & Management (IJAIEM)
Web Site: www.ijaiem.org Email: editor@ijaiem.org, editorijaiem@gmail.com
Volume 3, Issue 2, February 2014
ISSN 2319 - 4847
Stage 3: Extract records having unique version Id for each AB experiment
Stage 4: Calculate total users exposed for a version (TUV) for each unique Experiment as:
Where DVV refers to distinct users visiting the site identified by the unique first party cookie, unique experiment Id and
unique version Id
Stage 5: Calculate User exposure for each segment (USV) for each version having unique Experiment as:
Where DSVV refers to distinct users visiting the site identified by the unique first party cookie, unique experiment Id
unique version Id and unique segment name.
Stage 6: Calculate Segment Impact Score (SIS) as:
Stage 7: Normalize the SIS score for each segment
SIS gives traffic distribution ratio for each behavioral segment exposed to different AB experiment variants. The segment
with the highest Impact Score is tagged as the Top Impacted Segment exposed to a unique experiment with different
variants
3.3 Calculating Interestingness Score
Interestingness Score helps in mining the interesting patterns among different versions of a specific AB experiment.
Interestingness score is calculated by taking the standard deviation of the Segment Impact Score for each version of a
specified experiment. The greater the number of positive standard deviations the more interesting the design patterns are.
The interestingness score is calculated as:
Where N is the number of unique version of a specific experiment
Fig 5: IS, SER, SIS values stored in read-only database (Splout Sql)
Volume 3, Issue 2, February 2014
Page 69
International Journal of Application or Innovation in Engineering & Management (IJAIEM)
Web Site: www.ijaiem.org Email: editor@ijaiem.org, editorijaiem@gmail.com
Volume 3, Issue 2, February 2014
ISSN 2319 - 4847
4. CONCLUSION
The research presented, will thus provide Impact Analysis of AB experiment variants leading to the discovery of nontrivial, valid, and novel design patterns of E-commerce site that play an important role in driving the growth of an
organization. The proposed work has been successfully tested on an E-commerce site Artsya (www.artsya.com) and it was
observed that the system performs better, providing programmable actions by performing deeper analysis of the available
metrics that can be used as feedback for personalization.
References
[1]
[2]
[3]
[4]
Web Analytics Association (http://www.webanalyticsassociation.org)
Optimzely: AB Testing Simplified [online Article] Available: https://www.optimizely.com/ab-testing
The complete guide to AB Testing [online Article] Available: https://visualwebsiteoptimizer.com/ab-testing
The AB Test : Inside the technology that’s changing the rules of Business [online Article] Available:
http://www.wired.com/business/2012/04/ff_abtesting
[5] Advanced Ab Testing that you should know [online Article] Available: http://online-behavior.com/testing/advancedab-testing-tactics-1356
[6] How to build Strong AB Test Plans that gets better Results [online Article] Available: http://conversionxl.com/howto-build-a-strong-ab-testing-plan-that-gets-results
[7] Occam’s razor: Experimentation and Testing a Premier Avinash Kaushik [Blog] Available:
http://www.kaushik.net/avinash/experimentation-and-testing-a-primer
Volume 3, Issue 2, February 2014
Page 70
International Journal of Application or Innovation in Engineering & Management (IJAIEM)
Web Site: www.ijaiem.org Email: editor@ijaiem.org, editorijaiem@gmail.com
Volume 3, Issue 2, February 2014
ISSN 2319 - 4847
[8] Buyer
Behaviour:
A
Basis
For
Market
Segmentation
(online
Article)
Available:
http://www.topentrepreneur.co.za/Pages/Marketing/Consumer%20Behaviour%20A%20Basis%20for%20Segmentatio
n.html
[9] Retail Analytics – Emmit Cox (Book)
[10] Web Analytics An Hour A Day - Avinash Kaushik (Book)
[11] Audience
Engagement
:
Google
Analytics
[online
Article]
Available:
https://support.google.com/analytics/answer/1144430?hl=en&ref_topic=1007027
[12] New
vs.
Returning
User
:
Google
Analytics
[online
Article]
Available
:
https://support.google.com/analytics/answer/1144430?hl=en&ref_topic=1007027
[13] Frequency
and
Recency
:
Google
Analytics
[online
Article]
Available:
https://support.google.com/analytics/answer/1144430?hl=en&ref_topic=1007027
[14] Web
Analytics
:
Behavioral
Segmentation
[online
Article]
Available:
http://semphonic.blogs.com/semangel/2008/06/web-analytics-.html
[15] Segmentation in Web Analytics A Fundamental approach by Jacques Warren [White paper]
[16] Infosys- Web Analytics for enterprise growth [White Paper]
[17] Segmenting consumers by E-Shopping Behavior and online purchase intention : Ruiz mafe Carla, Lassala Lavare
[18] Mubbashir Nazir and Umar Mir, “Actionable Web Analytics: Customer Segmentation Approach Based on
Behavioral Patterns in E-Commerce Industry”, International Journal of Emerging Trends & Technology in Computer
Science (IJETTCS) ,pp 227-229, Vol 2, Issue 6,November-December 2013.
Volume 3, Issue 2, February 2014
Page 71
Download