Digitala Vetenskapliga Arkivet

Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
IT’S IN THE DATA 2: A study on how effective design of a digital product’s user onboarding experience can increase user retention
Linköping University, Department of Computer and Information Science.
2021 (English)Independent thesis Advanced level (degree of Master (Two Years)), 20 credits / 30 HE creditsStudent thesis
Abstract [en]

User retention is a key factor for Software as a Service (SaaS) companies to ensure long-term growth and profitability. One area which can have a lasting impact on a digital product’s user retention is its user onboarding experience, that is, the methods and elements that guide new users to become familiar with the product and activate them to become fully registered users.

Within the area of user onboarding, multiple authors discuss “best practice” design patterns which are stated to positively influence the user retention of new users. However, none of the sources reviewed showcase any statistically significant proof of this claim. Thus, the objective of this study was to:

Design and implement a set of commonly applied design patterns within a digital product’s user onboarding experience and evaluate their effects on user retention

Through A/B testing on the SaaS product GetAccept, the following two design patterns were evaluated:

  1. Reduce friction – reducing the number of barriers and steps for a new user when first using a digital product; and
  2. Monitor progress – monitoring and clearly showcasing the progress of a new user’s journey when first using a digital product.

The retention metric used to evaluate the two design patterns was first week user retention, defined as the share of customers who after signing up, sign in again at least once within one week. This was tested by randomly assigning new users into different groups: groups that did receive changes related to the design patterns, and one group did not receive any changes. By then comparing the first week user retention data between the groups using Fisher’s exact test, the conclusion could be drawn that with statistical significance, both of the evaluated design patterns positively influenced user retention for GetAccept.

Furthermore, due to the generalizable nature of GetAccept’s product and the aspects evaluated, this conclusion should also be applicable to other companies and digital products with similar characteristics, and the method used to evaluate the impact of implementing the design patterns should be applicable for evaluating other design patterns and/or changes in digital products.

However, as the method used for data collection in the study could not ensure full validity of it, the study could and should be repeated with the same design patterns on another digital product and set of users in order to strengthen the reliability of the conclusions drawn.

Place, publisher, year, edition, pages
2021. , p. 40
Keywords [en]
saas, software as a service, software-as-a-service, data-driven, data driven, analytics, A/B testing, A/B-testing, AB testing, AB-testing, bucket testing, split-run testing, web, web development, application, design, design pattern, programming, software, Fisher's exact test, Fisher exact test, growth, onboarding, onboarding experience, retention, user retention, churn, digital product, digital, statistics, user onboarding, customer, profitability
National Category
Computer and Information Sciences Probability Theory and Statistics Human Computer Interaction Software Engineering Business Administration
Identifiers
URN: urn:nbn:se:liu:diva-177591ISRN: LIU-IDA/LITH-EX-A--21/055--SEOAI: oai:DiVA.org:liu-177591DiVA, id: diva2:1575295
External cooperation
GetAccept AB
Subject / course
Computer Engineering
Presentation
2021-06-18, 13:15 (English)
Supervisors
Examiners
Available from: 2021-07-01 Created: 2021-06-29 Last updated: 2022-03-02Bibliographically approved

Open Access in DiVA

fulltext(1457 kB)232 downloads
File information
File name FULLTEXT01.pdfFile size 1457 kBChecksum SHA-512
14310b13a08291a7bbddbfbb91aec05c71a01a6e29eb8bb7d8bfb1a78bc9415100443647ec11cd8420687694857d57454a5919452d93c84b4fbba0c6822877ee
Type fulltextMimetype application/pdf

By organisation
Department of Computer and Information Science
Computer and Information SciencesProbability Theory and StatisticsHuman Computer InteractionSoftware EngineeringBusiness Administration

Search outside of DiVA

GoogleGoogle Scholar
Total: 232 downloads
The number of downloads is the sum of all downloads of full texts. It may include eg previous versions that are now no longer available

urn-nbn

Altmetric score

urn-nbn
Total: 2093 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf