Using Web analytics to improve information architectures

Download Report

Transcript Using Web analytics to improve information architectures

Data Driven Design
Using Web Analytics to
Improve Information
Architectures
Andrea Wiggins
IA Summit 2007
SCHOOL OF INFORMATION
UNIVERSITY OF MICHIGAN
Motivation: What Information
Architects Want to Know

Interviewees said:
– Context for making design decisions
– Validation of heuristic assumptions
– Understand why visitors come to the site & what
they seek
SCHOOL OF INFORMATION
UNIVERSITY OF MICHIGAN
Agenda



Overview for Context
Insert show of hands here! (topic, tools, data)
What is web analytics (WA)? How is it done?
– major WA concepts
– what the data look like



IA questions to answer
Rubinoff’s user experience audit
Some WA measures for heuristic validation
SCHOOL OF INFORMATION
UNIVERSITY OF MICHIGAN
What is web analytics?

Data mining from web traffic logs
– Web server log files
– Page tag logs from client-side data collection (end
up in server logs)
– Cookies to identify “unique visitors”

What for?
–
–
–
–
–
Proving web site value (ROI)
Marketing campaign evaluation
Executive decision making - markets & products
Web site design parameters
More…
SCHOOL OF INFORMATION
UNIVERSITY OF MICHIGAN
How do you do it?

Vendor analysis solutions
• Hosted ASP
– Currently most popular model
– Provides traffic stats “on-demand”
• Software
– Runs on dedicated servers
– Scalability: requires significant data storage space and
data maintenance
• Costs
– Starts at FREE for Google Analytics and goes way, way up
– Large organizations spend $50K/yr and up

Open source: not a robust option
SCHOOL OF INFORMATION
UNIVERSITY OF MICHIGAN
Very Quick Major Concepts




Sessionizing (cookie > IP & UA)
Hits: all server requests
Pageviews: all server requests for page
filetypes, variously defined
Visits & Visitors: stronger measures from
sessionizing, sensitive to time periods
SCHOOL OF INFORMATION
UNIVERSITY OF MICHIGAN
Sample Logs
#Software: Microsoft Internet Information Services 6.0
#Version: 1.0
#Date: 2005-08-01 00:00:35
#Fields: date time cs-method cs-uri-stem cs-username c-ip cs-version cs(User-Agent) cs(Referer) scstatus sc-bytes
2005-08-01 00:10:05 GET /index.htm - 216.xx.76.7 HTTP/1.1
Mozilla/4.0+(compatible;+MSIE+6.0;+Windows+98)
http://search.yahoo.com/search?p=purple+rose+theater&sm=Yahoo%21+Search&fr=FP-tab-webt-280&toggle=1&cop=&ei=UTF-8 200 13099
2005-08-01 00:10:29 GET /current.html - 216.xx.76.7 HTTP/1.1
Mozilla/4.0+(compatible;+MSIE+6.0;+Windows+98) http://www.purplerosetheatre.org/ 200 17985
2005-08-01 00:11:24 GET /tickets.html - 216.xx.76.7 HTTP/1.1
Mozilla/4.0+(compatible;+MSIE+6.0;+Windows+98) http://www.purplerosetheatre.org/current.html
200 15689
2005-08-01 00:18:06 GET /index.htm - 152.xxx.100.11 HTTP/1.0
Mozilla/4.0+(compatible;+MSIE+6.0;+AOL+9.0;+Windows+NT+5.1;+SV1;+.NET+CLR+1.1.4322)
http://www.guide2detroit.com/arts/stage-calendar.shtml 304 300
2005-08-01 00:20:18 GET /index.htm - 68.xx.117.55 HTTP/1.1
Mozilla/4.0+(compatible;+MSIE+6.0;+Windows+NT+5.1;+SV1;+.NET+CLR+1.1.4322)
http://www.google.com/search?hl=en&q=purple+rose+theatre 200 13099
2005-08-01 00:20:21 GET /classes.html - 68.xx.117.55 HTTP/1.1
Mozilla/4.0+(compatible;+MSIE+6.0;+Windows+NT+5.1;+SV1;+.NET+CLR+1.1.4322)
http://www.purplerosetheatre.org/ 200 15296
SCHOOL OF INFORMATION
UNIVERSITY OF MICHIGAN
Spiders


2005-08-01 00:49:32 GET /robots.txt - 68.xxx.251.159 HTTP/1.0
Mozilla/5.0+(compatible;+Yahoo!+Slurp;+http://help.yahoo.com/help/us/
ysearch/slurp) - 200 319
2005-08-01 00:49:32 GET /plays/completing_dahlia.html 68.xxx.249.67 HTTP/1.0
Mozilla/5.0+(compatible;+Yahoo!+Slurp;+http://help.yahoo.com/help/us/
ysearch/slurp) - 200 3507
SCHOOL OF INFORMATION
UNIVERSITY OF MICHIGAN
A Few Good Metrics

Information Architects want to know:
– Confirmation of heuristics
• Do users leave at first glance of this awful page?
• Where do they click?
• What position on the screen or layout produces the most
clicks for the same content?
• Do the users “pogo-stick” back and forth between pages?
What are they comparing?
– Ambient findability measures
• At what hierarchy depth do visitors enter the site? How
do they get in on deep pages?
• Do they ever see the home page?
• Can they find their way to where we want them to go?
SCHOOL OF INFORMATION
UNIVERSITY OF MICHIGAN
Searching for IA Answers

On-site search behaviors
–
–
–
–
–
–
How many searches do users make?
Do users refine their search results?
What type of queries do users make?
How often are search results the last page?
From what pages are searches initiated?
Do the search terms have context in the page from
which the search is initiated?
– Why are users querying about chimpanzees?!?
SCHOOL OF INFORMATION
UNIVERSITY OF MICHIGAN
What IAs Want

Good navigation and content make the online
world go ‘round
– Where in a process do users leave? Where do
they go? Do they re-enter the process?
– How do users move through the site? Is there a
better route?
– What pages don’t get visited? What pages get
unexpectedly high visits?
– What prompts conversion?
– Where do search engine spiders go in the site? Is
the best content being indexed?
SCHOOL OF INFORMATION
UNIVERSITY OF MICHIGAN
Everybody Loves Rubinoff

UX audit quantifies subjective measures
– Offers structure for comparing properties of the
site
– Completely customizable, use strategically

In a perfect world:
– Analyst & IA work together to set key performance
indicators (KPI) and measurable heuristics
– Each independently evaluates the site on the
same points and compare the IA’s heuristics to
user data for validation
– They set before-and-after measures to prove
value for the entire project
SCHOOL OF INFORMATION
UNIVERSITY OF MICHIGAN
Rubinoff’s Four Categories

Using a sample of statements from
Rubinoff’s model:
1. Branding
a) Engaging, memorable brand experience
b) Value of multimedia & graphics
2. Functionality
a) Server response time & technical errors
b) Security & privacy practices
3. Usability
a) Error prevention & recovery
b) Supporting user goals & tasks
4. Content
a) Navigation & site structure
b) Search & referrals
SCHOOL OF INFORMATION
UNIVERSITY OF MICHIGAN
1a: Branding
Memorable & Engaging
Experiences






Ratio of new to returning visitors is key; set
target KPI specific to site business goals
Track trends over time and in relation to
cross-channel marketing
Median visit length in minutes
Average visit length in pages viewed
Depth, breadth of visits
Segment new and returning visitors to
examine visit trends for different audiences
SCHOOL OF INFORMATION
UNIVERSITY OF MICHIGAN
1b: Branding
Value of Multimedia & Graphics




Flash & AJAX require deciding upon what to
measure, programming appropriate data
collection, and configuring analysis tools
Plan to include measures when designing
multimedia applications to prove value
Compare clickthrough rates for clickable
graphics to rates for standard navigation links
Great tools like Crazy Egg’s heatmap - easy!
(also relevant to navigation, of course)
SCHOOL OF INFORMATION
UNIVERSITY OF MICHIGAN
Crazy Egg Heatmap Example
SCHOOL OF INFORMATION
UNIVERSITY OF MICHIGAN
Crazy Egg Overlay Example
SCHOOL OF INFORMATION
UNIVERSITY OF MICHIGAN
Crazy Egg List Example
SCHOOL OF INFORMATION
UNIVERSITY OF MICHIGAN
2a: Functionality
Response Time & Technical Errors



Response time is a default log field, easy to
measure
Check at peak load time to make sure site is
responding quickly enough
Monitor the rate of 500 (server) errors: this
should be an extremely low number
SCHOOL OF INFORMATION
UNIVERSITY OF MICHIGAN
2b: Functionality
Security & Privacy Practices

A matter of design for measurement, not
measurement of design: considerations for
designing a site that will be measured
– Privacy best practices:
• Give a short, accurate, easy to understand privacy
statement and stand by your word
• True first-party cookie
– Security best practices: (from an IA/analytic POV)
• SSL encryption on any transactional forms: lead
generation, ecommerce, surveys
• Secure file transfer for & restricted access to raw web
analytic data; password restrictions at minimum
SCHOOL OF INFORMATION
UNIVERSITY OF MICHIGAN
3a: Usability
Error Prevention & Recovery



Percentage of visits experiencing 404 and
500 errors: errors should be < 0.5% of all hits
Percentage of visits including an error, that
end with an error - frustrated into leaving
Where do 404 errors occur?
– Use to build a redirect page list to ensure
(temporary) continuity of service to bookmarked
URLs
– Path/navigation analysis: how did users arrive at
404? What did they do after?

User errors: identify problems & re-enact or
test
SCHOOL OF INFORMATION
UNIVERSITY OF MICHIGAN
3b: Usability
Supporting User Goals & Tasks

Scenario/conversion analysis
– Define tasks and procedures supporting user
goals
– Examine completion rates, step by step, intervals
& overall
• A to B, B to C, C to D; A to C, B to D; A to D
– Look at leakage points
• Where did they go when they left the process? Did they
come back later?
– Shopping cart analysis
• Keep in mind that users shop online for offline purchases
• Do behaviors suggest a need for a tool like a shipping
calculator or product comparison?
– Online form completion
SCHOOL OF INFORMATION
UNIVERSITY OF MICHIGAN
4a: Content
Navigation & Site Structure

Pogo-sticking: jumping back & forth between
content or hierarchy levels (what about tabs?)
– Need a comparison tool, can’t identify product: not
enough detail at the right level of site hierarchy or
step of the purchase decision process


Compare page-level traffic statistics for larger
trends, broad navigation analysis: the usual
#s
Path analysis on navigation tools (by type) to
pinpoint navigation and labeling problems
– Extensive use of supplemental navigation may
indicate need for updates to global navigation
SCHOOL OF INFORMATION
UNIVERSITY OF MICHIGAN
4b: Content
Mining Search & Referrals

Popularity = value? What about findability? If
it’s not findable, it probably won’t be popular.
– Compare the content’s value (against similar
content) with proportions of returning visitors,
average page viewing length, external referrals especially search referrals

Search log analysis: what do your users
value?
– Does user query language match site contents?
Are users searching for panties when you’re
selling pants?
SCHOOL OF INFORMATION
UNIVERSITY OF MICHIGAN
Validate the Match Between
the Site & the Real World

More ways to use search log analysis:
– Does user vocabulary match site vocabulary?
– Do different audiences have different
vocabularies, and does the site support them
equally?
– Brand measurement returns
• product and industry terminology usage
• “accuracy” of brand queries: spelling, inclusion of
competitor’s brands, advertising slogans
– Did users find what they expected? How many
visits end on search results? Null results are
revealing.
SCHOOL OF INFORMATION
UNIVERSITY OF MICHIGAN
Language Validation
SCHOOL OF INFORMATION
UNIVERSITY OF MICHIGAN
Conclusions





Not much out there in the academic literature
on using web analytics (hopefully to change!)
WA data is flawed and tough to handle, but
ultimately pays off in developing holistic
understanding of user behavior
Best-suited to case studies
WA is ripe for adoption into formal usability
frameworks, particularly for persona design
and determining design parameters
Best used iteratively: beginning, middle, end,
annual follow-up…
SCHOOL OF INFORMATION
UNIVERSITY OF MICHIGAN
Thanks! Questions?
SCHOOL OF INFORMATION
UNIVERSITY OF MICHIGAN