Shedd Aquarium, as a client of Pratt DX Center, wants to know the potential problem with their ticketing purchase pathway. Our team conducted 16 eye-tracking tests and provided 3 insightful recommendations on the navigation, pricing, and add-on issues.
Info
- Timeline: Feb 2025 - May 2025
- Team Kelp Desk: Conor M, Hridya N, Iris S(me), Saskia S.
- My Role: Eye-tracking testing, Data collection & analysis
- My takeaways: Eye-tracking study, Quantitative + Qualitative data analysis, Insightful recommendations that value business impact
Define the Research Question
“How might we make a better Shedd Aquarium's ticket purchasing process for individual and family visitors to support their visits?”
Deconstruct the Question
Still, the question was too vague for research.
So we looked into their ticket purchasing process and thought about the possible problems the users have, talked to the senior designer at Shedd, then deconstructed the questions into 4 parts:
-
Visual design: How does the use of visual indicators impact the task completion?
-
Content clarity: Do users easily understand and differentiate ticket categories?
-
Mobile Experience: How do mobile users interact with the system compared to desktop?
-
Process Efficiency: How does the number and sequence of steps impact decision-making?
Having the specific questions in mind, we made the interview script and post-interview questionnaire, and started the test recruitment process.
What is Our Eye-Tracking Test About? 👀
- A simple task: “Purchase a ticket to visit the Shedd Aquarium for next week. And additionally, see penguins in 4D!”
- Interview questions & Post-interview questionnaire: Participants’ feelings on clarity, design, add-ons, and check-out process (follow-up if possible)
- Participants: 16 participants in total, 4 for each member
We also spent time setting up the Tobii Lab.
How We Find Out the Insights and Prioritize Them
We made a rainbow spreadsheet to better organize the data and prioritize our major issues.
And we picked the problems that lack color, compared to the eye-tracking heat maps, fixation rates, and prioritized the 4 major issues.
Top 4 Issues & Recommendations
1. Add-on: High cognitive loads on understanding the different ticket options cause business loss
Severity level: 🟡 Mid
Emphasis level: 🔴 High
In addition, the add-on page contains too much text, making users get lost or overlook some of the options, which can cause business loss.
High cognitive load on reading the content
An example of failure in finding the add-on because of not scrolling down
💡Recommendation: Rename and combine the general ticket page and add-on page
-
Rename “Must See Tour” as “General Tickets with Exclusive Tour” to let the visitors know what’s the difference between general tickets and tour tickets
-
Combine the General Ticket page and the Add-on page, shorten the description, and use accordions to make the users browse all the options at one glance
- Add a comparison sheet or short explanation on the different types of tickets
2. Pricing: The pricing info appears too late in the flow
Severity level: 🔴 High
💡Recommendation:
- Redesign the visual hierarchy on the time slot page.
-
A/B Testing on the place of pricing
Hypothesis: If dynamic pricing is visible earlier in the flow, fewer visitors will be entering the flow to price check.
Key Metrics: Reduction in funnel entry from “Plan Your Visit” page
Secondary Metrics: Drop-off from calendar page, Average session duration for non-IL users, Price calendar interactions
3. Calendar: Users got confused on the calendar design and the price difference shown on the calendar
Severity level: 🟡 Mid
💡Recommendation:
-
Minimize content to reduce processing time
-
Dynamic pricing on the time slot page instead of the calendar page
4. Navigation: Users lack of freedom to go back and forth
Severity level: 🟡 Mid
Users felt a high cognitive load looking at the unclear visual designs and trying to go back and forth
💡Recommendation:
-
Consistent placement and color on the buttons
-
Add breadcrumbs to enhance clarity and user freedom
Reflection
-
Eye-tacking Testing method + Tobii lab operation
- Qualitative + Quantitative data analysis
- Good amount of testing + Thorough data analysis
- Insightful recommendations that value the business impact
- We cared about the same thing the client was focusing on (Add-on!)
- Good presentation
- More visual mockup on recommendations
- More detailed analysis on the quantitative data (friction point)