Testing the customer experience of Amazon's Prime Day

Posted on July 17, 2015
1 min read

Share

Prime Day was Amazon's global shopping event on July 15, 2015, that offered more deals than Black Friday, exclusively for Prime members. We decided to run a study to see how users would respond to Prime Day, and also to see if Amazon's design made sense to them or not. Here are three things we observed:

Observation #1: Lack of relevancy

Users were irritated by the random mashup of deals into big lists, and their inability to filter by categories that were personally relevant to them. Members—who theoretically use the site more often and cultivate wishlists, etc.—tended to be more irritated than non-members.

Observation #2: Unclear page elements

The problem above was amplified by the fact that once they left the big Prime Day lists, it was difficult to tell what did and didn't qualify for Prime Day. Users relied on the wrong page elements to determine if something was part of Prime Day. They either used:

  • The "Prime" icon, which just indicated if the product qualified for Prime 2-day shipping, or
  • The discounted price information in red, which, if it was a steep discount, indicated to users that it must be part of Prime Day

Observation #3: Social proof in action

Although users didn't respond particularly well to the Prime Day clock, the "% claimed" indicators, and the "Deal Over" information, many of them admitted that the "% claimed" indicator could influence their purchasing decision. Many users reported that they'd be less likely to purchase something with low % claimed (because it meant it was a bad deal), or they'd be more likely to purchase something with a high % claimed (because the deal must be incredible).

 

In this Article

    Read more

    • Learn how to measure campaign impact over time, track performance beyond KPIs, and uncover long-term marketing ROI with smarter testing and insights.

      Blog

      Your campaign didn’t end at launch—here’s how to measure its real impact

      Your campaign didn’t fail—you just stopped measuring too soon. Marketers love a clean ending...
    • Ranjitha Kumar shares how to use AI intentionally, avoid “AI fairy dust,” and balance automation with human insight in this Bloomberg interview.

      Blog

      Why “AI Fairy Dust” Isn’t a Strategy: Key Takeaways from Bloomberg Intelligence

      In a recent Bloomberg Intelligence interview, UserTesting’s Chief Scientist, Ranjitha Kumar, shared a grounded...
    • AI personalization: where it works and where it fails

      Blog

      AI personalization works, until it doesn’t— the human insight gap CMOs can’t afford to ignore

      AI personalization is redefining what customers expect from every interaction. From personalized recommendations to...