Amazon - Redesigning the Internal Data Collection Tool

Internal Data Tool Development | Generative Research

Project Overview

My Role: UX Researcher (with Product & Design Ownership)
Timeline: August 2021 - February 2022
Team: UX Designer, Engineer, TPMs (Seattle & San Jose)
Research Method: Concept testing, In-depth interviews (IDIs), and post prototype surveys.
My role:
As the sole UX Researcher on this project, I took ownership of both the research direction and early product thinking. I collaborated cross functionally with the UX Designer, Engineers, and TPMs to drive the redesign from field discovery to usability validation.

My responsibilities included:

  • Planning and executing end-to-end research (discovery → testing → validation)

  • Leading field visits and direct observation at testing sites

  • Conducting interviews and running 2 rounds of usability testing

  • Collaborating on wireframes, iterating designs based on feedback

  • Delivering insights and benchmarks to inform product readiness and rollout strategy

This was a high ownership role that blended UX research, product insight, and rapid iteration under tight deadlines.

Due to the nature of Amazon's non-disclosure agreement, I cannot share through image details about my research process on my website. More specific information on my images, visuals, and designs for Amazon can be shared privately in a portfolio presentation upon request by emailing me.

Project Journey & Impact Overview

30%

30%

Reduction in data collection enrollment and setup time

25%

25%

Fewer clicks

95%

95%

Satisfaction rate

Goal

Redesign Amazon’s internal ML Data Collection Tool to reduce the time and friction involved in onboarding participants and collecting video data, improving both usability for data collectors and operational efficiency for the ML model team.

Let's Dive Deeper!

The Problem Statement

The legacy tool used to onboard participants and collect ML training data was error prone, slow, and poorly designed.

  • Enrollment took over 90 minutes

  • Users couldn’t find basic setup options or measurements

  • UI required zooming and manual guessing

  • Equipment setup was physically intensive and inconsistent


One TPM said:

“This tool is actually s***. Someone please help rebuild it.”


And the Director of Ops warned:

“We are definitely not going to meet the June deadline. We are in deep trouble.”


The experience wasn't just frustrating, it was threatening operational goals. We had 6 months to fix it.

Research Process

Before jumping into research, I aligned with the TPM and engineers to fully understand how the current ML Data Collection Tool was being used across different environments. I quickly realized this wasn’t just a software issue, it was an operational bottleneck. Equipment setup was inconsistent, enrollment steps were unclear, and the UI wasn’t responsive across devices. There was also growing pressure to hit tight deadlines, which meant we needed research to surface critical usability issues fast, prioritize fixes, and validate improvements before rollout.

Discovery: Field Study & Site Observations

After aligning with our TPM and engineers, I needed firsthand insight into how the ML Data Collection Tool performed in the field. I prioritized in-person observation to understand environmental friction and UI struggles that remote testing might miss.

Method: Field Study & Observational Research

Why this method: To capture in context usage, environmental barriers, and unspoken frustrations across different data collection setups.

Participants: Data collectors at indoor and outdoor test sites

What happened:

  • I traveled to observe the tool in action: unpacking gear, onboarding participants, and navigating setup flows

  • Documented pain points like squinting, repetitive zooming, lack of clear next steps


Key Findings:

  • Small font size and touch targets caused interaction delays

  • UI lacked hierarchy, labels, and guided navigation

  • Setup ergonomics varied greatly and introduced safety risks


Transition:
• These insights revealed critical UI and physical design problems
• I transitioned into contextual interviews to better understand mental models and workflow logic.

In Depth Interviews (IDIs)

After witnessing usability and ergonomic issues firsthand, I needed to dive deeper into users' cognitive load, why they struggled with navigation, measurement input, and task confidence.

Method: Contextual 1:1 Interviews

Why this method & goal:
To gather qualitative insights on task confusion, expectation mismatches, and pain points that weren’t always visible during live sessions (during data collection).

Who was involved:
• 6 Data Collection Researchers
• Insights shared with UX designer and TPM

Process:
• Conducted remote interviews covering enrollment, measurement, and completion workflows
• Used task walkthroughs to pinpoint friction and get quotes in users’ own words

What came out of it:
• Drop downs had over 100 items with no search functionality
• Users couldn’t tell how to begin setup or validate successful entry
• Metric vs. imperial confusion created delays

Transition:
Interviews revealed a clear need for simplified flows, scalable interaction models, and visual hierarchy. I collaborated with the designer and engineers to develop low fidelity prototypes for fast iteration.

Low Fidelity Prototyping

With pain points validated and design priorities defined, I partnered with our UX Designer to start building and iterating on solutions that could meet field needs while staying within development constraints.


Method: Collaborative Wireframing + MVP Scoping


Why this method & goal:
To prototype ideas quickly and test layout, hierarchy, and functionality logic before investing engineering effort.


Who was involved:
• UX Designer, TPM, and engineers
• I led synthesis and structured collaborative creative sessions


Process:
• Created wireframes that added search, reduced visual noise, and clarified navigation
• Proposed collapsible UI, dark mode, and safer measurement prompts
• Prioritized MVP features based on feasibility and user urgency


What came out of it:
• Clear MVP scope for engineering
• Visual hierarchy and flows mapped to user needs
• Early alignment across product, design, and dev


Transition:
With low-fi wireframes in hand, we needed to validate them through usability testing to ensure real users could complete core workflows with ease.

Usability Testing (2 Rounds)

To validate wireframe direction and identify existing UX friction, I conducted two iterative rounds of usability testing with real users.


Method: Moderated Usability Testing


Why this method & goal:
To measure clarity, task completion, and user confidence while gathering confusion and reaction data.


Who was involved:
• 12 returning users from field sites
• I coordinated logistics, sessions, and synthesis


Process:
• Asked participants to complete full setup flow using new wireframes
• Used think aloud protocols and click tracking
• Iterated between rounds to resolve drop off points and ambiguities


What came out of it:
• Search bar significantly improved dropdown navigation
• Collapsible layout reduced scrolling and cognitive overload
• 95% said the process felt “easier” or “much easier” than before


Transition:
With validated flows and confidence in design performance, I launched benchmark testing and a post survey to quantify gains and secure stakeholder buy in.

Benchmark & Satisfaction Survey

Final testing confirmed usability improvements, but I needed hard numbers to justify rollout and measure success at scale.


Method: Benchmark Comparison + Post-Use Survey


Why this method & goal:
To quantify improvements in setup time, navigation effort, and satisfaction over the legacy tool.


Who was involved:
• 30 data collectors from across teams
• PM and TPM used results to inform rollout


Process:
• Timed task benchmarks across legacy vs. redesigned flows
• Collected survey data on ease, satisfaction, and confidence


What came out of it:
• Setup time reduced by 30% (90 → 40 minutes)
• Navigation clicks reduced by 25%
• 95% user satisfaction with the new tool


Transition:
These results were presented to the PM and engineering team to finalize launch priorities and inform future design ops at scale.

Overall Reflection

My POV

This was one of the most intense and rewarding projects I’ve worked on. With only 6 months, I had to move quickly, scope carefully, and prioritize what mattered most: clarity, communication, and operational usability.


Field research grounded me in real user pain, while close collaboration with TPMs and engineers helped ensure we scoped features that made sense. When a data collector said, “I’m still not sure where to start,” it reminded me that great design is about building trust, especially under pressure.


If I had more time, I would’ve:

  • Run longitudinal studies to track tool adoption

  • Explored data quality impact from better UX

  • Included accessibility testing for vision and physical strain


This project sharpened my ability to lead research under tight deadlines, collaborate across disciplines, and bring a product mindset to usability at scale.

View Other Case Studies

View Other Case Studies