User Testing Process and Results

According to Usability.gov, Usability Testing is a research technique used to evaluate a product or service. The tests have been performed with participants that are a representation of the target audience. Each participant tries to perform the tasks while an observer watches, listens and writes down notes about the test.

It is important to note that the focus of the Usability testing is to observe what people do instead of what they say.

The relevant sections of the checklist suggested by Loranger (2016) were followed for the Usability testing.

The goal of this research was to test the prototype and find usability problems in the screens that were not detected during the self-evaluations and the guerilla testing. Additionally, the aim was to gather behavioural data about how the participants felt about the proposed method of logging food via a chatbot.

The tests were performed in different locations. According to Nielsen (2012), it almost doesn’t matter where you conduct user testing. Therefore a travelling usability lab was set up.

The studies were moderated so the observations could provide richer design insights and opportunities to probe and ask for clarification. Finally, the tests were done with the participants and the observers in the same room to facilitate the detection of subtle cues.

Nielsen (1993) affirmed that with “5 users, you almost get close to user testing’s maximum benefit-cost ratio”. Therefore the tests were performed with five users in the first iteration and six users for the second iteration.

Participants were picked by identifying people that were within the results of the initial demographics research and people that matched the User Persona built previously.

Users were given specific tasks with a unique endpoint. The task was to log what they had for breakfast in 3 different ways: by text, by photo and by voice. The breakfast was set previously as 250 grams of avocado and 250 grams of toast (Fig. 2) to enable a timing comparison between different entry methods.

Figure 2. Avocado and toast picture that was placed in front of the participants.

A test pilot was run, and a few inconsistencies were found in the information sheet given to the participants.

After the tasks were performed, users were given a SUS questionnaire to subjectively measure the perceived usability of a product, followed by a short semi-structured interview before discharging the participants. 

A test plan was put together (Annex 11) to work as guidelines for the observers.

Figure 3. The initial screen of the prototype.

First test

The screens were generated using Sketch and uploaded to InVision to display the prototypes in the devices used by the participants. The sessions were recorded in different ways, some tests were recorded with QuickTime Player, and other tests were recorded with a camera phone pointing at the participant’s hands and screen.


Video 1. Usability test performed with the user using the first prototype.

A total of six participants were included in this test, but one participant could not complete the tasks because of the issues with the prototype. Consequently, the data regarding this particular participant was discarded.

After the tests were performed, a list of issues was compiled in a table (Fig. 4), and a proposed solution offered to each problem detected.

The videos and notes can be visualised in Annex 12 and Annex 13, respectively.

IssueSolution
Task unclear settingMaybe tell users they are going to experience a bot integration. But not sure this is best solution.
Confused ‘Why is the app asking about breakfast?’Change scenario state you're inputting info for breakfast, the time and situation.
Prototype reactions especially on the voice interactionInsure another person conducts the user test with you. One to record task and one to act as the bot ‘Rockie’ and to the automatic reactions.
Users didn’t know what the plus meantMake plus blue so it’s more clear and may add text ‘add’
People didn’t know they were talking to a botMake it clear they are talking to Rockie. Let ‘Hey! What’s for breakfast?’ take over the whole screen
Didn’t like the app was text and talking to themMake it clear they can input information through the diary
Felt the interaction was too personal texting someoneGive them the option no to talk to the coach “Rockie”
Prototype reactions on textParticipants felt that the prototype was jumping and auto-populating the information
Prototype reactions especially on the voice interactionParticipant felt that the voice feature was inconsistent with real-life systems and therefore couldn’t complete the task.
Reactions on photo inputThe simplest task to complete, however as an observer the participant accidentally clicked on "DONE".
Participants clicked on gets suggestionsHighlight button as the primary call to action within the screen
As the facilitator of the test it was difficult to guide the participant when it the prototype could not perform the task as expected.Highlight within the usability study that the app tested today is a prototype and that there may be minor things they may find do not work as expected, however they should be highlighted.

Figure 4. Compilation of errors detected in the first user test.

A new series of sketches based on the proposed solutions were made (Fig. 5).

Figure 5. Sketches after the first usability tests.

Second prototype

Based on the feedback from the first iteration a new set of screens were generated using Sketch and recorded (Video 2, Video 3 and Video 4).

Figure 6. Initial screen added to introduce the bot interaction.

Voice entry


Video 2. Voice entry example.

Text entry


Video 3. Text input example.

Photo entry


Video 4. Photo entry example.

User testing

The second prototype was tested with six users. A modified version of the information sheet (Annex 14) was given to the participants as some of the issues occurred because the high fidelity prototypes enticed the participants to believe that they were using a production version of MyFitnessPal.


Video 5. User test using InVision to display prototypes and Lookback to capture screen and image of the participants.

Even though there were issues during the tests, all users responded that they would prefer the proposed solution than the current MyFitnessPal. All videos and notes are available in Annex 12 and Annex 13, respectively.

A new table with problems detected and possible solutions for each item was compiled (Fig. 7).

IssueSolution
By text: user wanted to click on the educational tool and it to react. ie it to appear as text.Enable users to put in one piece of information at at time.
By text: People said they wouldn’t use ‘and’ but just input one piece of information at a timeSame as above
By photo: didn’t know you needed to click twice when confirming information.Use an overlay to explain reaction. (was in original prototypes)
User felt they wouldn’t use voiceThis input would need to be tested with the observer being in a different room as the user didn’t seem comfortable talking to the app in the front of another person.
Rockie’s character didn’t come across!I think our only solution would be to test over a longer period of time
User laughed at the idea of entering the information by voice in front of someone else.Voice test needs to be done remotely to have a good result
User had troubles with the microphone iconAdd the MICROPHONE icon beside the ADD NEW in the first screen so user don't have to make 2 clicks to find MIC icon.
User questioned if the voice feature would work with her accent.Technology needs to be ready and well tested.
User hesitates when adding by photo.Screen to add avocado and toast by photo needs to be clearer
Adding by photo, confirmation button is too far to reachMove confirmation to the right side of the screen
Don't like the use of voice in front of other peopleVoice test needs to be done remotely to have a good result
Participant expected to send in list format (Avocado > select weight > send, Toast > select weight > send)Allow user to input in list format, this would also be similar to the current input where items are input individually
Participant laughed and wasn’t sure why letter were appearing assumed it was autocomplete.Only make the letters related to the task tappable
Participant commented on allowing users to input the weight or amount manuallyProvide an option in the screen below to add weight manually ( i think we previously discussed this, however, didn’t add to wireframe)

Participant recognised that the data was inconsistent between the task and the prototypeInclude consistent data throughout the task/prototype
Participant commented that the weight of the bread isn't realistic ‘250g is a lot of bread’ would be better highlighted in slices ‘do people weigh bread?’Update g’s to slices

Participant expected feedback/indicator of his recording on the screen before sending it. He might have been tapping the screen to see if anything happened as he had previously seen in the text taskBreakdown and update the steps in the voice task, in particular, the screen below offers no indicator that a voice recording has been sent


Figure 7. Compilation of errors that occurred in the second user test.

Results

Most of the errors in both versions (Fig. 8 and Fig 9) were caused by issues with the fixed mental model set on the prototypes. In Test 2, this was the case for 11 out of 14 errors.

The times to perform the tasks (Fig. 11) increase in Task 2, the reason is that the second prototype a more realistic approach, the user had to type “avocado and toast”, and perform a full voice interaction with Rockie.

Figure 8. Errors per task.
Figure 9. Errors per participant.
Figure 10. Mean and median SUS Scores.
Figure 11. Mean completion times per task.
Figure 12. The preferred input method of the participants.

 SUS Scores
User 162.5
User 255
User 360
User 495
User 562.5
Average67
Median62.5

Figure 13. SUS Scores of Test 1.

 SUS Scores
User 190
User 290
User 390
User 492.5
User 592.5
User 663
Average86.25
Median90

Figure 14. SUS Scores of Test 2.

Test 1Text Input TimePhoto Input TimeVoice Input TimeTotal Time
User 100:45:0000:22:0002:05:0003:12:00
User 200:54:0000:53:0001:15:0003:02:00
User 300:40:0002:00:0001:00:0003:40:00
User 400:13:0000:53:0001:56:0003:02:00
User 500:33:0000:13:00* did not complete00:46:00
Average Time00:37:0000:52:1201:34:0002:44:24

Figure 15. The timing of Test 1.

Test 2Text Input TimePhoto Input TimeVoice Input TimeTotal Time
User 102:11:0000:27:0001:08:0003:46:00
User 200:26:0000:39:0001:25:0002:30:00
User 300:54:0002:14:0001:06:0004:14:00
User 400:53:0001:22:0001:57:0004:12:00
User 501:01:0000:53:0001:45:0003:39:00
User 601:35:0001:26:0001:10:0004:11:00
Average Time01:10:0001:10:1001:25:1003:45:20

Figure 16. The timing of Test 2.

Conclusion

According to Sauro (2011), a SUS score above 68 would be considered above average, and anything below 68 is below average (Fig. 17). A score above 80.3 is equivalent to an A. The mean SUS score of the first test was 67, and the median SUS score was 62.5. The second test received a mean SUS score of 86.25 and a median SUS score of 90. This excellent progress resulted in a new and improved user journey (Fig. 18).

Figure 17. A comparison of the adjective ratings, acceptability scores, and school grading scales, in relation to the average SUS score.
Figure 18. New journey map.
Figure 19. Comparison of old (in green) and new user journey map (in blue).

References

Usability Testing (n/a). Retrieved from https://www.usability.gov/how-to-and-tools/methods/usability-testing.html

Loranger, L. (2016). Checklist for Planning Usability Studies. [https://www.nngroup.com/articles/usability-test-checklist/]. Accessed 15 April 2018.

Nielsen, J. (2012). Travelling Usability Lab. [https://www.nngroup.com/articles/traveling-usability-lab/]. Accessed 15 April 2018.

Nielsen, J., & Landauer, T. K. (1993). A mathematical model of the finding of usability problems. Proceedings of the SIGCHI Conference on Human Factors in Computing Systems – CHI 93. doi:10.1145/169059.169166

Sauro, J. (2011). Measuring Usability with the System Usability Scale (SUS). [https://measuringu.com/sus/]. Accessed 15 April 2018.

Leave a Reply

Your email address will not be published. Required fields are marked *