UsabilityCluster/MonthlyMeeting/2018-06-06 MeetingNotes
From Earth Science Information Partners (ESIP)
Meeting Agenda - Usability Cluster - 2018-06-06 1PM EDT
http://wiki.esipfed.org/index.php/UsabilityCluster/MonthlyMeeting/2018-06-06
Attendees: Bob, Megan, Tamar, Annie, Madison, Connor, Sophie
Agenda:
- Presentation on "Measuring User Experience"
- Discussion of the Framework's "Post-Test Reflection" section
- Usability pilot testing - IEDA usability tasks
Notes:
- Sophie goes over the agenda for today
- Presentation on “Measuring User Experience”
- The following presentation was taken from a “Measuring User Experience” course. Contact Sophie about questions she will forward them to original presenter.
- Understanding the difference between quantitative and qualitative approaches to usability
- Qualitative (Formative)
- We use information from users to iterate over designs
- Heuristic evaluation: do we meet a criteria? How do we perform against a principal?
- Quantitative (Summative)
- Usually measured at the end of a SDLC
- Like NCAR data archive
- We need a calculable matrix; larger sample sizes for statistical soundness
- For Qualitative, 5 people can be sufficient - Qualitative should be 30-40 up to 100 people
- We have clear tests (likert scales with number values)
- Design Considerations
- Mitigate errors
- Internal vs. External Validity
- Internal: Randomizing the order of questions
- External: Ensuring that the user is correct (nurse v. educator with operating room app)
- Within-subjects design: Asking a user to go between multiple designs and evaluate the designs comparatively
- Can require less users
- Problems with experience carrying over from one design compared to other
- Randomize the order of designs
- Self-Reported vs. Performance
- Self reported: How did you feel you did?
- Performance: measure variables (number of clicks, time, etc.)
- In person v. remote moderated v remote unmoderated
- In person: most accurate, due to live feedback and presence of moderator
- Remote: Not ideal, but not necessarily worse; preferred to remote unmoderated;
- Remote Unmoderated: Worst. We don’t know what people are doing between tasks. Uncontrolled environment
- Rating scale
- Lickert - 1 to 5
- Semantic Differential - Completely Satisfied v Completely Outraged
- Well written tasks
- Controlled, Clear Singular Solution, internally valid
- Guidelines for writing tasks
- Clear success criterion
- Its important that people do be distracted by other items
- Make the task neutral - but also mindful of the persona
- Statistics
- We won’t go into significant detail
- It’s important to properly analyze results
- Applications
- Evaluating Task Success
- Connor asks about Likert Scale and Semantic differential
- Sophie explains that they should not be combined
- Consistency is key
- Discussion of the Framework's "Post-Test Reflection" section
- Recommendations on how to “interview” users after the test
- Madison : Quantitative questions could apply
- Bob: Important to understand if the experience was a “waste of their time”. Phrasing “This was an efficient use of my time”
- System usability testing questions can be applied here from
- Problem? Is 10 questions too many?
- Bob: It’s important to clearly express which 1 and 5 is (ie. worst to best or vice versa)
- We can ask about specific tasks in a more granular fashion to understand specific functional aspects of the system vs. the system as a whole
- Recommendations on how to “interview” users after the test
- Usability pilot testing - IEDA usability tasks
- Tamar will do the pilot testing
- Megan asks about if we would give the task taker a copy of the tasks
- Madison: It could be ok if they give out tasks one at a time
- Megan asks if it’s appropriate to prompt people about their general opinions and feeling about the site
- Sophie says no problem; its very appropriate
- Tamar
- First thoughts: simple interface, very text heavy, attention is directed to titles of search results, search facet list very straightforward not too cluttered, overall nice
- First task: “If you want data from Juan de Fuca ridge what would you do”
- Used the search function
- Was sucessful
- Next: Narrow results down? How would you do that
- Tamar: Started with another search term
- Moved to Search Facets - Not certain how to re-search or apply changes to original search
- Does new search with settings selected; doesn’t seem to work
- Tamar sees that she needs to click on small text to apply the filter
- Decides to move to author
- Sophie interjects at next task to explain explain expectations of next task (Essentially how would you view a dataset)
- Tamar clicks the DOI to the research site
- Megan - asks tamar to access the data set specifically
- Sophie clarifies that the task is to “obtain the files”
- Tamar clicks on download data - is now seeking a download link
- Next task - Perform another search?
- Easily done
- If you need help what would you do?
- Tooltips were cool
- New link - How would you find the IEDA integrated catalog?
- Via search data button on IEDA main page
- Sophie expresses approval at Megan’s distancing herself from guiding the tasks
- Sophie encourages her to ask more questions in the case that the user is quiet “What are you thinking what are you doing?”
- It’s ok to add more contextual guidance (assume a certain motivation as a user)
- Megan expresses personal frustrations with one of the search facets (considers removing it entirely)
- Sophie relates with Megan’s experiences
- Madison - There was the point where Tamar had no datasets and she was asked to look into a dataset
- It’s ok to restart them from that point
- Megan does not feel like she needs to adjust the tasks according to this pilot
- It was interesting to see how much the search box was the fixation
- Time Considerations - 20-25 mins (this is a good baseline)
- Tamar - Was unsure if she had succeeded at a task