仓库源文站点原文


layout: post title: "HCI笔记 | Week 08 Evaluation" date: "2018-07-07 07:07:07" categories: 计算机科学 auth: conge

tags: HCI OMSCS

3.6 Evaluation

Introduction to Evaluation

methods for evaluation depend on stages at design life cycle.

Three Types of Evaluation

Evaluation

  1. qualitative evaluation: get qualitative feedback from users. What do they like, what do they dislike, what's easy, what's hard. use the methods for need finding.
  2. empirical evaluation: do some controlled experiments and evaluate the results quantitatively. - need more participants, and need to address the big qualitative feedback first.
  3. predictive evaluation. evaluation without users.

Evaluation Terminology

  1. Reliability refers to whether or not some assessment of some phenomenon is consistent over time

  2. validity. Validity refers to how accurately an assessment measures reality.

  3. Generalizability is the extent to which we can apply lessons we learned in our evaluation to broader audiences of people.
  4. Precision: Precision is a measurement of how specific some assessment is.

5 Tips: What to Evaluate

tips

  1. Efficiency. Time, number of actions to accomplish certain tasks. [expert users]
  2. Accuracy. The number of errors users commit. [expert users]
  3. Learnability. learning curve.
  4. Memorability. The user's ability to remember how to use an interface over time. I
  5. Satisfaction. Users' enjoyment of the system, or the cognitive load experience while using the system. [Need to avoid social desirability bias]

It's important to articulate at the beginning what you're evaluating, what data you're gathering, and what analysis you will use.

Evaluation Timeline

time line Over time, the evaluation method we'll use will change.

purpose

Approach

Data

settings

Settings are where the evaluation takes place.

Evaluation Design

Evaluation Design

Steps to ensure that the evaluation is actually useful.

  1. First, clearly define the task.
  2. Second, define our performance measures to avoid confirmation bias.
  3. Third, develop the experiment.
  4. Fourth, recruit the participants.
  5. Fifth, do the experiment.
  6. Sixth, analyze the data.
  7. Seventh, summarize the data in a way that informs our on-going design process.

Design life cycle

Qualitative Evaluation

The methods we use for qualitative evaluation are very similar to the methods we used for need finding. Interviews, think-aloud protocols, focus groups, surveys, post-event protocols.

Capturing Qualitative Evaluation

How to capture qualitative evaluation sessions:

  1. record the session.
  2. Note Taking
  3. Logging the behavior inside the software.

5 Tips: Qualitative Evaluation

  1. run pilot studies. try your experiment with friends or family or coworkers before trying it out with real users.
  2. focus on feedback. Try not to teach this one user. Hear the critics.
  3. use questions when users get stuck.
  4. tell users what to do, but not how to do it.
  5. Capture satisfaction. the user should be able to use the interface and like it.

Empirical Evaluation

Designing Empirical Evaluations

Apply treatments to participants and investigate the comparison between the treatments to find out about how they’re different.

Hypothesis Testing

Types of Hypothesis Testing

5 Tips Empirical Evaluation

  1. control what you can, document what you can't.
  2. limit your variables. Focus on varying only one or two things and monitor only a handful of things in response.
  3. work backward in designing your experiment. Decide the research question first and then the analysis and data to gather.
  4. script your analyses in advance. Analyze data with all the method and selectively report the ones showed significant results might be misleading.
  5. pay attention to power. The size of a difference that a test can detect. a small effect needs a lot of participants. a big effect needs fewer.

Predictive Evaluation

Predictive evaluation: evaluation without actual users.

Types of Predictive Evaluation

heuristic evaluation: hand the interface and the design guidelines to a few experts to evaluate.

model-based evaluation

simulation-based evaluation: construct an artificially intelligent agent that interacts with our interface

Cognitive Walkthroughs:

Evaluating Prototypes

!](/assets/images/计算机科学/118382-55de9c91ccd9d26a.png)

  1. qualitative evaluation. ask questions about the interface
  2. quantitative analysis.
  3. predictive evaluation:

Quiz: Exercise: Evaluation Pros and Cons

Quiz

2018-07-07 初稿