Progress and Challenges in Evaluating Tools for Sensemaking Jean Scholtz Pacific Northwest National Laboratory Richland, WA [email protected] ABSTRACT

The goal of the evaluation efforts is to devise a number of user-centered metrics for researchers in visual analysis to use in evaluating their work. We use the term user-centered to describe metrics that are focused on the utility the software provided to the analysts. While performance metrics are certainly necessary, they are not sufficient to guarantee utility. Moreover, user-centered metrics are one way to convey to researchers the priorities of the analysts. This is extremely valuable as academics most often have little access to analysts.

In this paper we discuss current work and challenges for the development of metrics to evaluate software designed to help analysts with sensemaking activities. While much of the work we describe has been done in the context of intelligence analysis, we are also concerned with the general applicability of metrics and evaluation methodologies for other analytic domains. Author Keywords

Evaluation, metrics, intelligence analysis, sensemaking.

Information analysts carry out complex tasks. Complex tasks differ from well structured tasks in many ways, including [6,7]:

ACM Classification Keywords

H5.m. Information interfaces and presentation (e.g., HCI): Miscellaneous.



Information overload is endemic. People have to locate the relevant information/sources and focus only on those.



Data analysis and recursive decision-making are cognitively very burdensome; people have little cognitive workload available for dealing with unusable interfaces.



Information is often incomplete or even unreliable.



In some domains, there may be no way to know if the result one gets is right or wrong.



In many domains time may be critical. Good decisions made too late are bad decisions.



Domain experts may not be computer experts. The demands of their work may make it difficult for them to put much time or effort into the learning curve of new programs.



Software systems for complex problem solving often consist of several components and visualizations. We must evaluate the individual components and/or visualizations as well as the entire system.



Information for many of these tasks is not static but changes over time. This dynamic nature increases the complexity of the task.



Complex problems often require multiple domain experts to collaborate. This presents a challenge in combining multiple perspectives.

INTRODUCTION

The Pacific Northwest National Laboratory has been funded by the Department of Homeland Security (DHS) to run the National Visualization and Analysis Center (NVAC). A number of activities occur in this program. The Regional Visualization and Analysis Centers (RVACs) are engaged in research to help solve problems that analysts in DHS encounter. Other activities within NVAC include: development of visual analysis software, customizing and deploying software at various DHS agencies, and developing metrics and evaluation methodologies for the field of visual analysis. The latter effort includes: designing and conducting the Visual Analytics Science and Technology (VAST) 2006 and 2007 contests, developing metrics for the various software deployments and consulting with the RVACs on user-centered evaluation studies of their research (http://www.cs.umd.edu/hcil/VASTcontest07/).

1

In the rest of the paper we discuss various activities that are helping to define user-centered metrics and the challenges that are still ahead. THE VAST CONTESTS

We have to date conducted two VAST contests: VAST 2006 and VAST 2007. Material developed for these contests include: •

A data set containing group truth and some background information about the task



Metrics for evaluation



Specifications on materials to submit

Participants are given approximately five months to assemble and submit their entry. Then a panel of judges consisting of experts in visualization and usability and experts in intelligence analysis meet and review the submissions. Winning entries are then invited to a closed session held in conjunction with the VAST symposium. In this session, the participants work with a professional intelligence analyst to investigate a data set for possible threats. The data set for the interactive session is smaller than the contest data set to allow for sufficient progress to be made in a two hour session. The metrics for the VAST contest are refined every year. The different categories of metrics for the VAST 2007 contest were [4] •

Accuracy on the answers to who, what, where and when questions



Quality of the debrief



Quality of the visualizations



Appropriateness of the interactions



Overall utility of the components of the software

Clearly, quality measures are subjective. Each of the above categories had subcategories. Judges were given criteria and asked to assign points from 1-5 and to provide a rationale. It is not surprising that judges with different expertise had difficulty in rating different aspects. The analysts were very comfortable rating the debriefs while the visualization and usability professionals were not. And while the visualization and usability professionals were comfortable in rating the visualizations and interactions, the comments of the analysts were extremely helpful to rate these in the context of use. The overall utility of the different software components was the most difficult aspect to consider. Ideally these components should be examined for their contributions to different aspects of sensemaking but that was difficult to do from the static submissions (process description and screen shots).

The contest committee is interested in refining the metrics to reduce the ambiguity in judging the subjective measures. Observations of the interactive session help in this respect as do our other interactions with analysts in different NVAC activities. SENSEMAKING AND METRICS

Since sensemaking is an integral part of analysis it is logical that we focus on assessing how tools help with the different aspects of sensemaking. However, not all analytic tasks are the same. What aspects of the task do we need to consider in devising appropriate measures for assessing improvement? Let’s consider three different scenarios. Scenario 1 involves a law enforcement official who gets a tip that an individual might have been involved in a particular crime. The law enforcement agent will go to a number of databases to put together what is known about this individual including such things as outstanding warrants, past arrests, any information about whereabouts. Using this information the law enforcement official will determine whether or not to further investigate this individual. Scenario 2 involves an intelligence analyst who has been asked to prepare a report on some suspicious activity in a remote corner of the world. There seems to be some interaction with a foreign chemical plant and construction activities in this region. In addition, the small military in this rather volatile country have been actively moving troops to different parts of their country. This analyst will need to look through many reports and possibly images to understand and report on the situation. It is known that this country has a reputation for hiding activities both from the outside but also from citizens within the country. Scenario 3 also involves a law enforcement official. This analyst is aware of a particular crime that is being investigated and is looking through other unsolved cases to determine if there are similarities with this particular case. These three scenarios are only a small representation of the variability of analytic tasks. However, even in these three tasks we can see the activities included in foraging, hypothesis generation, and evidence gathering. In scenario 1 the analyst is looking at a number of data bases but focusing on a particular known individual. In scenario 2 the analyst is looking at heterogeneous data, needs to cover a number of possible scenarios, but has to consider misinformation and deception in interpreting the information as well as missing information. In scenario 3, the analyst is looking through many cases to determine if there are patterns that match a case currently under investigation. Analytic Methods

Moreover, there are a number of analytic methods that analysts employ. These include: •

Problem Restatement

• Link analysis • Social network analysis • Telephone toll analysis • Timelines • Decision Tree • Analysis of Competing Hypotheses • Deception detection • Red Cell • Devil’s advocate • Team A/Team B • Alternative futures • Role Playing A number of these are best done as collaborative exercises, such as Team A/Team B, Devil’s advocate and Role Playing. Other methods, such as Red cell, involve looking at the situation from another point of view to gain a different perspective [4].

Quality •

Insertion Integration of visualizations into current work practices Usability •



The workshop participants agreed that this is clearly only a beginning. A web site is being established and a list serve will be setup. We envision further workshops as more of us attempt evaluations and can report on our results.

A workshop on Evaluation for Visual Analysis Environments was held at InfoVis 2007. Participants were asked to submit papers outlining possible metrics. Prior to the workshop the organizers extracted metrics from these position papers plus added measures used in the VAST 2007 contest. The metrics were grouped into a number of different categories. The category labels were NOT discussed and therefore should only be considered as placeholders; however, it is clear that metrics in a number of different categories should be considered.

NEXT STEPS

A number of suggestions for evaluating visual analytic environments have been proposed. However, these are more general metrics and do not directly assess the impact the software has on sensemaking activities [1, 8, 12]. A number of different activities currently underway will contribute to better means of assessing new tools for analysis. Sensemaking research and task analysis will improve our understanding and identify measures. Current work [3] has been extremely useful for researchers developing tools to improve sensemaking. Working with different types of analytic tasks, different users, and different application domains will help us understand commonalities as well as differences. We also need to look at other research areas to understand the impact of different application domains and tasks. Human–robot interaction (HRI) is also struggling with this problem [11].

Discussions during the workshop lead to the following list of prioritized metrics: Analysis Was the analyst able to explore a comprehensive set of hypotheses?



Was the analyst able to generate a number of hypotheses; eliminate red herring hypotheses; track the hypotheses being followed?

Task analysis has been widely used in the HCI field as the basis for understanding user needs and designing software to integrate well into the current work practice. In information analysis, however, we need to consider that current work practices are NOT (in most cases) satisfactory to deal with the increasing amount of data types and data sources. We must design tools to facilitate transitions to or at least accommodate new work practices. The issue of how various aspects of problem solving should be divided between humans and computers will be of great importance here.

Collaboration •

Was the analyst able to obtain multiple view points?

Visualization- Data types What kinds and amounts of data can be considered by the tool? Visualization - usability •

We are working on an approach in the VAST contest to reduce the risk in entering and also to increase the reward. The goal is to have participants learn more about their software. We are investigating ways to use the evaluations to increase communication between participants. We are also investigating ways to increase participation by allowing partial entries. This might be done by setting up tracks for different sensemaking activities.

Effectiveness, efficiency, user satisfaction BUT in context of user’s work/goals Visualization - utility •



Flexibility/ adaptability of visualizations

Work Practices •

Number of steps needed to accomplish basic tasks

Participants represented several different domains – intelligence analysis; transportation; and law. It was agreed that productivity measures could differ based on the domain and even on the high level task.

WORKSHOP RESULTS



Report quality (accuracy, usability, relevance)

Productivity measure (Document decisions per hour was suggested in one particular domain) 3

We are working to develop an infrastructure for the VAST contest to make the data sets and metrics easily available for interested parties to use on their own. This will include ways for users to provide feedback on the results of their evaluations. This, in addition to workshops on evaluation, will help us understand which metrics work in which situations. We also need to incorporate information from other activities. Badalamente and Greitzer [2] reported on a workshop with analysts identifying the top needs for tools. Included in this list were very practical needs: • Seamless data access and ingest • Diverse data ingest and fusion • Imagery data resources • Intelligence analysis knowledge base • Templates for analysis strategies Hypothesis generation and tracking was also high on the list. We also need to work with different research programs in the analytic community to understand their goals and to develop ways to measure the impact of the research tools they seek to develop. The Novel Intelligence from Massive Data (NIMD) program was focused on user-centered metrics and provided researchers with a much improved understanding of users and user-centered evaluations [9,10]. Looking at software that has been transitioned into the various analytic domains from these research programs is also helpful. Looking at software that was NOT transitioned into actual use is also useful. It is important to remember that user- centered evaluation is a communication tool. The metrics and evaluation methods are useful to convey to researchers the priorities and problems of the analysts. This is extremely important in domains where we are designing for complex tasks for expert users. Evaluation is needed to make progress in a field. In the case of visual analysis environments, user-centered evaluation is also a research topic. It must be investigated in conjunction with sensemaking research and research in visual analysis tools and methods. ACKNOWLEDGMENTS

This work in supported in part by the National Visualization and Analytics Center (NVAC) led by the Pacific Northwest National Laboratory for the U.S. Department of Homeland Security and by National Science Foundation Collaborative Research IIS-0713198. REFERENCES

1. Amar, R. and Stasko, J. A Knowledge Task-Based Framework for Design and Evaluation of Information

Visualizations, Proceedings of IEEE InfoVis '04, Austin, TX, October 2004, pp. 143-149. 2. Badalamente, R. and Greitzer, F. Top Ten Needs for Intelligence Analysis Tool Development. In 2005 International Conference on Intelligence Analysis, May 2005. 3. Card, S., and Pirolli, P. Sensemaking Processes of Intelligence Analysts and Possible Leverage Points as Identified Through Cognitive Task Analysis. In 2005 International Conference on Intelligence Analysis, May 2005. 4. Grinstein, G., Plaisant, C., Laskowski, S., O’Connell, T., Scholtz, J., and Whiting, M. VAST 2007 Contest-Blue Iguanodon, IEEE Symposium on Visual Analytics Science and Technology 2007. October 30 – Nov. 1. Sacramento, CA, USA. 5. Jones, M. The Thinker’s Toolkit. 1998. Three Rivers Press: New York. 6. Redish, J. Expanding Usability Testing to Evaluate Complex Systems, 2007. Journal of Usability Studies, Vol. 2(3), 102-111. 7.

Redish, J. C. and Scholtz, J., 2007, Evaluating complex information systems for domain experts, paper presented at a symposium on HCI and Information Design to Communicate Complex Information, Memphis, TN: University of Memphis, February. (paper available by request to the first author at [email protected])

8.

Scholtz, J., 2006, Beyond usability: Evaluation aspects of visual analytic environments, Proceedings of the IEEE Symposium on Visual Analytics Science and Technology, 145-150.

9.

Scholtz, J., Morse, E., and Potts Steves, M., 2006, Evaluation metrics and methodologies for usercentered evaluation of intelligent systems, Interacting with Computers, 18, 1186-1214.

10. Scholtz, J., Morse, E., and Potts-Steves, M., Metrics and Methodologies for Evaluating Technologies for Intelligence Analysts, Poster presented at 2005 International Conference on Intelligence Analysis, May, 2005. 11. Steinfeld, A., Fong, T. Kaber, D. Lewis, M. Scholtz, J. Schultz, A., Goodrich. M. Common metrics for human-robot interaction. HRI 2006: 33-40. 12. Thomas, J. and Cook, K. Illuminating the Path: the Research and Development Agenda for Visual Analytics. IEEE Computer Society. 2005.

Progress and Challenges in Evaluating Tools for Sensemaking

research. (http://www.cs.umd.edu/hcil/VASTcontest07/). The goal of the evaluation efforts is to devise a number of user-centered metrics for researchers in visual analysis to use in evaluating ... Data analysis and recursive decision-making are cognitively ... The different categories of metrics for the VAST 2007 contest were [4].

67KB Sizes 0 Downloads 154 Views

Recommend Documents

Progress and Challenges in Evaluating Tools for ...
This is extremely valuable as academics most often have little access to .... Social network analysis .... Badalamente, R. and Greitzer, F. Top Ten Needs for.

Understanding and Supporting Sensemaking in ...
Collaborative sensemaking, collaborative Web search,. SearchTogether ... structure, and visualize task-related information [2, 9]. Few researchers have .... CoSense uses data from a user's SearchTogether session and provides alternate ...

Representational Change in Sensemaking
jumps, but the basic fact that people begin with an ... In this illustration, the possible representations considered are all unordered ... moves down the semi-lattice by incrementally creating new features and assigning them to more and more ...

weick sensemaking in organizations pdf
Page 1 of 1. File: Weick sensemaking in. organizations pdf. Download now. Click here if your download doesn't start automatically. Page 1 of 1.

Supporting Synchronous Sensemaking in Geo ...
College of Information Sciences and Technology, the Pennsylvania State University. University ... sensemaking in an emergency management situation, which.

HTR Progress in China
Apr 8, 2014 - 10. Development history in China. ▫ 1970s:Technology research. ▫ 1986:National Hi-Tech program (863 program), start the design of HTGR.

Challenges for Inquiry and Knowledge in Social Construction of Reality
University of Management and Technology, Lahore, Pakistan. KHURAM ..... management: a review of 20 top articles. Knowledge and Process ... In E. Eisner. & A. Peshkin (Eds.), Qualitative inquiry in education: The continuing debate (pp. 19-.

Representational Change in Sensemaking
example is the space of hierarchical tree structures, such as might be used for outlines or for ... representations, in a way that may be useful for predicting when collaborating ... 1 The task, though not this particular analysis, is described in [3

Individual and Group Work in Sensemaking: an ...
This group came together in 13 facilitated meetings to discuss and build an ontology ... In order to create these tools, we need to investigate. Permission to make ...

GLOBE-CONCEPTS-AND-CHALLENGES-IN-PHYSICAL-SCIENCE ...
There was a problem previewing this document. Retrying... Download. Connect more apps... Try one of the apps below to open or edit this item. GLOBE-CONCEPTS-AND-CHALLENGES-IN-PHYSICAL-SCIENCE-TEXT-4TH-EDITION-2003C.pdf.

Sensemaking: Building, Maintaining and Recovering ...
Contact: [email protected]; 937.873.8166. Overview ... time decision making environments, for individuals and teams (Klein et al., 2006a;. 2006b; Sieck et al, ... sensemaking requirements for system support (the methodology piece). Finally, we ...

Social Information Foraging and Sensemaking
information flowing to him or her from a social network of collaborators. ... computational ecologies [6, 9, 10], library science [17], and anthropological studies of ...

Progress in Oceanography
1A). There is no argument among scientists that the number of Steller sea lions in the ... declined dramatically over the last 30 years (Braham et al., 1980; Kenyon and Rice, ...... l(hi; yi) using the computer language R and minimized it using the B

Sensemaking and its Handoff
examined in the light of general research on collaboration ... impact its collaborative handoff. ..... help of information on the web was the task used here.

Challenges in Warsaw.pdf
Page 1 of 1. Challenges in Warsaw.pdf. Challenges in Warsaw.pdf. Open. Extract. Open with. Sign In. Main menu. Displaying Challenges in Warsaw.pdf.

A Vision for Progress in Community Health Partnerships
Building in Partnership, Inc; (7) Morgan State University, School of Public Health and Policy; ... public agencies, we believe that we are at a moment in time when.