ArticlePDF Available

Towards Joint Activity Design Heuristics: Essentials for Human-Machine Teaming

Authors:

Abstract

As machines increasingly behave more like active cognitive agents than passive tools, additional heuristics for supporting joint human-machine activity are urgently needed to complement existing usability heuristics. Despite the rich and extensive design guidance produced by forty years of cognitive systems engineering (CSE) and related fields, the lack of large-scale impact can be attributed, in part, to insufficient translation of CSE principles and guidelines to language and tools that are ready for designers and other decision-makers responsible for these automation-infused solutions. Towards this need, we synthesized a partial and preliminary list of ten machine requirements intended to capture some of the essentials of joint activity. We believe solidifying these essentials and their implications for machines is a first and necessary step towards deriving joint activity design heuristics that are valuable, practical, and sustainable for operational personnel. Through iterative refinement, we believe the combination of strong ideas and strong practicality in these tools can be the basis for a large-scale shift in the design and evaluation of human-machine teams.
https://doi.org/10.1177/21695067231193646
Proceedings of the Human Factors and
Ergonomics Society Annual Meeting
1 –6
Copyright © 2023 Human Factors
and Ergonomics Society
DOI: 10.1177/21695067231193646
journals.sagepub.com/home/pro
Special Issue: Proceedings of the 67th HFES International Annual Meeting
Introduction
The proliferation of artificial intelligence (AI) and machine
learning (ML) available to organizations and the increasing
variety of ways in which these technologies can interact
with people (Amershi et al., 2014; Arrieta et al., 2020)
emphasizes the pressing need for reliable, practical, and sus-
tainable tools to help stakeholders quickly identify which
technologies are likely to be successful for them. This is
nontrivial because people overestimate machine capabilities
and how they interact with people so frequently that it has
been called a law (see “Robin Murphy’s Law” in Woods &
Hollnagel, 2006), which could be feasibly measured as an
“over-promise rate” (NASEM, 2021). Organizations need
additional support to guide the development of AI/ML tech-
nologies early in the design process in order to evaluate an
increasing volume of AI/ML options, counteract the ten-
dency to overestimate capabilities, and quickly discard dis-
advantageous options before the cost of change becomes
prohibitively expensive.
For this needed intervention to be impactful, its imple-
mentation must be grounded in strongly supported scientific
ideas, aligned with stakeholders’ mental models about what
is relevant and valuable, pragmatic relative to the costs of
implementation, compatible with the resources and actors
available, and sustainable as organizational goals and priori-
ties continue to change over time (Fitzgerald, 2019).
However, there are currently no available tools and methods
for designing and evaluating how well technologies can inte-
grate into a joint human-machine cognitive system that fully
satisfy these requirements for impactful implementations.
Although empirical evaluations, either before (Roth et al.,
1987) or after (Beede et al., 2020) full-scale implementation,
are perhaps the most reliable methods to uncover the ways in
which technologies can interact with people in surprising
(and detrimental) ways, they are neither sufficiently prag-
matic nor sustainable to have a large-scale impact on their
own. High-fidelity testing is costly. It typically involves
solutions late in their development lifecycle, reserved time
from end-users, and months or more of planning and
preparation from a dedicated study team. Because of these
1193646PROXXX10.1177/21695067231193646Proceedings of the Human Factors and Ergonomics Society Annual MeetingMorey et al.
research-article20232023
1The Ohio State University, Columbus, OH, USA
2Charles River Analytics, Cambridge, MA, USA
Corresponding Author:
Dane A. Morey, The Ohio State University, 1971 Neil Ave, Columbus,
OH 43210, USA.
Email: morey.38@osu.edu
Towards Joint Activity Design Heuristics:
Essentials for Human-Machine Teaming
Dane A. Morey1, Prerana Walli1, Kenneth S. Cassidy1,
Priyanka K. Tewani1, Morgan E. Reynolds1, Samantha Malone1,
Mohammadreza Jalaeian1, Michael F. Rayo1, and Nicolette M.
McGeorge2
Abstract
As machines increasingly behave more like active cognitive agents than passive tools, additional heuristics for supporting
joint human-machine activity are urgently needed to complement existing usability heuristics. Despite the rich and extensive
design guidance produced by forty years of cognitive systems engineering (CSE) and related fields, the lack of large-scale
impact can be attributed, in part, to insufficient translation of CSE principles and guidelines to language and tools that are
ready for designers and other decision-makers responsible for these automation-infused solutions. Towards this need, we
synthesized a partial and preliminary list of ten machine requirements intended to capture some of the essentials of joint activity.
We believe solidifying these essentials and their implications for machines is a first and necessary step towards deriving joint
activity design heuristics that are valuable, practical, and sustainable for operational personnel. Through iterative refinement,
we believe the combination of strong ideas and strong practicality in these tools can be the basis for a large-scale shift in the
design and evaluation of human-machine teams.
Keywords
Joint Activity, Human-Machine Teaming, Design Heuristics, Cognitive Systems Engineering, Design Requirements
2 Proceedings of the Human Factors and Ergonomics Society Annual Meeting
constraints, tests are often limited to a focused subset of can-
didate technologies. Therefore, empirical evaluations alone
are usually too costly (in money, time, and human resources)
to be pragmatic for a broader exploration of candidate tech-
nologies, or else too slow and limited to sustainably keep
pace with changing organizational needs. Though some
empirical testing is almost certainly necessary to mitigate the
risks of implementation surprise, complementary analytical
evaluations (e.g., heuristic reviews) will also be necessary to
satisfy the pragmatic and sustainable requirements of organi-
zational interventions.
Despite widespread adoption, usability heuristics alone
are conceptually insufficient to reliably evaluate technolo-
gies that behave more like active cognitive agents than pas-
sive tools (Rayo, 2017). The prominence, popularity, and
consistency of principles like Nielsen’s 10 usability heuris-
tics (Nielsen, 1994, 2020) is a testament to their model align-
ment, pragmatism, and sustainability relative to the interface
design community. Nevertheless, as technologies become
capable of making inferences and interjections, there remains
concern that usability alone does not sufficiently address the
critical need to support macrocognitive and coordinative
functions like event detection, sensemaking, replanning,
forming a basic compact, and common grounding (Rayo,
2017). There is strong agreement among experts that concep-
tualizing AI/ML technologies like teammates or team play-
ers is valid and valuable beyond conceptualizing these
technologies simply like tools (NASEM, 2021). Therefore,
tool-like usability heuristics alone do not appear sufficiently
grounded in the supported scientific ideas to be impactful
interventions.
A growing body of literature from cognitive systems engi-
neering and other related fields is providing guidance beyond
usability alone for designing and evaluating AI/ML technol-
ogies for human-machine teaming (NASEM, 2021); how-
ever, despite forty years of research, this literature has yet to
have the large-scale impact desired beyond the human fac-
tors community (Dominguez et al., 2021). This shortcoming
may in part stem from the wide variety of perspectives, each
with their own terminology, from which relevant AI/ML
guidelines can be derived, including: joint activity (Klein
et al., 2005), teamwork (Feigh & Pritchett, 2014), interde-
pendence (Johnson et al., 2014), situation awareness
(Endsley, 2017), macrocognition (Klein et al., 2003; Rayo,
2017), distributed problem solving (Smith, 2018), joint cog-
nitive systems (Woods & Hollnagel, 2006), human-AI inter-
action (Amershi et al., 2019), and resilience (Woods, 2019).
Furthermore, most of this guidance was written for human
factors professionals, often too vague or inaccessible for a
broader audience (e.g., Rayo, 2017). In either case, the
inconsistency and inaccessibility of guidance suggests that
this body of literature as a whole, regardless of its potential
value, has insufficient alignment and/or pragmatics to engage
the broader set of organizational actors (beyond the limited
number of human factors professionals) required to keep
pace with the scale and tempo of AI/ML proliferation.
The aim of this work is to begin to bridge the gap between
the scientific literature on joint human-machine activity and
the actors available to make large-scale impacts. Our end
goal is to compile, synthesize, and translate guidance rele-
vant to designing for and evaluating joint activity in a way
that is understandable, pragmatic, and sustainable for opera-
tional personnel. We chose to specifically target operational
personnel because they are likely to be the people in organi-
zations who (1) are most familiar with the technologies and
demands of work, (2) have a direct stake in the outcome, and
(3) comprise a large workforce capable of a large-scale
impact. However, before we can address this end, we first
seek to solidify the essentials of joint activity from which
subsequent design heuristics can be derived. We report our
partial and preliminary findings, focused specifically on the
subset of joint human-machine activity, as a set of ten
machine requirements stemming from joint activity essen-
tials. We expect this partial list to grow and change as we
expand beyond human-machine activity to joint activity in
general and as we shift from machine-focused requirements
to system-focused essentials.
Methods
We began with a broad review of the literature relevant to
human-machine joint activity, including but not limited to
the literature bases mentioned previously. For each article,
we compiled a list of all takeaways relevant to designing for
and/or evaluating joint human-machine activity and catego-
rized each takeaway as either a problem (to be avoided) or
guideline (to be design for). We used these problems and
guidelines as the basis for thematic analysis. In alignment
with the constant comparative analysis method (Fram, 2013),
we first constructed open codes to reflect the individual lan-
guage of the authors of each article. We then proceeded
through iterative axial coding to group our open codes into
clusters of common concepts. Informed by what we were
finding, we continued to review the literature until we had
reached saturation.
From these axial codes, we next constructed a set of
machine requirements for satisfying joint activity essentials.
First, we composed one or more concise evocative state-
ments to represent each of our axial codes. Importantly, we
chose language intended to resonate with operational person-
nel, which was not necessarily the language found in the aca-
demic literature, to better align with their mental models.
Second, we expounded upon each of these evocative state-
ments with a short 1-3 sentence synopsis. Finally, we elabo-
rated on each heuristic even more with 1-3 paragraphs of
further details, a toy example to illustrate the core concept(s),
and a list of tips and tricks for effective use. We modeled this
progressive description of each heuristic after Nielsen (2020)
Morey et al. 3
and expected that this would help to engage operational per-
sonnel in a way that is meaningful, but not overwhelming.
Results
Our literature review covered a total of 113 articles. From
these articles, we identified 590 problems and/or guidelines
relevant to the design and evaluation of joint human-machine
activity. From these open codes, we identified nine axial
codes: jointness or teamwork, escalation and resilience,
observability, directability, limits and boundaries, informa-
tiveness, redirecting attention, interjection strength, and
diversity. The following section includes the partial and pre-
liminary ten machine requirements we composed to convey
to operational personnel the core concepts underlying these
nine axial codes. Due to space constraints, we report only the
tagline (in bold) and a consolidated synopsis of the remain-
der of related materials for each requirement.
Ten Machine Requirements To Satisfy
Essentials Of Joint Activity
The machine changes and augments what
people can do, rather than replaces them
Machines can never replace people, but they can change
people’s roles and the kinds of the work people do (Woods &
Hollnagel, 2006). No matter how capable machines become
to take actions on their own, people will ultimately be held
responsible for the actions of machines (Murphy & Woods,
2009). At minimum, this requires people to monitor or super-
vise machines, which often creates additional problems
(Sarter et al., 1997). Not only do machines that attempt to
replace people fail to do so, they also miss out on the benefits
of being part of a team. Effective teams enable all members
to contribute to the work so that the team accomplishes more
than any of its members could accomplish alone. Likewise,
machines should augment, amplify, or contribute in a way
that aids, not replaces, the work of people (Woods, 1985).
The machine interacts with people: they
cannotremain separate or invisible
Effective teams are well-coordinated, which requires team
members to interact with others in ways that help the team
align their activities. These interactions include making
actions and intentions visible to others, monitoring others’
actions and intentions, synchronizing actions, aligning goals,
shifting roles, and minimizing the burden that these interac-
tions have on others in the team (Klein et al., 2005). Machines
attempting to remain separate from or invisible to people do
not eliminate the need for people to coordinate with machines,
but rather make this coordination difficult or impossible to
accomplish (Roth et al., 1987). Instead, machines should
expand, rather than reduce, the ways in which they can inter-
act with people to coordinate activities (Johnson et al., 2020).
The machine clearly indicates to people its past,
present, and future status and intentions
To be well-coordinated, teams need to have a shared under-
standing of the overall task and the progress being made
towards achieving it (Klein et al., 2005). Team members
need to clearly communicate their activities, status, and
intentions so that others can seamlessly and efficiently coor-
dinate their current and future actions without surprises. For
machines, this requires more than simply making actions or
computations visible. People must understand what the
machine has done, what it is doing, and what it is going to do
next (Sarter et al., 1997).
The machine conveys to people the ways it is
misaligned to the world, even when it is unaware
that it is misaligned to the world
Machines are literal-minded, meaning that they will only
function in accordance with their programming and resultant
models of the world. However, a machine cannot tell if its
model of the world is in fact the world it is in (Woods &
Hollnagel, 2006). Consequently, machines will often take the
right actions (according to their models of the world) in the
wrong world. People need to help keep machines aligned to
the current situation and make sure they are not operating
outside their limits (Hoffman et al., 2002). To help people do
this, machines must send signals or clues to people that con-
vey when, how, and why they are operating outside their lim-
its. However, like any team member, machines have limits to
how well they can understand their own limits (Woods,
2018). Therefore, machines will be unreliable at communi-
cating their own limits and need help from people. At mini-
mum, this requires some way for people to simultaneously
understand the world, how the machine is “seeing” the world
through its model, and how well the machine’s view of the
world is aligned to the world itself (Rayo et al., 2020).
The machine changes its actions, plans, and
priorities based on people’s directions
Effective teammates are able to direct the actions of others
and be directed by others (Klein et al., 2005). However,
unlike human teams which share responsibility, people are
always responsible for the outcomes of machine actions
(Murphy & Woods, 2009). Therefore, people must always be
in control. Otherwise, people will find ways to indirectly exert
their influence, like turning off the machine (Christoffersen &
Woods, 2002), or else machines can cause catastrophic acci-
dents, like the two Boeing 737 MAX crashes which killed a
total of 346 people (FAA, 2020). The ability to comply with
4 Proceedings of the Human Factors and Ergonomics Society Annual Meeting
people’s directions is a compulsory machine design require-
ment (Johnson et al., 2020).
The machine helps people adapt how it works to
keep the system working
As situations escalate from normal to exceptional, machines
often get in the way more than they help (Woods & Patterson,
2000). Machines stick to their set of rules, but successfully
responding to these exceptional situations often requires
changing procedures, sacrificing some goals, or otherwise
breaking the normal rules (Chuang et al., 2019). Machines
must allow (or even help) people break the machine’s rules
during exceptional circumstances; otherwise, machines are
likely to exacerbate the situation with additional burden
(Woods & Patterson, 2000).
The machine helps guide, but does not force,
people’s flow of attention to what is important
Attention is limited. Effective teams send signals to help
each other shift and focus their attention on what is impor-
tant; however, machines often struggle to understand what
is important (Woods, 1995a). Machines that frequently
send unhelpful alerts can annoy or even impair people’s
ability to figure out what is important (Rayo & Moffatt-
Bruce, 2015). Therefore, people must be able to determine
whether the signals machines send are important without
having to fully shift their attention from what they are cur-
rently doing (Woods, 1995a).
The machine conveys to people changes or
events without disrupting their workflow
Effective teammates assess whether their message or action
is important enough to interrupt what another is doing, which
depends upon both the importance of the interruption and the
importance of the other’s current actions (Klein et al., 2005).
However, machines are often poor at gauging interruptibility,
which can exacerbate periods of high workload with addi-
tional unhelpful disruptions (Woods, 1995a). Machines
should alert people to important changes in the situation in
the least disruptive manner possible, without loss of informa-
tion or urgency, so that managing the interruptions them-
selves does not become an additional burden.
The machine simultaneously helps people
understand the details and the bigger picture
Though machines can detect, process, and display massive
volumes of data, they struggle to reliably make sense of what
that data means, especially when something is novel or unex-
pected. For example, a substantial proportion of AI/ML
research continues to focus on image classification (Arrieta
et al., 2020), something so basic for people that we do not
consider it to be a decision. Machines still largely rely upon
people to understand the bigger picture, but data overload
becomes a problem. The volume of data machines display
can overwhelm and inhibit people from seeing the bigger
picture, yet machines that reduce this volume risk removing
crucial data, which also prevents people from seeing the big-
ger picture (Woods et al., 2002). Effective machines must
organize, but not reduce, the data available in a way that
helps people simultaneously see the bigger picture without
getting lost in the details and see the details without losing
sight of the bigger picture (Woods, 1995b).
The machine enables and encourages people to
consider multiple different perspectives
Every perspective is limited, both revealing and hiding cer-
tain aspects of the current situation (Hoffman & Woods,
2011). Effective teams overcome the limits of each perspec-
tive by supporting people in seeing and contrasting multiple
viewpoints. Machines need to be explicitly designed to sup-
port switching, comparing, and combining different view-
points. Otherwise, switching between viewpoints can be too
costly or disruptive (Woods, 1984). Machines must provide
low-cost ways for people to shift perspectives and highlight
when it is valuable to do so.
Discussion
Synthesizing the rich and extensive literature on joint human-
machine cognitive systems in a way that is valuable, practi-
cal, and sustainable for operational personnel is ambitious.
We chose colloquial (rather than academic) language and
phrases to align and connect with practitioner mental mod-
els. We limited the scope and level of detail in communicat-
ing concepts to enhance the pragmatics and explainability of
these requirements. This operationalization, like any, risks
oversimplification of the true complexity underlying these
concepts (Feltovich et al., 2004). However, we tried to miti-
gate this risk by composing statements that would evoke
rather than mask the underlying complexity. Ultimately, we
believe this operationalization of concepts specifically for
operational personnel is a necessary endeavor if the literature
on joint cognitive systems is to have the large-scale impact
that is desired (Dominguez et al., 2021).
However, the utility of these machine requirements,
essentials of joint activity, and eventual design heuristics can
only be fully evaluated by the impact they have for opera-
tional personnel. To this end, each should be empirically
evaluated to assess how and how well operational personnel
can utilize them. This involves, at minimum, the reliability,
validity, and pragmatism of using these tools to design and/
or evaluate candidate technologies. An effective intervention
like joint activity design heuristics will require both strong
Morey et al. 5
ideas and strong practicality to have a large-scale and benefi-
cial impact.
Acknowledgments
This material is based upon work supported by the US Air Force
Research Laboratory 711th Human Performance Wing (US AFRL
711 HPW) under Contract No. FA8650-22-C-6453. Any opinions,
findings, and conclusions or recommendations expressed in this
material are those of the authors and do not necessarily reflect the
view of the US AFRL.
ORCID iDs
Dane A. Morey https://orcid.org/0000-0002-9859-7719
Prerana Walli https://orcid.org/0009-0007-6036-4999
References
Amershi, S., Cakmak, M., Knox, W. B., & Kulesza, T. (2014).
Power to the People: The Role of Humans in Interactive
Machine Learning. AI Magazine, 35(4), Article 4. https://doi.
org/10.1609/aimag.v35i4.2513
Amershi, S., Weld, D., Vorvoreanu, M., Fourney, A., Nushi, B.,
Collisson, P., Suh, J., Iqbal, S., Bennett, P. N., Inkpen, K.,
Teevan, J., Kikin-Gil, R., & Horvitz, E. (2019). Guidelines
for Human-AI Interaction. Proceedings of the 2019 CHI
Conference on Human Factors in Computing Systems, 1–13.
https://doi.org/10.1145/3290605.3300233
Arrieta, A. B., Díaz-Rodríguez, N., Del Ser, J., Bennetot, A.,
Tabik, S., Barbado, A., Garcia, S., Gil-Lopez, S., Molina, D.,
Benjamins, R., Chatila, R., & Herrera, F. (2020). Explainable
Artificial Intelligence (XAI): Concepts, taxonomies, opportuni-
ties and challenges toward responsible AI. Information Fusion,
58, 82–115. https://doi.org/10.1016/j.inffus.2019.12.012
Beede, E., Baylor, E., Hersch, F., Iurchenko, A., Wilcox, L.,
Ruamviboonsuk, P., & Vardoulakis, L. M. (2020). A Human-
Centered Evaluation of a Deep Learning System Deployed in
Clinics for the Detection of Diabetic Retinopathy. Proceedings
of the 2020 CHI Conference on Human Factors in Computing
Systems, 1–12. https://doi.org/10.1145/3313831.3376718
Christoffersen, K., & Woods, D. D. (2002). How to make auto-
mated systems team players. In E. Salas (Ed.), Advances in
Human Performance and Cognitive Engineering Research
(Vol. 2, pp. 1–12). Emerald Group Publishing Limited. https://
doi.org/10.1016/S1479-3601(02)02003-9
Chuang, S., Chang, K.-S., Woods, D. D., Chen, H.-C., Reynolds,
M. E., & Chien, D.-K. (2019). Beyond surge: Coping with
mass burn casualty in the closest hospital to the Formosa Fun
Coast Dust Explosion. Burns, 45(4), 964–973. https://doi.
org/10.1016/j.burns.2018.12.003
Dominguez, C., Rayo, M. F., Dorton, S., Morey, D. A., Naikar, N., Roth,
E., & Allspaw, J. (2021). Panel Discussion Cognitive Engineering:
Will They Know Our Name When We Are 40? Proceedings of
the Human Factors and Ergonomics Society Annual Meeting,
65(1), 257–261. https://doi.org/10.1177/1071181321651042
Endsley, M. R. (2017). From Here to Autonomy: Lessons Learned
From Human–Automation Research. Human Factors, 59(1),
5–27. https://doi.org/10.1177/0018720816681350
FAA. (2020). Preliminary Summary of the FAA’s Review of the
Boeing 737 MAX (Preliminary Summary version 1). Federal
Aviation Administration.
Feigh, K. M., & Pritchett, A. R. (2014). Requirements for Effective
Function Allocation: A Critical Review. Journal of Cognitive
Engineering and Decision Making, 10.
Feltovich, P. J., Hoffman, R. R., Woods, D., & Roesler, A. (2004).
Keeping it too simple: How the reductive tendency affects
cognitive engineering. IEEE Intelligent Systems, 19(3), 90–94.
https://doi.org/10.1109/MIS.2004.14
Fitzgerald, M. C. (2019). The IMPActS Framework: The neces-
sary requirements for making science-based organizational
impact [The Ohio State University]. https://etd.ohiolink.edu/
apexprod/rws_olink/r/1501/10?clear=10&p10_accession_
num=osu1557191348657812
Hoffman, R. R., Feltovich, P. J., Ford, K. M., & Woods, D. D.
(2002). A rose by any other name. . . Would probably be given
an acronym [cognitive systems engineering]. IEEE Intelligent
Systems, 17(4), 72–80.
Hoffman, R. R., & Woods, D. D. (2011). Beyond Simon’s Slice:
Five Fundamental Trade-Offs that Bound the Performance
of Macrocognitive Work Systems. IEEE Intelligent Systems,
26(6), 67–71. https://doi.org/10.1109/MIS.2011.97
Johnson, M., Bradshaw, J. M., Feltovich, P. J., Jonker, C. M., van
Riemsdijk, M. B., & Sierhuis, M. (2014). Coactive design:
Designing support for interdependence in joint activity.
Journal of Human-Robot Interaction, 3(1), 43–69. https://doi.
org/10.5898/JHRI.3.1.Johnson
Johnson, M., Vignatti, M., & Duran, D. (2020). Understanding
Human-Machine Teaming through Interdependence Analysis.
In Contemporary Research. CRC Press.
Klein, G., Feltovich, P. J., Bradshaw, J. M., & Woods, D. D.
(2005). Common Ground and Coordination in Joint Activity.
In W. B. Rouse & K. R. Boff (Eds.), Organizational
Simulation (pp. 139–184). John Wiley & Sons, Inc. https://doi.
org/10.1002/0471739448.ch6
Klein, G., Ross, K. G., Moon, B. M., Klein, D. E., Hoffman,
R. R., & Hollnagel, E. (2003). Macrocognition. IEEE
Intelligent Systems, 18(3), 81–85. https://doi.org/10.1109/
MIS.2003.1200735
Murphy, R., & Woods, D. D. (2009). Beyond Asimov: The Three
Laws of Responsible Robotics. IEEE Intelligent Systems,
24(4), 14–20. https://doi.org/10.1109/MIS.2009.69
NASEM. (2021). Human-AI Teaming: State-of-the-Art and
Research Needs. National Academies of Sciences, Engineering,
and Medicine. https://doi.org/10.17226/26355
Nielsen, J. (1994). Heuristic evaluation. In J. Nielsen & R. L.
Mack (Eds.), Usability Inspection Methods. John Wiley &
Sons.
Nielsen, J. (2020, November 15). 10 Usability Heuristics for
User Interface Design. https://www.nngroup.com/articles/
ten-usability-heuristics/
Rayo, M. F. (2017). Designing for collaborative autonomy:
Updating user-centered design heuristics and evaluation
methods. Proceedings of the Human Factors and Ergonomics
Society Annual Meeting, 61, 1091–1095. https://doi.org/10.11
77/1541931213601877
Rayo, M. F., Fitzgerald, M. C., Gifford, R. C., Morey, D. A.,
Reynolds, M. E., D’Annolfo, K., & Jefferies, C. M. (2020).
The Need for Machine Fitness Assessment: Enabling Joint
Human-Machine Performance in Consumer Health
Technologies. Proceedings of the International Symposium on
Human Factors and Ergonomics in Health Care, 9, 40–42.
https://doi.org/10.1177/2327857920091041
6 Proceedings of the Human Factors and Ergonomics Society Annual Meeting
Rayo, M. F., & Moffatt-Bruce, S. D. (2015). Alarm system manage-
ment: Evidence-based guidance encouraging direct measure-
ment of informativeness to improve alarm response: Table 1.
BMJ Quality & Safety, 24(4), 282–286. https://doi.org/10.1136/
bmjqs-2014-003373
Roth, E. M., Bennett, K. B., & Woods, D. D. (1987). Human inter-
action with an “intelligent” machine. International Journal of
Man-Machine Studies, 27(5), 479–525. https://doi.org/10.1016/
S0020-7373(87)80012-3
Sarter, N. B., Woods, D. D., & Billings, C. E. (1997). Automation
Surprises. Handbook of Human Factors & Ergonomics, 2.
https://doi.org/10.1201/9780849375477.ch587
Smith, P. J. (2018). Making Brittle Technologies Useful. In P. J.
Smith & R. R. Hoffman (Eds.), Cognitive systems engineer-
ing: The future for a changing world (pp. 181–208). CRC
Press.
Woods, D. D. (1984). Visual momentum: A concept to improve
the cognitive coupling of person and computer. International
Journal of Man-Machine Studies, 21(3), 229–244. https://doi.
org/10.1016/S0020-7373(84)80043-7
Woods, D. D. (1985). Cognitive Technologies: The Design of Joint
Human-Machine Cognitive Systems. AI Magazine, 6(4), 1–7.
https://doi.org/10.1609/aimag.v6i4.511
Woods, D. D. (1995a). The alarm problem and directed attention in
dynamic fault management. Ergonomics, 38(11), 2371–2393.
https://doi.org/10.1080/00140139508925274
Woods, D. D. (1995b). Toward a Theoretical Base for Representation
Design in the Computer Medium: Ecological Perception and
Aiding Human Cognition. In Global Perspectives on the
Ecology of Human-Machine Systems. CRC Press.
Woods, D. D. (2018). The theory of graceful extensibility: Basic
rules that govern adaptive systems. Environment Systems and
Decisions, 38(4), 433–457. https://doi.org/10.1007/s10669-
018-9708-3
Woods, D. D. (2019). Essentials of resilience, revisited. In M. Ruth
& S. Goessling-Reisemann (Eds.), Handbook on Resilience
of Socio-Technical Systems (pp. 52–65). Edward Elgar
Publishing. https://doi.org/10.4337/9781786439376.00009
Woods, D. D., & Hollnagel, E. (2006). Joint Cognitive Systems:
Patterns in Cognitive Systems Engineering. CRC Press. https://
doi.org/10.1201/9781420005684
Woods, D. D., & Patterson, E. S. (2000). How Unexpected
Events Produce an Escalation of Cognitive and Coordinative
Demands. In P. A. Hancock & P. A. Desmond (Eds.), Stress,
Workload, and Fatigue (pp. 290–302). CRC Press. https://doi.
org/10.1201/b12791-2.3
Conference Paper
Full-text available
Deep learning algorithms promise to improve clinician workflows and patient outcomes. However, these gains have yet to be fully demonstrated in real world clinical settings. In this paper, we describe a human-centered study of a deep learning system used in clinics for the detection of diabetic eye disease. From interviews and observation across eleven clinics in Thailand, we characterize current eye-screening workflows, user expectations for an AI-assisted screening process, and post-deployment experiences. Our findings indicate that several socio-environmental factors impact model performance, nursing workflows, and the patient experience. We draw on these findings to reflect on the value of conducting human- centered evaluative research alongside prospective evaluations of model accuracy.
Article
Full-text available
In the last few years, Artificial Intelligence (AI) has achieved a notable momentum that, if harnessed appropriately, may deliver the best of expectations over many application sectors across the field. For this to occur shortly in Machine Learning, the entire community stands in front of the barrier of explainability, an inherent problem of the latest techniques brought by sub-symbolism (e.g. ensembles or Deep Neural Networks) that were not present in the last hype of AI (namely, expert systems and rule based models). Paradigms underlying this problem fall within the so-called eXplainable AI (XAI) field, which is widely acknowledged as a crucial feature for the practical deployment of AI models. The overview presented in this article examines the existing literature and contributions already done in the field of XAI, including a prospect toward what is yet to be reached. For this purpose we summarize previous efforts made to define explainability in Machine Learning, establishing a novel definition of explainable Machine Learning that covers such prior conceptual propositions with a major focus on the audience for which the explainability is sought. Departing from this definition, we propose and discuss about a taxonomy of recent contributions related to the explainability of different Machine Learning models, including those aimed at explaining Deep Learning methods for which a second dedicated taxonomy is built and examined in detail. This critical literature analysis serves as the motivating background for a series of challenges faced by XAI, such as the interesting crossroads of data fusion and explainability. Our prospects lead toward the concept of Responsible Artificial Intelligence , namely, a methodology for the large-scale implementation of AI methods in real organizations with fairness, model explainability and accountability at its core. Our ultimate goal is to provide newcomers to the field of XAI with a thorough taxonomy that can serve as reference material in order to stimulate future research advances, but also to encourage experts and professionals from other disciplines to embrace the benefits of AI in their activity sectors, without any prior bias for its lack of interpretability.
Conference Paper
Full-text available
Advances in artificial intelligence (AI) frame opportunities and challenges for user interface design. Principles for human-AI interaction have been discussed in the human-computer interaction community for over two decades, but more study and innovation are needed in light of advances in AI and the growing uses of AI technologies in human-facing applications. We propose 18 generally applicable design guidelines for human-AI interaction. These guidelines are validated through multiple rounds of evaluation including a user study with 49 design practitioners who tested the guidelines against 20 popular AI-infused products. The results verify the relevance of the guidelines over a spectrum of interaction scenarios and reveal gaps in our knowledge, highlighting opportunities for further research. Based on the evaluations, we believe the set of design guidelines can serve as a resource to practitioners working on the design of applications and features that harness AI technologies, and to researchers interested in the further development of human-AI interaction design principles.
Article
Full-text available
Purpose: To provide an insight into the challenges faced by the closest hospital to the Formosa Fun Coast Dust Explosion (FFCDE) disaster scene, and to examine how the hospital staff adapted to cope with the mass burn casualty (MBC) in their overcrowded emergency department (ED) after the disaster. Material and methods: The critical incident technique was used for the investigation. Data was gathered through in-depth individual interviews with 15 key participants in this event. The interview data was combined with the medical records of the FFCDE patients and admission logs to build a detailed timeline of ED workload. Process tracing analysis was used to evaluate how the ED and other units adapted to deal with actual and potential bottlenecks created by the patient surge. Results: Fifty-eight burn patients were treated and registered in approximately six hours while the ED managed 43 non-FFCDE patients. Forty-four patients with average total body surface area burn 51.3% were admitted. Twenty burn patients were intubated. The overwhelming demand created shortages primarily of clinicians, ED space, stretchers, ICU beds, and critical medical materials for burn care. Adaptive activities for the initial resuscitation are identified and synthesized into three typical adaptation patterns. These adaptations were never previously adopted in ED normal practices for daily surge nor in periodical exercises. The analysis revealed adaptation stemmed from the dynamic re-planning and coordination across roles and units and the anticipation of bottlenecks ahead. Conclusion: In the hospital closest to the FFCDE disaster scene, it caused an overwhelming demand in an already crowded, beyond-nominal-capacity ED. This study describes how the hospital mobilized and reconfigured response capacity to cope with overload, uncertainty, and time pressure. These findings support improving disaster planning and preparedness for all healthcare entities through organizational support for adaptation and routine practice coping with unexpected scenarios.
Article
This panel discussion will examine the societal awareness of cognitive engineering today. Cognitive engineering celebrated its 30 th anniversary in 2018 at the HFES annual meeting. Still, some would say that cognitive engineering is not as well-known as it should be, and that it is applied in an ad hoc manner in the many high-stakes, high-risk technology modernization efforts where it would be useful. As technology advances proliferate for sharp end of the spear decision makers, we are at risk of catastrophic results if CE remains in the shadows; these results are arguably emerging on a daily basis. Each panelist will describe, from their vantage point, CE’s state of the art today, thoughts on barriers to acceptance and application, and how they envision we act towards a future in 2028 in which cognitive engineers engage systematically in complex systems’ development.
Article
We introduce the concept of machine fitness assessment, which is the process of correctly determining the degree of fit between a machine’s inferences on a specific world and the world itself. We describe its importance in complex, high-stakes worlds, including healthcare, and how it will be critically important to realize the potential of consumer health technologies that promise institutional-quality health diagnosis and planning in decidedly non-institutional settings (e.g., our homes, offices, or anywhere else).