Systems Engineering and Artificial Intelligence instant download
Systems Engineering and Artificial Intelligence instant download
install download
https://ebookmeta.com/product/systems-engineering-and-artificial-
intelligence/
https://ebookmeta.com/product/applications-of-artificial-
intelligence-in-process-systems-engineering-1st-edition-
jingzheng-ren/
https://ebookmeta.com/product/artificial-intelligence-expert-
systems-sourcebook-hunt-v-daniel/
https://ebookmeta.com/product/artificial-intelligence-for-
renewable-energy-systems-artificial-intelligence-and-soft-
computing-for-industrial-transformation-1st-edition-ajay-kumar-
vyas/
https://ebookmeta.com/product/a-meeting-in-the-devil-s-house-
other-stories-1st-edition-richard-dansky/
Enthrall Whispers From the Bayou 3 1st Edition Sandra R
Neeley
https://ebookmeta.com/product/enthrall-whispers-from-the-
bayou-3-1st-edition-sandra-r-neeley/
https://ebookmeta.com/product/designing-smart-manufacturing-
systems-1st-edition-chaudhery-mustansar-hussain/
https://ebookmeta.com/product/filthy-jealous-heir-part-one-heirs-
of-all-hallows-1-1st-edition-caitlyn-dare/
https://ebookmeta.com/product/social-networks-modeling-and-
analysis-1st-edition-niyati-aggrawal/
https://ebookmeta.com/product/inventive-minds-marvin-minsky-on-
education-the-mit-press-marvin-minsky/
Freedom from Health Anxiety: Understand and overcome
obsessive worry about your health or someone else's 1st
Edition Karen Lynn Cassiday
https://ebookmeta.com/product/freedom-from-health-anxiety-
understand-and-overcome-obsessive-worry-about-your-health-or-
someone-elses-1st-edition-karen-lynn-cassiday/
William F. Lawless · Ranjeev Mittu ·
Donald A. Sofge · Thomas Shortell ·
Thomas A. McDermott Editors
Systems
Engineering
and Artificial
Intelligence
Systems Engineering and Artificial Intelligence
William F. Lawless · Ranjeev Mittu ·
Donald A. Sofge · Thomas Shortell ·
Thomas A. McDermott
Editors
Systems Engineering
and Artificial Intelligence
Editors
William F. Lawless Ranjeev Mittu
Paine College Information Technology Division
Augusta, GA, USA United States Naval Research Laboratory
Washington, DC, USA
Donald A. Sofge
United States Naval Research Laboratory Thomas Shortell
Washington, DC, USA Lockheed Martin Space Systems
King of Prussia, PA, USA
Thomas A. McDermott
Stevens Institute of Technology
Hoboken, NJ, USA
This Springer imprint is published by the registered company Springer Nature Switzerland AG
The registered company address is: Gewerbestrasse 11, 6330 Cham, Switzerland
Preface
1 https://aaai.org/Symposia/Spring/sss20symposia.php#ss03.
2 Michael Wollowski designed and built our supplementary website ([email protected]),
found at https://sites.google.com/view/scienceofinterdependence.
v
vi Preface
fatality, to best protect society, we need to know what happens if the context shared
by human-machine teams is incomplete, malfunctions, or breaks down.
This book also includes one of the first, if not the very first, chapters coauthored by
an artificially intelligent coauthor. Her name is Charlie. Her fellow coauthors address
the value of recognizing Charlie and treating her with respect to build a context that
is shared by all participants. For autonomous teams and systems involving humans
and machines, constructing a shared context is fundamental, meaning that joint inter-
pretations of reality must be addressed, requiring the interdisciplinary approach that
we have adopted, so that we too can learn from Charlie, a significant moment for us,
our fellow contributors, and we hope for you the reader, too.
We had several more participants than the speakers who attended our symposium.
We wanted speakers and participants who could assess the foundations, metrics, or
applications of autonomous AI/ML, human-machine teams, and systems and how
these teams and systems affect or may be affected themselves. We kept both of
the symposia open-ended for the topics and for this book. We considered all papers
submitted for the two symposia and several afterwards for the book as long as they had
a systems perspective. Accompanied by contributions from non-symposium partici-
pants, too, our goal then and now is to advance AI theory and concepts to improve the
performance of autonomous human-machine teams and systems to improve society.
viii Preface
After the AAAI-Spring and Fall Replacement Symposia in 2020 were completed,
speakers were asked to revise their talks into manuscripts for the chapters in this
book. After the symposium, other authors who did not participate in the symposium
were also invited and they agreed to participate. The following individuals were
responsible for the proposal submitted to Springer for the book before the symposia,
for the divergence between the topics considered by the two, and for editing this
book that has resulted.
ix
x Contents
Abstract In this introductory chapter, we first review the science behind the two
Association for the Advancement of Artificial Intelligence (AAAI) Symposia that
we held in 2020 (“AI welcomes Systems Engineering. Towards the science of inter-
dependence for autonomous human-machine teams”). Second, we provide a brief
introduction to each of the chapters in this book.
Presently, the United States is facing formidable threats from China and Russia. In
response to these threats, the Director of the Defense Intelligence Agency (Ashley,
2019) and DNI stated:
China ... [is] acquiring technology by any means available. Domestic [Chinese] laws forced
foreign partners of Chinese-based joint ventures to release their technology in exchange
for entry into China’s lucrative market, and China has used other means to secure needed
technology and expertise. The result … is a PLA on the verge of fielding some of the
most modern weapon systems in the world. ... China is building a robust, lethal force with
capabilities spanning the air, maritime, space and information domains which will enable
China to impose its will in the region. (p. V) ... From China’s leader, Xi Jinping, to his 19th
Party Congress (p. 17) “We must do more to safeguard China’s sovereignty, security, and
development interests, and staunchly oppose all attempts to split China or undermine its
ethnic unity and social harmony and stability.”
W. F. Lawless (B)
Paine College, Augusta, Georgia
e-mail: [email protected]
R. Mittu · D. A. Sofge · T. Shortell · T. A. McDermott
Systems Engineering Research Center, Stevens Institute of Technology, Hoboken, NJ, USA
e-mail: [email protected]
D. A. Sofge
e-mail: [email protected]
T. Shortell
e-mail: [email protected]
To address these and other competitive threats, artificial intelligence (AI), espe-
cially machine learning (ML) that we discuss with fusion next, is a major factor. The
U.S. Department of Defense (DoD), industry, commerce, education, and medicine
among many other fields are seeking to use AI to gain a comparative advantage for
systems. From the perspective of DoD (2019):
AI is rapidly changing a wide range of businesses and industries. It is also poised to change
the character of the future battlefield and the pace of threats we must face.
Simultaneously, the DoD recognizes the disruptive nature of AI (Oh et al., 2019).
To mitigate this disruption while taking advantage of the ready-made solutions AI
already offers to commerce, the current thinking appears to first use AI in areas
that are less threatening to military planners, the public, and potential users; e.g.,
back-office administration; finance (e.g., Airbus is using AI to cut its financial costs
by increasing efficiency, reducing errors, and freeing up humans for more strategic
tasks such as planning, analysis, and audits; in Maurer, 2019); data collection and
management; basic personnel matters; virtual assistants for basic skills training (i.e.,
Military Occupational Specialties, or MOSs); personal medical monitoring (e.g.,
drug compliance, weight reduction, sleep cycles); military maintenance; and simple
logistics (e.g., ordering, tracking, maintaining supplies).
Second, when the DoD and other fields address the more disruptive aspects of AI,
like autonomy and autonomous human–machine teams, many more social changes
and impacts will arise, including the adverse threats posed by the use of AI, such as
the “consequences of failure in autonomous and semi-autonomous weapon systems
that could lead to unintended engagements” (DoD, 2019).
Machine Learning (ML) and Fusion: Machine learning has already had an extraor-
dinary economic impact worldwide estimated in the trillions of dollars with even
more economic and social impact to come (Brynjolfsson & Mitchell, 2017). The
basic idea behind traditional ML methods is that a computer algorithm is trained
with data collected in the field to learn a behavior presented to it as part of previous
experience (e.g., self-driving cars) or with a data set to an extent that an outcome can
be produced by the computer algorithm when it is presented with a novel situation
(Raz et al., 2019).
Autonomy is changing the situation dramatically in the design and operational
contexts for which future information fusion (IF) systems are evolving. There are
many factors that influence or define these new contexts but among them are:
movement to cloud-based environments involving possibly many semi-autonomous
functional agents (e.g., the Internet of Things or IoT; Lawless et al., 2019b), the
employment of a wide range of processing technologies and methods spread across
agents and teams, an exceptional breadth of types and modalities of available data,
and diverse and asynchronous communication patterns among independent and
distributed agents and teams. These factors describe the contexts of complex adap-
tive systems (CAS) for “systems in which a perfect understanding of the individual
parts does not automatically convey a perfect understanding of the whole system’s
behavior” (Raz et al., 2019).
1 Introduction to “Systems Engineering and Artificial Intelligence” and the Chapters 3
Managing these disruptions must justify the need for speedy decisions; a systems
approach; the commonality of interdependence in systems and social science; social
science, including trust; the science of human–human teams (HHT); and human–
machine teams (HMT). We discuss these topics in turn.
Now is the time when decisions may need to be made faster than humans can
process (Horowitz, 2019), as with the military development of hypersonic weapons
by competitor nations (e.g., China; in Wong, 2018); the push for quicker command,
control, and communication upgrades for nuclear weapons (NC-3; in DoD, 2018);
and the common use of AI in public conveyances like self-driving cars, trucks, ships,
or subways.
Many systems are approaching an operational status that use AI with humans
“in-the-loop,” characterized by when a human can override decisions by human–
machine or machine–machine teams in combat, such as the Navy’s new Ghost fleet
(LaGrone, 2019); the Army’s autonomous self-driving combat convoy (Langford,
2018); and the Marine Corps’ remote ordinance disposal by human–machine teams
(CRS, 2018).
Even more dramatic changes are to occur with human “on-the-loop” decisions,
characterized by when decisions must be made faster than humans can process and
take action based on the incoming information. Among the new weapon systems,
these decisions may be made by a human–machine team composed of an F-35
teaming with the Air Force’s aggressive, dispensable “attritable” drones flying in
a wing or offensive position (Insinna, 2019); moreover, hypersonic weapons are
forcing humans into roles as passive bystanders until a decision and its accompa-
nying action have been completed. From an article in the New York Times Magazine
(Smith, 2019),
One of the two main hypersonic prototypes now under development in the United States is
meant to fly at speeds between Mach 15 and Mach 20 ... when fired by the U.S. submarines
or bombers stationed at Guam, they could in theory hit China’s important inland missile
bases ... in less than 15 minutes ...
By attacking the United States at hypersonic speeds, however, these speeds would
make ballistic missile interceptors ineffective (e.g., Aegis ship-based, Thad ground-
based, and Patriot systems). If launched by China or Russia against the United States
(Smith, 2019), these missiles:
would zoom along in the defensive void, maneuvering unpredictably, and then, in just a few
final seconds of blindingly fast, mile-per-second flight, dive and strike a target such as an
aircraft carrier from an altitude of 100,000 feet.
by human error (Lawless et al., 2017), self-directed machines may save more lives.
But an editorial in the New York Times (Editors, 2019) expressed the public’s concerns
that AI systems can be hacked, suffer data breaches, and lose control to adversaries.
The Editors quoted the UN Secretary General, Antonio Guterres, that “machines
with the power and discretion to take lives without human involvement … should
be prohibited by international law.” The editorial recommended that “humans never
completely surrender life and decision choices in combat to machines.” (For a review
of the U.N.’s failure to manage “killer robots,” see Werkhäuser, 2019.)
Whether or not a treaty to manage threats from the use of “on the loop” decisions is
enacted, the violations of existing treaties by nuclear states (e.g., NATO’s judgment
about suspected Russian treaty violations; in Gramer & Seligman, 2018) suggest
the need to understand the science of autonomy for “on the loop” decisions and to
counter the systems that use them.
Furthermore, the warning by the Editors of the New York Times is similar to
those that arose during the early years of atomic science, balanced by managing the
threats posed while at the same time allowing scientists to make numerous discov-
eries leading to the extraordinary gifts to humanity that have followed, crowned by
the Higgs (the so-called “God”) particle and quantum computing. The science of
autonomy must also be managed to balance its threats while allowing scientists to
make what we hope are similar advances in the social sphere ranging from Systems
Engineering and social science to international affairs.
SE is also concerned about whether AI and ML will replace humans in the decision
loop (Howell, 2019). System engineers prefer that humans and machines coexist
together, that machines be used to augment human intelligence, but that if decisions
by machines overtake human decision-making as is happening with “on-the-loop”
decisions, at least humans should audit the machine decisions afterward (viz., see the
Uber car fatality case below). SE also raises a series of other concerns and questions.
In addition to the public’s concerns about AI expressed by the Editors in the New
York Times, the application of AI/ML raises several concerns and questions for SE.
One concern is whether or not to use a modular approach to build models (Rhodes,
2019). System engineers note that safety is an emergent property of a system (Howell,
2019). When a team “emerges,” the whole has become more than the sum of its parts
(Raz et al., 2019); in contrast, when a collective fails, as appears to be occurring in
Europe today, it creates “a whole significantly less than the sum of its parts” (Mead,
2019). But if SE using AI/ML is to be transformed through model-centric engineering
(Blackburn, 2019), how is that to be accomplished for autonomous teams? Systems
often do not stand alone; in those cases where systems are a network of networks,
how shall system engineers assure that the “pieces work together to achieve the
objectives of the whole” (Thomas, 2019)? From retired General Stanley McCrystal’s
book, Team of teams, “We needed to enable a team operating in an interdependent
1 Introduction to “Systems Engineering and Artificial Intelligence” and the Chapters 5
For more concerns, along with metrics for autonomous AI systems, formal verifi-
cation (V&V), certification and risk assessments of these systems at the design, oper-
ational, and maintenance stages will be imperative for engineers (Lemnios, 2019;
Richards, 2019). Is there a metric to assess the risk from collaboration, and if so, can it
be calculated (Grogan, 2019)? The risk from not deploying AI systems should also be
addressed (DeLaurentis, 2019); while an excellent suggestion, how can this concern
be addressed?1 Measured in performance versus expectations, when will these risks
preclude humans from joining teams with machines; and what effect will machine
redundancy have in autonomous systems (Barton, 2019)? Because data are dumb,
how will the operational requirements and architectures be tested and evaluated for
these systems over their lifecycle (Dare, 2019; Freeman, 2019)?
Boundaries and deception: AI can be used to defend against outsiders, or used
with deception to exploit vulnerabilities in targeted networks (Yampolskiy, 2017). A
team’s system boundaries must be protected (Lawless, 2017a). Protecting a team’s
networks is also a concern. In contrast, deception functions by not standing out (i.e.,
fitting in structurally; in Lawless, 2017b). Deception can be used to compromise
a network. From the Wall Street Journal (Volz & Youssef, 2019), the Department
of Homeland Security’s top cybersecurity official, Chris Krebs, issued a statement
warning that Iran’s malicious cyberactivities were on the rise. “What might start as
an account compromise … can quickly become a situation where you’ve lost your
whole network.”
Caution: In the search for optimization, tradeoffs occur (Long, 2019); however,
an optimized system should not tradeoff resilience.
1 One possibility is to use global metrics. In the case of the Uber car accident that killed a pedestrian
discussed below, the industry’s first pedestrian fatality, the company’s self-driving section did not
suffer until the accident, and then Uber and the rest of the self-driving industry have been significantly
slowed by the fatality (Gardner, 2019).
6 W. F. Lawless et al.
with social scientists and by studying their own SE teams and organizations (DeLau-
rentis, 2019). Their own teams and organizations, however, are systems of social
interdependence.
Systems Engineering addresses the interactions of systems too complex for an
analysis of their independent parts without taking a system as a whole into account
across its life cycle. System complexity from the “interdependencies between …
constituent systems” can produce unexpected effects (Walden et al., 2015, p. 10),
making the management of systemic interdependence critical to a system’s success.
For example, the interactions for complex systems with numerous subsystems, like
the International Space Station (ISS), interact interdependently (i.e., interdependence
affected how the ISS modules were assembled into an integrated whole, how module
upgrades affected each other, how interfaces between ISS modules were determined
to be effective, how the overall configuration of the modules was constructed, how
modules were modeled, etc.; in Stockman et al., 2010). From the ISS, in SE, we
can see that interdependence transmits the interactions of subsystems. The study of
interdependence in systems is not a new idea. For example, Llinas (2014, pp. 1, 6)
issued a:
call for action among the fusion, cognitive, decision-making, and computer-science commu-
nities to muster a cooperative initiative to examine and develop [the] … metrics involved in
measuring and evaluating process interdependencies … [otherwise, the design of] modern
decision support systems … will remain disconnected and suboptimal going forward.
2 On December 7th, Uber sold its self-driving unit to Aurora Innovation Inc. (Somerville, 2020).
1 Introduction to “Systems Engineering and Artificial Intelligence” and the Chapters 7
went an average of nearly 5,600 miles before the driver had to take control from the computer
to steer out of trouble. As of March [2018, when the accident happened], Uber was struggling
to meet its target of 13 miles per “intervention” in Arizona …
It must be kept in mind, however, that as incompletely and poorly trained as was
the Uber car, it still responded to the situation as it had been designed; further, its
response was faster than its human operator.
The National Academy of Sciences (2019) Decadal Survey of Social and Behav-
ioral Sciences finds that the social sciences want to be included in research using
computational social science for human and AI agents in teams. In their thinking,
social scientists are concerned about ethical and privacy issues with the large digital
databases being collected. For systems of social networks, they recommended further
study on:
how information can be transmitted effectively … [from] change in social networks …
network structure of online communities, the types of actors in those communities …
More importantly, while social scientists want to be included in the AI/ML revo-
lution, they have had setbacks in their own disciplines with the reproducibility of
experiments (e.g., Nosek, 2015; also, Harris, 2018). For our purposes, unexpect-
edly, research has indicated that the poorest performing teams of scientists were
interdisciplinary teams (Cummings, 2015).3 In addition, however, Cummings added
that the best scientist teams maximized interdependence. Based on Cummings and
our research (e.g., Lawless, 2019), we conclude that for interdisciplinary teams to
function optimally, their team members must also be operating under maximum
interdependence (Lawless, 2017a). By extension, for the optimum size of a team
to maximize interdependence, a team’s size must be the minimum size to solve a
targeted problem (Lawless, 2017a), contradicting the Academy’s two assertions that
“more hands make light work” (Cooke & Hilton, 2015, Chap. 1, p. 13) and that the
optimal size of a scientific team is an open problem (p. 33).
The advent of human–machine teams has elevated the need to determine context
computationally, yet social science has offered little guidance for their design, oper-
ation, or to prevent accidents (see the Uber self-driving car accident described above
that killed a pedestrian in 2018), let alone the means to construct a computational
3 Cummings studied about 500 teams of scientists in the National Science Foundation’s data base.
8 W. F. Lawless et al.
context (Lawless et al., 2019a). Recognizing their plight, social scientists argue, and
we agree, that their science is the repository of an extraordinary amount of statistical
and qualitative experience in determining and evaluating contexts for humans and
human teams (NAS, 2019). Nonetheless, this situation leaves engineers to seek a
quantitative path on their own. Instead, we foresee an integrated path as the better
course going forward (Lawless, 2019).
Trust and machine autonomy: In the rapid decision-making milieux where trust
between machine and human members of a team becomes a factor (Beling, 2019),
to build trust, each member of a human–machine team must be able not only to
exchange information about their status between teammates but also to keep that
information private (Lawless et al., 2019a). In that humans cause most accidents
(Lawless et al., 2017), trust can be important outside of the team, as when a human
operator threatens passengers being transported, which happened with the crash of
GermanWings Flight 9525 in the Alps in March 2015, killing all 150 aboard at the
hands of its copilot who committed suicide (BEA, 2016); or the engineer on the
train in the Northeast Corridor in the United States who allowed his train rounding
a curve to speed above the track’s limits (NTSB, 2016); or the ship’s captain on
the bridge of the McCain at the time the destroyer was turning out of control in a
high-traffic zone (NTSB, 2019). In these and numerous other cases, it is possible
with current technology and AI to authorize a plane, train, other public vehicle or
military vehicle or Navy ship as part of a human–machine team to take control from
its human operator (the bet that a machine will be authorized to take control from a
dysfunctional human operator, Sofge et al., 2019).
From our review of human teams, Proctor and Vu (2019) conclude that the best
forecasts improve with competition (Mellers & Tetlock, 2019). They also conclude
that teams are formed by “extrinsic factors, intrinsic factors, or a combination of
both.” Extensive motivation is often generated from the collective consensus of many
stakeholders (the public, researchers, and sponsoring agencies) that there is an urgent
problem that needs to be solved. But they asserted that solutions require “a multi-
disciplinary team that is large in score … [with] the resources required to carry
out the research … to appropriate subject-matter experts, community organizations
and other stakeholders … [and] within an organization, administrative support for
forming, coordinating, and motivating multidisciplinary teams …”.
Salas and his colleagues (Bisbey et al., 2019) conclude that “Teamwork allows
a group of individuals to function effectively as a unit by using a set of interre-
lated knowledge, skills and attitudes (KSAs; p. 279). [On the other hand] … poor
teamwork can have devastating results … plane crashes, … friendly fire, … surgical
implications … When the stakes are high, survival largely depends on effective team-
work.” One of the first successes with human teams was: “Crew resource manage-
ment [CRM] prompted by not “human error,” but crew phenomena outside of crew
1 Introduction to “Systems Engineering and Artificial Intelligence” and the Chapters 9
member competencies such as poor communication in United Flight 173 led the
Captain to disregard fuel state. … CRM required the crew to solve its problems
as a team” (p. 280). Another success for team science occurred in the attempts to
understand the shoot-down of an Iranian commercial airliner by the USS Vincennes
in 1988, leading to the study of stress in decision-making. Subsequently, following
the combination of a significant number of unrelated human errors that led to new
research after President Clinton’s Institute of Medicine (IOM) review of medical
errors in hospitals; the coordination errors with the BP/Deepwater Horizon oil spill
in 2011; Hurricane Katrina in 2005; and the NASA accidents Columbia in 2003
and Challenger in 1986 space shuttle accidents. Based on this new research, human
team scientists separated task-work from teamwork. Task work dealt with skills or
a skills’ domain (flying a plane), teamwork skills with team effectiveness across
contexts (e.g., how to communicate with others; p. 282).
Viacom during 2016–18). Finally, orthogonality figures into our proposed next study
on fundamental decision processes and emotion for a model of a social harmonic
oscillator where we hypothesize that the best teams operate in a ground state while
underperforming teams operate in excited states (Lawless, 2019).
Artificial intelligence has already brought significant changes to the world; will the
impact of human–machine teams be even greater? The first of the contributed chap-
ters, Chap. 2, “Recognizing Artificial Intelligence: The Key to Unlocking Human
AI Teams,” was written by a team at Aptima, Inc., headquartered in Woburn, MA.
The authors consist of Patrick Cummings, Nathan Schurr, Andrew Naber, Charlie,
and Daniel Serfaty (Aptima’s CEO and Founder). Readers, please recognize that
one of the coauthors from Aptima, “Charlie,” has no last name; she is an artificial
embodiment. Charlie has made contributions to public before (e.g., at a workshop
and a panel), but her contributions to Chap. 2 may be one of the first, if not the
very first, chapters contributed to or co-authored by, as she is aptly described by
her fellow coauthors, an “intelligent coworker.” Interacting with Charlie in public
over the past year has produced several insights signified and discussed by all of the
authors in their chapter. Interestingly, several of these insights are based on the treat-
ment of Charlie’s spoken ideas and written contributions with deep respect, which
they have described as “recognizing” Charlie as an equal contributor. The authors
provide details about how Charlie came into existence and how she operates in public
(e.g., her architecture, her public persona, her ability to brainstorm). The stated goal
of all of the authors of Chap. 2 is to bring human and intelligent coworkers together
to build an effective system in the future, not only one that recognizes human and
artificial coworkers but also one that can be influenced by both human and artificial
coworkers and by the contributions from both. We add: “Welcome, Charlie!”.
Chapter 3 was written by three Systems Engineers, namely by Thomas A. McDer-
mott and Mark R. Blackburn at the Stevens Institute of Technology in Hoboken, NJ;
and by Peter A. Beling at the University of Virginia in Charlottesville, VA. (McDer-
mott is one of the co-editors of this book.) Their chapter is titled, “Artificial Intel-
ligence and Future of Systems Engineering.” In it, the authors address the major
transformation of their profession now occurring that is being driven by the new
digital tools for modeling, data and the extraordinary “digital twins” resulting in
the integration of data and modeling. These new tools include the artificial intelli-
gence (AI) and machine learning (ML) software programs that are becoming key to
the new processes arising during this period of transformation. Yes, Systems Engi-
neering (SE) is being transformed, but the hope of the authors is that SE is able to
guide these new tools and their applications to increase the benefits so that society
welcomes this transformation. To help guide this transformation, the authors provide
a roadmap being developed by the Systems Engineering Research Center (SERC);
SERC is a University-Affiliated Research Center of the US Department of Defense.
1 Introduction to “Systems Engineering and Artificial Intelligence” and the Chapters 11
The roadmap sets out a series of goals in the attempt by SERC to identify the oppor-
tunities and the risks ahead for the research community to guide Systems Engineers
in preparation for the journey to the emergence of autonomy safely and ethically.
The fourth chapter, “Effective Human-Artificial Intelligence Teaming,” was
written by Nancy J. Cooke and William Lawless. Cooke is a Professor of Human
Systems Engineering and Director of the Center for Human, Artificial Intelligence,
and Robot Teaming at Arizona State University. Lawless is a Professor of Mathe-
matics and Psychology at Paine College; he is also on two Navy Boards (the Science
of AI and Command Decision Making); and he is a new Topics Editor of the journal
Entropy (“The entropy of autonomy and shared context. Human–machine teams,
organizations and systems”). They begin their chapter with a review of the history
of interdependence. It has long been known to be present in every social interaction
and central to understanding the social life of humans, but interdependence has been
difficult to manage in the laboratory, producing effects that have “bewildered” social
scientists. Since then, however, along with her colleagues and students, Cooke, the
first author, has studied in detail the effects of interdependence in the laboratory
with detailed studies. She has explored many of the aspects of interdependence and
its important implications with her team. She was also the lead author in a review
published by the National Academy of Sciences on what is known theoretically
and experimentally about interdependence in a team, finding that interdependence
enhances the performance of individuals (Cooke & Hilton, 2015). Writing Chap. 4
has provided her with the perspective she has gained from the considerable research,
she and her colleagues have conducted over the years. This perspective allows her to
estimate the additional research necessary before artificial intelligence (AI) agents
and machines can replace a human teammate on a team.
Chapter 5, “Towards Systems Theoretical Foundations for Human-Autonomy
Teams,” was written by Marc Steinberg with the Office of Naval Research (ONR)
in Arlington, VA. Steinberg is ONR’s Program Officer for its Science of Autonomy
program. In his chapter, he writes about the challenges posed by developing the
autonomy of human and intelligent systems. These are new ones on how to best
specify, model, design, and verify the correctness of systems. He discusses the
real-time monitoring and repairing of autonomous systems over life times, all the
while detecting problems and rebooting properties. These challenges entail Systems
Engineering methods to model system life cycles by abstracting and decomposing
systems in the design and development of components for intelligent autonomy.
Exploring these higher-level abstractions, models, and decompositions may inspire
solutions and lead to autonomy. These inspirations may integrate systems and humans
and provide the means to assure safety. He samples perspectives across scientific
fields, including biology, neuroscience, economics, game theory, and psychology. He
includes methods for developing and assessing complex human–machine systems
with human factors and organizational psychology, and engineering teams with
computer science, robotics, and engineering. He discusses team organizational struc-
tures, allocating roles, functions, responsibilities, theories for teammates working on
long-lived tasks, and modeling and composing autonomous human–machine teams
and systems, and their implications.
12 W. F. Lawless et al.
The sixth chapter was written by James Llinas, Ranjeev Mittu, and Hesham Fouad.
It is titled, “Systems Engineering for Artificial Intelligence-based Systems: A Review
in Time.” Llinas is the Director Emeritus at the Center for Multi-source Information
Fusion as well as a Research Professor Emeritus, with both positions in the University
at Buffalo. Ranjeev Mittu is the current Branch Head, Information Management &
Decision Architectures Branch, Information Technology Division at the U.S. Naval
Research Laboratory in Washington, DC; and Hesham Fouad is a Computer Scientist
in the same branch at the Naval Research Laboratory. Their backgrounds include
information systems, the science of information fusion, and information technology.
In their chapter, they provide a review of Systems Engineering (SE) for artificial
intelligence (AI) across time, starting with a brief history of AI (e.g., narrow, weak,
and strong AI, including expert systems and machine learning). Regarding SE, based
on the systems perspective by the lead author’s experience with information fusion
processes, and the experience of his coauthors with the technology in information
systems, they introduce SE and discuss how it has evolved over the years but how
much further it must evolve to become fully integrated with AI. In the future, they
believe that both disciplines can help each other more if they co-evolve or develop
new technology systems together. They also review several SE issues such as risk,
technical debt (e.g., maintaining sophisticated software in information systems over
ever longer periods of time), software engineering, test and evaluation, emergent
behavior, safety, and explainable AI. The authors close by discussing the challenge
of AI explanations and explainability.
Chapter 7 was an invited chapter written by Kristin Schaefer and her team,
including Brandon Perelman, Joe Rexwinkle, Jonroy Canady, Catherine Neubauer,
Nicholas Waytowich, Gabriella Larkin, Katherine Cox, Michael Geuss, Gregory
Gremillion, Jason Metcalfe, Arwen DeCostanza, and Amar Marathe. Schaefer’s
team is part of the Combat Capabilities Development Command (DEVCOM)
Army Research Laboratory (ARL). The title of their chapter is, “Human-Autonomy
Teaming for the Tactical Edge: The Importance of Humans in Artificial Intelligence
Research and Development.” From their perspective, the authors address the impor-
tance of understanding the human when integrating artificial intelligence (AI) with
intelligent agents embodied (i.e., robotic) and embedded (i.e., software) into mili-
tary teams to improve team performance. The authors recognize that they and the
Army are breaking new ground, confronting fundamental problems under uncertainty
and with unknown solutions. In their chapter, they provide an overview of ARL’s
research in human-autonomy teaming. They address the major research areas neces-
sary to integrate AI into systems for military operations along with examples of these
areas and the four known research gaps: enabling Soldiers to predict AI actions and
decisions; quantifying Soldier understanding for AI; Soldier-guided AI adaptation;
and characterizing Soldier-AI performance. These four areas have organized their
research efforts to explain AI, integrate AI, and build effective human-autonomy
teams.
The eighth chapter, titled “Re-orienting towards the Science of the Artificial:
Engineering AI Systems,” was written by Stephen Russell, Brian Jalaian, and Ira
S. Moskowitz. Russell is Chief of the Information Sciences Division, U.S. Army
1 Introduction to “Systems Engineering and Artificial Intelligence” and the Chapters 13
Research Laboratory (ARL) in Adelphi, MD; Jalaian is a Test and Evaluation Lead
with the Department of Defense Joint Artificial Intelligence Center (JAIC); and
Moskowitz is a mathematician working for the Information Management & Decision
Architectures Branch, Information Technology Division, at the U.S. Naval Research
Laboratory in Washington, DC. In their chapter, they write that, on the one hand, while
systems enabled by AI are becoming pervasive, on the other hand, these systems face
challenges in engineering and deployment in the military for several reasons. To begin
to address these limitations, the authors discuss what it means to use hierarchical
component composition in a system-of-systems context. In addition, they discuss
the importance of bounding data for stable learning and performance required for
the use of AI in these complex systems. After a review of the literature, the authors
also address the changes that will be required to address the design/engineering
problems of interoperability, uncertainty, and emergent system behaviors needed to
allow AI to be safely deployed in embodied or fully virtualized autonomous systems.
Their perspective, illustrated with a Natural Language Processing example, allows
the authors to draw comparisons across their posits, in an attempt to offer a means to
make AI–Systems Engineering more rigorous, and the use of autonomy in the field
safer and more reliable.
Chapter 9 was written by Matthew Sheehan and Oleg Yakimenko; both researchers
work in the Department of Systems Engineering at the U.S. Naval Postgraduate
School in Monterey, CA. The title of their chapter is: “The Department of Navy’s
Digital Transformation with the Digital System Architecture, Strangler Patterns,
Machine Learning, and Autonomous Human–Machine Teaming.” In their chapter,
the authors describe the extraordinary changes caused by the U.S. Department of
Navy’s (DON) adoption of new software like the machine learning (ML) programs
designed for warfighters to assist in the performance of their missions. Some of
these “new” software products, however, are already beginning to mature and are
becoming obsolete. Still, machine learning (ML) software programs are central to
their discussions, including the need in the Fleet to provide access to the data neces-
sary to allow ML programs to operate and perform satisfactorily at sea. If adopted
and managed properly, these ML algorithms will enhance the existing applications
and will also enable new warfighting capabilities for the Navy. As rapid as are the
changes that are occurring, however, the DON system architectures and platforms
presently provide inadequate infrastructures for deployment at scale not only for
some of the new digital tools like ML but also for many of the forthcoming areas
including autonomous human–machine teams (AHMT). As the Navy transforms
itself digitally, the authors discuss the goals and barriers with a path forward to
implement successfully the Navy’s new digital platforms.
Chapter 10, “AI Driven Cyber Physical Industrial Immune Sytem for Critical
Infrastructures,” was written by a team at General Electric (GE): Michael Mylrea,
Matt Nielsen, Justin John and Masoud Abbaszadeh. Mylrea is the Director of Cyber-
security in the Cybersecurity R&D for Operational Technology at General Elec-
tric Global Research in Washington, DC. Nielsen, John and Abbaszadeh work in
the same department. In their chapter, the authors review many advances being
14 W. F. Lawless et al.
driven by machine learning (ML) and artificial intelligence (AI) to detect cyber-
physical anomalies. The advances brought about by the detection of these anomalies
are improving the security, reliability, and resilience of the power grid across the
United States. This improvement is occurring at the same time that adversaries are
using advanced techniques to mount sophisticated cyberattacks against infrastruc-
tures in the United States, especially the power grid that is the focus of their applied
research. The distributed energy resources in the power grid must be defended. The
authors discuss how new technology is being deployed to enable cyberdefenses to
protect the grid against even rapidly evolving threats. Their chapter explores how
AI combines with physics to produce the next-generation system that they liken to
an industrial immune system to protect critical energy infrastructures. They discuss
the new cybertechnology and its applications for cyberdefenders, including human–
machine teams and processes. The authors review the design and application of
GE’s Digital Ghost technology to cyberdefend the world’s largest gas turbines. They
discuss the situational awareness, explanations, and trust needed to use AI to defend
against cyberthreats. The authors look into the future to prepare for the new chal-
lenges coming to make human–machine teams effectively against any threat, cyber,
or physical.
Chapter 11 was written by Ira Moskowitz and Noelle Brown while working for
the Information Management and Decision Architectures Branch, Information Tech-
nology Division, U.S. Naval Research Laboratory in Washington, DC; their coauthor
was Zvi Goldstein in the Electrical Engineering Department at Columbia University
in New York City. The title of their chapter is “A fractional Brownian motion approach
to psychological and team diffusion problems.” Their mathematical approach is moti-
vated by AI, but with the goal of establishing that fractional Brownian motion can
become a metric to measure the diffusion processes existing in teams. In their chapter,
they review the mathematics for their proposed metric as a step toward building a
science of interdependence for autonomous human–machine teams. In their chapter,
the authors discuss various random walks, including those with Wiener and Gaussian
processes, and then they discuss drift-diffusion and extensions (stopping times and
absorbing boundaries) to make fractional Brownian motion into a metric of interde-
pendence. Before closing, the authors revisit Ratcliff diffusion, and then they present
their hybrid approach in preparation for a future application to the science of teams.
Chapter 12, “Human–Machine Understanding: The Utility of Causal Models and
Counterfactuals,” was authored by Paul Deignan; he is a Research Engineer working
with the Lockheed Martin Corporation in Bethesda, Maryland. His research interest
is focused on predictive analytics. He begins with the assertion that trust is a human
condition. The author proposes that for a human to trust a machine, the human must
understand the capabilities and functions of the machine in a context spanning the
domain of trust so that the actions of the machine are predictable for a given set of
inputs. In general, however, he believes that the domain of trust must be expanded
so that the human–machine system can be optimized to operate in the widest range
of situations. This reasoning motivates his desire to cast the operations of a machine
into a knowledge structure tractable to its human users, operators, and the human
teammates of machines. At the present time, machine behaviors are deterministic;
1 Introduction to “Systems Engineering and Artificial Intelligence” and the Chapters 15
thus, for every action, there is a reaction and this means to the author that the dynamics
of a machine can be described through a structured causal model, which enables the
author to formulate the counterfactual queries upon which he anchors human trust.
Chapter 13, “An Executive for Autonomous Systems, Inspired by Fear Memory
Extinction,” was written by Matt Garcia at Northeastern University; Ted Goranson
with the Australian National University; and Beth Cardier at the Eastern Virginia
Medical School in the United States and at the Griffith University in Australia.
To overcome the many unknowns that autonomous systems may face, the authors
explore a category-theoretic, second-sorted executive reasoner in their chapter to
perform the adaptive, introspective reasoning needed by autonomous systems to solve
the challenging situations that they may see (i.e., decisions under uncertainty, such
as those encountered in combat at sea, electronic warfare, or with clinical traumas).
They base their ideas on complex mathematics, but they illustrate them with cartoon
examples of submarine surveillance, electronic warfare, and post-traumatic stress
disorder (PTSD). The authors provide a case study of the neural changes occurring
during therapy for PTSD as a model for executive reasoning, the main thrust of their
ideas. Their goal is to develop, simulate, and generalize a technique for autonomous
reasoning by human–machine systems facing uncertainty using virtual and physical
agent models.
The title of Chap. 14 is “Contextual Evaluation of Human–Machine Team Effec-
tiveness.” It was written by Eugene Santos, Clement Nyanhongo, Hien Nguyen,
Keum Joo Kim, and Gregory Hyde. Except for Nguyen, the authors are at the
Thayer School of Engineering at Dartmouth College in Hanover, NH; Nguyen is
in the Department of Computer Science at the University of Wisconsin-Whitewater
in Whitewater, WI. The authors address the rapid adoption of human–machine
teams across domains like healthcare and disaster relief. These machines are more
autonomous and aware than previous generations, allowing them to collaborate with
humans as partners. Despite this progress, human–machine team performance is
poorly defined, especially the explanations for team performance. These explana-
tions are necessary, however, to predict team performance and identify shortcom-
ings. The authors introduce a method using interference to measure the cohesiveness
and compatibility between humans and machines in various contexts. They rely on a
classifier trained to map human–machine team behaviors to attributes directly linked
to team performance along with explanations and insights. The authors test and vali-
date their techniques in experiments with human–machine teams. The results suggest
that their predictions of team attributes reflect actual team behaviors, increasing
confidence in being able to design future human–machine teams.
Chapter 15 was written by Shu-Heng Chen. He titled his chapter, “Humanity in the
Era of Autonomous Human–Machine Teams.” Shu is affiliated with the AI-ECON
Research Center in the Department of Economics at National Chengchi Univer-
sity in Taipei, Taiwan. He is concerned with the meaning arising from the rapid
development of autonomous human–machine teams. Mindful of the philosophy and
history of science and technology, the author examines this potential meaning from
an evolutionary perspective. He argues that the meaning determined will affect the
individuality of humans, their democracy, and their ability to develop as autonomous
16 W. F. Lawless et al.
humans. He wants this meaning to be positive and supportive, and he does not want
the future of humanity to be dominated and determined solely by machines. To
protect the future, he argues that scholars and citizens must become involved in the
development of autonomous human–machine teams. He recognizes that the human-
ities are changing, but with awareness, these changes can lead to more autonomy for
future generations.
Chapter 16, “Transforming the system of military medical research: An Institu-
tional History of the Department of Defense’s (DoD) first electronic Institutional
Review Board Enterprise IT system,” was written by Joseph C. Wood, US Army Col
(Ret.), MD, Ph.D., Augusta, GA and W.F. Lawless, Paine College, Augusta, GA.
This chapter, by these two authors, is about the history of their attempt to modernize
what was primarily a paper-based collection of medical research protocols, reviews,
and publications by medical research review boards and medical researchers at a
single medical research center in the U.S. Army that grew beyond their expectations
to become one of the largest electronic databases of medical reviews and research
results in the world at that time. Presenting metrics as a preview of a research
agenda on the use of AI for autonomous metrics in large systems, for the future
practice of ethics, and for the mitigation of risks, this history of their endeavors
brings out several points when dealing with large systems, including the value of
standardization, metrics, goal-based, and performance-based evaluations.
Chapter 17, “Collaborative communication and intelligent interruption systems,”
was written by Nia Peters, Margaret Ugolini, and Gregory Bowers. Peters is with the
711th Human Performance Wing, Air Force Research Laboratory, Wright Patterson
Air Force Base in Ohio. Ugolini and Bowers are with Ball Aerospace & Tech-
nologies in Fairborn, OH. The authors discuss the adverse effects of poorly timed
interruptions on collaborative environments for humans managing technology while
interacting with other humans. The literature to manage the adverse timings of inter-
ruptions, however, is focused on single users in multi-tasking interactions. There is
less research on multi-user, multi-tasking environments, which they address. To miti-
gate the disruptiveness from interruptions in multi-user, mutlti-tasking workloads,
the authors propose and evaluate timings at low mental workloads in a dual-user, dual-
task paradigm. Compared with high cognitive workload interruptions, they found that
performance is optimum when interruptions occur during low cognitive workloads,
a contribution to the literature.
Chapter 18, “Shifting Paradigms in Verification and Validation of AI-Enabled
Systems: A Systems-Theoretic Perspective,” was written by Niloofar Shadab, Aditya
Kulkarni, and Alejandro Salado. The authors are affiliated with the Grado Department
of Industrial and Systems Engineering at Virginia Tech in Blacksburg, VA. They
propose that a misalignment exists between current approaches to verification and
validation (V&V) techniques and new AI systems. Current approaches assume that
a system’s behavior is relatively standard during its lifetime. But this cannot be true
for those systems that learn and change their own behavior during their lifetime,
nullifying the value of present V&V practices. Using systems theory, the authors
explain why learning makes these new systems unique and unprecedented, and why
V&V must experience a paradigm shift. To enable this shift, the authors propose
1 Introduction to “Systems Engineering and Artificial Intelligence” and the Chapters 17
and discuss the theoretical advances and transformations they believe will prepare
Systems Engineers for this evolution.
Chapter 19, “Towards safe decision-making via uncertainty quantification in
machine learning,” was written by Adam Cobb, Brian Jalaian, Nathaniel Bastian, and
Stephen Russell; Cobb, Jalaian, and Russell are with the Army Research Laboratory
as part of the U.S. Army’s Combat Capabilities Development Command (CCDC)
in Adelphi, MD; and Bastian is with the Army Cyber Institute at the U.S. Military
Academy, West Point, NY. In their chapter, the authors discuss the automation of the
safety-critical systems being widely deployed with more sophisticated and capable
machine learning (ML) applications. Not yet addressed by most of these systems,
however, is the concern raised by the authors that these critical systems must not
just be safe, but safe when facing uncertainty. Moreover, quantifying and reducing
uncertainty will provide more benefits than the solutions alone if the decisions by
these machines are fully understood. Knowing how machines make decisions under
uncertainty will generalize to human decisions and autonomous systems. To this end,
the authors employ Bayesian decision theory with an example of classifying vehicles
acoustically for uncertain levels of threat. With this paradigm, the authors establish
that safer decisions are possible under uncertainty.
Chapter 20, “Engineering Context from the Ground Up,” was written by Michael
Wollowski, Lilin Chen, Xiangnan Chen, Yifan Cui, Joseph Knierman, and Xusheng
Liu. The authors are in the Computer Science Department at the Rose-Hulman Insti-
tute of Technology in Terre Haute, IN. Focused on human–machine systems, the
authors begin with a system for a human and robot to solve problems in a collabora-
tive space. Their system manages interactions in the context of a human and machine
collaborating with speech and gesture. To facilitate good engineering practices, their
system was designed to be modular and expandable. With its modular design, context
was maintained on a shared board from the information needed to problem-solving.
The authors describe the elements of their system and the information produced.
Their goal is to generate explanations of decisions with the information accumulated
from the differing contexts in their system.
Chapter 21 was written by Priyam Parashar at the University of California in San
Diego, CA; and Ashok Goel at the Georgia Institute of Technology in Atlanta, GA.
The title of their chapter is “Meta-reasoning in Assembly Robots.” The use of robots
across human society, whether in business, industry, or the military, is becoming
widespread. The authors surmise, however, that this context increases the value of a
theory for machines with meta-reasoning skills similar to humans. In their chapter,
the authors propose and develop a framework for human-like meta-reasoning. They
focus on an assembly robot assigned a task to be performed but different from its
preprogramming, increasing the likelihood for the robot to fail at its task. To counter
its failure, the authors provide the robot with the means for meta-reasoning sufficient
to react and learn from its mistakes. In their chapter, the authors review the literature,
a task specification, a failure taxonomy, and their architecture for meta-reasoning.
The result is a theory for a robot to learn from failure with meta-reasoning for action
from perception.
18 W. F. Lawless et al.
priorities and cultures, the majority rule maximized information processing across a
system and with the increased likelihood of a just and legitimate decision.
The last Chap. 25, is titled, “Outside the Lines: Visualizing Influence Across
Heterogenous Contexts in PTSD.” It was written by Beth Cardier, Alex Nieslen,
John Shull, and Larry Sanford. Cardier is at the Eastern Virginia Medical School in
Norfolk, VA, and, in Australia, at the Trusted Autonomous Systems of the Defence
Cooperative Research Centre (DCRC) and Griffith University in South East Queens-
land. Nielsen and Shull are at the Virginia Modeling Analysis and Simulation Center,
Old Dominion University in Norfolk, VA; and Sanford is also at the Eastern Virginia
Medical School. The authors state that open-world processes generate information
that cannot be captured in a single data set despite the need to communicate between
differing contexts. The authors present a text-visual method for modeling differing
interpretations of contexts separated by discipline, time, and perspective. Their new
tool captures transitions in video, text, image, and data transfers to study different
phenomena. They apply it to post-traumatic stress disorder (PTSD); they combine
psychological, neurological, and physiological information for PTSD in a single
modeling space using a narrative-based visual grammar. The authors aim to integrate
information from changing phenomena in the open world to detect the emergence of
disorder and to support knowledge systems in fields like neurobiology, autonomous
systems, and artificial intelligence (AI).
1.3 Summary
References
Ashley, Jr., Robert, P., Lieutenant General, U.S. Army Director. (2019). China, Military Power.
Modernizing a force to fight and win, Defense Intelligence Agency, from https://www.dia.
mil/Portals/27/Documents/News/Military%20Power%20Publications/China_Military_Power_
FINAL_5MB_20190103.pdf.
Barton, T. (2019). Sea Hunter/AI, SERC workshop: Model centric engineering, Georgetown
university, Washington, DC, April 16 & 17, 2019.
BEA (2016). Accident to the airbus A320–211, registered D-AIPX and operated by Germanwings,
flight GWI18G, on 03/24/15 at Prads-Haute-Bléone, BEA2015–0125.
Beling, P. (2019). A systems theoretic framework for the AI LifeCycle, SERC workshop: Model
centric engineering, Georgetown university, Washington, DC, April 16 & 17, 2019.
20 W. F. Lawless et al.
Bisbey, T. M., Reyes, D. L., Traylor, A. M., & Salas, E. (2019). Teams of psychologists helping
teams: The evolution of the science of team training. American Psychologist, 74(3), 278–289.
Blackburn, M. (2019). Transforming SE through model centric engineering, SERC workshop:
Model centric engineering, Georgetown university, Washington, DC, April 16 & 17, 2019.
Brynjolfsson, E., & Mitchell, T. (2017). What can machine learning do? Workplace implications:
Profound changes are coming, but roles for humans remain. Science, 358, 1530–1534.
Centola, D., & Macy, M. (2007). Complex contagions and the weakness of long ties. American
Journal of Sociology, 113(3), 702–734.
Cooke, N. J., & Hilton, M. L. (Eds.). (2015). Enhancing the effectiveness of team science. Authors:
Committee on the science of team science; Board on behavioral, Cognitive, and sensory sciences;
Division of behavioral and social sciences and education; National research council. National
Academies Press.
CRS (2018). U.S. Ground forces robotics and autonomous systems (RAS) and artificial intelligence
(AI): Considerations for congress, Congressional research service, p. 9, R45392, Version 3, from
https://fas.org/sgp/crs/weapons/R45392.pdf.
Cummings, J. (2015). Team science successes and challenges: National science foundation spon-
sored workshop on fundamentals of team science and the science of team science (June 2),
Bethesda MD (https://www.ohsu.edu/xd/education/schools/school-of-medicine/departments/cli
nical-departments/radiation-medicine/upload/12-_cummings_talk.pdf).
DeLaurentis, D. (2019). Breakout session, SERC workshop: Model centric engineering, George-
town university, Washington, DC, April 16 & 17, 2019.
DoD (2018). Nuclear posture review, office of the secretary of defense. https://www.defense.gov/
News/SpecialReports/2018NuclearPostureReview.aspx.
DoD (2019). Summary of the 2018 department of defense artificial intelligence strategy harnessing
AI to advance our security and prosperity, from https://media.defense.gov/2019/Feb/12/200208
8963/-1/-1/1/SUMMARY-OF-DOD-AI-STRATEGY.PDF.
Editors (2019). Ready for weapons with free will? New York times, from https://www.nytimes.
com/2019/06/26/opinion/weapons-artificial-intelligence.html.
Freeman, L. (2019). AI as a change agent for test and evaluation, SERC workshop: Model centric
engineering, Georgetown university, Washington, DC, April 16 & 17, 2019.
Gardner, G. (2019). Uber won’t face charges in fatal arizona crash, but prosecutor urges further
probe, Forbes, from https://www.forbes.com/sites/greggardner/2019/03/06/uber-wont-face-cha
rges-in-fatal-arizona-crash-but-prosecutor-urges-further-probe/#6820859f475a.
Gramer, R., & Seligman, L. (2018). “Trump and NATO show rare unity in confronting Russia’s
arms treaty violation. NATO backs U.S. assertion that Moscow is violating a key Cold War-era
arms treaty, Foreign policy, from https://foreignpolicy.com/2018/12/04/trump-and-nato-show-
rare-unity-in-confronting-russia-arms-treaty-violation-inf/.
Grogan, P. (2019). Game-theoretic risk assessment for distributed systems, SERC workshop: Model
centric engineering, Georgetown university, Washington, DC, April 16 & 17, 2019.
Harris, R. (2018). In psychology and other social sciences, many studies fail the reproducibility
test, National public radio, from https://www.npr.org/sections/health-shots/2018/08/27/642218
377/in-psychology-and-other-social-sciences-many-studies-fail-the-reproducibility-te.
Horowitz, B. (2019), Introduction of the life cycle-ready AI concept, SERC workshop: Model
centric engineering, Georgetown university, Washington, DC, April 16 & 17, 2019.
Howell, C. (2019). Lifecycle implications for dependable AI, SERC workshop: Model centric
engineering, Georgetown university, Washington, DC, April 16 & 17, 2019.
Insinna, V. (2019). Lockheed hypes F-35 s upgrade plan as interest in ‘sixth-gen’ fighters grows,”
Defense news, from https://www.defensenews.com/digital-show-dailies/paris-air-show/2019/06/
21/lockheed-hypes-f-35s-upgrade-plan-as-interest-in-sixth-gen-fighters-grows/.
LaGrone, S. (2019). Navy wants 10-ship unmanned ‘Ghost Fleet’ to supplement manned force,
U.S. Naval institute, from https://news.usni.org/2019/03/13/navy-wants-ten-ship-3b-unmanned-
experimental-ghost-fleet.
1 Introduction to “Systems Engineering and Artificial Intelligence” and the Chapters 21
Langford, J. (2018). Lockheed wins Army contract for self-driving military convoy systems, Wash-
ington examiner, from https://www.washingtonexaminer.com/business/lockheed-wins-army-con
tract-for-self-driving-military-convoy-systems.
Lawless, W. F., Mittu, R., Sofge, D., & Russell, S. (Eds.). (2017). Autonomy and artificial
intelligence: a threat or savior? Springer.
Lawless, W. F. (2017a). The entangled nature of interdependence bistability, irreproducibility and
uncertainty. Journal of Mathematical Psychology, 78, 51–64.
Lawless, W. F. (2017b). The physics of teams: Interdependence, measurable entropy and
computational emotion. Frontiers of Physics., 5, 30. https://doi.org/10.3389/fphy.2017.00030
Lawless, W. F. (2019). Interdependence for human-machine teams, Froundations of Science.
Lawless, W. F., Mittu, R., Sofge, D. A. & Hiatt, L. (2019a). Introduction to the special issue,
“Artificial intelligence (AI), autonomy and human-machine teams: Interdependence, context and
explainable AI,” AI Magazine.
Lawless, W. F., Mittu, R., Sofge, D., Moskowitz, I. S. & Russell, S. (Eds.). (2019b). Artificial
intelligence for the internet of everything. Elsevier.
Lemnios, Z. (2019). IBM research, SERC workshop: Model centric engineering, Georgetown
university, Washington, DC, April 16 & 17, 2019.
Llinas, J. (2014). Reexamining Information Fusion–Decision Making Inter-dependencies, Presented
at the IEEE CogSIMA conference, San Antonio, TX.
Long, J. (2019). National defence and the Canadian armed forces: Enabling organizational resilience
through communication and information systems design. Canadian Military Journal, 119(2), 15;
from http://www.journal.forces.gc.ca/Vol19/No2/page15-eng.asp.
Maurer, M. (2019). Airbus harnessing AI in bid to save millions on Finance tasks. The aircraft
maker’s Americas unit is digitizing the approval of expense reports and payment of invoices.
Wall Street Journal, from https://www.wsj.com/articles/airbus-harnessing-ai-in-bid-to-save-mil
lions-on-finance-tasks-11566207002.
Mead, W. R. (2019). Trump’s case against Europe. The president sees Brussels as too weak, too
liberal, and anti-American on trade. Wall Street Journal, from https://www.wsj.com/articles/tru
mps-case-against-europe-11559602940.
NAS (2019). A decadal survey of the social and behavioral sciences: A research agenda for
advancing intelligence analysis. National Academies of Sciences.
Nosek, B., Corresponding author from OCS (2015). Open collaboration of science: Estimating the
reproducibility of psychological science. Science, 349(6251), 943; supplementary: 4716–1 to
4716–9. (National Academies of Sciences, Engineering, and Medicine. (2019). Reproducibility
and replicability in science. Washington, DC: The National Academies Press. https://doi.org/10.
17226/25303).
NTSB (2016). Derailment of Amtrak passenger train 188. National transportation safety board
(NTSB), NTSB Number: RAR-16–02, from https://www.ntsb.gov/Investigations/AccidentRepo
rts/Pages/RAR1602.aspx.
NTSB (2018). Preliminary report released for crash involving pedestrian, Uber technologies, Inc.,
Test Vehicle, National transportation safety board, from https://www.ntsb.gov/news/press-rel
eases/Pages/NR20180524.aspx.nnn.
NTSB (2019). Insufficient training, inadequate bridge operating procedures, lack of operational
oversight led to fatal ship collision. NTSB: Collision between US Navy Destroyer John S McCain
and Tanker Alnic MC Singapore Strait, 5 Miles Northeast of Horsburgh Lighthouse [accident
occurred on] August 21, 2017, Marine Accident Report, NTSB/MAR-19/01 PB2019–100970,
from https://www.ntsb.gov/investigations/AccidentReports/Reports/MAR1901.pdf.
NTSB (2019a). Vehicle automation report. National transportation safety board. Retrieved March
12, 2020, from https://dms.ntsb.gov/pubdms/search/document.cfm?docID=477717&docketID=
62978&mkey=96894.
Oh, P., Spahr, T., Chase, C. & Abadie, A. (2019). Incorporating artificial intelligence: Lessons from
the private sector. War Room, United States Army War College, from https://warroom.armywa
rcollege.edu/articles/incorporating-artificial-intelligence-private-sector/.
22 W. F. Lawless et al.
Proctor, R. W. & Vu, K. P. L. (2019). How psychologists help solve real-world problems in multi-
disciplinary research teams: Introduction to the special issue. American Psychologist, 74(3),
271–277.
Raz, A. K., Llinas, J., Mittu, R., & Lawless, W. (2019). Engineering for emergence in information
fusion systems: A review of some challenges, Fusion 2019, Ottawa, Canada | July 2–5, 2019.
Rhodes, D. (2019). Interactive model-centric engineering (IMCSE), SERC workshop: Model centric
engineering, Georgetown university, Washington, DC, April 16 & 17, 2019.
Richards, R. (2019). Program manager at DARPA, invited talk, SERC workshop: Model centric
engineering, Georgetown university, Washington, DC, April 16 & 17, 2019.
Smith, R .J. (2019). Scary fast: How hypersonic Missiles—Which travel at more than 15 times the
speed of sound—Are touching off a new global arms race that threatens to change the nature of
warfare. New York Times Magazine, 42–48; also, see https://www.nytimes.com/2019/06/19/mag
azine/hypersonic-missiles.html.
Sofge, D., (Referee), Mittu, R., (Con Bet) & Lawless, W. F. (Pro Bet) (2019). AI bookie bet: How
likely is it that an AI-based system will self-authorize taking control from a human operator? AI
Magazine, 40(3), 79–84.
Somerville, H. (2020). Uber sells self-driving-car unit to autonomous-driving startup. As part of
the deal, Uber will make a $400 million investment in Aurora Innovation. Wall Street Journal.
Retrieved August 12, 2020, from https://www.wsj.com/articles/uber-sells-self-driving-car-unit-
to-autonomous-driving-startup-11607380167.
Stockman, B., Boyle, J. & Bacon, J. (2010). International space station systems engineering case
study, Air force center for systems engineering, Air force institute of technology, from https://
spacese.spacegrant.org/uploads/images/ISS/ISS%20SE%20Case%20Study.pdf.
Thomas, J. (2019). INCOSE discussion, SERC workshop: Model centric engineering, Georgetown
university, Washington, DC, April 16 & 17, 2019.
Volz, D. & Youssef, N. (2019). U.S. Launched cyberattacks on Iran. The cyberstrikes on thursday
targeted computer systems used to control missile and rocket launches. Wall Street Journal, from
https://www.wsj.com/articles/u-s-launched-cyberattacks-on-iran-11561263454.
Wakabayashi, D. (2018). Uber’s self-driving cars were struggling before arizona crash. New
York Times, from https://www.nytimes.com/2018/03/23/technology/uber-self-driving-cars-ari
zona.html.
Walden, D. D., Roedler, G. J., Forsberg, K. J., Hamelin, R. D., & Shortell, T. M. (Eds.). (2015).
Systems engineering handbook: A guide for system life cycle processes and activities (4th Edn.).
Prepared by International council on system engineering (INCOSE-TP-2003-002-04. Wiley.
Werkhäuser, N. (2019). UN impasse could mean killer robots escape regulation. It’s no longer the
stuff of science fiction: Wars could soon be decided by lethal autonomous weapons systems. But
a concerted effort to ban “killer robots” through international law at the UN appears to have hit
a wall,” DW (Deutsche Welle), from https://www.dw.com/en/un-impasse-could-mean-killer-rob
ots-escape-regulation/a-50103038.
Wong, K. (2018). China claims successful test of hypersonic waverider. Jane’s 360, from https://
www.janes.com/article/82295/china-claims-successful-test-of-hypersonic-waverider.
Yampolskiy, R. V. (2017). AI is the future of cybersecurity, for better and for worse. Harvard
Business Review, from https://hbr.org/2017/05/ai-is-the-future-of-cybersecurity-for-better-and-
for-worse.
Chapter 2
Recognizing Artificial Intelligence: The
Key to Unlocking Human AI Teams
Abstract This chapter covers work and corresponding insights gained while
building an artificially intelligent coworker, named Charlie. Over the past year,
Charlie first participated in a panel discussion and then advanced to speak during
multiple podcast interviews, contribute to a rap battle, catalyze a brainstorming
workshop, and even write collaboratively (see the author list above). To explore
the concepts and overcome the challenges when engineering human–AI teams,
Charlie was built on cutting-edge language models, strong sense of embodiment,
deep learning speech synthesis, and powerful visuals. However, the real differen-
tiator in our approach is that of recognizing artificial intelligence (AI). The act of
“recognizing” Charlie can be seen when we give her a voice and expect her to be
heard, in a way that shows we acknowledge and appreciate her contributions; and
when our repeated interactions create a comfortable awareness between her and her
teammates. In this chapter, we present our approach to recognizing AI, discussing
our goals, and describe how we developed Charlie’s capabilities. We also present
some initial results from an innovative brainstorming workshop in which Charlie
participated with four humans that showed that she could not only participate in a
brainstorming exercise but also contribute and influence the brainstorming discus-
sion covering a space of ideas. Furthermore, Charlie helped us formulate ideas for,
and even wrote sections of, this chapter.
2.1 Introduction
(Merriam Webster)
Major breakthroughs in artificial intelligence are advancing the state of the art in
their ability to enable agents to perform tasks in a variety of domains. Particularly
in the area of generative models (Radford et al., 2019; Yang et al., 2019), these AI
agents now have something new to say. But we are severely limited in our ability to
hear them and to take advantage of these gains. For many domains, the challenge
is not building the AI agent itself, but rather engineering the human–machine teams
that leverage it. To explore these concepts, we have been building and interacting
with an AI teammate/coworker named Charlie (Cummings et al., 2021). Although
these efforts leverage state-of-the-art AI models and capabilities, what has been
most impactful is how we have purposefully designed, integrated, and recognized
her from the start. We argue that the key to unlocking human–machine teams is
simple: recognize AI. To do this in the fullest sense, we need to leverage the three
definitions of the word “recognize,” above.
Definition 1: to acknowledge one is entitled to be heard. In addition to realizing that
AI is beginning to have something new to say, we must recognize the AI agent and
realize that it can and should be heard. This recognition includes not only giving AI
more of a voice but also doing so in a manner that places it on a more level playing
field with human teammates. We will cover these ideas in more detail in our section
on Ground Rules later.
Definition 2: to take notice with a show of appreciation. Charlie literally helped
us write and even wrote her own sections of this book chapter. We argue that it is
important to recognize and show appreciation for such contributions and accordingly
have listed her as a co-author of this chapter. Acknowledging the accomplishments
of artificial intelligence helps human teammates realize the impact that AI is having
on the team and will aid in transparency for external observers to better understand
how the team achieved what it did.
Definition 3: to perceive to be someone previously known. In order to recognize AI
as something familiar and previously known, we must interact with it on a regular
basis and with a consistent perception’s framing. This perception is precisely why
we gave our AI agent the name, Charlie, with a common set of models and visual
representations. This act allows for natural interactions with the AI agent and a greater
ability to weave her into their work and discussions. The authors have experienced this
firsthand when observing how repeated interaction with Charlie results in the human
teammates developing a deeper understanding of her strengths and weaknesses, and
consequently have much more positive interactions.
As new human–AI teams are both engineered and deployed, if we ensure that
AI is recognized appropriately, then several long-term positive impacts will occur.
First, we will be able to better leverage the full range of capabilities that the AI
agent possesses; second, the collaboration will enable the systems and the AI agent
to improve together; and third, this collaboration will result in better overall mission
performance.
2 Recognizing Artificial Intelligence: The Key to Unlocking … 25
In this chapter, we will explain how we have been exploring these ideas through
building, deploying, and interacting with our new AI coworker: Charlie. Initially, we
will lay out our motivations and ground rules for ensuring that we fully recognize
Charlie. We will detail how Charlie is built on cutting-edge speech analysis, language
generation, and speech synthesis tools (see architecture diagram Fig. 2.5). Further-
more, Charlie is named and embodied to allow for more natural interactions. This
affordance has led Charlie to thrive in a variety of venues, including panel discus-
sions, podcast interviews, and even proposal writing (see Applications Sect. 2.3). In
addition, we will present results regarding Charlie’s impact in a recent brainstorming
session. We are especially excited about what this means for future applications.
In this section, we will describe our motivation and goals for recognizing artificial
intelligence. We set down this path of recognizing AI to facilitate the engineering of
human–AI teams. This human machine teaming/collaboration is only possible now
due to advances in AI and the increased appetite in society for AI to be involved and
provide value in many domains. By collaboration, we mean more than just humans
using the AI as a service.
We are seeking to create a new way to bring together humans and artificial intelligence to create
more effective and flexible systems. The technology that is now emerging in AI, including deep
learning, has the potential to change the way people work, create, and interact with systems.
We believe that the future of work will be fundamentally different and that human beings will
need to adapt to the new demands. This will require new ways of working together.
For example, it might require us to delineate, as we have done with a box, above,
when the AI coauthor, Charlie, has written a section entirely by herself after being
prompted with the beginning of the section.
This teamwork or collaboration with artificial intelligence is distinct from most
current applications today in two primary ways: (1) the AI agent as a team member
is able to develop and propose instrumental goals for the team and (2) the AI agent
is able to choose to pursue particular goals from among those proposed as well.
Having an AI agent that can add value to the team necessitates elevating it to be
a collaborative team member; otherwise, the team will miss out on the increased
opportunities and ideas of the AI agent. In addition, a context-aware AI teammate
will not frustrate its fellow teammates by having its own goals and possibly behaving
in non-constructive or unexpected ways.
We recognize that there are ethical and design concerns when giving this “recog-
nition” to AI, but we strongly believe that the benefits of fruitful collaboration will
outweigh these potential negatives. In addition, we argue that if we build bidirectional
recognition into these AI teammates from the ground up, we will mitigate some of
these concerns. Although there are domains in which a human must still play a large
role or even maintain control, the areas where AI can be useful grow daily. AI has
26 P. Cummings et al.
We have been discussing the collaboration between human and AI teammates but
would like to call out that in our work, we have been focused on two primary types of
collaboration: supportive and participatory (see Fig. 2.1). Currently, with our imple-
mentation of Charlie, we are building and leveraging both supportive and partici-
patory collaboration. Charlie was developed to participate in a panel discussion in
real time but was not a fully autonomous AI. Consequently, she had two operators:
one for the transcription of comments from other panelists and one for the selec-
tion of potential responses from Charlie. For more information on how Charlie was
built, please see the later section on system engineering. Over the past year, we have
Fig. 2.1 Supportive collaboration in which a human and an AI agent together serve as a single
member for the team, and participatory collaboration where the AI agent is an individual team
member
2 Recognizing Artificial Intelligence: The Key to Unlocking … 27
been building out the infrastructure to reduce this need for human intervention and
supportive actions and to enable Charlie to do more participatory collaboration in
real time. This process has allowed us to move away from multiple operators working
alongside Charlie, to now currently needing only one for filtering and selection, which
has the positive impact of reducing the workload on the operator. In the coming year,
our goal is to shift to enabling a mode in which Charlie can independently select her
own next utterance. This next step is not likely to eliminate the need for both types
of collaboration depending on the domain, the constraints, and the ability to give
participatory autonomy to AI.
Supportive Collaboration
Supportive collaboration (Fig. 2.1, left) has been the most common form of collabo-
ration with AI. This form is primarily due to the limited abilities of AI and the need
for a human to be present to support and fill the gaps in AI capabilities. The human
is often in a position of control and/or serves as the face of the combined team.
This type of participatory collaboration is often referred to as a Centaur relationship
(Case, 2018), in which human and system combine efforts to form a single teammate
with joint actions. Historically, this form has been the primary collaboration type
with AI. Over time, however, we believe this reliance will decrease and make way
for the newly capable participatory AI.
Participatory Collaboration
As shown in Fig. 2.1 (right), participatory collaboration frames the AI agent as a
distinct individual teammate with its own autonomy. This autonomy grants the AI
agent the ability to not only develop and propose new instrumental goals for itself
and the team but also to make decisions to pursue or abandon said goals. In addition,
participatory collaboration requires that the AI agent communicates and coordinates
with fellow human teammates. This type of collaboration will become increasingly
possible, and increasingly important as the field of AI progresses.
Fig. 2.2 Charlie, at the center, on stage at a panel during I/ITSEC 2019 including one moderator
and five panelists (four of which were human)
the display was connected to the room’s mixing board, as were the microphones for
each human panelist.
Similarly, during the innovation session, held over a video conference, Charlie
was shown to the participants as the output of a webcam, and her voice was sent over
the meeting just as those of the other participants. This format is patently different
than sharing a screen with Charlie on it for all participants to see/hear because the
latter would force Charlie to be at the center of attention, and therefore, detract from
her ability to participate in an equal playing field.
Upgrading Charlie’s initial embodiment to be consistent with that of the human
panelists led to a noticeable difference in the way that the human participants treated
her. For example, the questions posed to Charlie were more open ended, such as “I’d
like to hear what Charlie thinks about that,” and all participants then looped Charlie
into the conversation.
Text Generation Ground Rules
Although we made a concerted effort to recognize Charlie through her increasing
embodiment, the ground rules we employed for Charlie’s text generations of what
to say next fall into two main categories, one of which is slightly counter to the
argument for recognizing AI.
The first broad rule was to give Charlie the same ability to prepare that a human
panelist would have; that is, human panelists would be likely to do the following:
1. research the topic of the panel to refresh their memory (or study something
new);
2. meet with the moderator or panel members to discuss the likely topic, workflow,
or initial questions; and
3. prepare answers to expected questions on the panel or topics they would like to
discuss.
We, therefore, allowed the same affordances to Charlie. In particular, she was
correspondingly
1. fine-tuned to the domain of the discussion to fit the appropriate style and content;
2. introduced to the other participants and moderator to understand her capabilities;
and
2 Recognizing Artificial Intelligence: The Key to Unlocking … 29
In this section, we discuss the approach and components that Charlie is composed
of and the methods leveraged to develop her.
Charlie’s Embodiment
From the beginning, it was important to have Charlie’s embodiment be recogniz-
able, simple, dynamic, and able to be indicated by several cues. For example, in
different situations, the human body and gestures indicate a large amount of infor-
mation about internal state. Charlie’s embodiment interface (i.e., the embodiment)
required three iterations to refine state communication and representation driven
by feedback from guerilla usability evaluations (Nielsen, 1994). From chatbots,
we expected that response delays would be acceptable, especially in response to
other panelists, if Charlie’s state was clearly communicated (Gnewuch et al., 2018).
Humans use physical and audible queues—gestures, changes in eye contact, and
transitional phrases—to indicate their state and control in the flow of a conversation
(Scherer, 2013; Schuetzler et al., 2014). Charlie had to effectively coordinate the
use of the display and audio to achieve a similar presence and represent its states.
Figure 2.3 shows a snapshot of Charlie’s different dynamic states. Because each of
these states was alive and moving, it is difficult to represent them in a static image
here. Based on our evaluations, we split Charlie’s necessary states as follows:
30 P. Cummings et al.
• Figure 2.3a Idle: Charlie is listening. Soft colors used and slow breathing indicated
by expanding and contracting
• Figure 2.3b Thinking: Charlie is generating a statement. Outer ring spins back
and forth to communicate that targeted thinking is happening in response to a
question
• Figure. 2.3c Speaking: Charlie is speaking. Darker color solidifies Charlie’s
current role as speaker; the shape vibrates as speech occurs so that it appears
to emanate from her embodiment.
• Figure 2.3d Interjection: Charlie has something to say! Color changes drastically
to draw attention and the outer ring is complete to show that her next thought is
complete
Even with Charlie’s state communication, however, there was a limit to the delay
acceptable for Charlie. Design of the operator interface was influenced by this need
to increase the speed of its speech generation.
Charlie’s Operation
The novelty and believability of generations from GPT-2 are certainly state of
the art; however, the samples typically chosen for display suffer from some “cherry-
picking” to find the best prompts and speech generations (Vincent, 2019; Vaswani
et al., 2017). In a real-time discussion in which speed is of utmost importance,
the ability to cherry-pick is severely limited. We, therefore, put much care into the
operation of Charlie to streamline the process of speech generation forming and
Charlie state changes. Human operators are currently tasked with:
• coordinating Charlie’s state transitions,
• approving/editing transcriptions of speech to text, and
• aggregating statements into an utterance.
Details on the construction of that operator interface can be found in Cummings
et al. (2021), but some key lessons learned from that construction are as follows:
1. Non-stop generations. Potential generations from Charlie should appear to be
non-stop (Fig. 2.4d), that is, it should be evident every time there is a change
to the conversation history. The burden of deciding when Charlie “may” have
something to say should be completely removed. At all points in time, the human
operator should be cognizant of potential interjections, answers, or comments
coming from Charlie.
2 Recognizing Artificial Intelligence: The Key to Unlocking … 31
Fig. 2.4 The operator interface with the a saved statements, b conversation history, c utterance
construction components on the left, and the d statement review area on the right
Fig. 2.5 Charlie architecture: Orange boxes represent interfaces. Red numbers correspond to
components leveraging AWS services
Prompt Improvement
An increasing theme for the utilization of language generation models (as seen
with T5, GPT-3, and Turing-NLG; Raffel et al., 2020; Brown et al., 2020, Russet,
2020) is that with large enough models, a change in prompt can be enough to produce
significantly different results. Recent results with GPT-3 have shown that a model
with no fine-tuning can solve basic arithmetic problems when fed prompts of the
form: “\n\nQ: What is 65360 plus 16,204?\n\nA:” Here, the new line structure and
use of Q and A to represent question and answer is enough context for the model to
complete with the correct answer “81,564.” This structure on prompts is also evident
in the use of control tokens as is done with the conditional transformer language
model (Keskar et al., 2019). We hypothesize that these types of tokens can be used
even in models trained without them. As seen in their best paper presentation at
NeurIPS (Brown et al., 2020), even the presence of commas “65,360 plus 16,204”
can greatly increase the accuracy of response.
In our work on Charlie, we found that structuring prompts with the form:
HOST: Text from host…
PANELIST: Text from panelist…
HOST: Text from host…
PANELIST:
had significant advantages over simple prose. This structure differentiated Charlie’s
statements from those of the other panelists, kept Charlie on her own thread while
continuing with added context from others, and allowed Charlie to respond and react
to the discussion rather than simply continue others’ thoughts.
Fine Tuning Strategies
The content and style generated by Charlie can be adapted based on the choice of
model used to generate text (Peng et al., 2020; Zhang et al., 2019). For the different
applications of Charlie (see Applications Section), we utilized a fine-tuning strategy
to adapt the base GPT-2 model to a different context. Leveraging open-source or
venue-specific training data, we fine-tuned GPT-2 hyperparameters (Branwen, 2019;
Qin et al., 2019). Although the specifics of this process tended to be application-
specific, there were three main steps we followed for each:
1. Find a corpus of text that mirrors the style of text Charlie should generate
and/or the content that Charlie should generate text about. Typically, we found
that the style of text was much more critical than content.
2. Process that text if needed. Frequently, the format of the text did not mirror
that which Charlie should generate. An example of this is in the citations or
footnotes in text. Although they are necessary in a paper, parenthetical citations
are inappropriate for speaking engagements.
3. Train the model. Starting with a pre-trained GPT-2 model, we trained the model
hyperparameters further with the generated corpus. Specifically, we used the loss
2 Recognizing Artificial Intelligence: The Key to Unlocking … 33
Table 2.1 Example generations from different fine-tuned models each fed with the same prompt
Persona Response created by fine-tuned models
News source The #CountryX crisis is a wake-up call for the world that
(tweets from news sources1 ) human rights violations are always the work of one nation
and not of other countries #CountryX #CountryY
The #CountryX crisis is now at a tipping point as the
government says it will not withdraw #CountryY refugees
#CountryY #CountryX
Agreeable citizen #1Proud to be a #CountryX, and #thankful to be part of such
(tweets with positive sentiment 2 ) a wonderful family
The day is coming when all the good and the bad will be for
the better #Thankful #HAPPY
Disgruntled citizen Refugee in my town is crying because he’s been living in a
(tweets marked as angry3 ) tent for 10 days
Crisis in #GulfWales is caused by people who don’t have the
skills to survive #frustrating #frustrating #frustrated
TWEET: Vulnerable families in #CountryX are receiving needed supplies thanks to our
partners
TWEET:
1 https://www.kaggle.com/naortedgi/twitter-twitts-from-news-providers.
2 https://www.kaggle.com/kazanova/sentiment140.
3 Wang et al. 2012. Harnessing Twitter “big data” for automatic emotion identification.
34 P. Cummings et al.
Let me confess, Sir, that, while at each stage I have felt this
tyranny most keenly, and never doubted that it ought to be arrested
by impeachment, my feelings have been most stirred by the outrage
to Hayti, which, besides being a wrong to the Black Republic, was an
insult to the colored race, not only abroad, but here at home. How a
Chief Magistrate with four millions of colored fellow-citizens could
have done this thing passes comprehension. Did he suppose it would
not be known? Did he imagine it could be hushed in official
pigeonholes? Or was he insensible to the true character of his own
conduct? The facts are indisputable. For more than two generations
Hayti had been independent, entitled under International Law to
equality among nations, and since Emancipation in our country
commended to us as an example of self-government, being the first
in the history of the African race and the promise of the future. And
yet our President, in his effort to secure that Naboth’s Vineyard on
which he had set his eyes, not content with maintaining the usurper
Baez in power, occupying the harbors of Dominica with war-ships,
sent other war-ships, being none other than our most powerful
monitor, the Dictator, with the frigate Severn as consort, and with
yet other monitors in their train, to strike at the independence of the
Black Republic, and to menace it with war. Do I err in any way, am I
not entirely right, when I say that here was unpardonable outrage to
the African race? As one who for years has stood by the side of this
much-oppressed people, sympathizing always in their woes and
struggling for them, I felt the blow which the President dealt, and it
became the more intolerable from the heartless attempts to defend
it. Alas, that our President should be willing to wield the giant
strength of the Great Republic in trampling upon the representative
government of the African race! Alas, that he did not see the infinite
debt of friendship, kindness, and protection due to that people, so
that instead of monitors and war-ships, breathing violence, he had
sent a messenger of peace and good-will!
This outrage was followed by an incident in which the same
sentiments were revealed. Frederick Douglass, remarkable for his
intelligence as for his eloquence, and always agreeable in personal
relations, whose only offence is a skin not entirely Caucasian, was
selected by the President to accompany the Commissioners to San
Domingo,—and yet on his return, and almost within sight of the
Executive Mansion, he was repelled from the common table of the
mail-steamer on the Potomac, where his companions were already
seated; and thus through him was the African race insulted and their
equal rights denied. But the President, whose commission he had
borne, neither did nor said anything to right this wrong, and a few
days later, when entertaining the Commissioners at the Executive
Mansion, actually forgot the colored orator whose services he had
sought.[155] But this indignity is in unison with the rest. After insulting
the Black Republic, it is easy to see how natural it was to treat with
insensibility the representative of the African race.
Two apologies I hear. The first is that he means well, and errs
from want of knowledge. This is not much. It was said of Louis the
Quarreller, that he meant well; nor is there a slate head-stone in any
village burial-ground that does not record as much of the humble
lodger beneath. Something more is needed for a President. Nor can
we afford to perpetuate power in a ruler who errs so much from
ignorance. Charity for the past I concede, but no investiture for the
future.
The other apology is, that his Presidency has been successful.
How? When? Where? Not to him can be attributed that general
prosperity which is the natural outgrowth of our people and country;
for his contribution is not traced in the abounding result. Our golden
fields, productive mines, busy industry, diversified commerce, owe
nothing to him. Show, then, his success. Is it in the finances? The
national debt has been reduced, but not to so large an amount as by
Andrew Johnson in the same space of time. Little merit is due to
either, for each employed the means allowed by Congress. To the
American people is this reduction due, and not to any President. And
while our President in this respect is no better than his predecessor,
he can claim no merit for any systematic effort to reduce taxation or
restore specie payments. Perhaps, then, it is in foreign relations that
he claims the laurels he is to wear. Knowing something of these from
careful study and years of practical acquaintance, I am bound to say
that never before has their management been so wanting in ability
and so absolutely without character. With so much pretension and so
little knowledge, how could it be otherwise? Here the President
touches nothing which he does not muddle. In every direction is
muddle,—muddle with Spain, muddle with Cuba, muddle with the
Black Republic, muddle with distant Corea, muddle with Venezuela,
muddle with Russia, muddle with England,—on all sides one
diversified muddle. If there is not muddle with Germany and France,
it must be from their forbearance. To this condition are we reduced.
When before in our history have we reached any such bathos as that
to which we have been carried in our questions with England? Are
these the laurels for a Presidential candidate?
But where else shall we look for them? Are they found on the
Indian frontier? Let the cry of massacre and blood from that distant
region answer. Are they in reform of the civil service? But here the
initial point is the limitation of the President to one term, so that he
may be placed above temptation; yet this he opposes. Evidently he
is no true reformer. Are these laurels found in the administration of
the Departments? Let the discreditable sale of arms to France in
violation of neutral duties and of municipal statute be the answer;
and let the custom-houses of New York and New Orleans, with their
tales of favoritism and of nepotism, and with their prostitution as
agencies, mercenary and political, echo back the answer; while
senatorial committees, organized contrary to a cardinal principle of
Parliamentary Law as a cover to these scandals, testify also. And
again, let the War Department recall the disappearance of important
archives bearing on an important event of the war, so that empty
boxes remain like a coffin without a corpse. Where, then, are the
laurels? At last I find them, fresh and brilliant, in the harmony which
the President has preserved among Republicans. Harmony, do I say?
This should have been his congenial task; nor would any aid or
homage of mine have been wanting. But instead he has organized
discord, operating through a succession of rings, and for laurels we
find only weeds and thistles.
But I hear that he is successful in the States once in rebellion.
Strange that this should be said while we are harrowed by the
reports of Ku-Klux outrages. Here, as in paying the national debt,
Congress has been the effective power. Even the last extraordinary
measure became necessary, in my judgment, to supplement his little
efficiency. Had the President put into the protection of the colored
people at the South half the effort and earnest will with which he
maintained his San Domingo contrivance, the murderous Ku-Klux
would have been driven from the field and peace assured. Nor has
he ever exhibited to the colored people any true sympathy. His
conduct to Frederick Douglass on his return from San Domingo is an
illustration; and so also was his answer to the committee of colored
fellow-citizens seeking his countenance for the pending measure of
Civil Rights. Some thought him indifferent; others found him
insulting. Then came his recent letter to the great meeting at
Washington, May 9, 1872, called to assert these rights, where he
could say nothing more than this: “I beg to assure you, however,
that I sympathize most cordially in any effort to secure for all our
people, of whatever race, nativity, or color, the exercise of those
rights to which every citizen should be entitled.”[171] Of course
everybody is in favor of “the rights to which every citizen should be
entitled.” But what are these rights? And this meaningless juggle of
words, entirely worthy of the days of Slavery, is all that is
vouchsafed by a Republican President for the equal rights of his
colored fellow-citizens.
I dismiss the apologies with the conclusion, that in the matters to
which they invite attention his Presidency is an enormous failure.
I will say to the North, Give up; and to the South, Keep not
back.—Isaiah, xliii. 6.
The immediate occasion of the present Letter appears in the following, from
colored citizens of Washington to Mr. Sumner:—
Washington, D. C., July 11, 1872.
A. T. Augusta, m. d.
Samuel Proctor.
David Fisher, sr.
J. J. Ketchum.
Jno. H. Smith.
Chas. N. Thomas.
Edward Crusor.
Wm. H. Shorter.
Wm. H. A. Wormley.
Henry Hill.
William P. Wilson.
Furman J. Shadd.
R. W. Tompkins.
Geo. D. Johnson.
John H. Brown.
Chris. A. Fleetwood.
Henry Lacy.
Chas. F. Bruce.
W. H. Bell.
David Fisher, jr.
J. L. N. Bowen.
David King.
Jacob De Witter.
Wm. Polkeny.
LETTER.
Washington, July 29, 1872.
QUESTIONS PROPOSED.
You are right in saying that the choice for the
Presidency is now “narrowed down” to President Grant
or Horace Greeley. One of these is to be taken, and,
assuming my acquaintance with both and my
observation of their lives, you invite my judgment
between them, asking me especially which of the two,
“judging from their antecedents as well as present
position,” would enforce the Constitution and laws
securing your civil and political rights “with the most
heart-felt sympathy and the greatest vigor.” Here I
remark that in this inquiry you naturally put your rights
in the foreground. So do I,—believing most sincerely
that the best interests of the whole country are
associated with the completest recognition of your
rights, so that the two races shall live together in
unbroken harmony. I also remark that you call
attention to two things,—the “antecedents” of the
candidates, and their “present position.” You wish to
know from these which gives assurance of the most
heart-felt sympathy and greatest vigor in the
maintenance of your rights,—in other words, which,
judging by the past, will be your truest friend.
The communication with which you have honored
me is not alone. Colored fellow-citizens in other parts
of the country, I may say in nearly every State of the
Union, have made a similar request, and some
complain that I have thus far kept silent. I am not
insensible to the trust reposed in me. But if my opinion
is given, it must be candidly, according to my
conscience. In this spirit I answer your inquiries,
beginning with the antecedents of the two candidates.