Software Incident Reporting
Software Incident Reporting
Chris W. Johnson,
School of Computing Science, University of Glasgow, Glasgow, G12 8RZ.
Johnson@dcs.gla.ac.uk,
http://www.dcs.gla.ac.uk/~johnson
The
EU
Telecoms
directive
(2009/140/EC)
places
an
obligation
on
the
providers
of
networks
and
services
to
manage
security
risks
and
to
take
appropriate
measures
to
guarantee
the
security
and
integrity
of
their
op-‐
erations.
In
particular,
Article
13a
requires
providers
to
report
significant
security
breaches
and
losses
of
integrity
to
competent
national
authori-‐
ties.
The
proposed
Network
and
Information
Security
Directive
extends
this
obligation
to
‘market
operators’
responsible
for
critical
national
in-‐
frastructures,
across
the
energy,
banking,
health,
transport,
financial
ser-‐
vices
and
food
sectors.
These
initiatives
mirror
trends
in
the
United
States,
where
for
example
the
Security
and
Exchange
Commission
already
expects
its
members
to
file
information
about
cyber
incidents,
which
are
“significant
factors
that
make
an
investment
in
the
company
speculative
or
risky”.
In
consequence,
many
organizations
are
establishing
reporting
systems
for
gathering
information
about
cyber-‐incidents.
These
initia-‐
tives
typically
build
on
reconstruction,
causal
analysis
and
pattern
match-‐
ing
techniques
that
were
initially
intended
to
support
existing
Safety
Management
Systems
[1].
In
contrast,
this
paper
identifies
differences
that
complicate
the
use
of
conventional
safety
reporting
techniques
to
mitigate
cyber
threats.
For
instance,
it
is
important
to
communicate
safety
lessons
as
widely
as
pos-‐
sible
to
avoid
any
recurrence
of
previous
accidents.
However,
disclosing
the
details
of
a
security
incident
can
expose
vulnerabilities
or
assets
that
motivate
further
attacks.
Further
insights
include
the
need
to
unify
cyber-‐
reporting
mechanisms
–
in
some
sectors
companies
must
report
to
indus-‐
try
regulators,
to
national
telecoms
agencies,
to
local
and
national
law
enforcement,
to
Computer
Emergency
Response
Teams
(CERTs),
to
na-‐
tional
infrastructure
protection
agencies.
In
contrast
to
more
established
safety-‐reporting
systems,
these
different
security
reporting
schemes
have
inconsistent
data
and
disclosure
requirements.
Their
utility
is
further
undermined
by
the
lack
of
data
mining
or
other
information
retrieval
tools
that
might
be
used
to
identify
common
attack
patterns
on
software
used
both
within
and
across
critical
infrastructures.
Other
concerns
fo-‐
cus
on
the
need
for
companies
to
secure
their
supply
chain
–
especially
where
they
may
not
know
about
the
COTS
products
being
used
by
sub-‐
contractors
within
their
own
facilities.
Following
a
safety-‐related
inci-‐
dent,
sub-‐contractors
will
often
work
closely
with
service
providers
to
identify
and
rectify
potential
problems.
The
safety-‐maturity
level
of
most
organizations
is
usually
more
developed
than
any
comparable
secu-‐
rity
culture.
In
consequence,
the
response
to
security
related
incidents
is
often
to
deny
responsibility
and
avoid
blame
[2].
This
compromises
the
utility
of
incident
reporting
when
commercial
confidentiality
and
IP
bar-‐
riers
frustrate
attempts
to
diagnose
the
vulnerabilities
that
compromise
cyber-‐security.
The
insights
from
a
cyber-‐security
incident
reporting
system
help
to
fo-‐
cus
and
justify
subsequent
investments.
Just
as
in
safety,
considerable
resources
can
be
deployed
to
address
threats
that
are
never
realised.
In
such
circumstances,
without
evidence
of
attacks
or
of
safety
incidents
in
other
organisations,
management
can
be
under
pressure
to
cut
safety
and
security
budgets.
Incident
reporting
can;
therefore,
help
to
validate
se-‐
curity
threat
analysis
and
safety
risk
assessments.
The
US
Department
of
Homeland
Security
(DHS)
National
Cyber
security
and
Communications
Integration
Center
(NCCIC)
coordinates
the
collec-‐
tion
and
dissemination
of
threat
information.
The
NCCIC
uses
this
data
to
improve
situation
awareness
across
public
and
the
private
organisa-‐
tions.
Increasing
national
resilience
and
informing
longer-‐term
threat
assessments.
They
also
disseminate
information
to
coordinate
the
na-‐
tional
response
to
any
attack.
There
have
been
few
attempts
to
integrate
safety
and
security
reporting
systems.
One
reason
for
this
is
a
traditional
distinction
between
the
safe-‐
ty
and
security
teams
in
software-‐related
organisations.
They
follow
dif-‐
ferent
standards
–
typified
in
the
distinction
between
the
ISO
27k
securi-‐
ty
series
and
IEC61508
in
the
process
industries.
Although
there
are
common
concepts,
such
as
risk
assessment
and
mitigation,
the
detailed
definitions
and
the
techniques
used
to
perform
similar
tasks
can
be
very
different.
This
is
lack
of
integration
raises
significant
concerns
given
that
safety-‐
critical
systems
are
increasingly
vulnerable
to
cyber
attacks
[5,
6].
In
the
past,
safety
related
applications
were
hard
to
attack
because
few
people
had
the
technical
skills
necessary
to
identify
vulnerabilities
in
bespoke
operating
systems
and
network
protocols.
In
contrast,
many
industries
now
rely
on
Commercial
Off-‐The
Shelf
(COTS)
infrastructures,
including
Linux
and
the
IP
stack.
There
are
three
main
cyber
threats
to
safety-‐critical
infrastructures.
The
first
stems
from
disaffected
insiders.
Companies
increasingly
rely
on
a
wider
range
of
sub-‐contractors
with
correspondingly
less
knowledge
of
their
underlying
technical
infrastructures.
Incident
reporting
is
im-‐
portant
because
it
can
help
national
security
agencies
and
regulators
to
determine
the
extent
of
this
problem
and
identify
changes
over
time,
for
instance
the
growing
use
of
service
oriented
service
provision.
A
second
source
of
attack
is
from
the
creators
of
mass-‐market
malware.
These
attackers
typically
have
little
interest
in
safety-‐related
applications.
Instead,
their
focus
is
on
intercepting
credit-‐card
details
or
account
passwords.
They
are
not
trying
to
halt
electricity
generation
or
aviation
operations.
These
attackers
have
little
idea
of
the
impact
their
code
can
have
on
safety-‐related
systems.
For
example,
‘second
generation’
denial
of
service
attacks
force
operators
to
suspend
critical
processes
following
the
detection
of
mass-‐market
malware.
It
is
unlikely
that
the
develop-‐
ment
techniques
used
to
create
a
Trojan
would
meet
the
requirements
of
an
industry
regulator
within
safety-‐critical
industries!
Finally,
a
growing
number
of
attacks
and
of
active
surveillance
programs
stem
from
state
sponsored
agencies.
The
detection
of
W32.Stuxnet
and
its
predecessors
has
shown
what
is
possible.
Sniffer
applications,
which
exfiltrate
network
and
infrastructure
information,
have
been
found
in
national
critical
infrastructures.
It
is,
therefore,
important
that
we
learn
as
much
as
possible
from
previous
cyber-‐attacks,
not
only
those
that
are
intended
to
inflict
direct,
deliberate
damage
but
also
those
that
disclose
sensitive
information
that
compromise
the
longer-‐term
safety
of
complex
systems.
In
previous
work,
with
support
from
ENISA,
we
have
identified
a
range
of
techniques
that
can
be
used
to
integrate
safety
and
security
incident
re-‐
porting
[2].
There
are
a
number
of
motivations
for
this.
The
most
obvi-‐
ous
is
to
reduce
the
complexity
of
maintaining
two
orthogonal
reporting
architectures.
This
has
implications
for
reporting
bias
when
staff
are
un-‐
sure
how
to
submit
to
either
system.
In
many
cases,
it
can
be
difficult
to
determine
whether
an
incident
has
been
triggered
by
a
more
routine
software
bug
or
by
a
deliberate
attack
–
especially
given
that
the
state
machine
used
to
drive
the
Stuxnet
attack
was
deliberately
intended
to
mask
the
attack
as
a
PLC-‐STEP
7
software
failure.
Different
architectures
have
been
developed
to
support
the
integration
of
safety
and
cyber-‐security
incident
reporting
[2].
Company
reporting
sys-‐
tems
have
been
established
to
collate
manual
submissions
or
the
results
of
automated
anomaly
detection.
Incident
investigators
gather
further
evidence,
including
system
logs,
to
map
out
the
events
leading
to
an
in-‐
trusion
or
safety
concern.
Reconstruction
and
analysis
support
more
detailed
studies
of
the
causal
and
contributory
factors.
From
there
it
is
possible
to
identify
those
actions,
which
are
intended
to
reduce
the
likeli-‐
hood
or
mitigate
the
consequences
of
any
recurrence.
The
findings
are
then
distributed
to
other
stakeholders
within
the
organisation
so
that
corrective
actions
can
be
implemented.
Internal
safety
and
security
reporting
systems
have
a
number
of
limita-‐
tions.
The
corrective
actions
often
focus
on
disciplining
staff
rather
than
on
identifying
underlying
causes
–
such
as
inadequate
procedures
or
training.
Internal
reporting
systems
also
limit
the
dissemination
of
secu-‐
rity
information
across
an
industry.
In
consequence,
the
Telecom
Di-‐
rective
and
the
proposed
EC
Network
and
Information
Security
Directive
introduce
interfaces
between
internal
security
reporting
systems
and
external
organisations
just
as
existing
regulatory
provisions,
such
as
the
European
Railway
Safety
Directive
(2008/110/EC)
create
interfaces
for
the
exchange
of
safety
related
information.
The
previous
section
has
identified
the
concerns
that
motivate
the
inte-‐
gration
of
reporting
systems
for
gathering
information
about
safety
and
cyber
security
concerns.
However,
the
following
paragraphs
identify
a
host
of
differences
that
complicate
the
development
of
common
ap-‐
proaches.
This
paper
is
the
result
of
having
helped
in
the
design
and
im-‐
plementation
of
reporting
systems
across
a
number
of
industries
–
in-‐
cluding
the
ENISA
reporting
system
fro
Article
13a
and
commercial
sys-‐
tems
run
by
energy
distribution
companies
as
well
as
several
European
Air
Navigation
Service
Providers.
Confidentiality
concerns
prevent
di-‐
rect
references
to
the
architecture
of
particular
systems
and
this
in
itself
illustrates
both
the
sensitivity
of
security
incident
reporting
tools
and
also
the
need
to
identify
an
appropriate
security
culture
to
match
the
safety
culture
that
enables
the
exchange
of
information
that
can
help
avoid
future
accidents.
4.1 The Imbalance between Safety and Security Incidents
One
of
the
most
striking
differences
is
the
relative
frequency
of
safety
and
security
concerns
that
are
submitted
to
integrated
reporting
systems.
Typically,
cyber
security
reporting
functions
are
later
added
to
an
exist-‐
ing
safety
management
system.
Most
reports
focus
on
safety
concerns.
One
interpretation
is
that
safety-‐related
organisations
suffer
a
compara-‐
tively
low
level
of
security
incidents.
Alternatively,
it
can
be
argued
that
the
imbalance
reflects
a
reporting
bias
–
with
staff
either
failing
to
detect
or
deliberately
not
submitting
cyber
security
reports.
Further
work
is
required
to
investigate
these
different
hypotheses.
Until
then,
it
seems
likely
that
the
reporting
imbalance
stems
from
a
mixture
of
these
con-‐
cerns.
Management
implicitly
improve
many
security
violations,
for
ex-‐
ample
the
use
of
USB
devices
by
sub-‐contractors,
because
they
are
anx-‐
ious
to
maintain
operations.
These
incidents
would
not
normally
be
re-‐
ported
even
though
they
threaten
the
integrity
of
an
infrastructure.
At
the
same
time,
many
safety-‐related
systems
lack
automated
intrusion
detection
that
might
identify
potential
malware
–
this
is
particularly
true
for
SCADA
systems
using
PLCs
and
Smart
Controllers
where
these
are
very
limited
forensic
tools.
Inside
the
company,
safety
and
security
management
groups
provide
an
interface
between
the
team
running
the
reporting
system
and
external
agencies,
including
national
regulatory
bodies,
industry
associations,
CERTS
and
national
security
agencies.
The
management
group
plays
a
critical
role
in
the
success
of
the
reporting
system;
many
companies
are
reluctant
to
provide
their
competitors
with
a
commercial
advantage
by
disclosing
information
about
previous
attacks.
There
are
justified
con-‐
cerns
about
the
potential
impact
of
disclosure
on
market
confidence
and
on
potential
litigation.
Without
legal
protection,
many
companies
only
provide
anonymous
summaries
security
incidents,
especially
if
they
have
a
potential
impact
on
safety.
One
of
the
most
immediate
concerns
from
the
integration
of
safety
and
cyber-‐security
reporting
systems
is
to
determine
the
appropriate
inter-‐
faces
with
external
agencies
that
must
be
notified
after
an
incident
has
occurred.
It
can
be
unclear
whether
reports
should
be
sent
to
an
indus-‐
try
regulator,
such
as
the
US
Federal
Aviation
Administration
or
Nuclear
Regulatory
Commission,
to
a
security
agency,
such
as
the
Department
of
Homeland
Security
or
the
US
CERT,
or
to
telecoms
regulators
who
have
responsibility
for
collating
information
about
wider
cyber-‐security
con-‐
cerns
such
as
the
Federal
Communications
Commission.
In
some
cases,
a
single
incident
must
be
reported
to
more
than
one
agency.
For
example,
a
cyber-‐attack
with
safety
related
consequences
must
be
reported
to
the
national
industry
safety
regulator
and
potentially
also
to
a
subset
of
the
National
Crime
Agency,
the
National
Cyber
Crime
Unit,
GOVCERT,
the
UK
Information
Commissioner
as
well
as
the
CESG/Centre
for
the
Protection
of
National
Critical
Infrastructure
via
providers
registered
under
the
Cyber
Incident
Response
(CIR)
or
the
Cyber
Security
Incident
Response
Scheme
(CSIR).
There
is
widespread
confusion
over
who
to
consult.
In
the
United
States,
there
are
Federal
investigatory
agencies
(e.g.,
the
Fed-‐
eral
Bureau
of
Investigation
[FBI]
and
the
U.S.
Secret
Service),
district
attorney
offices
as
well
as
both
state
and
local
(e.g.,
county)
law
enforce-‐
ment.
Regulatory
agencies
introduce
an
additional
layer
of
complexity
in
safety-‐related
industries.
It
is
for
this
reason
that
NIST
urge
organisa-‐
tions
to
contact
the
legal
agencies
that
must
respond
to
cyber
incidents;
“one
reason
many
security-‐related
incidents
do
not
result
in
convictions
is
that
some
organizations
do
not
properly
contact
law
enforcement...
incident
response
team
should
become
acquainted
with
its
various
law
enforcement
representatives
before
an
incident
occurs
to
discuss
condi-‐
tions
under
which
incidents
should
be
reported
to
them,
how
the
report-‐
ing
should
be
performed,
what
evidence
should
be
collected,
and
how
it
should
be
collected”
[7].
The
main
distinction
between
these
different
systems
is
that
some
explic-‐
itly
understand
the
safety-‐related
domain
in
which
an
incident
occurs
but
have
little
cyber
security
competence
while
others
have
little
or
no
indus-‐
try
specific
safety-‐related
competence
but
do
understand
the
general
class
of
cyber-‐threats.
In
particular,
industry
safety
regulators
have
al-‐
most
no
expertise
in
cyber-‐security;
many
suffer
from
a
long
legacy
of
physical
security
specialists
whose
talents
provide
little
help
in
mitigat-‐
ing
new
generations
of
advanced
persistent
threats.
These
problems
have
been
recognised,
for
example
in
recent
reports
from
the
US
Gov-‐
ernment
Accountability
Office:
“A
Better
Defined
and
Implemented
Na-‐
tional
Strategy
Is
Needed
to
Address
Persistent
Challenges”
[8]
and
“Cyber
security:
National
Strategy,
Roles,
and
Responsibilities
Need
to
Be
Better
Defined
and
More
Effectively
Implemented”
[9].
There
is
a
need
to
draft
letters
of
agreement
between
regulatory
and
state
security
agencies
to
establish
roles
and
responsibilities
in
the
aftermath
of
a
cyber-‐attack.
These
help
to
avoid
“turf
wars”
when
engineers
are
working
to
ensure
the
safety
of
compromised
systems.
These
conflicts
and
confusions
between
safety
and
security
lines
of
re-‐
porting
are
compounded
because
many
companies
operate
across
na-‐
tional
borders.
Each
local
division
must
respect
the
very
different
re-‐
porting
mechanisms
and
jurisdictions
that
have
been
implemented
in
each
country.
There
are
further
complexities,
especially
in
Europe
where
there
are
increasing
requirements
to
report
to
national
agencies
and
also
to
share
recommendations
with
other
member
states,
for
example
within
the
proposed
Network
and
Information
Security
Directive.
If
an
incident
is
detected
in
a
system
in
one
national
jurisdiction,
there
is
often
a
re-‐
quirement
to
share
lessons
with
groups
in
other
countries
without
com-‐
promising
national
sovereignty
in
any
area
of
operation.
Appropriate
means
of
achieving
this
remain
the
subject
both
of
academic
research
and
of
heated
debate
within
the
European
Commission
[9].
The
legal
and
regulatory
framework
for
the
investigation
of
cyber
inci-‐
dents
is
likely
to
become
more
and
more
complex
with
the
increasing
cross-‐border
integration
of
national
critical
infrastructures.
The
Europe-‐
an
Commission
has
encouraged
the
interoperability
of
national
rail
net-‐
works.
They
have
promoted
the
development
and
integration
of
smart
grids
and
the
creation
of
the
single
European
skies
network.
We
have
already
glimpsed
the
confusion
that
this
can
create
in
the
investigation
of
conventional
system
failures,
including
the
Viareggio
accident
that
has
taken
years
not
months
to
complete.
The
US
Department
of
Justice
makes
it
clear
that
the
web
of
international
interdependencies
is
far
more
com-‐
plex
in
cyber
incidents,
for
example
when
a
company
in
one
country
is
attacked
through
the
exfiltration
of
operational
data
on
servers
in
a
se-‐
cond
state
that
are
attacked
by
systems
in
a
third
nation
remotely
con-‐
trolled
by
attackers
in
a
fourth
state
[12].
In
the
UK,
the
Association
of
Chief
Police
Officers
[13]
has
published
guidelines
for
the
handling
of
evidence
following
cyber
incidents:
1. No
action
taken
by
law
enforcement
agencies
or
their
agents
should
change
data
held
on
a
computer
or
storage
media
which
may
subsequently
be
relied
upon
in
court;
2. In
circumstances
where
a
person
finds
it
necessary
to
access
original
data
held
on
a
computer
or
on
storage
media,
that
person
must
be
competent
to
do
so
and
be
able
to
give
evidence
explaining
the
relevance
and
the
implications
of
their
actions;
3. An
audit
trail
or
other
record
of
all
processes
applied
to
computer-‐based
elec-‐
tronic
evidence
should
be
created
and
preserved.
An
independent
third
party
should
be
able
to
examine
those
processes
and
achieve
the
same
result;
4. The
person
in
charge
of
the
investigation
(the
case
officer)
has
overall
respon-‐
sibility
for
ensuring
that
the
law
and
these
principles
are
adhered
to.
These
principles
were
drafted
to
support
cyber
incidents
in
‘convention-‐
al’
or
office-‐based
systems.
They
do
not
address
the
conflicts
between
a
desire
to
meet
forensic
guidelines
and
the
need
to
maintain
levels
of
safe-‐
ty
and
levels
of
service.
In
practice,
it
is
likely
that
senior
investigators
will
have
to
work
closely
with
systems
and
safety
engineers.
We
urgently
need
more
guidance
on
the
roles
and
responsibilities
of
these
key
staff
in
the
aftermath
of
cyber
incidents
affecting
safety-‐critical
infrastructures
[14].
The
US
Department
of
Justice
[12]
reinforce
these
arguments
when
they
suggest
that
first
responders
must:
It
would
be
hard
to
enforce
such
cyber-‐incident
response
requirements
in
safety-‐critical
systems.
Removing
“all
persons
from
the
crime
scene”
could
be
catastrophic
in
a
crowded
Air
Traffic
Control
centre
or
Nuclear
control
room
where
recovery
from
a
cyber-‐incident
requires
cooperation
between
multiple
teams.
To
“leave
a
computer
or
electronic
device
off
if
it
is
already
turned
off”
would
prevent
the
use
of
redundant
protection
sys-‐
tems.
Such
conflicts
illustrate
the
hose
of
legal
and
forensic
concerns
that
stem
from
the
integration
of
cyber
and
safety
related
reporting
require-‐
ments.
In
safety-‐critical
systems,
there
has
been
a
growing
focus
on
the
systemic
factors
that
create
the
context
in
which
an
incident
or
accident
is
likely
to
occur.
In
contrast,
security
investigations
tend
to
focus
more
narrowly
on
the
policies
and
procedures
that
have
been
violated
as
a
precursor
to
an
incident.
There
is
often
a
disciplinary
aspect
to
the
causal
analysis
that
is
less
apparent
in
the
immediate
aftermath
of
safety
related
events.
This
difference
between
safety
and
cyber-‐security
incident
reporting
can
be
explained
in
terms
of
the
links
between
cyber-‐security,
physical
security
and
policing
associating
blame
with
legal
transgressions.
There
are
other
potential
interpretations.
The
differences
in
causal
analysis
and
in
rec-‐
ommendations
between
safety
and
security
incidents
may
be
due
to
the
relative
maturity
of
the
two
areas.
Most
security
reporting
systems
are
being
integrated
with
existing
safety
management
systems.
The
perfec-‐
tive
approach
focussing
on
the
attribution
of
human
error
in
previous
generations
of
safety
reporting
is
still
to
be
seen
in
the
relatively
imma-‐
ture
field
of
security
investigation.
In
this
interpretation,
we
might
ex-‐
pect
that
the
focus
of
security
investigations
may
shift
over
time
from
individual
violations
to
systemic
factors
through
development
in
the
un-‐
derlying
‘security
culture’.
Irrespective
of
the
reasons
behind
these
dif-‐
ferences,
questions
remain
about
the
tools
and
techniques
that
investiga-‐
tors
might
use
to
improve
the
coherence
and
consistency
of
causal
analy-‐
sis
in
the
aftermath
of
cyber
incidents.
In
particular,
further
work
is
re-‐
quired
to
determine
whether
the
existing
application
of
safety-‐related
root
cause
analysis
techniques,
including
those
using
counter-‐factual
rea-‐
soning
and
systemic
models,
can
be
extended
to
support
the
analysis
of
cyber-‐attacks
[3].
For
instance,
it
is
hard
to
be
sure
that
any
security
measures
in
isolation
will
be
sufficient
to
have
prevented
a
directed
in-‐
trusion
–
especially
when
we
cannot
be
certain
that
we
have
detected
all
active
malware
in
the
aftermath
of
an
attack.
A
final
area
of
concern
illustrates
both
the
need
for
integration
between
safety
and
cyber-‐security
incident
reporting
as
well
as
the
technical
prob-‐
lems
in
achieving
this
integration.
Many
existing
safety
recommenda-‐
tions
create
security
vulnerabilities
and
vice
versa.
As
a
specific
exam-‐
ple,
redundancy
is
typically
used
to
increase
the
dependability
of
critical
systems.
However,
this
provides
few
benefits
in
software
related
sys-‐
tems
without
some
level
of
diversity.
Leaving
aside
the
issue
of
common
requirements
failure,
two
redundant
versions
of
the
same
code
are
likely
to
contain
the
same
bugs
and
hence
fail
in
the
same
way.
In
conse-‐
quence,
N-‐version
programming
techniques
rely
on
using
two
or
more
teams/contractors
to
develop
redundant
versions
of
the
same
program.
In
the
event
of
one
program
failing,
it
is
hoped
that
the
diverse
supply
chain
used
to
deliver
the
redundant
version
will
ensure
that
they
do
not
share
common
bugs.
This
approach
to
safety-‐critical
software
develop-‐
ment
has
been
widely
accepted
and
used,
for
instance
in
the
deployment
of
Air
Data
Inertial
Reference
Units
across
most
commercial
aircraft.
Unfortunately,
such
techniques
create
immediate
problems
for
security
management.
Companies
must
devote
considerable
expense
to
security
two
diverse
supply
chains
–
increasingly
this
involves
audit
processes
to
ensure
that
sub-‐contractors
meet
agreed
security
policies
during
the
de-‐
velopment
of
diverse,
redundant
code.
In
the
past,
these
security
con-‐
cerns
were
not
factored
in
the
costs
associated
with
such
safety-‐critical
software
development
practices.
A
converse
set
of
problems
arises
from
the
introduction
of
intrusion
de-‐
tection
systems
in
the
aftermath
of
security
related
incidents.
White
list
approaches
create
problems
because
in
most
safety-‐related
systems
the
use
of
sub-‐contractors
and
significant
legacy
systems
prevents
operators
from
enumerating
all
of
the
processes
that
should
be
running
within
a
particular
application.
Conversely,
there
are
concerns
over
black-‐list
approaches
because
the
malware
signatures
that
are
downloaded
every
24
hours
in
most
desktop
systems
might
themselves
cause
a
safety
relat-‐
ed
system
to
fail
–
and
cannot
be
subjected
to
the
extensive
verification
and
validation
processes
recommended
by
safety
related
standards
with-‐
out
creating
a
lengthy
delay
during
which
the
application
process
would
be
unpatched
and
vulnerable
to
any
attack.
Alternatives
such
as
the
use
of
data
diodes
that
enable
monitoring
of
intrusions
without
altering
net-‐
work
traffic
raise
further
questions
when
false
positives
force
the
closure
of
safety-‐related
services1.
The
meta-‐level
point
is
that
integrating
safety
and
security
critical
computation
reveals
a
host
of
tensions,
which
can
only
be
addressed
through
integrated
research
in
venues
such
as
SAFECOMP.
This
paper
has
identified
a
number
of
national
and
international
initia-‐
tives
that
extend
reporting
requirements
from
safety
related
events
to
include
cyber-‐incidents.
Attempts
have,
therefore,
been
made
to
integrate
security
events
into
existing
safety
reporting
architectures.
This
is
moti-‐
vated
by
superficial
similarities
in
terms
of
the
processes
involved
in
re-‐
cording
and
analysing
these
incidents.
In
both
safety
and
security,
it
is
important
to
obtain
‘sufficient’
evidence,
to
reconstruct
an
adverse
event,
to
conduct
causal
analysis,
to
identify
corrective
actions
etc.
Many
of
the-‐
se
tasks
are
already
well
supported
within
safety
management
systems.
However,
it
is
important
to
recognise
that
malware
also
poses
some
unique
challenges
–
for
example,
the
forensic
analysis
of
a
cyber-‐attack
is
very
different
from
most
accident
investigations.
An
active
adversary
will
deliberately
seek
to
disguise
their
actions.
Further
complexities
stem
from
the
need
to
safeguard
an
application
and
the
public
in
the
immedi-‐
ate
aftermath
of
a
cyber
attack.
For
instance,
how
do
we
land
the
aircraft
1
These issues are addressed in a companion paper submitted to SAFECOMP on Barriers to the
Use of Intrusion Detection Systems in Safety-Critical Applications.
in
flight
when
we
fear
that
an
Air
Traffic
Management
infrastructure
is
compromised
and
yet
at
the
same
time
meet
legal
requirements
to
safe-‐
guard
forensic
evidence?
This
paper
has
argued
that
to
address
such
questions,
we
must
encourage
the
integration
of
expertise
across
both
safety
and
security
management.
In
particular,
we
must
identify
appropriate
communications
interfaces
with
both
internal
and
external
stakeholders
–
to
establish
clear
jurisdic-‐
tions,
roles
and
responsibilities.
There
is
significant
scope
for
further
work.
In
particular,
tools
and
techniques
are
urgently
required
to
sup-‐
port
the
safety
analysis
of
security
incidents
and
vice
versa.
This
is
par-‐
ticularly
important
given
that
most
safety-‐critical
companies
lack
in-‐
house
forensic
expertise.
As
mentioned,
gathering
evidence
in
the
after-‐
math
of
a
security
incident
poses
unique
challenges
within
safety-‐critical
applications
[2].
Existing
guidance
from
the
US
Department
of
Justice
focuses
on
conventional
office
based
systems.
They
stress
the
need
to
isolate
compromised
networks
both
to
preserve
evidence
and
to
prevent
cross-‐contamination.
Such
guidance
cannot
easily
be
applied
in
many
safety-‐critical
systems
when,
for
instance,
passengers
and
aircrew
rely
on
continued
service
from
air
navigation
service
providers.
Additional
sup-‐
port
is
also
required
for
the
causal
analysis
of
security
incidents;
in
safe-‐
ty-‐related
systems
there
has
been
a
focus
on
systemic
approaches
that
look
for
the
detailed
managerial
and
organisational
precursors
to
acci-‐
dents
and
incidents.
These
ideas
have
had
little
impact
on
security
man-‐
agement
where
many
recommendations
continue
to
focus
on
the
imme-‐
diate
violation
of
rules
and
regulations
by
operational
staff.
6 References
1. C.W.
Johnson,
Supporting
the
Exchange
of
Lessons
Learned
from
Cyber-‐Security
Inci-‐
dents
in
Safety-‐Critical
Systems.
In
D.
Swallom
(ed.),
Proceedings
of
the
32nd
Inter-‐
national
Systems
Safety
Society,
Louisville,
International
Systems
Safety
Society,
Un-‐
ionville,
VA,
USA.
2. C.W.
Johnson,
Inadequate
Legal,
Regulatory
and
Technical
Guidance
for
the
Forensic
Analysis
of
Cyber-‐Attacks
on
Safety-‐Critical
Software.
In
D.
Swallom
(ed.),
Proceed-‐
ings
of
the
32nd
International
Systems
Safety
Society,
Louisville,
International
Sys-‐
tems
Safety
Society,
Unionville,
VA,
USA.
3. C.W.
Johnson,
Failure
in
Safety-‐Critical
Systems:
A
Handbook
of
Accident
and
Incident
Reporting,
University
of
Glasgow
Press,
Glasgow,
Scotland,
2003.
4. NIST,
Computer
Security
Incident
Handling
Guide:
Recommendations
of
the
National
Institute
of
Standards
and
Technology,
NIST
Special
Publication
800-‐61
Revision
2,
August
2012.
5. C.W.
Johnson,
CyberSafety:
On
the
Interactions
Between
CyberSecurity
and
the
Soft-‐
ware
Engineering
of
Safety-‐Critical
Systems.
In
C.
Dale
and
T.
Anderson
(eds.),
Achieving
System
Safety,
85-‐96,
Springer
Verlag,
London,
UK,
Paper
to
acompany
a
keynote
address,
20th
Annual
Conference
of
the
UK
Safety-‐Critical
Systems
Club,
ISBN
978-‐1-‐4471-‐2493-‐1,
2012.
6. C.W.
Johnson,
The
Telecoms
Inclusion
Principle:
The
Missing
Link
between
Critical
In-‐
frastructure
Protection
and
Critical
Information
Infrastructure
Protection.
In
P.
The-‐
ron
and
S.
Bologna
(eds.),
Critical
Information
Infrastructure
Protection
and
Resili-‐
ence
in
the
ICT
Sector,
IGI
Global,
Pennsylvania,
USA,
2013.
7. U.S.
National
Institute
of
Standards
and
Technology
(NIST)
(2006),
Guide
to
Integrat-‐
ing
Forensic
Techniques
into
Incident
Response,
Special
Publication
800-‐86,
Gaithersburg,
Maryland,
2006.
http://csrc.nist.gov/publications/nistpubs/800-‐
86/SP800-‐86.pdf
10. U.S.
National
Institute
of
Standards
and
Technology
(NIST)
(2006),
Guide
to
Integrat-‐
ing
Forensic
Techniques
into
Incident
Response,
Special
Publication
800-‐86,
Gaithersburg,
Maryland,
2006.
http://csrc.nist.gov/publications/nistpubs/800-‐
86/SP800-‐86.pdf
11. U.S.
National
Institute
of
Standards
and
Technology
(NIST)
(2012),
Computer
Securi-‐
ty
Incident
Handling
Guide
(Draft),
Special
Publication
800-‐61
Revision
2
(Draft),
Gaithersburg,
Maryland,
2012.
http://csrc.nist.gov/publications/drafts/800-‐61-‐
rev2/draft-‐sp800-‐61rev2.pdf
12. U.S.
Department
of
Justice
(2004),
Forensic
Examination
of
Digital
Evidence:
A
Guide
for
Law
Enforcement,
2004.
13. Association
of
Chief
Police
Officers
(2011),
Managers
Guide:
Good
Practice
and
Advice
Guide
for
Managers
of
e-‐
Crime
Investigation,
2011,
ACPO.
http://www.acpo.police.uk/documents/crime/2011/201103CRIECI14.pdf
14. C.W.
Johnson,
Inadequate
Legal,
Regulatory
and
Technical
Guidance
for
the
Forensic
Analysis
of
Cyber-‐Attacks
on
Safety-‐Critical
Software.
In
D.
Swallom
(ed.),
Proceed-‐
ings
of
the
32nd
International
Systems
Safety
Society,
Louisville,
International
Sys-‐
tems
Safety
Society,
Unionville,
VA,
USA.