Aviation Training Debrief Analysis
Aviation Training Debrief Analysis
March 1997
Space Administration
Kimberly K. Jobe, San Jose State University/Ames Research Center, Moffett Field, California
Lori K. McDonnell, San Jose State University/Ames Research Center, Moffett Field, California
March 1997
PREFACE
This study originated from requests from several airline training departments for
help in analyzing the effectiveness of LOFT debriefings. Doug Daniel and Steve
Gregorich helped identify crucial issues and ways to study these issues.
The study could not have been conducted without the generous willingness of
instructors and line crews to allow us to observe their debriefings. We are
impressed with their high standards of professionalism. Training department
managers from each of the airlines that participated in the study provided a
wealth of background information and made valuable suggestions on early drafts
of this manuscript.
This study was funded by the FAA's Office of the Chief Scientist and Technical
Advisor for Human Factors (AAR-100). Eleana Edens, the program manager,
provided support, encouragement and helpful suggestions.
TABLE OF CONTENTS
Preface iii
1.0 OVERVIEW 1
2.0 INTRODUCTION 3
2.1 Background 3
2.2 What is Facilitation and Why Use It? 4
2.3.1 Introductions 6
2.3.3 Questions 6
2.3.4 Silence 6
2.3.5 Videos 6
3.0 METHODS 7
3.1 Participants 7
3.2 Procedures 7
3.3 Measures 8
4.0 RESULTS 9
4.2.1 Participation 10
4.2.4 Interruptions 11
4.2.5 Videos 11
4.2.6 Crew participation 11
4.3.1 Scores 12
4.3.2 Correlations 12
5.0 DISCUSSION 14
5.1.1 Duration 15
5.1.2 Content 15
References 49
Appendix A. Coding 51
Table 10. Percent of Total Crew Words & Utterances Coded R, S1, S & Q
Table 18. Relationship of High and Low Introduction Scores to Crew Analysis &
Evaluation and Depth of Activity
SUMMARY
This study analyzes techniques instructors use to facilitate crew analysis and
evaluation of their LOFT performance. A rating instrument called the Debriefing
Assessment Battery (DAB) was developed which enables raters to reliably
assess instructor facilitation techniques and characterize crew participation.
Thirty-six debriefing sessions conducted at five U.S. airlines were analyzed to
determine the nature of instructor facilitation and crew participation. Ratings
obtained using the DAB corresponded closely with descriptive measures of
instructor and crew performance. The data provide empirical evidence that
facilitation can be an effective tool for increasing the depth of crew participation
and self-analysis of CRM performance. Instructor facilitation skill varied
dramatically, suggesting a need for more concrete hands-on training in facilitation
techniques. Crews were responsive but fell short of actively leading their own
debriefings. Ways to improve debriefing effectiveness are suggested.
1.0 OVERVIEW
How much crews learn in Line-Oriented Flight Training (LOFT) and take back to
the line depends on the effectiveness of the debriefing that follows the LOFT.
The Crew Resource Management (CRM) literature and the Federal Aviation
Administration's (FAA) advisory circular (AC) 120-35C recommend that in the
debriefing instructors should facilitate self-discovery and self-critique by the crew
rather than lecture on what they did right and wrong. Self discovery by the crew
is believed to provide deeper learning and better retention. Also, crews are more
likely to enhance their performance of CRM in line operations if they develop
their ability to analyze flight operations in terms of CRM and debrief themselves
after line flights.
In this study 36 LOFT debriefings conducted at five major U.S. airlines were
analyzed. Audiotape recordings of each session were made with the permission
of instructors and crews. The recordings were subsequently deidentified, coded,
and analyzed for more than 70 variables. The Debriefing Assessment Battery
was developed to systematically characterize instructor effectiveness at
facilitation and the nature of crew participation in debriefings. The data indicate
that the Debriefing Assessment Battery is a reliable and valid instrument for
assessing instructors' skill in facilitation and for analyzing crew participation. The
battery was designed to be used by researchers, however a short form of the
battery that can be used by training departments to evaluate debriefings in real
time is currently being developed and evaluated.
Most instructors at all five airlines followed a similar general format for debriefing.
However, within each airline both instructors and crews varied widely on many of
the specific variables observed. There were also substantial differences among
airlines on several variables for both instructors and crews, though most of these
differences were not statistically significant due to the large variability within each
airline.
The debriefings lasted an average of 31 minutes, with a range of 8 to 82 minutes.
However, 31 minutes may not allow adequate time for crews to analyze their
performance thoroughly or learn and practice the skills of self-debriefing. This
study provides no data on the optimal length for debriefings, however an hour
may be a useful rough target, with adjustments for the needs of individual crews.
This suggestion must, of course, be considered in the context of other demands
on instructors' time.
The content of the debriefings was driven almost exclusively by the instructors;
crew members rarely brought up topics on their own initiative. Also, discussions
revolved around the instructor, even when the instructor succeeded in getting the
crew to do most of the talking: there was little back-and-forth discussion directly
between crew members. The data indicate that crews were responsive but not
very proactive. This may be in part because few of the instructors explicitly told
crews they should take a proactive role and perform their own analysis without
depending on the instructor to lead them step by step. It may also be that
instructors themselves either do not fully accept or understand the concept of
crews taking initiative and responsibility for the content of the debriefing.
The wide range of instructor effectiveness in facilitation indicates that the airlines
face an issue of standardization of this aspect of debriefing. The distribution of
facilitation scores was distinctly bimodal, with one group of instructors scoring in
the good to very good range and another group of instructors scoring in the
marginal range. Also, instructors who did well in one aspect of facilitation typically
did well in all aspects (except stating expectations for crew participation), and
those who did poorly in one aspect tended to do poorly in all aspects. These data
suggest instructors' ability to use various techniques is determined at least in part
at the conceptual level: Do they grasp the underlying concept of facilitation? Do
they accept the concept? Is facilitation the type of approach for which they have
ability?
The CRM literature states that debriefings should be led by the crews
themselves, using the instructor as a resource. Our data suggest that this goal,
although worthwhile, is rather idealistic. Instructors become discouraged when,
after a brief and rather abstract course in facilitation, they attempt to facilitate
debriefings and discover that crews often do not immediately respond. We
suggest that it would be more effective to teach instructors that facilitation should
be adapted to the level at which the particular crew is able to respond.
Facilitation can be conducted at levels ranging from high, which approaches the
ideal of the debriefing being led by the crew, to low, in which the instructor leads
the crew substantially, but in all cases debriefings should emphasize as much
self-discovery by the crew as possible.
Instructors are encouraged to attempt to facilitate at the highest level possible for
a particular crew. Realistically, however, most crews do not yet have the skills
and motivation needed to lead their own debriefings without substantial
assistance from the instructor. It may be possible to change this situation over
time if LOFT instructors consistently encourage crews to take a proactive role in
debriefing their own training.
2.0 INTRODUCTION
2.1 Background
Line Operational Simulation (LOS) is widely used to provide opportunities for
crews to practice CRM concepts in realistic and challenging simulated flight
situations. As indicated in the FAA's AC 120-35C (1995), LOS includes LOFT,
Line Operational Evaluation (LOE), and Special Purpose Operational Training
(SPOT). LOFT is the original "non-jeopardy" form of simulation training in which
crews are not graded on their performance. Like LOFT, SPOT is used for training
rather than evaluative purposes. In LOE crews are graded, which is required in
those airlines that participate in the FAA's Advanced Qualification Program
(AQP). Both LOFT and LOE are full-mission simulations that include all phases
of flight, whereas SPOT may be full-mission or only a segment of a flight tailored
to focus on a particular training point.
How much crews learn in LOFT and take back to the line depends on the
effectiveness of the debriefing that follows the LOFT (Helmreich & Foushee,
1993). The simulation itself is a busy, intense experience, and thoughtful
discussion afterward is necessary for the crew to sort out and interpret what
happened and why. Instructors are expected to lead debriefings in a way that
encourages crew members to analyze their LOFT performance for themselves.
Rather than lecturing to the crew on what they did right and wrong, the instructor
is expected to facilitate self-discovery and self-critique by the crew (Butler, 1993;
Hawkins, 1987; Smith, 1994).
CRM and LOFT programs have developed considerably since their inception
almost twenty years ago. The concepts and the value of CRM are now generally
accepted by both airline managers and pilots. However, it is not clear whether
crews consistently think about and practice CRM in line operations (see
discussion in Helmreich & Foushee, 1993). AQP is bringing to fore the issue of
how well crews are actually able to practice CRM, because poor CRM can cause
crews to fail a LOE (Birnbach & Longridge, 1993; FAA, 1991). In order for LOE
programs to be effective and accepted, pilots must believe they are being graded
on performance dimensions they understand and by criteria that seem
appropriate and achievable. The ability of crews to analyze and evaluate their
own performance in LOFT may predict their acceptance of LOE grading.
The facilitator should not handle the debrief in a "teacher tell" manner but,
instead, operate as a resource to crew members by highlighting different portions
of the LOS that may be suitable for review, critique, and discussion. The
discussion should be led by the crew themselves, using the facilitator and the
videotape as resources for use during their critique...Self-criticism and self-
examination are almost always present in these situations, and in many cases
they are much more effective than facilitator criticism...Thus, the facilitator should
do everything possible to foster this sort of self-analysis, while at the same time
keep the debrief at a constructive level. In the role of moderator, the facilitator
can guide the discussion to areas that he or she has noted...However, unless
absolutely necessary, the facilitator should avoid "lectures" about what is right
and wrong.
The concept of facilitated debriefings appears to have been part of the early
inception of LOFT (Lauber & Foushee, 1981). The origin of this concept is not
clear, but it appears to have been derived from the use of facilitation in other
business settings, such as retreats in which managers discuss their
organizational goals and issues (e.g., Gibb, 1982; Mills & Roberts, 1981).
The primary rationale for facilitating rather than lecturing is that crews can learn
and remember much more when they participate actively and make their own
analyses than when they listen passively to the instructor (Duvall & Wicklund,
1972; Smith, 1994). Another potential benefit of crew-centered LOFT debriefings
is that they can help crews develop the habits of analyzing their own CRM
performance on the line and conducting their own crew debriefings following line
operations (Butler, 1993). In practice, crew debriefings on the line in civil
operations are as yet rare, although military crews often debrief their missions.
Thus, the LOFT debriefing is an important tool for showing crews how to debrief
and for illustrating the benefits of self-debriefing.
The general literature on facilitation in settings other than LOFT is also rather
sketchy. This is a trade literature rather than a scientific literature, and very little
empirical evidence is provided to support assertions, validate specific techniques,
or qualify the range of settings in which advocated techniques may be effective.
However, the general concept of facilitation has considerable face validity as a
way to encourage self-discovery by crew members. Both the adult learning
literature and the cognitive research literature suggest that self-discovery
improves learning, retention, and the ability to apply knowledge in diverse
settings.
Facilitation can help individuals develop problem solving and critical thinking
skills (Gow & Kember, 1993). Research in several areas of expertise suggests
that individuals are better at solving problems and applying their knowledge in
diverse situations if they have a good metacognitive perspective of their technical
skills (see Metcalfe & Shimamura, 1994). Metacognition refers to knowledge of
one's own thought processes and the ability to keep track of what one is doing
while analyzing problems and managing tasks. Debriefings that emphasize self-
analysis and self-discovery help crews develop metacognitive skills for managing
cockpit situations. One could argue that the concept of metacognition is implicit in
the philosophy of CRM; for example, CRM teaches crews to establish priorities
and keep track of how they are managing their priorities during abnormal line
situations.
2.3.2 Active listening. Good listening skills enable the facilitator to work with
what the participants are saying and to encourage further participation. Active
listening shows that the facilitator is attending to the speaker, understands what
is being said, and wants to hear more. Active listening can range from a simple
"uh-huh" or "okay" to echoing or reflecting in one's own words what a speaker is
trying to communicate.
2.3.5 Videos. Most airlines videotape the LOFT. Although the use of video is not
a facilitation technique per se, it can aid facilitation. Instructors select segments
of the videotape to show during the debriefing to help the crew observe and
discuss their performance. The video can help the crew view their performance
from a third-party perspective (FAA, 1995); it may also help the crew remember
what happened.
2) What techniques do instructors use to facilitate and how effective are these
techniques?
5) How much variation occurs among instructors and among airlines in the
conduct of debriefings?
3.0 METHODS
3.1 Participants
Thirty-nine LOFT debriefings conducted at five major U.S. airlines between June
1994 and May 1995 were observed. All five airlines are large, well-established
national companies; four are passenger airlines and one is a cargo company. At
each of the airlines the first author observed four to eleven debriefings. (At the
first company visited, a second research observer was also present at the
debriefings and interviews.) The training department managers who arranged the
observations were asked not to preselect which instructors and crews would be
observed; rather, the selection was driven by the schedules of who was
instructing during the three to five days each airline was visited. The observed
debriefings represented all or most of the fleets operated by each airline, and at
least one LOFT simulation of each scenario flown in each fleet was observed.
Generally, one debriefing was observed per instructor and crew; however, four of
the instructors were observed debriefing a second crew for the purpose of
comparison.
Permission to attend the debriefing and to audio tape the session was obtained
from each instructor and each crew member, and assurance was provided that
all data collected would be completely deidentified to assure anonymity for all
participants.
3.2 Procedures
Prior to observation of the debriefings, the written scenarios for each LOFT were
reviewed and managers in the CRM departments were interviewed. After each
debriefing the instructor was interviewed and asked to rate the crew's CRM
performance and technical performance on separate five-point Likert scales
ranging from poor (1) to exemplary (5). Instructors were also asked for comments
about the debriefing process.
The audio recordings of 36 of the 39 debriefings were transcribed into text in their
entirety and all references to individuals and organizations were deleted. (Two of
the recordings were not sufficiently intelligible for transcribing and the tape
recorder failed during another debriefing.) Of the 36 debriefings that were
transcribed, 25 were from two-person crews, and eleven were from three-person
crews (Table 1).
3.3 Measures
3.3.1 Descriptive measures. Each instructor and crew utterance was coded for
nine factors and the coding was checked during data entry. (The factors and the
coding rules are described in Appendix A.) From these nine factors 72 utterance
variables were calculated (see Appendix B). Data were also extracted on the
instructors' use of videotapes to illustrate the crews' performance in the LOFT,
including the number of video segments played for crew discussion, the length of
the segments played, and the extent to which the segments were discussed. The
above data will be referred to as "descriptive" to distinguish them from the data
generated using the Debriefing Assessment Battery described below.
Two of the authors independently rated the instructors and crews from each of
the debriefing sessions after listening to each LOFT session audio tape while
reading the verbatim transcript. For each of the first 10 debriefings, the ratings on
the individual battery items were compared and discussed before rating the next
debriefing. During each discussion, if either believed any ratings needed to be
changed based on issues raised by the other, the scores were revised
accordingly, although no effort was made to reach consensus on each item. For
the remaining 26 debriefings, ratings were not systematically discussed.
Aside from reliability coefficients, data from the battery are based on the average
of the two raters' scores for each item. Composite scores for each of the five
instructor and two crew categories were calculated by averaging the scores for
the four items in each category.
Four instructors conducted two debriefings; thus, each of these four instructors
received two measurements for each of the variables associated with their
performance. These two measurements were averaged to obtain a single data
point (n = 32) for (i) calculation of means and standard deviations, and (ii) the
analysis described below. The means with duplicate instructors' scores averaged
(n = 32) are reported for scores on the Debriefing Assessment Battery. However,
since differences between the two methods of calculating the means were minor
for the descriptive variables, these means are reported for the full data set (n =
36).
Data from these four instructors were used to explore the question of whether the
large variability observed among instructors reflected stable differences among
the instructors. Five variables were selected for this analysis: session duration,
percent of group words uttered by the instructor, percent of instructor words
addressing CRM, percent of instructor words addressing crew performance, and
instructor scores on a composite QEF variable created by combining the
Questions, Encouragement, and Focus categories of the assessment battery.
For each of these variables the difference between the values for the two
debriefings given by the same instructors was obtained, providing a delta score.
The average of the delta scores for these four instructors was compared to delta
scores obtained by 448 random pairings among instructors who gave only one
briefing.
4.0 RESULTS
4.1 General Observations
At all five airlines most debriefings were not conducted immediately after the
LOFT. Instead, after a short break, the instructor and crew first returned to the
simulator to conduct about two hours of "batting practice" as rehearsal for the
proficiency check that would follow the next day. A few instructors, apparently on
their own initiative when scheduling allowed, reversed the order so they could
debrief the LOFT before batting practice.
At all airlines most debriefings followed the same general format. The instructor
would either give a very short introduction or no introduction at all, and then lead
discussion of segments of the LOFT in the chronological order in which they
occurred. Rarely did the instructor engage the crew in setting an agenda for
discussion, although some instructors invited general comments on the LOFT
before starting the discussion of specific segments. In the four airlines with video
equipment, the instructor generally used a video segment to begin the discussion
of related portions of the LOFT. A few instructors varied this general format; for
example, one instructor systematically went through the CRM categories
displayed on a wall poster, asking the crew to identify places in the LOFT in
which they had employed each category.
For most variables large differences occurred among debriefings within each
airline. For some variables substantial differences also occurred in the average
values between airlines, although in most cases the within-airline variability
prevented the differences between airlines from being statistically significant.
Across airlines, instructors' ratings of crew performance averaged 3.6 (SD = .90)
for CRM and 3.5 (SD = .89) for technical on a 1 to 5 scale in which 1 = poor, 3 =
average, and 5 = exemplary. No statistically significant differences were found
among airlines.
The percentage of the talking done by instructors was negatively correlated (p <
.01) with the percentage of the talking done by each category of crew member
(CA: r = -.62; FO: r = -.83; FE: r = -.77). In contrast, the percentage of talking by
captains was not significantly correlated with the percentage of talking by first
officers or flight engineers, but the percentage of talking by first officers was
positively correlated with the percentage of talking by flight engineers (r = .68, p <
.05).
On average, 41% of instructor words and 52% of crew words were directed to the
performance of the crew in the LOFT (Table 6). Instructors emphasized positive
aspects of crew performance (18%) over negative aspects (3%) and ways to
improve performance (4%). Most of the crews' words concerning performance
were neutral descriptions of what they did (33%), compared to positive aspects
(8%), negative aspects (6%), and ways to improve (5%).
The content of the crews' remarks mirrored the content of the instructors'
remarks. The percentages of crew words directed to discussion of CRM,
technical, positive performance, negative performance, and ways to improve
performance were all significantly positively correlated with the percentages of
instructor words directed to these topics (Tables 7a and 7b).
4.2.3 Instructor questions. Most instructors asked a large number of questions,
averaging 48 per hour among two-person crews (Table 8a). Among two-person
crews, 60% of these questions were directed to specific crew members. Similar
results were observed with three-person crews (Table 9a). No significant
differences were found in either the proportion of questions directed to each crew
member or in the proportion of non-directed questions answered by each crew
member (Tables 8b & 9b), although the proportion answered by the flight
engineer was substantially lower, falling just short of statistical significance (p <
.06).
4.2.5 Videos. On average, instructors showed 8.8 (SD = 5.0) video segments per
hour, each averaging 150 (SD = 113) seconds in duration. No significant
differences were found among airlines.
On average, individual crew members asked about six questions per hour. To
analyze the character of crew questions, the set of all crew questions from
airlines Y and Z (n = 98) were divided into three categories. Proactive questions
address the content of the debriefing, raising new issues or bringing new
information into the discussion (e.g., Did you realize I had not finished the
checklist?). Reactive questions respond to a prompt without adding new
information, usually to disambiguate what was said or meant (e.g., Do you mean
the taxi checklist or the predeparture checklist?). Miscellaneous questions are
generally extraneous (e.g., "Do I have time for a coke?") or meta-conversational
(e.g., "You know what I mean?").
Thirty-five percent of crew questions were proactive, 34% were reactive, and
30% were miscellaneous (Table 11). Sixty percent of the proactive questions
addressed CRM, technical, or mixed topics, but only 12% of the reactive
questions, and 7% of the miscellaneous questions addressed CRM, technical, or
mixed topics.
Three other measures of crew participation were also examined: the number of
analyzing utterances per hour, the number of words per utterance, and the
number of words per response to the instructor's questions (Table 13). Analyzing
utterances were defined as those that go beyond simple description of events
and actions to examine underlying factors and how those factors influenced the
outcome (see coding rules in Appendix A). The number of analyzing utterances
per hour averaged 6.2 (SD = 4.7), with no significant differences among airlines
or among the three crew member positions. The number of words per utterance
and the number of words per response averaged 22 (SD = 10) and 30 (SD = 17),
respectively, with no significant differences among airlines or among the crew
member positions.
4.3.2 Correlations. Crew scores on Analysis & Evaluation and Depth of Activity
were significantly positively correlated with instructor Questions, Encouragement,
and Focus, with coefficients ranging from .51 to .78 (Table 16 and Figure 2).
Instructor Introduction and Use of Videos were not significantly correlated with
crew scores on the battery. However, the third item in the Introduction category
was significantly positively correlated with Crew Analysis & Evaluation (r = .45, p
< .006), and the third item in the Use of Videos category was significantly
positively correlated with Crew Analysis & Evaluation (r = .45, p < .02) and fell
just short of significant positive correlation with Depth of Activity (r = .38, p <
.055).
4.3.3 Effect of introductions. The ten debriefings for which the instructor
Introduction scores were 1.0 (the lowest possible score) and the nine debriefings
for which the Introduction scores were the highest (ranging from 1.8 to 4.9) were
analyzed further. Crew Analysis & Evaluation scores for the latter group were
significantly higher than for the former group (Table 18). No significant difference
between the two groups was found for Depth of Activity.
4.4.2 Instructor battery with crew descriptive. The correlations between the
five instructor battery variables and seven crew descriptive variables involving
the nature of crew participation were examined (Table 20). The Introduction
category was significantly positively correlated with crew words per utterance,
words per response, and percent CRM. Encouragement was significantly
positively correlated with crew percent participation, words per utterance, words
per response, self-initiated words, analyzing utterances, and percent CRM.
Questions and Focus showed a pattern of correlations similar to that of
Encouragement, except that the correlations with words per response and self-
initiated words were smaller and not statistically significant. The Use of Videos
category was significantly positively correlated with percent CRM only.
4.4.3 Crew battery with crew descriptive. Table 21 displays the correlations
between the two crew battery categories and the seven crew descriptive
variables. Both Analysis & Evaluation and Depth of Activity were significantly
positively correlated with all seven descriptive variables except proactive
questions.
5.0 DISCUSSION
The five companies studied appear to be representative of large, well-established
U.S. airlines. Although some differences occur, debriefings at these five
companies show many common patterns. These findings, however, may not be
representative of smaller, regional, or newly-started airlines, some of which have
not developed CRM and LOFT programs to the extent that major airlines have.
The large variability observed among instructors at each airline has important
implications. For some variables the average values differed substantially among
some of the airlines, although given the large variability, few of these differences
were statistically significant. At airlines W and X, only four and five debriefings,
respectively, were observed because not many LOFT sessions were run during
our visits. With this small sample size and the variance observed, the standard
errors for some of the mean values are large; thus, especially for these two
airlines, the representativeness of these mean values is uncertain.
For the reasons discussed above, one cannot conclude from these data whether
real differences exist among the airlines on most dimensions (one major
exception is emphasis on CRM, discussed below). What is clear is that individual
instructors at each airline differed enormously in their effectiveness as facilitators
and in their emphasis on CRM topics and crew participation. This large variability
within all five airlines overshadows any differences that might exist among the
airlines. This finding reveals an urgent need for additional training and
standardization within each airline (see section 5.4).
Instructors who rated the crews' LOFT performance as high tended to conduct
shorter debriefings. During interviews with instructors after each debriefing, some
instructors indicated that some of them feel there is less to discuss with a crew
that has performed well, and these instructors wanted to avoid "nit-picking" good
performance. We suspect this attitude may shortchange high performing crews. It
is important for these crews to analyze why things went well in order to help them
make explicit the factors and behaviors that led to success. These behaviors may
have been intuitive and may have depended on the compatibility of the particular
two or three crew members involved. In order to take the lessons learned back to
the line and apply then in situations in which the crew may not be so compatible,
it would be helpful for the crew members to explicitly discuss what makes certain
behaviors effective. Also, even high-performing crews need a chance to practice
the as yet infrequently used skill of self-debriefing.
The content of the instructors' utterances and the content of the crews'
utterances were highly correlated along most dimensions examined. Although
correlation does not necessarily imply causality, our subjective impression is that
the general content and emphasis of the debriefings was driven almost
exclusively by the instructors. This impression is supported by the pattern of
discourse, discussed below.
One might wonder if the percent of participation by the instructor might be driven
by the crew; an instructor might be forced to do more of the talking if he or she
tried unsuccessfully to induce the crew to participate substantially. However, the
data suggest otherwise: the battery variable Encouragement was strongly
negatively correlated with percent instructor participation, which is not consistent
with instructors resorting to lecturing only after seriously attempting to facilitate
crew participation. Also, our subjective impression is that instructors seemed
predisposed to whatever level of facilitation they used.
The large number of questions asked by most instructors suggests that they are
attempting to elicit crew participation. The number of questions asked by
instructors was not significantly correlated with any measures of crew
participation, but this might reflect a limitation of the across subjects design of
this study. An instructor might increase the participation of a given crew by
asking more questions, but this may be confounded by the possibility that
instructors increase the number of questions they ask when they encounter a
crew that participates inadequately. The crew prone to low participation may
increase its activity in response to questions but still may remain below average.
The battery category Questions, which addresses the way in which instructors
ask questions and takes into account the crew with which the instructor is
confronted, appears to be a much more useful measure than the simple number
of questions the instructors ask. Instructors' scores on the battery category were
significantly positively correlated with several descriptive measures of crew
participation and both battery categories of crew participation.
Many instructors frequently interrupted crew utterances, and in many cases the
crew members never completed their comment after the interruption.
Surprisingly, the frequency of interruption was not correlated with any of the
descriptive or battery measures of crew participation. Nevertheless, it is hard to
believe that crew members find frequent interruptions encouraging.
On average, individual crew members made only about six utterances per hour
that were characterized as "analyzing". For coding purposes the definition of
"analyzing" was necessarily arbitrary, and other definitions might have yielded
numbers substantially larger or smaller. Nevertheless, this rough characterization
suggests substantial room for improvement toward one of the major goals of the
debriefing.
Properly speaking, the use of the video of the crews' LOFT performance is not
technically a component of facilitation, but it is widely regarded as an important
tool that can help the crews understand their performance. The nature of the data
(transcribed audio tapes of the debriefing) limited the types of items that could be
used to asses the instructors' Use of Videos. For example, what may be one of
the most important aspects of the video clips, their content, could not be
measured. The items in Use of Videos showed little predictive power for any
aspect of crew performance except percent CRM, and this correlation may only
reflect the fact that instructor scores on Use of Videos were fairly strongly
correlated with instructor percent CRM. Thus we are inclined to delete this
category from the battery.
The most common problem, failing to state explicitly the expectation for crew
participation, is discussed above. Twenty-eight percent of instructors made no
statement at all about expectations and only one instructor gave an explicit
rationale for why the crew should take an active role. Other common mistakes
included failing to pause when the crew did not respond immediately to
questions, keeping the discussion centered on the instructor instead of
encouraging the crew to interact with each other, making long soliloquies,
evaluating crew performance before eliciting crew self-evaluation, failing to push
beyond superficial description of events, and not getting the crew to analyze why
things went well.
A companion to this report describes in detail specific techniques instructors
used and suggests ways to integrate these techniques for effective facilitation
(McDonnell, Jobe, & Dismukes, in press). This companion report, written as a
training manual for instructors, also suggests ways to avoid common facilitation
mistakes.
The fact that instructor scores on Questions, Encouragement, and Focus were
distinctly bimodal and highly intercorrelated suggests that the instructors either
grasped the concept of facilitation and were able to put it into practice or did not
grasp the concept and were therefore unable to practice it effectively. Alternately,
the instructors who were not effective facilitators may not have "bought into" the
concept of facilitation or might simply have lacked the ability for this type of
approach.
These findings suggest that the airlines face an issue of standardization and
quality control of debriefings. Although no attempt was made to measure these
characteristics, it was clear that the great majority of instructors were highly
competent technically, were conscientious, and displayed strong interpersonal
skills. All seemed comfortable with and committed to the concepts of CRM. Thus,
the variability may reflect inadequate training of instructors in the techniques of
facilitation. When interviewed, several instructors spontaneously volunteered that
they did not feel adequately trained to facilitate. To date, in most airlines with
which we are familiar, training in facilitation is vague, consisting mainly of general
concepts and adages (e.g., "Don't insist on closure"). However, facilitation,
especially because it departs radically from the instructional techniques
traditionally used in aviation, requires hands-on training in which instructors
observe expert facilitators, practice facilitating, and receive feedback.
As this report is being written, several airlines are expanding their training in
facilitation, and this can be expected to improve the conduct of debriefings.
Currently, an industry group, the ATA AQP LOFT/Instructor Focus Group, is
preparing a paper that will provide guidance on training instructors in facilitation,
evaluation of crew performance, and related topics.
These findings also suggest that the concept of crews debriefing themselves
using the instructor as a resource (a concept expressed frequently in the CRM
literature and in AC 120-35C), though a worthwhile goal, is rather idealistic. Only
one of the instructors observed attempted to have the crew lead their own
debriefing. Though that debriefing was one of the better ones in terms of the level
of crew participation, the crew only partially understood what constituted a good
debriefing and needed considerable help. In order for crews to take greater
responsibility for the debriefing they must first be told how to conduct one. It
would also help if crews could observe another crew debriefing themselves
effectively; this could be the subject of classroom training that precedes the
LOFT. Crews may need to practice self-debriefing of several LOFTs before they
become proficient.
Instructors are encouraged to attempt to facilitate at the highest level possible for
each crew. Realistically, however, most crews do not yet have the skills and
motivation needed to lead their own debriefings without substantial assistance
from the instructor. It may be possible to change this situation over time if LOFT
instructors consistently encourage crews to take a proactive role in debriefing
their own training and to consider the benefits of debriefing line operations.
3. This study provides empirical evidence that facilitation, when used effectively,
substantially increases the depth of crew participation and the quality of crew
analysis and evaluation of their performance.
5. Within each of the five airlines, instructors varied widely in their conduct of
debriefings, especially in terms of emphasis on CRM, emphasis on crew
participation, and effectiveness in facilitation. Not surprisingly, the character of
crew participation varied similarly, and consequently it seems likely that how
much the crews learned from the LOFT experience may also have varied
considerably. This suggests a need for better standardization within companies.
The great variability within individual airlines obscured the statistical significance
of differences observed among the airlines.
6. These findings suggest that instructors need better training in facilitation. One
way to enhance training would be to emphasize hands-on practice and to follow
up with mentoring by instructors who are themselves expert facilitators. The
current literature on facilitation is rather idealistic, and instructors may become
discouraged when they discover that crews sometimes do not immediately
respond as desired. Instructor training should address obstacles to effective
facilitation and should provide specific techniques to use when crews do not
initially respond. Training should explain to instructors that facilitation can be
conducted at different levels ranging from predominantly crew-led, with instructor
assistance, to predominantly instructor-led, but still emphasizing self-discovery
by the crew as much as possible. Instructors should adapt their level of
facilitation in response to the skill and responsiveness of the particular crew.
Instructor Scores
IP
Introduction 35a .91
Questions 36 .78
Encouragement 36 .80
Focus 36 .84
Use of Videosb 18c .77
Crew
Analysis & Evaluation 36 .78
Depth of Activity 36 .73
bReported reliability for Videos is for crews Y and Z only. Reliability could not be
calculated for all crews because one item was changed after scoring was
completed, and that item was recoded by only one rater.
cThe video equipment was not working for one of the 19 crews in Airlines Y and
Z.
Mean (SD)
Combined
Airline V Airline W Airline X Airline Y Airline Z
Airlines
28.1 (14.8) 29.2 (2.9) 40.3 (25.5) 36.9 (15.6) 23.1 (7.3) 30.7 (15.2)
Mean (SD)
Combined
Airline V Airline W Airline X Airline Y Airline Z
Airlines
Instructor:
58(15) -- 61(18) 54(16) 67(14) 61(15)a
2-person crews
50(3.5) 58(27) -- 40(16) 41 49(20)a
3-person crews
Captain:
19(6.9) -- 24(8.2) 22(8.1) 19(8.6) 21(7.8)b
2-person crews
23(17) 16 (8.9) -- 22(7.9) 21 20(9.4)
3-person crews
First Officer:
23(9.4) -- 15(10) 23(13) 14(7.0) 18(9.7)
2-person crews
16(12) 13(9.2) -- 27(14) 23 19(13)c
3-person crews
Flight Engr:
12(2.8) 14(11) -- 12(7.9) 15 13(7.8)
3-person crews
Note: Differences among airlines were not statistically significant. Significant differences among
participants:
a Instructor > captain, first officer, flight engineer (p<.01); b captain > first officer
(p<.01); c first officer > flight engineer (p<.03).
Mean (SD)
Note. Statistically significant differences were found among airlines: a Y>W; Z>V,W,X. b X>Y,Z. c
V>Y,Z; W>X,Y,Z. d not statistically different. e Y>V,W,X; Z>V,W,X. f V>Y,Z; X>Y,Z. g V>X,Y,Z;
W>X,Y,Z. h X>Z.
Mean (SD)
Airline Combined
Airline V Airline X Airline Y Airline Z Airlines
W
Positive aspects
19(11) 5.8(5.1) 15(9.3) 16(13) 24(12) 18(12)
% of IP words
6.5(7.3) 3.8(5.6) 7.4(13) 9.9(8.9) 9.5(12) 8.0(9.6)
% of crew
words
Negative
aspects
3.8(2.7) 3.3(2.5) 9.4(13) 1.1(2.1) 1.6(2.6) 3.2(5.5)
% of IP words
6.6(4.1) 8.0(7.9) 9.8(12) 5.1(3.8) 3.4(7.2) 5.9(6.7)
% of crew
words
Ways to
improve
5.0(4.4) 4.5(5.3) 6.8(6.7) 3.0(3.2) 2.7(4.4) 4.1(4.6)
% of IP words
3.6(4.3) 5.0(8.7) 5.6(4.0) 4.6(5.1) 5.6(8.6) 4.8(6.1)
% of crew
words
Neutral
description
18(14) 17(9.6) 9.4(4.5) 21(7.0) 15(8.1) 17(9.5)
% of IP words
40(15) 36(15) 25(18) 28(15) 33(26) 33(19)
% of crew
words
Performance
total
46(21) 30(14) 41(15) 41(13) 43(13) 41(15)
% of IP words
56(22) 53(19) 47(17) 48(21) 56(27) 52(21)
% of crew
words
Mean (SD)
Combined
Airline V Airline W Airline X Airline Y Airline Z
Airlines
Mean (SD)
Combined
Airline V Airline W Airline X Airline Y Airline Z
Airlines
Mean (SD)
Combined
Airline V Airline W Airline X Airline Y Airline Z
Airlines
bquestions directed to FE: V>WY; c total number of questions per hour: V>WY.
Mean (SD)
Combined
Airline V Airline W Airline X Airline Y Airline Z
Airlines
statements that raise issues, introduce topics, or add information to an existing topic. Statements = all
utterances that do not
fit the criteria for R, S1, or Q. Question = any utterance that explicitly asks a question.
Table 11. Distribution of Crew Questions (number per category)
Non-
CRM Technical Mixed Total
specific
Proactive 7 11 3 14 35
Reactive 4 3 0 26 33
Miscellaneous 0 2 1 27 30
Total 11 16 4 67 98
Mean (SD)
Combined
Airline V Airline W Airline X Airline Y Airline Z
Airlines
CA 4.9(3.6) 1.7(2.1) 7.5(8.5) 1.5(1.7) 1.2(1.7) 3.0(4.3)
Note. No statistically significant differences were observed between two and three person crews.
Statistically significant differences found among airlines: CA: X>Z; FE: V>WY.
Mean (SD)
Captain First Officer Flight Engineer Crew Average
Analyzing utterances
7.0 (6.2) 6.4 (6.1) 3.4 (2.8) 6.2 (4.7)
per hour
Words per utterance 21 (10) 24 (13) 17 (9.2) 22 (10)
Note. No statistically significant differences were found between airlines or crew positions.
Mean(SD)
Combined
Airline V Airline W Airline X Airline Y Airline Z
Airlines
Instructor Profile:
1.5(.65) 1.4(.73) 1.1(.13) 2.1(1.3) 1.4(.42) 1.6(.83)
Introduction
3.9(1.7) 3.1(1.9) 3.4(1.5) 5.0(.66) 4.2(2.0) 4.1(1.6)
Questions
3.8(1.7) 3.5(2.4) 3.3(1.7) 5.1(.66) 3.9(2.0) 4.1(1.7)
Encouragement
3.2(1.8) 2.9(1.0) 3.0(1.3) 5.0(.69) 4.0(1.7) 3.8(1.6)
Focus
-- 4.3(.85) 2.9(.62) 4.5(1.4) 5.1(1.0) 4.4(1.2)
Use of Videos
Crew Profile:
3.3(1.3) 3.4(1.2) 3.3(1.1) 4.8(.87) 4.2(1.8) 3.9(1.4)
Analysis & Eval.
4.0(1.0) 4.2(1.5) 4.0(1.5) 5.1(1.1) 4.4(1.9) 4.4(1.4)
Depth of Activity
Note: Numbers are average scores of two independent raters (except Video scores for airlines W
& X, which were coded
by only one rater) on a 7-point Likert scale: 1 = poor, 2 = marginal, 3 = needs improvement, 4 =
adequate, 5 = good,
Crew
Analysis &
36 1 6 8 4 13 3 1
Eval.
Depth of
36 1 2 8 5 11 7 2
Activity
Mean (SD)
Analysis &
Introduction Scores N Depth of Activity
Evaluation
1.0 10 3.2 (1.3)* 4.1 (1.4)
Note. The ten debriefings for which instructor Introduction scores were lowest were compared
with the nine debriefings for which Introduction scores were highest.
Descriptive variables
% words
% words
Words ## non-
Battery % total Total # addressing
per directed
directed addressing
Variables questions
participation questions performance
utterance questions CRM
Introduction -.07 .12 .41* -.20 .42* .05 .35*
Questions -.49** -.38* .56*** .10 .60*** .05 .35*
Encourage -.75*** -.58*** .38* .15 .43** -.04 .25
Focus -.40* -.31 .50** .08 .52*** .12 .45**
Use of Videos -.06 .09 .24 .17 .38 .25 .69***
a See Debriefing Assessment Battery (Appendix C)
b See Appendix E
Table 20. Correlations Between Instructor Battery Variables and Crew Descriptive Variables
Instructor
Self-
Percent Words per Analyzing Proactive
Words per initiated Percent
Battery
utterance CRM
variables participation response utterances questions
words
Descriptive variables
# of
Self-
Percent words Analyzing Proactive
Battery Words per initiated Percent
variables utterance CRM
participation per utterances questions
words
response
Instructor variables
S1 statements
-.79*** -.62*** -.07 .20 -.07 -.06
(# words per hour)
Analysis &
-.67*** -.39* .23 .05 .12 .40*
Evaluation
Mean (SD)
of variable
different
same instructor instructor
REFERENCES
Baddeley, A. (1990). Human memory: Theory and practice. Massachusetts:
Simon & Schuster, Inc.
Casey, D., Roberts, P., & Salaman, G. (1992). Facilitating learning in groups.
Leadership and Organization Development Journal, 13(4), 8-13.
Duval, S., and Wicklund, R. A. (1972). A theory of objective self awareness. New
York: Academic Press.
Eitington, J. E. (1989). The winning trainer. Houston: Gulf Publishing Company.
Gibb, P. (1982, July). The facilitative trainer. Training and Development Journal,
14-19.
Gow, L., & Kember, D. (1993). Conceptions of teaching and their relationship to
student learning. British Journal of Educational Psychology, 63, 20-33.
Jacobsen, D., Eggen, P., & Kauchak, D. (1989). Methods for teaching: A skills
approach (3rd ed.). Columbus: Merrill Publishing Company.
Lauber, J. K., & Foushee, H. C. (1981). Guidelines for line-oriented flight training.
(NASA Conference Publication 2184). Moffett Field, CA: NASA-Ames Research
Center.
McDonnell, L. K., Dismukes, R. K., & Jobe, K. K. (In preparation). A short form
battery for assessing LOS debriefings.
McDonnell, L. K., Jobe, K. K., & Dismukes, R. K. (In press). Facilitating LOS
debriefings: A training manual.
Mills, P., & Roberts, B. (1981, March). Learn to guide and control discussion.
Successful Meetings, 96-97.
Moos, R. H. (1994). Group environment scale manual: Development,
applications, research. Palo Alto, CA: Consulting Psychologists Press, Inc.
Ornstein, A. C. (1990). Strategies for effective teaching. New York: Harper &
Row.
Zemke, R., & Zemke, S. (1981). 30 things we know for sure about adult learning.
In P. G. Jones (Ed.) (1982), Adult learning in your classroom: The best of training
magazine's strategies and techniques for managers and trainers. Minneapolis,
MN: Lakewood Books.
Appendix A. Coding
Utterance factors coded
Utterance length: number of words
Instructor (IP), 2nd Instructor in role of Flight
Speaker: Engineer (FEI). Captain (CA), First Officer
(FO), or Flight Engineer (FE)
Completed (C), Unfinished (U), Interrupted (I),
Interruptions/Interjections: Interrupted and Unfinished (I/U), Active
listening interjection (I/AL)
Question, Command, Response, or Statement
Utterance type: (Statements self-initiated by crew further coded
as S1)
Target of Question (if
Captain (CA), First Officer (FO), or Flight
clearly directed to a
Engineer (FE)
particular crew member):
Crew Proactive Questions: "P" if crew question is proactive, "O" (Other) if
it is a reactive or miscellaneous question
CRM, Technical, Mixed (CRM & Technical),
Topic type:
or Non-Specific
"A" if crew analyzes situation/performance,
Analysis:
"O" (Other) if not
Evaluation of crew
Positive, Negative, Improve, or Neutral
performance:
Video factors coded
All video segments are coded by indicating
ON ( ): segment number with duration in parentheses
[e.g., ON #1 (:45)]
Code end of video segments by indicating
OFF:
(OFF)
Time spent searching in silence [e.g., SEARCH
SEARCH ( ):
(:30)]
CODING RULES
1. Fill in a word count for every utterance for which a speaker and content are
identified. Do not count utterances in which speaker is identified but the words
are unintelligible; or words are transcribed but speaker cannot be identified.
Speaker (SPKR)
Identify the speaker of each utterance using one of the following; IP, CA, FO, FE,
or FEI.
3. Type titles in parentheses [e.g., (CA) or (FO)] in place of spoken names and
type (XX) in place of spoken name of airline.
9. Record length of video silent search time (no one speaks while IP tries to find
a specific video segment) in bold type.
1. Code all utterances that are not completed (whether the speaker is interrupted
or trails off) as "U" and code all completed utterances as "C"
2. Code all utterances that interrupt or interject the preceding speaker as "I"
(code as "I/U" if the interruption is not completed, either because the preceding
speaker keeps talking or another speaker interrupts the interruption)
3. Code all active listening as "AL" (code interjections of active listening as "I/AL")
1. Record a "P" in the crew PAQ column if crew question is proactive, or an "O"
(other) if the question is not proactive (i.e., reactive or misc.)
CRM = Pertains to the coordination and interaction of the crew and specifically
relates to one or more CRM issues or topics.
Mixed = Has between 1/3 and 2/3 of both CRM and technical.
(ANALYSIS)
Code all utterances that indicate the speakers are Analyzing the situation &/or
their performance in the LOFT by considering any of the following issues (both
explicit and implicit) as A (Analyzes). Code all utterances which are not analytical
as O (Other).
Generally, analyzing utterances are those that go beyond just describing what
happened to discussing why it happened and identifying what factors contributed
to the situation and/or how these factors influenced the outcome.
explanations of why something was done and/or done a certain way, or what
could have been done differently. Key words include because, should have,
could have , and might have (e.g., "I think we could have performed faster in
holding because we had to take a couple of turns in holding just to make sure we
got set up." and "I felt a little disorganized pushing off and taxiing out and doing
all of that and then having to de-ice; that breaks your flow because you don't put
the flaps down")
how & why factors influenced decisions, actions, and outcomes (e.g., "The
reason this influenced my decision/actions was ..." and "I was thinking this, so I
did this").
contingencies (e.g., "It might have been a lot different if we had asked for more
time before we took that turn. Maybe I should have asked for one more minute.")
(EVALUATION)
Code all utterances which do not fit into the above categories as O (other)
(VIDEO)
(COMMENTS)
Appendix B.
Appendix C.
INSTRUCTOR PROFILE
The Instructor Profile is a summary of the strategies and techniques IP's use to assist crews in
conducting their own debriefings while giving direction and focus as necessary. The two main
goals of the debriefing are to 1) get the crew to perform an in-depth analysis of the situation that
confronted them, how they understood and managed the situation, the outcome, and ways to
improve, and 2) get the crew to participate in a proactive, rather than reactive, manner in which
they initiate discussion and elaborate beyond the minimal. These goals are based on the
assumption that active participation by the crew will result in a higher level of learning and
increased likelihood of transfer to the line.
Directions:
Use the scale below to rate the instructors on each of the following elements, then total the
scores to get the overall rating for each category
1234567
Introduction
One purpose of the introduction is to let the crew know that participation and self-evaluation are
expected of them, and why it is important.
Makes clear that his role is guide/facilitator and that crew should do most of the talking
Clearly conveys that crew should take an active role, initiating discussion rather than just
responding to him
Clearly conveys that he wants crew to dig deep, critically analyzing the LOFT and their
performance
Gives a persuasive rationale for the crew to participate actively and make their own analysis
Questions
The purpose of asking questions is to get the crew to participate, focus the discussion on
important topics, and enlist the crew in discussing the topics in depth.
Asks an appropriate number of questions to get crew talking & lead them to issues
Avoids answering for the crew when they do not respond immediately or correctly and uses a
pattern of questioning that keeps the focus on the crew
Uses probing and follow-up questions to get crew to analyze in depth and to go beyond yes/no
and brief factual answers
Uses questioning techniques to encourage interaction and sharing of perspectives among crew
members
Encouragement
Encouragement refers to the degree to which the instructor encourages and enables the crew to
actively and deeply participate in the debriefing.
Conveys sense of interest in crew views and works to get them to do most of the talking
Encourages continued discussion through active listening, strategic pauses, avoiding disruptive
interruptions, and/or following up on crew-initiated topics
Encourages all members to participate fully, drawing out quiet members if necessary
Refrains from giving long soliloquies or giving his own analysis before crew has fully analyzed
Overall rating of Encouragement
Encourages crew to analyze along CRM dimensions the situation that confronted them, what they
did to manage the situation, and why they did it
Encourages crew to evaluate their performance and/or ways they might improve
Encourages crew to explore CRM issues and how they specifically affect LOFT performance and
line operations
Encourages crew to analyze issues, factors, and outcomes in depth, going beyond simply
describing what happened and what they did
Use of Videos
One stated purpose of showing videotaped segments of the LOFT is to enable the crew members
to see how they performed from an objective viewpoint so they can better evaluate their
performance. More realistically, perhaps, the video reminds the crew of the situation, aiding their
memory and providing a focus for discussion.
Uses video equipment efficiently: is able to find desired segment without wasting time and pauses
the video if substantial talk begins while playing
Consistently discusses video segments, using them as a springboard for discussion of specific
topics
Has a point to make and uses the video to make that point.
CREW PROFILE
The crew profile measures the degree and depth of participation by the crew.
Directions:
Use the scale below to rate the crew on each of the following elements, then total the scores to
get the overall rating for each category
1234567
Analyze along CRM dimensions the situation that confronted them, what they did to manage the
situation, and why they did it
Explore CRM issues and how they affect LOFT performance and line operations
Analyze issues, factors, and outcomes in depth, going beyond simply describing what happened
and what they did
Initiate dialogue rather than just responding to questions, and/or interact with each other rather
than only with the IP
Behave in a predominantly proactive rather than reactive manner, being actively involved rather
than just passing through the training
Appendix D.
- Very specifically and thoroughly explains that his role is guide/facilitator and that crew should do
most of the talking and lead the discussion
- Sets strong expectations for proactive crew participation, explicitly stating they should initiate
discussion rather than just responding to IP questions
- Explicitly and emphatically states that crew should dig deep, critically analyzing the LOFT and
their performance
- Gives a persuasive rationale for the crew to participate actively and make their own analysis and
makes a strong case for why it is important to do it this way.
Very Good:
- Clearly conveys that his role is guide/facilitator and that crew should do most of the talking and
lead the discussion
- Clearly conveys that crew should take an active role, initiating discussion rather than just
responding to IP
- Clearly conveys that crew should dig deep, critically analyzing the LOFT and their performance
- Clearly conveys the general rationale for the crew to participate actively and make their own
analysis
Good:
- Conveys that his role is guide/facilitator and that crew should do most of the talking, but not
specifically that they should lead their own discussion.
- Conveys that crew should take an active role, initiating discussion rather than just responding to
IP
- Conveys that crew should dig deep, critically analyzing the LOFT and their performance
- Makes a general statement of the rationale for the crew to participate actively and make their
own analysis
Adequate:
- Conveys that his role is guide/facilitator and that crew should do most of the talking, but does
not emphasize strongly
- Conveys that crew should take an active role and initiate discussion
- Conveys that crew should analyze the LOFT and their performance
- Gives a clear, though implicit rationale for the crew to participate actively and make their own
analysis
Needs Improvement:
- Implies that his role is guide/facilitator and that crew should do most of the talking, but does not
emphasize strongly
- Implies that crew should take an active role and initiate discussion
- Implies that crew should analyze the LOFT and their performance
- Gives a vague, implicit rationale for the crew to participate actively and make their own analysis
Marginal:
- Implies that his role is guide/facilitator and that the crew should talk, but does not emphasize
- Implies that crew should take an active role, but does not specify what they should do.
- Implies that crew should discuss the LOFT and their performance
Poor:
- Does not make clear that his role is guide/facilitator or that crew should do most of the talking
- Does not make clear that crew should take an active role or initiating discussion
- Does not make clear that crew should dig deep or critically analyze the LOFT and their
performance
- Does not give rationale for the crew to participate actively and make their own analysis
IP Questions
Outstanding:
- Consistently asks questions as appropriate to get crew talking & lead them to issues
- Consistently rewords questions or otherwise avoids answering for the crew when they do not
respond immediately or correctly, and consistently uses a pattern of questioning that keeps the
focus on the crew
- Consistently uses probing and follow-up questions as a tool to evoke in-depth discussion and
optimize crew self-discovery, while forcing crew to go beyond yes/no and brief factual answers
- Frequently asks questions when appropriate to get crew talking & lead them to issues
- Predominantly rewords questions or otherwise avoids answering for the crew when they do not
respond immediately or correctly and predominantly uses a pattern of questioning that keeps the
focus on the crew
- Frequently uses probing and follow-up questions as a tool to evoke in-depth discussion and
optimize crew self-discovery, pushing crew to go beyond yes/no and brief factual answers
Good:
- Generally asks questions as necessary to get crew talking & lead them to issues
- Generally rewords questions or otherwise avoids answering for the crew when they do not
respond immediately or correctly and generally uses a pattern of questioning that keeps the focus
on the crew
- Generally uses probing and follow-up questions to get crew to analyze in depth and to go
beyond yes/no and brief factual answers but may steer crew to predetermined answers while
emphasizing self-discovery.
Adequate:
- About half of the time asks questions when necessary to get crew talking & lead them to issues
- Generally avoids answering for the crew when they do not respond immediately or correctly, but
may not reword the questions. On average uses a pattern of questioning that keeps the focus on
the crew
- On average uses probing and follow-up questions to get crew to analyze in depth and to go
beyond yes/no and brief factual answers but steers crew to predetermined answers as much as
emphasizes self-discovery.
Needs Improvement:
- Sometimes asks questions when necessary to get crew talking & lead them to issues
- To some extent avoids answering for the crew when they do not respond immediately or
correctly and uses a pattern of questioning that keeps the focus on the crew
- Sometimes uses probing and follow-up questions to get crew to analyze in depth and to go
beyond yes/no and brief factual answers but steers crew to predetermined answers more than
emphasizes self-discovery.
Marginal:
- Occasionally asks questions to get crew talking & lead them to issues
- Occasionally avoids answering for the crew when they do not respond immediately or correctly
but generally answers for them rather than keeping focus on the crew.
- Occasionally uses probing and follow-up questions to get crew to analyze in depth but generally
settles for yes/no and brief factual answers
Poor:
- Usually answers for the crew when they do not respond immediately or correctly.
- Rarely uses probing and follow-up questions to get crew to analyze in depth. Usually settles for
yes/no and brief factual answers
IP Encouragement
Outstanding:
- Consistently communicates an interest in crew views and actively strives to get them to do most
of the talking and lead their own discussion.
- Consistently uses active listening and pauses, avoids interrupting, and follows up on crew
topics.
- Consistently encourages all members to participate and draws out quiet members as necessary.
- Consistently refrains from lecturing and giving own analysis before crew.
Very Good:
- Clearly communicates to the crew that their views are important and works to get them to do
most of the talking and to lead their own discussion.
- Frequently uses techniques such as active listening and pauses, avoids interrupting, and follows
up on crew topics to encourage continued discussion.
- Frequently encourages all members to participate and attempts to draw out quiet members as
necessary.
- Usually refrains from lecturing and giving own analysis before crew.
Good:
- Shows a clear interest in crew views and attempts to get them to do most of the talking. Makes
an effort to get crew to lead their own discussion.
- Often uses active listening and pauses, avoids interrupting, and follows up on crew topics.
- Generally encourages all members to participate, drawing out quiet members as necessary.
- Sometimes lectures, but generally gets crew to analyze situation before giving own analysis.
Adequate:
- On average demonstrates a desire to have crew participate and discuss their views.
- Uses some facilitation techniques to encourage crew discussion and generally avoids
interrupting them. Acknowledges crew topics but may not follow up on them thoroughly.
- On average gets the crew to analyze the situation themselves before evaluating and lecturing to
them.
Needs Improvement:
- Shows interest in crew views but does not push them to do most of the talking.
- Sometimes uses active listening and pauses, and follows up on crew topics, but also sometimes
interrupts.
- Expresses a desire for crew to participate but does not put a lot of effort into getting all members
actively involved.
Marginal:
- Only occasionally uses active listening, pauses, and/or follows up on crew topics, and often
interrupts.
- Expresses a desire for crew to participate but puts minimal effort into actively encouraging them
to do so.
- Tends to lecture and analyze for crew without encouraging them to discuss what happened
themselves.
Poor:
- Frequently hinders rather than encourages crew talk and does not follow up on topics initiated
by crew.
- Frequently lectures to crew about what they did and how to improve.
- Continually encourages and pushes crew to analyze along CRM dimensions the situation that
confronted them, what they did to manage the situation, and why they did it.
- Consistently encourages and pushes crew to evaluate their performance and/or ways they
might improve.
- Consistently encourages crew to explore CRM issues and how they specifically affect LOFT
performance and line operations.
- Continually encourages crew to analyze issues, factors, and outcomes in depth, going beyond
simply describing what happened and what they did.
Very Good:
- Frequently encourages and pushes crew to analyze along CRM dimensions the situation that
confronted them, what they did to manage the situation, and why they did it.
- Frequently encourages crew to evaluate their performance and/or ways they might improve.
- Frequently encourages crew to explore CRM issues and how they specifically affect LOFT
performance and line operations.
- Frequently encourages crew to analyze issues, factors, and outcomes in depth, going beyond
simply describing what happened and what they did
Good:
- Generally encourages crew to analyze along CRM dimensions the situation that confronted
them, what they did to manage the situation, and why they did what they did, but may settle for
less than extensive discussion.
- Generally encourages crew to evaluate their performance and/or ways they might improve.
- Generally encourages crew to explore CRM issues, and attempts to get crew to discuss how
they specifically affect LOFT performance and line operations.
- Generally encourages crew to analyze issues, factors, and outcomes in depth. Generally
encourages crew to go beyond simply describing what happened and what they did.
Adequate:
- On average encourages crew to analyze along CRM dimensions the situation that confronted
them and what they did to manage the situation. Encourages but does not push crew to analyze
why they did what they did.
- Tends to encourage crew to evaluate their performance and/or ways they might improve, but
may not pursue thoroughly.
- On average encourages crew to explore CRM issues but tends not to get crew to discuss how
they specifically affect both LOFT performance and line operations.
- Generally encourages crew to analyze issues, factors, and outcomes, but settles for moderate
depth, sometimes letting crew simply describe what happened and what they did.
Needs Improvement:
- Sometimes encourages crew to analyze along CRM dimensions the situation that confronted
them and what they did to manage the situation but does not push crew to discuss why they did
what they did.
- Verbally requests but does not pursue getting the crew to evaluate their performance and/or
ways they might improve.
- Encourages crew to explore CRM issues but does not ask crew to discuss how they specifically
affect LOFT performance and line operations.
- Tends not to push crew to analyze issues, factors, and outcomes in depth. Often settles for
letting the crew simply describe what happened and what they did.
Marginal:
- Only minimally encourages crew to analyze along CRM dimensions the situation that confronted
them and/or what they did to manage it. Does not push crew to discuss why they did what they
did.
- Only occasionally encourages crew to evaluate their performance and/or ways they might
improve.
- Occasionally encourages crew to explore CRM issues, and does not encourage crew to discuss
how they affect LOFT performance or line operations.
- Only occasionally encourages crew to analyze issues, factors, and outcomes in depth. Content
for crew to describe what happened and what they did.
Poor:
- Does not encourages crew to analyze along CRM dimensions the situation that confronted
them, what they did to manage the situation, or why they did it.
- Rarely encourages crew to evaluate their performance or ways they might improve.
IP Use of Videos
Outstanding:
- Consistently uses video equipment efficiently: is able to find desired segment without wasting
time and pauses the video if talk begins while playing.
- Actively evokes and consistently pursues thorough crew discussion of each video segment or
topic.
- Consistently has a point to make and uses the video to make that point.
Very Good:
- Usually uses video equipment efficiently: is able to find desired segment without wasting much
time and pauses the video if substantial talk begins while playing.
- Works to get crew to discuss most of the video segments or topics in detail.
- Usually has a point to make and uses the video to make that point.
Good:
- Tends to use video equipment efficiently: is generally able to find desired segment without
wasting much time and generally pauses the video if substantial talk begins.
- Encourages crew to discuss most video segments or topics and refrains from lecturing to crew
or hindering their discussion.
- Generally has a point to make and usually uses the video to make a point.
Adequate:
- On average shows an appropriate number of videos, usually of appropriate duration, to illustrate
and introduce topics.
- On average uses video equipment somewhat efficiently, finding desired segment without
wasting too much time and generally pausing the video if substantial talk begins while playing.
- Generally encourages crew to discuss video segments or topics, but may also lecture to crew,
thereby somewhat discouraging thorough crew discussion.
- Generally has a point to make, but the point is not always clearly tied to the video.
Needs Improvement:
- Shows somewhat too few or too many videos. Sometimes shows very short and/or very long
segments while trying to illustrate/introduce topics.
- Tends to use video equipment inefficiently: tends to waste some time trying to find desired
segments and is slow to pause the video if substantial talk begins while playing.
- Sometimes encourages crew to discuss video segment or topic, but may lecture, interrupt crew
discussion, and/or not consistently pursue crew discussion.
- Sometimes has a predetermined point to make, and sometimes uses the video to make a point.
Marginal:
- Clearly shows too few or too many videos, sometimes of much too long and/or short a duration.
Many videos not used to illustrate/introduce topics.
- Uses video equipment inefficiently, wasting significant time trying to find desired segments while
rarely pausing the video if substantial talk begins while playing.
- Tends not to discuss video segments, and when they are discussed tends to lecture to crew
about what occurred, only minimally encouraging crew to participate in a discussion.
- Only occasionally has a point to make or uses the video to make a point.
Poor:
- Shows way too few or too many videos which are often much too long and/or short. Does not
use videos to illustrate/introduce topics.
- Uses video equipment very inefficiently: wastes substantial time trying to find desired segments
and fails to pause the video if substantial talk begins while playing.
- Usually does not discuss video segments, and when discussed usually lectures to crew without
encouraging (and often hindering) crew participation.
- Consistently analyze along CRM dimensions the situation that confronted them, what they did to
manage the situation, and why they did it.
- Consistently explore CRM issues and how they affect LOFT performance and line operations.
- Consistently analyze issues, factors, and outcomes in depth, going beyond simply describing
what happened and what they did.
Very Good:
- Frequently analyze along CRM dimensions the situation that confronted them, what they did to
manage the situation, and why they did it.
- Often explore CRM issues and how they affect LOFT performance and line operations.
- Frequently analyze issues, factors, and outcomes in depth, going beyond simply describing
what happened and what they did.
Good:
- Generally analyze along CRM dimensions the situation that confronted them and what they did
to manage the situation. Briefly discuss why they did what they did.
- Generally explore CRM issues and how they affect LOFT performance and/or line operations.
- Generally analyze issues, factors, and outcomes in moderate depth, usually going beyond
simply describing what happened and what they did.
Adequate:
- On average analyze along CRM dimensions the situation that confronted them and what they
did to manage the situation. Briefly discuss why they did what they did.
- On average explore CRM issues and how they affect LOFT performance and/or line operations.
- Analyze some issues, factors, and outcomes in some depth, often going beyond simply
describing what happened and what they did.
Needs Improvement:
- Only part of the time analyze along CRM dimensions the situation that confronted them, what
they did to manage the situation, or why they did it.
- Only sometimes evaluate their performance and ways they might improve.
- Sometimes explore CRM issues but give little discussion of how they affect LOFT performance
or line operations.
- Analyze only a few issues, factors, and outcomes in any depth, sometimes going beyond simply
describing what happened and what they did.
Marginal:
- Occasionally analyze along CRM dimensions the situation that confronted them. Occasionally
discuss what they did to manage the situation or why they did it.
- Only occasionally evaluate their performance and do not discuss ways they might improve.
- Only occasionally explore CRM issues and do not discuss how they affect LOFT performance
and line operations.
- Analyze issues, factors, and outcomes in very little depth, rarely going beyond simply describing
what happened and what they did.
Poor:
- Do little to analyze along CRM dimensions the situation that confronted them, what they did to
manage the situation, or why they did it.
- Rarely explore CRM issues and how they affect LOFT performance and line operations.
- Do not analyze issues, factors, and outcomes in depth; only briefly describe what happened.
- Continually initiate dialogue and pursue issues to completion rather than just responding to
questions, and consistently interact with each other rather than only with the IP.
- Behave in a consistently proactive rather than reactive manner, being actively involved rather
than just passing through the training.
Very Good:
- Frequently go substantially beyond minimal responses to IP questions.
- Frequently initiate dialogue rather than just responding to questions, and often interact with each
other rather than only with the IP.
- Usually behave in a proactive rather than reactive manner, being actively involved rather than
just passing through the training.
Good:
- Tend to initiate dialogue rather than just responding to questions and generally interact with
each other rather than only with the IP.
- Generally behave in a proactive rather than reactive manner, being actively involved rather than
just passing through the training.
Adequate:
- On average initiate dialogue rather than just responding to questions and interact with each
other rather than only with the IP.
- On average behave in a proactive rather than reactive manner, being actively involved rather
than just passing through the training.
Needs Improvement:
- Tend to just respond to questions rather than initiate dialogue. Tend to interact with the IP more
than with each other.
Marginal:
Poor:
- Rarely initiate dialogue; usually just respond to IP. Rarely interact with each other.
- Behave in a consistently reactive rather than proactive manner. Appear to just pass through the
training rather than being actively involved.
CRMPERF -.4881**
IPIMP .4395**
IPNS -.1320
IPDQ_FO .6965**
INTER_UN .8014**
FOPART .3255
CREWIMPR .4525**
CREWNS -.1818
CREWPERE .7368**
FOSIUTPH .5963**
CREWANUT .4429
NONDQ_FO .1322