0% found this document useful (0 votes)
201 views165 pages

MNL 13-2020

The second edition of the Manual on Descriptive Analysis Testing for Sensory Evaluation, edited by Rebecca N. Bleibaum, provides updated methodologies for sensory evaluation, including Consensus/Flavor Profile, Texture Profile, Quantitative Descriptive Analysis, and Spectrum, along with new chapters on Free Choice Profiling and Temporal Methods. This manual emphasizes the importance of descriptive analysis in identifying and quantifying product attributes through trained human assessors, applicable across various industries beyond food and beverages. The updates reflect the evolving applications of these methods in product development, quality control, and marketing strategies.

Uploaded by

ka15149081
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
201 views165 pages

MNL 13-2020

The second edition of the Manual on Descriptive Analysis Testing for Sensory Evaluation, edited by Rebecca N. Bleibaum, provides updated methodologies for sensory evaluation, including Consensus/Flavor Profile, Texture Profile, Quantitative Descriptive Analysis, and Spectrum, along with new chapters on Free Choice Profiling and Temporal Methods. This manual emphasizes the importance of descriptive analysis in identifying and quantifying product attributes through trained human assessors, applicable across various industries beyond food and beverages. The updates reflect the evolving applications of these methods in product development, quality control, and marketing strategies.

Uploaded by

ka15149081
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 165

Rebecca N.

Bleibaum, Editor

Descriptive Analysis Testing for


Sensory Evaluation
Second Edition

ASTM Stock Number: MNL13-2ND

ASTM International, 100 Barr Harbor Drive, PO Box C700, West Conshohocken, PA 19428-2959

Printed in the U.S.A.

BK-AST-MNL_13-200365-FM.indd 1 23/11/20 10:23 AM


Library of Congress Cataloging-in-Publication Data

Names: Bleibaum, Rebecca N., editor.


Title: Manual on descriptive analysis testing for sensory evaluation /
Rebecca N. Bleibaum.
Description: 2nd edition. | West Conshohocken, PA : ASTM International,
[2020] | ASTM stock number: MNL13-2ND | Includes bibliographical
references and index. | Summary: “The Second Edition of this Manual on
Descriptive Analysis Testing for Sensory Evaluation is sponsored by ASTM
International Committee E18 on Sensory Evaluation. Descriptive analysis
is a sensory method by which the attributes of a food or product are
identified and quantified, using human subjects who have been
specifically trained for this purpose. The analysis can include all
parameters of the product, or it can be limited to certain aspects, for
example, aroma, taste, texture, and aftertaste. Many descriptive
analysis methods and method variations are currently employed by sensory
professionals. This second edition contains updates on the four
previously published methodologies that are widely used:
Consensus/Flavor profile, Texture Profile, Quantitative Descriptive
Analysis (QDA), and Spectrum, along with two new chapters on Free Choice
Profiling and a Summary of Temporal Methods. Novel descriptive test
methods are developed and published every year, yet the commonly
utilized toolbox of the fundamental methods, such as those outlined in
this manual, have changed little over the past decades. What has changed
dramatically is the in-depth experience and applications of these
methods in industry. This allows the practitioners an opportunity to
fine-tune and apply the methodologies across a multitude of industries
and product categories. Hence, the need for an updated version of this
manual”– Provided by publisher.
Identifiers: LCCN 2020019032 (print) | LCCN 2020019033 (ebook) | ISBN
9780803171183 (paperback) | ISBN 9780803171190 (pdf)
Subjects: LCSH: Food–Sensory evaluation.
Classification: LCC TX546 .M36 2020 (print) | LCC TX546 (ebook) | DDC
664/.07–dc23
LC record available at https://lccn.loc.gov/2020019032
LC ebook record available at https://lccn.loc.gov/2020019033

Copyright © 2020 ASTM International, West Conshohocken, PA. All rights reserved. This material may
not be reproduced or copied, in whole or in part, in any printed, mechanical, electronic, film, or other
distribution and storage media, without the written consent of the publisher.

Photocopy Rights
Authorization to photocopy items for internal, personal, or educational classroom use, or the internal,
personal, or educational classroom use of specific clients, is granted by ASTM International provided that
the appropriate fee is paid to the Copyright Clearance Center, 222 Rosewood Drive, Danvers, MA 01923,
Tel: (978) 646-2600; http://www.copyright.com/

Publisher:
ASTM International
100 Barr Harbor Drive
PO Box C700
West Conshohocken, PA 19428-2959
Phone: (610) 832-9585 Fax: (610) 832-9555

ISBN 978-0-8031-7118-3
ISBN-EB: 978-0-8031-7119-0

ASTM Stock Number: MNL13-2ND


DOI: 10.1520/MNL13-2ND-EB

ASTM International is not responsible, as a body, for the statements and opinions advanced in the publi-
cation. ASTM International does not endorse any products represented in this publication.

Printed in Hanover, PA
December, 2020

BK-AST-MNL_13-200365-FM.indd 2 23/11/20 10:23 AM


Foreword
The second edition of Manual on Descriptive Analysis Testing for Sensory Evaluation is
sponsored by ASTM International, Committee E18 on Sensory Evaluation. Descrip-
tive analysis is a sensory methodology by which the attributes of products are identi-
fied and quantified, using human subjects who have been specifically screened,
qualified, and trained to use their senses for this purpose. The analysis can include all
modalities of the product, or it can be limited to certain modalities, for example,
aroma, taste, texture, and aftertaste. Many descriptive analysis methods and method
variations are currently used by sensory professionals around the world. This second
edition contains updates on the four previously published methodologies that are
widely used: Consensus/Flavor Profile, Texture Profile, Quantitative Descriptive
Analysis (QDA), and Spectrum, along with two new chapters on Free Choice Profiling
and a summary of Temporal Methods.
Novel descriptive test methods are developed and published every year, yet the
commonly utilized toolbox of the fundamental methods, such as those outlined in this
manual, have changed little over the past decades. What has changed dramatically is
the in-depth experience and applications of these methods in industry. This allows the
practitioners an opportunity to fine-tune and apply the methodologies across a multi-
tude of industries and product categories—hence the need for an updated version of
this manual.
The authors wish to thank the sensory professionals that contributed to the
first edition, including Patricia Keane, Herbert Stone, Alejandra M. Muñoz, Gail
Vance Civille, Alina S. Szczesniak, Margery Einstein, Naomi O. Schwartz, and Robert
C. Hootman. Without their contributions, there would be no second edition.

BK-AST-MNL_13-200365-FM.indd 3 18/11/20 4:13 PM


BK-AST-MNL_13-200365-FM.indd 4 18/11/20 4:13 PM
Contents

Foreword iii

Introduction and Overview vii

1 Consensus Profile Methods Derived from the Flavor Profile Method 1


Edgar Chambers IV and Delores H. Chambers

2 The Texture Profile Method 19


Alejandra M. Muñoz

3 Quantitative Descriptive Analysis 51


Rebecca N. Bleibaum, K. W. Clara Tao, and Heather Thomas

4 Spectrum Descriptive Analysis 77


Gail Civille and Katie Osdoba

5 Free Choice Profiling 99


Margaret J. Hinds

6 Summary of Temporal Methods 139


Suzanne Pecore and Chris Findlay

Index 145

BK-AST-MNL_13-200365-FM.indd 5 23/11/20 10:23 AM


BK-AST-MNL_13-200365-FM.indd 6 18/11/20 4:13 PM
Introduction and Overview
Descriptive analysis is a key methodological approach in sensory evaluation. It has
been called by many professionals the most sophisticated tool currently available to
measure human perception. Since its origin in the 1950s the field of descriptive analy-
sis has grown dramatically in methodology, applications, and adaptations.
The growth of descriptive analysis is evident in terms of the number and type of
organizations that have embraced this sensory technique as part of their core product
research methodologies. Many Fortune 500 companies, as well as many other compa-
nies and research organizations, use data gathered from descriptive analysis panels to
make business and technical decisions about products, including essential measures
such as identifying their key sensory signature and key quality attributes. These sen-
sory attributes can then be used to track product quality and help ensure consistency
for the intended consumer. Descriptive analysis is used to help bridge the gap between
development teams and marketing and is an essential component of brand building
and maintenance, along with providing numerous other benefits to the organization.
It can also create a more efficient product development process by providing insight
into consumer perceptions of the products’ sensory characteristics, similarities and
differences.
Descriptive information has value, among others, in the following key areas:
• Product improvement
• New product development
• Identifying unique sensory space/innovation opportunities
• Competitive assessments
• Ingredient and process sourcing and substitutions
• Mapping and controlling production variability for quality control
• Qualifying co-manufacturers
• Shelf-life studies
• Substantiating sensory claims for advertising (see ASTM E1958)
• Legal disputes
• Other
The product may be a physical “product” such as food or beverages, personal care and
home care products, hand-held devices and electronics, automotive products, and
textiles. It also can be a concept, package, graphics, illustration, or some other experi-
ence that can be described, defined, and measured. Descriptive analysis was originally
developed and has focused on food and beverages. However, its application has grown

BK-AST-MNL_13-200365-FM.indd 7 18/11/20 4:13 PM


viii Descriptive Analysis Testing for Sensory Evaluation: 2nd Edition

well beyond these product categories, and includes agriculture, personal care, home
care, auto care, pet care products, office supplies, paper and textiles, handheld devices
(e.g., shavers, toothbrushes, hair dryers, etc.), art supplies, perceptions of luxury prod-
ucts or services, currency, furniture, etc.
The application of descriptive analysis is limited only by one’s imagination, time,
and budget. The wide array of products currently being evaluated using descriptive
analysis techniques, or at a minimum, some form of attribute evaluations, illustrates
their diversity and value to corporations.
Descriptive analysis is indeed a sophisticated and complex tool from several per-
spectives, as it:
• is the sensory technique that provides qualitative descriptions and quantitative
measures (and temporal behavior) of products, based on the perceptions of a
group of qualified assessors,
• requires the screening and training of assessors, although the amount of training
varies greatly between descriptive methods. In addition, it requires in-depth
discussions for a panel to develop the most appropriate evaluation procedures and
comprehensive descriptive language (these requirements apply only to the
fundamental methods covered in this manual).
Our distinguished chapter authors for the second edition of this manual all have
extensive global experience working with multinational organizations through their
thirty-plus year careers across the above-mentioned categories. These authors volun-
teered their time and expertise in order to share these methods with interested readers,
especially our next generation of sensory scientists.
Since the last edition, it is fair to say that the techniques themselves and appli-
cations of the methods have expanded dramatically. Despite our fundamental
methods having been updated, in reality, their basic techniques have changed
slightly, because they are robust. Furthermore, these basic fundamental methods
represent the foundation of many new descriptive approaches. This second edition
focuses on the fundamental descriptive methods incorporating any updates and
new developments along with case studies. Our goal with this edition is to provide
the reader with each dedicated practitioner’s view of the descriptive methodology
covered.
This manual does not represent an exhaustive discussion of all of the published
descriptive analysis methods, as some other books do. However, members of ASTM
International Committee E-18 on Sensory Evaluation agreed that the newer methods
are all based on the core descriptive analysis fundamental methods, which are covered
in this manual.
To illustrate the uniqueness of the individual methods, each chapter contains the
background and core philosophy, the assessors’ qualification criteria, the panel mod-
erators’ qualifications, the language development process, along with data collection,
analysis, and reporting of results. The fundamental methods differ from each other in

BK-AST-MNL_13-200365-FM.indd 8 18/11/20 4:13 PM


Introduction and Overview ix

overall philosophy, the panel selection, training and language development process
and the scales used to collect data (e.g., scale length, type, anchors, and usage). Notably,
these methods also vary in the way data are collected and used; some use consensus
while others are data and statistics driven. Finally, methods differ in techniques and
statistical approaches to evaluate panel performance, along with data analysis, and
reporting. It is up to the practioner to determine the level of evidence or statistical
robustness required (e.g., exploratory versus legal dispute) when using descriptive
analysis. Each chapter also contains practical considerations for adaptation to other
categories.
In addition to the four methods previously published in the first edition, this
manual includes Free Choice Profiling and Temporal Methods. These were included
(1) to exemplify the nature and philosophy of a “nontraditional” method (i.e., Free
Choice Profiling) and (2) to illustrate the methodology that captures product percep-
tions over time (temporal methods), which has gained interest and value in the evalu-
ation of products.
All of these methods have tremendous applications in ways yet to be discovered.
Also, yet to witness are the new disciplines that will embrace and use these valuable
sensory tools!

Rebecca N. Bleibaum
Dragonfly SCI, Inc.
Santa Rosa, CA, USA

University of California, Davis Continuing and Professional Education

Alejandra M. Muñoz
IRIS: International Resources for Insights and Solutions, LLC
Mountainside, NJ, USA

For additional reading on this topic, the reader is directed to the following selected
references:

Standard Guide for Two Sensory Descriptive Analysis Approaches for Skin Creams and Lotions,
ASTM E1490 (2019) (West Conshohocken, PA: ASTM International, approved November 1,
2019), http://doi.org/10.1520/E1490-19
L. Stapleton, “Descriptive Analysis,” in Sensory Testing Methods, 3rd ed., ed. M. B. Wolf (West
Conshohocken, PA: ASTM International, 2020), 79–98.
Standard Guide for Time-Intensity Evaluation of Sensory Attributes, ASTM E1909-13 (West
Conshohocken, PA: ASTM International, approved October 1, 2013), http://doi.org/10.1520/
E1909-13
M. A. Brandt, E. Z. Skinner, and J. A. Coleman, “Texture Profile Method,” Journal of Food Science
28, no. 4 (1963): 404–409.
J. F. Caul, “The Profile Method of Flavor Analysis,” Advances in Food Research 7 (1957): 1–40.

BK-AST-MNL_13-200365-FM.indd 9 18/11/20 4:13 PM


x Descriptive Analysis Testing for Sensory Evaluation: 2nd Edition

M. C. Meilgaard, G. V. Civille, and B. T. Carr, Sensory Evaluation Techniques, 5th ed. (Boca Raton,
FL: CRC Press, 2015).
J. Delarue, B. Lawlor, and M. Rogeaux, Rapid Sensor Profiling Techniques and Related Methods:
Applications in New Product Development and Consumer Research (Cambridge, UK:
Woodhead Publishing, 2014).
M. C. J. Gacula, Descriptive Sensory Analysis in Practice (Washington, DC: Food and Nutrition
Press, 1997).
J. Hort, S. E. Kemp, and T. Hollowood, Time-Dependent Measures of Perception in Sensory
Evaluation (Oxford, UK: Wiley-Blackwell, 2017).
S. E. Kemp, J. Hort, and T. Hollowood, (eds.), Descriptive Analysis in Sensory Evaluation
(West Sussex, UK: John Wiley & Sons, Ltd., 2018).
H. T. Lawless and H. Heymann, Sensory Evaluation of Food: Principles and Practices, 2nd ed.
(New York: Springer, 2010).
M. Lindstrom, Small Data: The Tiny Clues That Uncover Huge Trends (New York, NY: St. Martin’s
Press, 2016).
H. Stone, J. L. Sidel, S. Oliver, A. Woolsey, and R. C. Singleton, “Sensory Evaluation by Quantitative
Descriptive Analysis,” Food Technology 28, no. 11 (1974): 24–34.
H. Stone, R. N. Bleibaum, and H. A. Thomas, Sensory Evaluation Practices, 5th ed. (London:
Elsevier/Academic Press, in press).
A. A. Williams and G. B. Arnold, “A New Approach to Sensory Analysis of Foods and Beverages,
in Progress in Flavour Research, Proceedings of the 4th Weurman Flavour Research
Symposium, ed. J. Adda (Amsterdam, The Netherlands: Elsevier, 1984), 35–50.
A. A. Williams and S. P. Langron, “The Use of Free-Choice Profiling for the Evaluation
of Commercial Ports,” Journal of the Science of Food and Agriculture 35, no. 5 (1984):
558–568.

BK-AST-MNL_13-200365-FM.indd 10 18/11/20 4:13 PM


Chapter 1 | Consensus Profile Methods
Derived from the Flavor Profile Method
Edgar Chambers IV1 and Delores H. Chambers1

Introduction
Consensus Profile Methods evolved from the original descriptive method, the Flavor
Profile,1,2 and describe methods that use a group decision-making descriptive process.
In a previous edition of this manual, the Flavor Profile Method was described, but its
adaptation to and more common use as a more generic Consensus Profile Method
resulted in a name change and a change of focus of this chapter.
The Consensus Method is based on the concept that a product’s profile consists of
identifiable sensory modalities (e.g., appearance, taste, odor) and a combination of
impressions not separately identifiable. The method consists of formal procedures for
describing and assessing the characteristics in a reproducible manner. The separate
characteristics contributing to the overall sensory impression of the product are iden-
tified and their intensity assessed to build a description of the product.
This “method” is not a single proscriptive tool for conducting a study but rather
describes a set of processes in which sensory descriptive analyses are performed by a
group of assessors to reach an agreed-on unified profile by consensus. The first adapta-
tion of the original Flavor Profile was to the consensus Texture Profile Method,3 and
consensus profiling has since been adapted for use with various scale lengths (7, 10,
and 15 points)4–6 and to other modalities such as vision7,8 and for such aspects as mea-
suring/comparing attributes among differing stages such as application, wearability,
and the removal of cosmetics.9
Consensus Profiling may be partial, meaning that it includes only a portion of the
process (e.g., determining attributes), or full, where the panel develops the full profile
on the basis of consensus methods, such as occurred with the Flavor Profile.
Partial consensus profiling is quite common in the initial development of lexicons
and sensory scorecards for later tests. In many cases, panels are asked to develop a set
of attributes for all of the modalities of interest that will be used later in testing either

1
Center for Sensory Analysis and Consumer Behavior, Kansas State University, 1310 Research Park Dr.,
Manhattan, KS 66502, USA E. C. https://orcid.org/0000-0002-2480-0200
DOI: 10.1520/MNL1320170037
Copyright © 2020 by ASTM International, 100 Barr Harbor Dr., PO Box C700, West Conshohocken, PA 19428-2959
ASTM International is not responsible, as a body, for the statements and opinions expressed in this chapter.
ASTM International does not endorse any products represented in this chapter.

BK-AST-MNL_13-200365-Chp01.indd 1 30/10/20 4:30 PM


2 Descriptive Analysis Testing for Sensory Evaluation: 2nd Edition

on a routine basis or for a set of products in further testing. Those attributes often are
developed on the basis of consensus methods, where assessors discuss potential terms,
come to some agreement on what terms will be included, and may even determine
definitions and references for those terms depending on the sensory methods to be
used. That list of terms then is made into a ballot and is used in other sensory methods
for individual measurement by assessors.

Core Philosophy
The essence of Consensus Profiling as embodied initially in the Flavor Profile is that it
allows a group of assessors to describe, in detail, the perceptual characteristics of prod-
ucts in a way that will be meaningful to the end user. The concept is that there is a way
to describe those characteristics that makes sense, that the “way” is not necessarily
determined in advance (i.e., there is not a set of standard attributes, a set order of
appearance, etc.), and that a group of trained assessors will discover and profile that
product through testing, discussion, and description. The format of the report may
take different forms such as tables, graphics, or descriptive paragraphs (e.g., storytell-
ing). Civille et al. state that “[c]onsensus evaluations allow highly trained, experi-
enced panelists to determine, discuss and conclude the best way to describe the
individual attributes of each sample.”10
Any number of different types of things can be “profiled.” Throughout this chap-
ter the term product is used to mean any product, service, information, or other stim-
ulus that can be evaluated using sensory descriptive methods. The product may not be
an actual physical product in the sense that many sensory analysts recognize a food,
lotion, cosmetic, automobile paint finish, etc., as a product. For example, profiles can
be developed for the sensory aspects of nutrition education campaigns (e.g., the colors,
fonts, graphics, crowdedness, layout) to compare with standards established by con-
sumers,11,12 instructions for recipes (e.g., fonts, types of instructions, graphics, style of
instructions),13 a library website (e.g., font, layout, readability, clickability, usability
aspects),14 or a concept for an app to take orders at a quick-service restaurant.
Consensus profiles may include a number of different aspects such as attributes;
intensities; order of appearance of the attributes; overall or integrated components
such as amplitude, balance, and blendedness; total impact; and other aspects that the
panelists or researchers believe are important.

Overview of Method
The process of Consensus Profiling as initiated by the Flavor Profile Method is essen-
tially the same regardless of whether a partial or full profile is completed using the
method. In a full consensus profile, the process simply is repeated or integrated
together to determine other aspects of the profile in addition to the attributes or what-
ever portion was to be developed in the partial profile.

BK-AST-MNL_13-200365-Chp01.indd 2 30/10/20 4:30 PM


CONSENSUS PROFILE METHODS DERIVED FROM THE FLAVOR PROFILE METHOD 3

The process usually consists of seven or eight steps depending on whether initially
recruiting the panel is included in the process. In this chapter, we consider recruiting the
panel a preliminary step and thus refer to seven steps of training, testing, and developing
profiles. Likewise, what happens to the profile once it is developed is considered outside the
process of gathering the data, although it is integral to the use and benefit of the method.
The seven steps are as follows:
1. Panel screening and selection
2. Panel training
3. Performance trials to validate the panel
4. Orientation to the defined product category
5. Product presentation—assessors begin individual evaluation
6. Assessors discuss and reevaluate to reach a consensus on the profile
7. Consensus profile created
Those steps (starting usually with Step 4) are repeated as needed to evaluate addi-
tional samples. A diagram of these steps is shown in figure 1. Aspects of these steps will
be described in more detail later in various parts of this chapter.

FIG. 1 Steps in Consensus Profiling Methods (including the Flavor Profile Method).

BK-AST-MNL_13-200365-Chp01.indd 3 23/11/20 10:34 AM


4 Descriptive Analysis Testing for Sensory Evaluation: 2nd Edition

Setting Appropriate Objectives


The Consensus Profile Method can be used for any number of objectives associated
with sensory testing, including short- and long-term objectives. However, it must be
stated that it usually takes considerable training to use the method most effectively
and, thus, is not suitable for situations in which a panel needs to be set up quickly and
a one-time evaluation of a product needs to be conducted. Other methods or outside
consulting panels are better suited for such tests. Note that it may be possible to repur-
pose an existing trained panel to a new category with minimal training depending on
the type of training the panel has received and the characteristics and objectives asso-
ciated with tests in the new category.
Because the Full Profile Method uses definition and reference standards, it can be
used for long-term studies and for reference back to studies that occurred years before.
It is ideal for developing gold-standard profiles of products that can be compared over
time. Because of the adaptability of the method, it is especially effective when some
attributes are unknown prior to beginning a study, such as with wide-ranging prod-
ucts or long-term studies for which additional attributes may develop over time. Con-
sensus methods also work well when the particular order of appearance or attributes
needs to be addressed or when attributes are intertwined and meshed together so that
panelists may need to discuss the product during evaluation to best understand its
profile. Consensus Profiles also may be used efficiently to compare differences in char-
acteristics when products are used in different ways or when different evaluation tech-
niques might affect the findings. The method also is well designed for screening
through large numbers of samples to narrow the list of samples to be tested in other
studies because the assessors can quickly assess whether samples meet the criteria for
inclusion in further testing. That is, do additional samples represent attributes that
already are included and, thus, not needed in the further test, are outside the range of
what is wanted in the further test, have not been noted yet and must be included, etc.?
One objective that typically is not suitable for the method is the determination of
statistically significant differences among specific samples using typical ANOVA tech-
niques. Although the data are appropriate for use in multivariate statistics15–17 and for
use in factorial experiments in which overall effects can be measured using ANOVA,18
because the data are consensus and invariate (they do not have variation [i.e., a stan-
dard deviation]) for an individual product, they cannot be used in a typical ANOVA
design. That does not mean that the data cannot be used to determine meaningful
differences. Many companies or researchers have determined that a difference of x
points on a scale (e.g., 0.5 points on a 15-point scale) is a statistically meaningful differ-
ence and therefore an important difference.

Assessors
Assessors may come from either a technical or nontechnical background; both will
need training, although that training may be of slightly different types. In the original

BK-AST-MNL_13-200365-Chp01.indd 4 30/10/20 4:30 PM


CONSENSUS PROFILE METHODS DERIVED FROM THE FLAVOR PROFILE METHOD 5

Flavor Profile, assessors with a technical background often were chosen because they
were deemed to be on site and available and have knowledge of flavors and ingredients
that could be helpful to the developer. However, later variations of the method often
chose to use assessors specifically hired to conduct the testing because scientists
already in the company were constrained by time and unavailable for training and
testing and there was the potential for biased knowledge about the products they were
testing. Regardless of background, assessors need the following major characteristics
to succeed as panelists: eagerness to participate, time to participate, physiological abil-
ity and innate aptitude, a willingness to learn the techniques necessary for the tasks,
and a disposition for working positively with others using unconditional positive
regard.
Assessors may be recruited from within an organization if the position of assessor
is clearly identified as part of the individual’s responsibilities and priorities. More fre-
quently, assessors are recruited from outside the organization and work as assessors
(often part time) as their primary job within the company. The reason for this is the
amount of time that is required for panel training, orientation, and testing, which for
many Consensus Profile panels is extensive. For example, Caul2 described the training
for the Flavor Profile in terms of 6 to 12 months, and other authors have described
training in terms of hours19 or panel sessions and time.20
It is essential that the individuals selected actually want to participate and are
selected on the basis of the criteria above. Selecting assessors simply because of their
job title, other responsibilities that suggest they “should” be on the panel, time avail-
ability, or their status within the organization is not acceptable. Assessors who are told
to participate, have many other responsibilities, do not have the aptitude for the testing
because they cannot differentiate among samples during screening sessions, or are
selected because they, as bosses, “have all the answers” (they do not when it comes to
sensory skills) make poor panelists.

Assessor Screening and Qualifications


First a determination of what the panel is intended to do needs to be assessed. For
Flavor Profiling, a series of taste and smell tests were used to screen panelists. If the
panel is to conduct testing over multiple modalities on multiple products, the variety
of qualification tests may need to be widened. Essentially, the screening tests need to
cover the range of products and sensory modalities being tested. For example, if the
panel is to test paint samples and the purpose of the panel is to test visual characteris-
tics, then visual acuity tests are the needed screening tools. If, however, the panel is to
test visual, odor, and tactile quality of the paint, a larger range of tests, including odor
and tactile texture screening, must be included.
Panelists are selected from among those assessors who have expressed an interest.
Those candidates are given a series of tests and are interviewed to provide assessments
of suitability for the panel.

BK-AST-MNL_13-200365-Chp01.indd 5 30/10/20 4:30 PM


6 Descriptive Analysis Testing for Sensory Evaluation: 2nd Edition

A typical battery of tests includes the following:


Basic perceptual test: Prospective panelists are not tested for threshold acuity but rather
for their ability to differentiate among fundamental perceptions of the modalities
they will be evaluating. For example, assessors for a range of tests on food should be
able to differentiate among tastes (sweet, sour, salty, bitter, and umami), colors (e.g.,
green, blue, yellow, blue-green), and tactile properties (e.g., gritty, fuzzy, smooth,
grainy) at low levels that are above threshold to determine that they can identify or
match the perceptions to the correct names. Modalities of interest in future testing
should be used. Blanks (if available for that modality) and duplicates may be pre-
sented to test consistency and discourage guessing. The number of correct answers
is counted. It is not necessary to get 100% correct, but it is expected that a high
percentage will be correct.
Ranking test: The candidates are asked to rank a series of intensities of a single percep-
tion in the modalities they will be evaluating (e.g., four solutions of sweet taste, five
levels of grittiness, seven levels of light to dark, four levels of orange odor). These
often are done in complex sensory systems to ensure that the perception that is being
evaluated is done against a background of other sensory perceptions. For example,
the light/dark continuum might be among different colors or pictures of similar, but
different, scenes. These tests simulate actual panel performance, in which panelists
must isolate and quantify elements from a complex product. The score is based on
the ranking of correct order.
Arrangement test: An important part of many Consensus Profile Methods is the concept
of amplitude, identity, wholeness, or other concepts that indicate the overarching
nature of the product. Thus, one test of the ability to understand and describe that
concept is to serve a group of five products in the same category but that differ in
relevant modalities or formulations. The potential assessors are asked to arrange the
products in order from the most complex to the least complex and describe their
rationale for doing so. The arrangement test seeks to measure a candidate’s ability
to perform this integrative measurement. For example, for the Flavor Profile, five
versions of an orange breakfast drink could be presented. Some are diluted and may
have sucrose or a flavor modifier added. The candidates are asked to rank the bev-
erages in some meaningful flavor (not preference or liking) order and describe the
basis for such ordering. The score in this test is based not on what the preconceived
ranking should be but rather the ability of the potential assessor to describe a rational
reason for ordering the samples in the way they did. Answers that better describe
a thought process based on the integration of attributes instead of simply saying a
product has more attributes receive higher scores. For consensus methods that incor-
porate other modalities, a test with relevant differences would be chosen.
Perceptual recognition series: For the Flavor Profile an odor recognition series is given
to determine a candidate’s aptitude for identifying and describing perceptions. This
can be used in other consensus methods as well. However, if the purpose of the panel

BK-AST-MNL_13-200365-Chp01.indd 6 30/10/20 4:30 PM


CONSENSUS PROFILE METHODS DERIVED FROM THE FLAVOR PROFILE METHOD 7

is not to describe flavors in particular, it need not be odors—it could be sounds,


tactile cues, etc. The point of the test is not to determine whether the person can
smell, hear, etc., but rather to determine their ability to verbalize and describe. In the
odor series, for example, 20 different odorants, most of which have been encountered
by the potential assessors, such as vanilla, benzaldehyde, anise, amyl acetate, methyl
salicylate, acetic acid, are presented. Odorants should be perceptible but not over-
whelming. The person is asked to identify, describe, or associate the odor. Scoring
typically is based on whether a person correctly names or describes the odor (e.g., 2
points) or a lessor score (e.g., 1 point) if the person associates the odor. For example,
suppose the odorant clove oil is used. If the person writes “cloves” or “brown spice”
or something similar they would get 2 points. If they say “baked ham” or “dentist’s
office” (both of which are reasonable associates with clove oil) they would get 1 point.
Higher scores indicate a better ability to name, describe, or appropriately associate
the perception.
Personal interview: After the candidates have taken the screening tests, they are inter-
viewed about their work, academic, or personal experiences. The interview is partic-
ularly focused on the person’s (a) interest in product evaluation, (b) experience and
ability to function cooperatively in a group, (c) ability to effectively communicate
opinions without being overbearing, (d) confidence (but not overconfidence) in their
ability to describe and report what one perceives, (e) experiences that may contribute
to their ability to function as a panelist, and (f) availability for panel work. In some
situations a group practice exercise, similar to a focus group, might be held to deter-
mine how individuals interact in a discussion setting.
  Remember, if the person is to be hired directly, all interviews must be conducted
with policies and procedures that apply to hiring in mind, making sure to provide
information on job expectations, attendance, and other aspects. The interviewer
should also discuss factors such as allergies to or moral and religious constraints
about any products that might be tested and address any health concerns that the
candidate may have.
Summary: Test scores and information from the interviews are the usual criteria for
panelist selection. The interviewer must rely on judgment, common sense, and expe-
rience to identify candidate potential. In addition to poor sensory acuity, lack of
interest or availability for panel work, poor interpersonal skills, or health/moral/reli-
gious concerns that would be impossible to reasonably accommodate during testing
can serve to disqualify individuals from further consideration.

Panel Training
After selection, assessors for full Consensus Profile Method panels are trained to
improve their abilities to (a) describe sensory perceptions, (b) score the perceptions
reliably on the designated scale, and (c) use consensus methods to profile actual prod-
ucts. Each of these skills is a key aspect of the training. Training increases reliability

BK-AST-MNL_13-200365-Chp01.indd 7 30/10/20 4:30 PM


8 Descriptive Analysis Testing for Sensory Evaluation: 2nd Edition

and the ability to differentiate small differences in key aspects.19 The duration of the
training will vary depending on the purpose of the panel. If the panel is expected to be
capable of describing a variety of products, a training period of 4 to 6 months is
required. This includes approximately 120 hours of training and 40–60 hours of prac-
tice for the panel. Training for a single type or category of product can be accomplished
in a shorter time.
The structured training consists of the following:
1. A basic course of instruction that includes lectures and demonstrations on the nature
of the senses, basic requirements for panel work, and techniques and procedures for
reproducible sensory testing. This usually requires 3 days.
2. Evaluation of products with a range of relatively easy perceptible differences that
have been selected for their particular teaching value. In this step three to four sets of
products usually containing about four products each are used. For each set, approx-
imately 2–4 hours (usually broken into two to three days or sessions) are spent on the
evaluation of the set of products, with a discussion of the development of attributes
and their references, discussion of scaling, and development of rapport among the
panelists. During this step in training three aspects are of particular interest:
• Describing perceptions in detail: It is essential that assessors grasp the
importance of using specific terms instead of general terms as descriptors. For
example, general terms such as “apple” (flavor), “red” (color), and “soft”
(texture) are shown to the assessors as being too broad. Instead, terms must be
described in much more specific ways. For example, is the apple flavor cooked
or raw? Is the red color cherry or brick, or can we use a specific color number
from a visual color system? Is a soft texture the result of fuzziness or the lack of
stiffness? Getting the assessors to effectively describe products is usually quite
simple once they conceptualize that they must describe the specific perceptual
properties (i.e., what is noticed) and not the ingredients, processes, mechanics,
or other preparation, chemical, or manufacturing processes.
• Comparing intensities: Assessors must begin to understand the types of scales
that will be used. For example, is it a universal-type scale or product-specific
scale? How many points are on the scale? Are there anchors present? How
many anchors are there? Assessors must become comfortable discerning and
stating that one product is higher or lower in intensity than another and begin
to easily compare intensities of individual attributes.
• The concept of unconditional positive regard:21 In consensus panel work this
concept means accepting that other assessors may provide conflicting
information during evaluations but that through meaningful discussion an
agreed-on working profile can be developed to provide the appropriate
description for the product. This does not mean that assessors must
immediately agree with everyone else. It simply means that no matter how
different the opinions are, each person trusts that the intent of everyone’s input
is to progress,22 (i.e., get to a better profile than anyone could have created

BK-AST-MNL_13-200365-Chp01.indd 8 30/10/20 4:30 PM


CONSENSUS PROFILE METHODS DERIVED FROM THE FLAVOR PROFILE METHOD 9

alone). Assessors have sometimes described heated arguments in the panel


room that end at the door, where everyone becomes best friends again. This is
an assessment we hope is a bit exaggerated but does illustrate the concept.
Chermack and Nimon23 determined that principles from unconditional
positive regard and similar communication theories resulted in a “humanistic,
team-driven, collaborative, learning approach” that positively benefited
decision making.
3. Evaluation of products of increasing difficulty that have been selected for their par-
ticular teaching values. Approximately 5–10 hours are spent on the evaluation of
these products, with discussion focused on the specific development of definitions
and references, further detailing of scaling, and continual development of rapport
and calibration among the panelists. During this time panelists begin to alternate
and practice as panel leaders. Typically, six to eight different product categories or
“projects” are conducted during this time to provide the assessors with time to learn
various strategies to learn the sensory methods, provide essential practice, and learn
to come to consensus. Three key concepts that need to be addressed specifically
during this step are as follows:
• Profiling approach: The panelists need to learn various techniques for
describing the profiles. These techniques include the use of a tabular format;
descriptive paragraphs that describe, in detail, the sensory aspects as they
happen during consumption, a form of storytelling; and perhaps graphical
formats such as various charting styles, including multivariate charts, to have
input into the best ways to explain the data.
• Consensus: Consensus is a simple but deceiving term that tends to be
misunderstood. Coming to agreement is not a simple “let’s meet in the middle”
sort of averaging. That would defeat the concept of the method, which is to
discuss and agree on each aspect of the sensory profile components. When
assessors do not agree, the discussion may require repeated sampling of the
product and discussion of the product attribute in terms of intensity, where the
attribute appears, whether it slowly builds or quickly disappears, and nuances
in product variation, testing techniques, and individual differences. This is
done to determine the final attribute, its intensity, order of appearance,
etc., and will be based on agreement and consensus and not a simple mean or
average.
• Panelist confidence: Descriptive panel training, as with many educational
processes, often causes anxiety about whether the assessor is performing well.
Assessors often go through peaks and valleys of confidence (one day all the
testing/training seems easy and the next day the testing/training seems
impossible), and those fluctuations may not coincide with other assessors. It is
critical that assessors understand how they are performing and that they be
given feedback providing them with confidence that they are progressing
appropriately in the training, while at the same time addressing any

BK-AST-MNL_13-200365-Chp01.indd 9 30/10/20 4:30 PM


10 Descriptive Analysis Testing for Sensory Evaluation: 2nd Edition

remediation in training that is needed. Building confidence in each assessor’s


ability to do the testing (assuming they are progressing appropriately) is
essential to building a positive collaborative relationship for consensus
building later on.
4. Additional practice using products similar to those the panel will work with after
training. During these sessions, the panel also practices applying and interpreting the
results. At this step, the panel may be integrated to work with established panels who
will help guide the new assessors.

Qualifications of Panel Leaders


In Flavor Profile and most Consensus Profile Methods, panel leaders are chosen from
among those on the panel who have shown an aptitude for high levels of panel work,
including the ability to effectively and efficiently work in the consensus process. These
assessors should be approached to determine their interest in serving as panel leaders
and their willingness to spend the extra time needed to work with the sensory analysts,
technicians, and other researchers outside of panel time. Time outside of the panel
often is needed to provide assistance in working to update definitions and references,
insight into the profiles when needed, or other information that can best be obtained
from the panel leader. Panel leaders for consensus panels usually are not the sensory
analyst, usually do not have expert knowledge of the product, and should be blinded to
some information about the products to ensure unbiased testing. However, the objec-
tive of the test may be given to both the panel leader and the panel, if appropriate and
not biasing, to help determine the best procedures and documentation necessary to
determine an answer.

Data Collection
For the Flavor Profile and several Consensus Profile derivatives, a single product is
evaluated at one time, and the complete profile is determined before moving to the
next product. In some Consensus Profile Methods, especially if there are many prod-
ucts to be profiled, the panel may examine a subset of products, establish tentative
profiles for those products, and then compare those products to each other to ensure
that the data provide the essential information before reevaluating all of the products
to develop final profiles. The process for profiling is as follows:
1. The development of consensus profiles begins with an understanding of the sensory
aspects of the category of product to be evaluated, if available. The sensory ana-
lyst may provide the panel with reports generated previously or reports that have
been published related to the attributes and testing of similar or related products.
If such materials are available, assessors are made aware that they provide informa-
tion that may, when needed, be adapted for the study in question. Note that in some
instances an end user may have some “standard terms” or “standard references” that
should not be changed. In that case the panel needs to be aware of that information

BK-AST-MNL_13-200365-Chp01.indd 10 30/10/20 4:30 PM


CONSENSUS PROFILE METHODS DERIVED FROM THE FLAVOR PROFILE METHOD 11

and will need to use those terms or references “as is” unless discussed with the end
user first. If the terms or references could result in misunderstandings (e.g., a term
that means one thing to a sensory panel but may be used differently by industry
experts), such information should be explained in notes attached to the profiles.
2. For evaluation, the assessors begin by looking, hearing, feeling, tasting, and s­ melling—
whatever is needed for the objectives of the evaluation. In the Flavor Profile, assessors
develop an individual profile, either partial or full, first on a blank paper or computer
screen. The assessor typically writes down a list of each attribute detected. Then the
assessor might score each of those attributes on a scale. Zero would not be used on
the scale because if the score were zero, the characteristic would not be listed and
would not need to be scored. In some derivative Consensus Profile Methods, the
panelists start with a sheet of terms and references and mark or score those they
believe are relevant to the product being evaluated. In fact, in some cases only a
subset of key attributes may be listed on the ballot to be evaluated. Either system is
appropriate. The blank-sheet method is considered the “purest” in that it does not
prompt assessors but may suffer from the exclusion of a particular attribute in the
quest to notice other attributes. It may not be appropriate if specific attributes are the
focus of the project. The list-of-attributes approach reminds panelists to look for each
attribute. However, it might encourage some panelists to find attributes that may not
really be present.
3. Regardless of whether a Flavor Profile or a more generic Consensus Profile Method is
being used, after all panelists have individually profiled the product, the panel leader,
who is a member of the panel, leads a discussion. First the assessors might discuss
what attributes were evaluated, and the leader would compose a list. Discussion con-
tinues, and depending on the objectives and further use of the terminology list (e.g.,
are these attributes part of a general list that will be used to create a ballot to evalu-
ate samples by individual assessors later or are they the specific list of attributes for
this product?), the group may quickly come to agree on a list of terms or may need
to spend more time developing a more focused lexicon. The panel then reevaluates
the product, examines the master list, discusses terms they did not agree on, and
describes the attribute in more detail (e.g., a description, where it appears, attributes
it follows, and other information). This is done to help other assessors determine
whether the attribute is one they omitted accidentally, is one they used a different
terms to describe, is one they did not find, or if they disagree with the other assessor
for some other reason. Usually, after a short discussion, the attributes can be identi-
fied in the product. If they cannot agree, additional evaluation of the product using
other samples, new reference materials, and further description and elaboration of
the attribute in question, including bringing in other examples of the attribute for the
assessors to examine, may be done to bring the group to consensus.
4. At that point, the panel typically addresses intensities of attributes that have not
already been determined, again using discussion as needed. Depending on the

BK-AST-MNL_13-200365-Chp01.indd 11 30/10/20 4:30 PM


12 Descriptive Analysis Testing for Sensory Evaluation: 2nd Edition

objectives of the study, such aspects as order of appearance of the attributes, the time
sequence of an attribute’s increase or decrease in intensity, dominance of attributes,
or other aspects such as how attributes interact with each other (e.g., amplitude in the
Flavor Profile) may be evaluated.
5. The panel writes the profile in the format determined either by the client or in the
format best suited to the use of the data. Multiple formats, such as a tabular format
with additional graphical components and paragraphs describing unique properties
or interactions that may need to be addressed, are fine. The key to the Consensus
Profile Method is understanding by the end user.

Pilot Test and Validation


Near the end of training, a validation test usually is conducted with several samples of
products that might be tested with the panel in actual studies. The test is conducted
using the protocol of an actual test, usually with trial samples from a previously con-
ducted study with a selected trained panel or with samples that have been specially
prepared to provide specific experiences for the panel and that have a known expected
outcome.
Data are collected using guidance from the sensory analyst as would be typical
in a normal panel. If the panel will be integrated with other assessors after training,
the trainer typically serves as the panel leader for this pilot test but is careful to
serve as a moderator rather than an actual panelist providing data in this case. If
the panel is to work independently after qualifying, it may be important to have one
of the trainees serve as a panel leader if the panel will stand on its own after train-
ing. However, the trainer should be in the room to observe. In both situations, the
trainer should watch for anomalies such as flipped intensities versus other panels
and also assess whether individual panelists need additional training to meet the
standards set for engagement and participation as panelists in the discussion phases
of any future testing.
Profiles from the test are compared to expected profiles to determine whether
they are similar enough to qualify the panel for work or if the panel needs additional
training. Similar does not mean exactly the same. Variations in profiles, as with any
test, can occur, and allowances should be made for slight differences in terminology
and references (if the panel were allowed to develop their own) and to a lesser extent in
intensity scores based on potential variation in the products themselves.
If information is desired on differences in individual assessors data, a study
should be conducted using a method such as Spectrum or Quantitative Descriptive
Analysis data (see chapters 3 and 4). It should be noted that assessors were not trained
in those methods, were not trained to mimic each other, and were trained to operate as
a team, not individually. However, many times their individual scores on ballots do
tend to be consistent with each other.

BK-AST-MNL_13-200365-Chp01.indd 12 30/10/20 4:30 PM


CONSENSUS PROFILE METHODS DERIVED FROM THE FLAVOR PROFILE METHOD 13

Panel Performance Measures


In general, panel performance measures relate to the ability to produce consistent
information rather than statistically comparing data from individual to individual.
Can the panel reproducibly determine differences in samples that are similar but not
the same and reliably note what the differences are? Is the panel able to provide the end
user with information that is informative and actionable? Does the data provide insight
beyond what the end user could provide to themselves simply by having a group of
consumers test the product? If the answers to these questions are yes, the panel is per-
forming as intended. If not, the panel is not performing meaningfully and needs to
either be retrained, disbanded, or realigned with the end user’s needs depending on
the problem.

Reporting of Results
As mentioned, Consensus Profiles can take many forms. Traditional Flavor Profiles
used a tabular form24,25 that included the attribute and intensity in order of appear-
ance. In addition to the tabular profile, the product often was described in paragraph
form to explain the profile.24 Other forms include graphical forms similar to any other
descriptive analysis such as spider plots, bar charts, trees, or in the case of data ana-
lyzed by multivariate methods, maps of products and attributes in multidimensional
spaces (e.g., principal component analysis biplots).4,26–29 In other cases, descriptive
paragraphs have been used to describe the product on the basis of results from the
consensus profile, such as the following summary of eating a manufactured potato
crisp provided by a panel:

For oiliness: Before even picking up the product one could tell that it likely would
be oily because of the 3–6 oil blotches on each chip giving them an inconsistent
greasy appearance. As the product was picked up one felt the need to get a napkin
to wipe the hands because the product had a high degree of oily/greasy feel on the
surface. On putting the chips in the mouth, there was an initially moderate oily/
greasy feel on whatever surfaces (e.g., lips, tongue) the chip touched, but interest-
ingly on biting, the chip seemed slightly dry. While chewing, the dryness grew to
a moderate to high intensity; a moderate starchy feel grew. The flavor did impart
some slight oiliness to it, but it was described as more of a “heated oil” note that
was just beginning to turn rancid, which appeared right at the end of chewing at
a slight level.

Practical Considerations
Consensus profiling is a relatively simple concept but may be difficult to execute in
practice without considerable training. As opposed to individual scoring, consensus
profiles use the concept that the panel of assessors must come to agreement on each of

BK-AST-MNL_13-200365-Chp01.indd 13 30/10/20 4:30 PM


14 Descriptive Analysis Testing for Sensory Evaluation: 2nd Edition

the aspects evaluated. The idea is simple enough. Instead of averaging scores over the
group, the group must come to an agreement on what the appropriate attributes,
scores, etc., are. Consensus methods usually allow the assessors to use whatever attri-
butes they believe are appropriate (i.e., they start with a “blank ballot”), but it differs
from methods, such as free-choice profiling, that allow all individuals to use any words
to describe their perceptions. With consensus methods all assessors must discuss and
come to agreement on the terms, whereas in methods such as free-choice profiling
statistics are used to attempt to relate information to determine similarity in
meaning.
Because Consensus Profiling requires a team decision, it is essential that one
member not dominate the group. A “bully” cannot be tolerated in any Consensus Pro-
file panel, either from the standpoint of a satisfactory end profile (it cannot be one
person’s viewpoint) or from the standpoint of pleasant and respectful group dynamics.
The key is open communication and the use of unconditional positive regard. Success-
ful consensus panels use a range of knowledge, experience, and decision-making skills
to facilitate agreement on the profile. This stimulates thinking and brings new ideas
(e.g., attributes that might not be noticed by everyone) to the group. This is common in
other fields such as psychology and medicine and often helps in successful business
decision making. The key is not to get distracted and bogged down in discussions that
are better held during external training or outside of the panel session; doing so frus-
trates assessors and derails an efficient profile process. When a good panel leader sees
that happening, the best course of action is to call a halt to that discussion, move to a
different aspect of the profile, and return to that aspect with refocused vigor later in the
discussion.
Much has been discussed over the years about the Consensus Profile Method and
its predecessor methods such as the Flavor Profile in terms of specific disadvantages. It
is essential that these discussions are addressed here. In essence, they have been related
to the time and cost of panel training, individual variation that is unaccounted for, bias
related to discussion, lack of statistical analyses, and panel size. The first and last of
these issues are related.
It is true that a consensus panel needs training—sometimes extensive training—
and requires time to calibrate intensities and develop cohesion for discussion and
consensus development, depending on the projects it conducts. Of course, much of
this can be said for many descriptive methods that are not based on consumer vocab-
ulary and are expected to provide a high level of detail to the end user. Likewise, the
number of panelists needed in a test is a function of the level of training and reproduc-
ibility, not a perceived “magic number” with no actual basis in science. Studies have
repeatedly shown that smaller numbers of highly trained panelists can perform as well
or better than a larger number of less well-trained panelists.19,30 A large study of deci-
sion making by teams concluded that groups of 5 to 6 members were optimal where
discussion and decision making are paramount.31

BK-AST-MNL_13-200365-Chp01.indd 14 30/10/20 4:30 PM


CONSENSUS PROFILE METHODS DERIVED FROM THE FLAVOR PROFILE METHOD 15

Individual differences, such as specific anosmias or thresholds to certain com-


pounds, can occur (e.g., boar taint, 6-n-propylthiouracil status) and must be accounted
for. However, the use of references as promulgated in the Flavor Profile and used in
Consensus Profile Methods tends to mitigate that challenge by helping assessors mod-
ify scale use32 and identify panelists who may have special “blank spots” and should
not be used for certain tasks. An advantage of consensus methods in the case of indi-
vidual differences is that they allow the assessors to note differences that cannot be
agreed on. If this represents an actual customer variation, it gives the end user knowl-
edge they would not otherwise get if the values had simply been averaged together,
providing means that may not represent the individual differences.
A major criticism of the Flavor Profile Method and thus of any Consensus Profile
Method is that data from consensus methods are biased because they are easily influ-
enced by a dominant member of the panel.33 Hall, who raised the original criticism,
also noted that just because a bias can take place does not mean that it does take place
and suggested that the possibility of it occurring can be minimized with appropriate
training.34 That aspect is critical and points to the essential focus of training related to
unconditional positive regard and ensuring all assessors have an equal voice. Although
the concept of influence potentially biasing answers and the use of unconditional pos-
itive regard is widely seen as a potential problem, this positive group dynamic is artic-
ulated and understood by fifth-grade children35 and used in the medical field for
making decisions on difficult clinical cases.36 This suggests that the concepts are not as
hard to understand and use to mitigate issues, as appropriate, as those who criticize the
method might have one believe.
The lack of statistical treatment has been cited as a major weakness of the Flavor
Profile37 and by default any Consensus Profile Method, but this topic has already been
discussed in the section “Setting Appropriate Objectives.”

Situational Adaptations
Time sequencing has been an integral part of the Flavor Profile from the beginning of
the method, as evidenced by the concept of order of appearance. Later Consensus Pro-
file Methods have adapted this concept to provide order of appearance of attributes in
single modalities, matched order of appearance in multiple modalities, or temporal
profiling as done in separate sensory tests. It has been included in some studies with
Consensus Profiles to characterize the rise and fall of attributes over time during the
profile. This use, which takes additional time, is typically conducted separately from
the initial profile. After the initial profile is established and the basic order of appear-
ance of the attributes is determined, the panelists reevaluate key attributes over time,
obtaining consensus data that are converted into a graph or line chart to show how the
characteristics interact with each other in terms of their appearance, rise, fall, and
disappearance in the profile. Such information potentially allows the time of appear-
ance and disappearance of certain key attributes to be adjusted for products as they are

BK-AST-MNL_13-200365-Chp01.indd 15 30/10/20 4:30 PM


16 Descriptive Analysis Testing for Sensory Evaluation: 2nd Edition

developed or modified. This information is important in some industries, such as the


liquid pharmaceutical and beverage industries, in which the order of appearance of
bitter and sweet tastes is essential to success.
One adaptation currently implemented in some software programs is the elec-
tronic collection and display of individual data used for discussion. Where attributes
and scores previously were read aloud and transcribed on a projector or display board
to promote discussion, that information can be captured with current computer sys-
tems and displayed as needed. This makes the process of discussion both quicker and
less able to impart bias to panelists who might be tempted to change their verbal score
to match those of others on the panel. The ability to capture data in real time, commu-
nicate face to face, and transfer simultaneously over long distances makes it possible
for the consensus process to occur in remote locations or be conducted by a single
panel with members in many disparate geographic locations as long as consistent
product and attribute reference materials are available to each assessor.

References
1. S. E. Cairncross and L. B. Sjöström, “Flavor Profiles—A New Approach to Flavor
Problems,” Food Technology 4 (1950): 308–311.

2. J. F. Caul, “The Profile Method of Flavor Analysis,” in Advances in Food Research, ed. E. M.
Mrak and G. F. Stewart (New York: Academic Press, 1957), 1–40.

3. M. A. Brandt, E. Z. Skinner, and J. A. Coleman, “Texture Profile Method,” Journal of Food


Science 28 (1963): 404–409.

4. P. A. Prell and F. M. Sawyer, “Flavor Profiles of 17 Species of North Atlantic Fish,” Journal
of Food Science 54 (1988): 1036–1042.

5. E. Chambers IV, E. C. Smith, L. M. Seitz, and D. B. Sauer, “Sensory Analysis of Musty


Compounds in Foods,” Developments in Food Science 40 (1998): 173–180.

6. T. R. Jaffe, H. Wang, and E. Chambers IV, “Determination of a Lexicon for the Sensory
Flavor Attributes of Smoked Food Products,” Journal of Sensory Studies 32 (2017):
e12262.

7. C. Maughan, E. Chambers IV, and S. Godwin, “A Procedure for Validating the Use of
Photographs as Surrogates for Samples in Sensory Measurement of Appearance: An
Example with Color of Cooked Turkey Patties,” Journal of Sensory Studies 31 (2016): 507–513.

8. L. Dooley, K. Adhikari, and E. Chambers IV, “A General Lexicon for Sensory Analysis of
Texture and Appearance of Lip Products,” Journal of Sensory Studies 24 (2009):
581–600.

9. C. Sun, K. Koppel, and E. Chambers IV, “An Initial Lexicon of Sensory Properties for Nail
Polish,” International Journal of Cosmetic Science 36 (2014): 262–272.

10. G. Civille, K. Lapsley, G. Huang, S. Yada, and J. Seltsam, “Development of an Almond


Lexicon to Assess the Sensory Properties of Almond Varieties,” Journal of Sensory
Studies 25 (2010): 146–162.

BK-AST-MNL_13-200365-Chp01.indd 16 30/10/20 4:30 PM


CONSENSUS PROFILE METHODS DERIVED FROM THE FLAVOR PROFILE METHOD 17

11. D. H. Chambers, M. M. Higgins, C. Roeger, and A-M. A. Allison, “Nutrition Education


Displays for Young Adults and Older Adults,” Health Education 104, no. 1 (2004): 45–54.

12. D. H. Chambers and A. M. Munoz, “Focus-Group Evaluation of Nutrition Education


Displays by Hispanic Adults Who Live in the USA,” Health Education 109 (2009):
439–450.

13. A. Lezama-Solano and E. Chambers IV, “Development and Validation of a Recipe Method
for Doughs,” Foods 7 (2018), https://doi.org/10.3390/foods7100163

14. Y. Chung, D. H. Chambers, and L. Lo, “Evaluation of Tablets for Undergraduate


Schoolwork,” in Proceedings SPISE 2016: From Product Formulation to Consumer
Experience, ed. D. H. Nguyen, D. Valentin, S. Le, D. A. Lam, S. Chollet, and H. Abdi, H. (Ho
Chi Minh City, Vietnam: Ho Chi Minh City Publishing House, 2016), 74–85.

15. H. Wang, X. Zhang, H. Suo, X. Zhao, and J. Kan, “Aroma and Flavor Characteristics of
Commercial Chinese Traditional Bacon from Different Geographical Regions,” Journal of
Sensory Studies 34 (2019), https://doi.org/10.1111/joss.12475

16. E. Chambers IV, J. Lee, S. Chun, and A. Miller, “Development of a Lexicon for
Commercially Available Cabbage (Baechu) Kimchi,” Journal of Sensory Studies 27 (2012):
511–518.

17. S. Suwonsichon, E. Chambers IV, V. Kongpensook, and C. Oupadissakoon, “Sensory


Lexicon for Mango as Affected by Cultivars and Stages of Ripeness,” Journal of Sensory
Studies 27 (2012): 148–160.

18. K. Sanchez and E. Chambers IV, “How Does Product Preparation Affect Sensory
Properties? An Example with Coffee,” Journal of Sensory Studies 30 (2015): 499–511.

19. D. H. Chambers, A. A. Allison, and E. Chambers IV, “Effects of Training on Performance of


Descriptive Sensory Panelists,” Journal of Sensory Studies 19 (2004): 486–499.

20. H. Kim, J. Lee, and B. Kim, “Development of an Initial Lexicon for and Impact of Forms
(Cube, Liquid, Powder) on Chicken Stock and Comparison to Consumer Acceptance,”
Journal of Sensory Studies 32 (2017), https://doi.org/10.1111/joss.12251

21. C. R. Rogers, Client-Centered Therapy: Its Current Practice, Implications and Theory
(Boston: Houghton Mifflin, 1951).

22. C. Rogers and B. F. Skinner, “Some Issues Concerning the Control of Human Behavior,”
Science 124 (1956): 1057–1065.

23. T. J. Chermack and K. Nimon, “Drivers and Outcomes of Scenario Planning: A Canonical
Correlation Analysis,” European Journal of Training and Development 37 (2013): 811–834.

24. E. Chambers IV and A. Robel, “Sensory Characteristics of Selected Species of Freshwater


Fish in Retail Distribution,” Journal of Food Science 58 (1993): 508–512.

25. C. Y. Chang and E. Chambers IV, “Flavor Characterization of Breads Made from Hard Red
Winter Wheat and Hard White Winter Wheat,” Cereal Chemistry 69 (1992): 556–559.

26. E. Chambers IV, K. Sanchez, U. X. T. Phan, R. Miller, G. V. Civille, and B. Di Donfrancesco,


“Development of a ‘Living’ Lexicon for Descriptive Sensory Analysis of Brewed Coffee,”
Journal Sensory Studies 31 (2016): 465–480.

BK-AST-MNL_13-200365-Chp01.indd 17 30/10/20 4:30 PM


18 Descriptive Analysis Testing for Sensory Evaluation: 2nd Edition

27. C. Oupadissakoon, E. Chambers IV, V. Kongpensook, S. Suwonsichon, R. Yenket, and A.


Retiveau, “Sensory Properties and Consumer Acceptance of Sweet Tamarind Varieties
Grown in Thailand,” Journal of the Science of Food and Agriculture 90 (2010): 1081–1088.

28. T. Pujchakarn, S. Suwonsichon, and T. Suwonsichon, “Development of a Sensory Lexicon


for a Specific Subcategory of Soy Sauce: Seasoning Soy Sauce,” Journal of Sensory
Studies 31 (2016): 443–452.

29. V. Lotong, D. H. Chambers, C. Dus, E. Chambers IV, and G. V. Civille, “Matching Results of
Two Independent Highly Trained Sensory Panels Using Different Descriptive Analysis
Methods,” Journal of Sensory Studies 17 (2007): 429–444.

30. E. Chambers IV, J. A. Bowers, and A. D. Dayton, “Statistical Designs and Panel Training/
Experience for Sensory Analysis,” Journal of Food Science 46 (1981): 1902–1906.

31. J. S. Muller and E. Wittenberg, “Is Your Team Too Big? Too Small? What Is the Right Number?”
The Wharton School, University of Pennsylvania, 2006, http://web.archive.org/save/
https://knowledge.wharton.upenn.edu/article/is-your-team-too-big-too-small-whats-the-
right-number-2/

32. A. M. Muñoz and G. V. Civille, “Universal, Product and Attribute Specific Scaling and the
Development of Common Lexicons in Descriptive Analysis,” Journal of Sensory Studies
(1998): 57–75.

33. M. Meilgaard, G. V. Civille, and B. T. Carr, Sensory Evaluation Techniques, 2nd ed. (Boca
Raton, FL: CRC Press, 1991).

34. R. L. Hall, “Flavor Study Approaches at McCormick & Company Inc.,” in Flavor Research
and Food Acceptance, ed. A. D. Little (New York: Reinhold, 1958), 224–240.

35. K. S. Evans, “Fifth-Grade Students’ Perceptions of How They Experience Literature


Discussion Groups,” Reading Research Quarterly 37 (2002): 46–69.

36. M. K. Murphy, N. A. Black, D. L. Lamping, C. M. McKee, C. F. Sanderson, J. Askham, and


T. Marteau, “Consensus Development Methods and Their Use in Clinical Guideline
Development,” Health Technology Assessment 2, no. 3 (1998): 1–7.

37. J. J. Powers, “Current Practices and Application of Descriptive Methods,” in Sensory


Analysis of Foods, 2nd ed., ed. J. R. Piggott (London: Elsevier Applied Science, 1988),
187–266.

BK-AST-MNL_13-200365-Chp01.indd 18 30/10/20 4:30 PM


Chapter 2 | The Texture Profile Method
Alejandra M. Muñoz1

Introduction
The Texture Profile Method was developed in the 1960s by General Foods researchers
and published by Brandt et al.1 This sensory descriptive method’s philosophy and
approach were based on two important methods/contributions: the Flavor Profile
Method2,3 and the work by Szczesniak4 and Szczesniak et al.5 on the classification of
textural characteristics and the development of rating/intensity scales.
The use and adaptation of these two technically strong methodological approaches
provide the Texture Profile Method its fundamental core characteristics and philoso-
phy, specifically:
• A descriptive method whose objective, similar to that of the Flavor Profile Method,
is to develop well trained sensory panels, highly skilled in the evaluation of both the
texture qualitative and quantitative descriptive components. Trained Texture
Profile panels evaluate the products’ perceived sensory rheological properties in a
technical, rigorous, and comprehensive way.
• A method that incorporates rheological concepts into a sensory descriptive
approach in order to analytically define the perceived texture attributes, establish
controlled and detailed evaluation procedures, and quantify the perceived sensory
texture properties.
The Texture Profile Method was developed to evaluate the texture attributes of
foods. However, its fundamental principles have been the basis for the development of
methods for evaluating sensory rheological/texture properties of nonfoods. This
chapter discusses several of these adaptations.

Background and Core Philosophy


DEVELOPMENT OF THE TEXTURE PROFILE METHOD
The Texture Profile Method was developed by General Foods researchers1,4,5 to address
business needs and to support/complement the pioneering technical work being

1
IRIS: International Resources for Insights and Solutions, LLC, 234 Robin Hood Rd., Mountainside, NJ 07092, USA
https://orcid.org/0000-0001-5506-742X

DOI: 10.1520/MNL1320160028
Copyright © 2020 by ASTM International, 100 Barr Harbor Dr., PO Box C700, West Conshohocken, PA 19428-2959
ASTM International is not responsible, as a body, for the statements and opinions expressed in this chapter.
ASTM International does not endorse any products represented in this chapter.

BK-AST-MNL_13-200365-Chp02.indd 19 30/10/20 4:31 PM


20 Descriptive Analysis Testing for Sensory Evaluation: 2nd Edition

developed in food rheology by these researchers. The Flavor Profile Method was well
established by the 1960s but mainly addressed flavor evaluation. Thus, there was a
need for developing another sensory approach that focused on the sensory texture
evaluations of foods. The core characteristics of the Texture Profile Method, which
were established on the basis of the work mentioned above, are described in the
two sections that follow.

Adaptation of the Principles of the Flavor Profile Method


The philosophy and approach of the Flavor Profile Method constituted the “model to
follow,”1 and several of its practices were incorporated into the Texture Profile Method
for the evaluation of texture characteristics in foods, such as (a) the development and
implementation of well-designed panel selection and training approaches to form the
technical descriptive profile panel; (b) the development and use of analytical descrip-
tive language; (c) the evaluation of the products’ attributes/“character notes” and
intensities; and (d) the adaptation of “the order of appearance” Flavor Profile Method’s
concept. As it relates to texture, the order of appearance is taken into account in the
evaluation of texture properties as they are progressively perceived in the different
phases of a product’s manipulation (e.g., from the initial phase/product contact
through the residual stage).

Use/adaptation of the rheological concepts developed by Szczesniak and


co-workers
Szczesniak4 and Szczesniak et al.5 established the fundamental principles and philoso-
phy of the Texture Profile Method and all current texture and sensory rheological
evaluations, specifically,
• the fundamental classification of texture sensory characteristics, formed by three
(3) groups: mechanical, geometrical, and “other” characteristics (mainly related to
the product’s moisture/fat content). Szczesniak4 defined these texture sensory
categories, explained how they are manifested/perceived and suggested sensory
terms for core attributes.
• the use of standard rating scales in order to provide a defined quantitative method
of evaluation of the texture parameters.
As Szczesniak et al.5 explained, the use of intensity references and scales would
(a) aid in rating the intensity of a texture attribute relative to the known standard/
attribute intensity, (b) eliminate panel drift and aid in calibration, and (c) facilitate
the correlation between sensory and instrumental measurements.
The use of intensity references is a core characteristic and contribution of the Texture
Profile Method. This practice has been used ever since by many practitioners and
has been adapted by other profile derivative descriptive methods.
The contributions of Szczesniak and coworkers in 1963 extend beyond the develop-
ment of the Texture Profile Method and food texture evaluation. The philosophy behind
the classification of texture attributes, the procedures to develop technical/precise

BK-AST-MNL_13-200365-Chp02.indd 20 30/10/20 4:31 PM


THE TEXTURE PROFILE METHOD 21

language and detailed definitions and evaluation procedures, and all other con-
cepts established in 1963 have been the basis for many applications and for the
development of texture language and evaluation procedures for diverse products,
including nonfoods, ever since.6 Examples of the impact of the Texture Profile
Method on other profile derivative descriptive methods, and other methodologies
and developments, are covered later under the section “Adaptations.”

CORE PHILOSOPHY
The adaptation and use of the contributions and principles of the Flavor Profile Method
and the work by Szczesniak and coworkers in 1963 established the core philosophy and
characterizing features of the Texture Profile Method:
• A technical descriptive method that follows a carefully planned panel screening
and training process designed to establish well-trained texture panels
• A rheologically based sensory approach that focuses on developing and using
accurate and rheologically based sensory language and controlled product
presentation and evaluation procedures to ascertain high-quality and thorough
texture measurements
• A comprehensive descriptive method that provides information on the mechanical,
geometrical, and fat/moisture-related products’ properties and the products’
structural changes/transformation when manipulated; thus capturing the
perceived texture changes throughout all manipulation stages (from initial through
residual)
• A descriptive method that establishes the practice and delineates the advantages of
developing and using quantitative/intensity references to score the attributes’
intensities

THE ORIGINAL TEXTURE PROFILE METHOD AND THE EVOLUTION


OF THE METHOD’S CORE PHILOSOPHY
The well-designed and scientifically based principles published in the 1960s and high-
lighted in previous sections continue to be used. Additionally, over the years other
researchers and sensory practitioners have incorporated several modifications and
improvements, as explained by Muñoz et al.,7 who highlighted the following advances
in the Texture Profile Method:
• Development of additional attribute definitions and evaluation procedures
• Update of texture intensity reference scales and the development of new intensity
scales
• Use of various scaling procedures
• Collection of individual scores without consensus
• Statistical analysis of data
• Adaptation of the method to nonfood products

BK-AST-MNL_13-200365-Chp02.indd 21 30/10/20 4:31 PM


22 Descriptive Analysis Testing for Sensory Evaluation: 2nd Edition

There are also other notable publications that have discussed the growth of the Texture
Profile Method.8–10 Thus, it is important that this publication covers the evolution of
the Texture Profile Method and the developments listed above. In applicable sections
of this chapter the advancement and modifications of the original Texture Profile
Method are highlighted and discussed, specifically how the methodology has been
adapted and is currently being used in food and nonfood texture evaluation.

Overview of the Texture Profile Method


The main features and principles of the Texture Profile Method, contributing to this
method’s very analytical nature, are that it (a) is a technical descriptive method based
on rheological principles, (b) is a profile-based descriptive approach, (c) provides a
comprehensive texture evaluation, and (d) uses intensity references for teaching and
calibration purposes. Each of these respective features and principles is covered in
detail in the sections that follow.

TECHNICAL DESCRIPTIVE METHOD BASED ON RHEOLOGICAL


PRINCIPLES
Tables 1–3 demonstrate the analytical nature of the Texture Profile evaluation proce-
dure and language. Tables 1 and 2 show the Texture Profile evaluation procedures,
attributes, and definitions of a snack (chip) and caramels, respectively. Table 3 is a thor-
ough compendium of texture terms and definitions inherent to the Texture Profile
Method summarized by Muñoz et al.7 Similar to a rheologically based analytical/
instrumental procedure, this method’s sensory protocols, as shown in these three
tables, are technical and precise. Specifically, (a) attributes are individually defined and
evaluated in a precise way, and detailed evaluation procedures are developed and used,
and (b) attributes cover the three texture categories proposed by Szczesniak4: mechanical
(e.g., hardness, crispness), geometrical (e.g., graininess), and moisture/fat-related (e.g.,
oiliness, moisture absorption) characteristics.4

PROFILE-BASED DESCRIPTIVE APPROACH


Tables 1 and 2 also exemplify how the Texture Profile Method adapted the previously
mentioned fundamental bases of the Flavor Profile Method, such as (a) the detailed,
rheologically based, and comprehensive nature of both components—the qualitative
element (characteristics or notes) and the quantitative element (perceived intensity or
level)—and (b) the order of appearance concept1 by defining the different stages of the
product’s manipulation (e.g., initial, manipulation through residual). Tables 1 and 2
show the evaluation phases applicable to each product category: initial surface proper-
ties, first bite (or first chew), mastication/chew-down, and residual phases. This evalu-
ation structure establishes the order in which the different attributes are evaluated.

COMPREHENSIVE TEXTURE EVALUATION


The Texture Profile Method’s objective is to provide a comprehensive sensory evalua-
tion of the perceived products’ rheological characteristics. Thus, the texture evaluation

BK-AST-MNL_13-200365-Chp02.indd 22 30/10/20 4:31 PM


THE TEXTURE PROFILE METHOD 23

TABLE 1 The texture profile evaluation procedure of a snack (chip)

I. Surface Properties Phase


Holding the chip between the thumb and index finger, gently manipulate to evaluate:
Surface oiliness: The degree to which the chip feels oily
(not oily to oily)
Place chip in mouth and gently drag the tongue across the chip’s surface to evaluate:
Roughness: The degree of abrasiveness of the surface as perceived by the tongue
(smooth to rough)

II. First Bite Phase


Place half of the chip between the incisors. Using a steady force, bite through the chip to
evaluate:
Hardness: The force required to bite through the sample
(soft to hard)

Crispness: The amount of snap, as measured by force and noise, released from the chip upon the
first bite
(soggy to crispy)

Place another half between the molars. Using a steady force, bite through the chip twice to
evaluate:
Denseness: The compactness of the cross section of the sample while biting completely through
with the molars
(airy to dense/compact)

III. Masticatory Phase


Place one chip in the mouth and chew approximately 10 to 12 times to evaluate:
Uniformity of mass: The degree to which the mass is uniform after chewing
(not uniform to uniform)

Cohesiveness of mass: The degree to which the mass holds together after chewing
(not cohesive/loose mass to cohesive mass)

Graininess of mass: The amount of small particles perceived by the tongue when the mass is gently
compressed between the tongue and palate
(not grainy to grainy)

Moistness of mass: The amount of moistness of the bolus after prescribed number of chews
(dry mass to moist mass)

IV. Residual Phase


The following parameters are measured after the sample has been swallowed or expectorated:
Oily film: The amount of oily coating left on the surfaces of the oral cavity
(not oily to oily)

Tooth packing: The amount of sample left within the crevasses of the teeth after swallowing/
expectorating
(no tooth pack to product packs in teeth)

BK-AST-MNL_13-200365-Chp02.indd 23 30/10/20 4:31 PM


24 Descriptive Analysis Testing for Sensory Evaluation: 2nd Edition

 he texture profile evaluation procedure of caramels7


TABLE 2 T

I. First Chew
Place sample between molar teeth, bite, and evaluate for:
1. Hardness: Force required to bite through sample.
2. Adhesiveness: Degree sample sticks to teeth.
3. Cohesiveness: Degree to which sample deforms rather than ruptures.
4. Smoothness: Degree to which sample is free of grits and/or grains.
II. Chewdown
Place sample between molar teeth, chew, and evaluate for:
1. Chewiness: Number of chews necessary to prepare sample for swallowing.
2. Gumminess: Amount of energy required to disintegrate sample to a state ready for swallowing.
3. Adhesiveness: Degree to which sample sticks to (a or b) during chewing.
a. Roof of mouth (10–15 chews)
b. Teeth
4. Cohesiveness of mass: Degree to which sample holds together.
5. Denseness: Compactness of sample.
6. Moisture absorption: Degree to which sample absorbs saliva.
a. Rate
b. Amount
7. Crystalline: Degree to which sample is granular.
III. Breakdown
Describe changes occurring during breakdown.
IV. Residual
After swallowing the sample evaluate for:
1. Ease: Degree to which prepared sample is readily swallowed.
2. Chalkiness: Degree to which mouth feels dry or chalky after all of the sample has been swallowed.
3. Grittiness: Degree to which mouth contains small particles after all of the sample has been
swallowed.
4. Toothpacking: Degree to which sample remains in teeth.

With permission.

shown in Tables 1 and 2 is thorough because it captures all perceived texture attributes
throughout the complete product manipulation, providing information on the prod-
uct’s textural changes throughout its manipulation.
In the case of foods, the Texture Profile evaluation includes the complete evalua-
tion of texture perceptions, from (a) initial contact with the product (i.e., initial surface
characteristics as perceived manually or when in contact with lips or mouth surfaces),
to (b) how the product behaves and disintegrates/melts down in the mouth, to (c) how

BK-AST-MNL_13-200365-Chp02.indd 24 30/10/20 4:31 PM


THE TEXTURE PROFILE METHOD 25

TABLE 3 E
 xamples of texture terms used in the Texture Profile Method and other
sensory texture evaluations7

Texture Term Definition

Adhesiveness Force required to remove the material that adheres to a specific surface.
Adhesiveness to lips Degree to which the product adheres to the lips following slight
compression.
Adhesiveness to palate Force required to remove the product completely from the palate with the
tongue following complete compression between tongue and palate.
Adhesiveness to teeth Amount of product adhering to the teeth after mastication.
Self-adhesiveness in the Force required to separate individual pieces with the tongue.
mouth outside the mouth Force required to separate individual pieces with the back of a spoon
(contents of a standard cup placed on a plate).
Bounce Resilience, rate at which the sample returns to the original shape after
partial compression.
Chewiness Number of chews (at I chew/s) needed to masticate the sample to a
consistency suitable for swallowing.
Coarseness Degree to which the mass feels coarse during product mastication.
Cohesiveness Degree to which the sample deforms before rupturing when biting with
molars.
Cohesiveness of mass Degree to which the bolus holds together after product mastication.
Denseness Compactness of cross section of the sample after biting completely
through with the molars.
Dryness Degree to which the sample feels dry in the mouth.
Fracturability Force with which the sample crumbles, cracks or shatters.
Fracturability encompasses crumbliness, crispness, crunchiness, and
brittleness.
Graininess Degree to which a sample contains small grainy particles.
Gumminess Energy required to disintegrate a semi-solid food to a state ready for
swallowing.
Hardness Force required to deform the product a given distance, that is, force to
compress between molars, bite through with incisors, compress between
tongue and palate.
Heaviness Weight of product perceived when first placed on tongue.
Moisture absorption Amount of saliva absorbed by product.
Moisture release Amount of wetness/juiciness released from sample.
Mouthcoating Type and degree of coating in the mouth after manipulation (for example,
fat/oil).
Roughness Degree of abrasiveness of product’s surface perceived by the tongue.
Slipperiness Degree to which the product slides over the tongue.
Smoothness Absence of any particles, lumps, bumps, etc. in the product.
Springiness Degree to which the product returns to its original size/shape after partial
compression (without failure) between the tongue and palate or teeth.
(continued)

BK-AST-MNL_13-200365-Chp02.indd 25 30/10/20 4:31 PM


26 Descriptive Analysis Testing for Sensory Evaluation: 2nd Edition

TABLE 3 E
 xamples of texture terms used in the Texture Profile Method and other
sensory texture evaluations7 (continued)

Texture Term Definition

Swallow, ease of Degree to which the chewed mass can be readily swallowed.
Tooth packing Degree to which the product sticks in the teeth.
Uniformity Degree to which the sample is even throughout.
Uniformity of chew Degree to which the chewing characteristics of the product are even
throughout mastication.
Uniformity of bite Evenness of force through bite.
Viscosity Force required to draw a liquid from a spoon over the tongue.
Wetness Amount of moisture perceived on product’s surface.

With permission.

the residual texture characteristics left in the mouth once the product is swallowed or
expectorated are perceived.
In the case of nonfoods, such as in the evaluation of creams/lotions, the texture/
skinfeel profile approach11 also captures the perceived texture/skinfeel attributes
throughout the entire product manipulation: from initial contact with the product
(i.e., pickup/initial characteristics as perceived manually), to how the product is per-
ceived on the skin during application, to the residual (“after feel”) sensory character-
istics left by the product on the skin once it is absorbed or the manipulation cycle
ends.
While many research and development (R&D) projects require this complete
evaluation, often a simpler assessment is sufficient. Cases for which a reduced
and focused texture profile evaluation is warranted is discussed later under
“Adaptations.”

USE OF INTENSITY REFERENCES FOR TEACHING AND


CALIBRATION PURPOSES
As previously mentioned, Szczesniak et al.5 stated the two main objectives of intensity
reference scales: to teach/train the panel the texture attributes’ evaluation procedures
and definitions, as described later under “Language Development/Panel Training”;
and to calibrate the panel and aid in minimizing panel drifts in the quantitative evalu-
ation of texture attributes. The original intensity reference scales5 were successfully
used for many years. The first update since their development was completed through
the work of the Best Foods Texture Profile panel.12 These revised scales are currently
used by texture panels, and several researchers have described the use of these updated
scales in their training programs and work.13–15 In discussing the growth of the Texture
Profile Method and the knowledge that had been built since the method was pub-
lished, Szczesniak8 highlighted that several parameters had not been considered in her

BK-AST-MNL_13-200365-Chp02.indd 26 30/10/20 4:31 PM


THE TEXTURE PROFILE METHOD 27

work on scales in 1963 but that they had been addressed by Muñoz.12 The original
intensity reference scales5 have also been adapted for their use in other countries.16–18

Assessors
TARGET PANEL SIZE
Among the Flavor Profile Method’s techniques adapted by the Texture Profile Method
was the use of consensus for data collection. In a consensus setup, and based on the
Flavor Profile Method’s techniques, a small panel size is used. Thus, in its early years,
the Texture Profile Method used only six to nine panel members.1 However, larger
numbers have been used, and currently groups of 15 to 20 assessors are usually trained
as texture profilists.12,16,19,20

BACKGROUND AND CHARACTERISTICS OF ASSESSORS


Back in the 1960s and subsequent years, assessors were recruited from within the
company/organization and were mainly employees. As the method evolved and as
employees’ schedules turned more demanding, the source of assessors became mainly
community residents. As currently done by most descriptive methods, “consumers”/
residents from the community are recruited following the procedures described below
to form Texture Profile panels.21,22 In technical descriptive methods, and especially in
the Texture Profile Method, assessors participate in an involved training program and
become expert/descriptive assessors and should no longer be considered consumers.
Muñoz et al.7 and Civille and Szczesniak23 provided information regarding the screen-
ing of texture profilists. A complete description of the process follows:

Availability
The training of a Texture Profile panel is an involved and relatively long process.
Therefore, it is important to communicate the time and commitment requirements to
prospective assessors. They must be available for all training and practice sessions, as
well be able to participate in projects, as needed. Therefore, the training and practice
schedules and product evaluation needs must be communicated. When employees
participate in the program, it is important that the schedule requirements also be com-
municated to the assessors’ management.

Interest in the program and acknowledgment of the program’s importance


As in all descriptive programs, candidates should display a keen interest in being
trained as texture profilists. This characteristic is very much related to the candidate’s
personality, and it can be investigated through a questionnaire or in a one-on-one
interview.
When employees are trained, it is also important that both the candidates and
their supervisors understand the importance of the descriptive program in the com-
pany. It is also the sensory practitioner’s responsibility to emphasize the program’s
importance and the impact of the candidates’ participation in its success.

BK-AST-MNL_13-200365-Chp02.indd 27 09/11/20 11:49 AM


28 Descriptive Analysis Testing for Sensory Evaluation: 2nd Edition

Personality
A positive attitude and ability to work with others are key factors to explore when
selecting candidates.
In the training and operation of a Texture Profile panel, a great deal of interaction
occurs among panel members. As described below in the panel training process, asses-
sors work as a group in the completion of many tasks. These tasks include effectively
reaching agreement in the development of the texture protocol (defining attributes and
developing evaluation procedures), in the review and discussion of references and
individual results, etc. Therefore, people with extreme personalities, such as exces-
sively shy or dominating candidates, should be avoided. Extremely shy candidates do
not actively participate in panel discussions. On the other hand, extremely dominating
assessors are controlling and negatively affect the group dynamics and the interaction
with the panel leader.
Assessors should have a positive attitude, be open to all input provided by other
panel members, positively share their opinions, actively participate in group discus-
sions to address differences, and effectively try to reach agreement in establishing tex-
ture evaluation procedures and protocols. Avoiding extreme personalities is more
crucial when the panel results are collected through consensus. An effective and posi-
tive group interaction must occur in order to be able to successfully reach consensus.
The best way to assess the candidates’ personalities and hopefully unveil extreme
personalities is by conducting one-on-one interviews.

Ability to apply abstract concepts


The Texture Profile Method is a technical descriptive method that requires asses-
sors to learn and apply abstract concepts, such as the learning and use of analytical/
rheological-based evaluation procedures and of quantitative reference scales in
rating intensities. Abstract thinking is best assessed in having assessors complete
scoring exercises, which might include visual or tasting or product application (for
nonfoods) stimuli.

General good health and regard to factors affecting texture perception


Candidates must be free of any major medical condition. In particular, they must have
no issues that interfere with texture evaluations. For example, candidates should have
natural dentition and exhibit good oral health. Issues such as excessive teeth fillings
and orthodontia should also be assessed. Likewise, in the evaluation of nonfoods, fac-
tors affecting the perception of the sensory attributes being measured should be
assessed. For example, for skinfeel profile evaluations, skin types, skin condition, and
age may have to be considered when recruiting assessors.11,24

Acuity
When the Texture Profile Method was developed and used in the 1960s and 1970s,
only one texture acuity test was used in the panel screening process: hardness.

BK-AST-MNL_13-200365-Chp02.indd 28 30/10/20 4:31 PM


THE TEXTURE PROFILE METHOD 29

Reference materials from the original standard hardness scale were presented to can-
didates, who were asked to arrange them in an increasing order of hardness.18,23
The screening of texture profilists currently includes more than one texture
attribute.6 The attributes are selected to cover the most important texture attributes
of the product categories in the program. Moreover, in most cases a panel is trained
not only in the Texture Profile Method’s procedures but also in other sensory
dimensions. Even when the Texture Profile Method was developed, panels were
also trained in the Flavor Profile Method to evaluate flavor in addition to
texture.1, 23 In these instances, the acuity screening should include flavor or other
sensory exercises to explore acuity in flavor or in other sensory dimensions. In the
case of nonfoods, acuity exercises to explore the acuity/perception of the sensory
attributes being measured should be conducted.24

Other current techniques used in screening assessors


In previous sections, modifications to the screening process followed by the original
Texture Profile Method were highlighted (e.g., screening a larger pool of assessors,
inclusion of more attributes in the screening process, modifications of techniques
when training a panel for nonfood texture, such as for skinfeel). Current additional
considerations and techniques in the screening of food and nonfood modified/
derivative Texture Profile Methods include regard to legal issues involving questions
and criteria included in the screening process; additional acuity exercises, such as exer-
cises involving scaling and use of references; and participation in a mock panel to
assess candidates’ abilities and behavior in a group session.
Muñoz and Keane 6 present a thorough discussion of the differences between
the assessors’ screening techniques used in the original Texture and Flavor Profile
Methods and current techniques used in derivative/modified Profile Methods.

Qualifications of Panel Leader/Moderator


Given the technical nature of the Texture Profile Method and its panel training approach,
the panel leader should have background and knowledge in texture evaluation, texture
attributes, and evaluation procedures. For example, in the first phase of training, the
panel leader teaches the concepts of the Texture Profile Method and the core texture
attributes through the use of reference scales. The panel leader has this experience if
he/she has conducted Texture Profile training programs, has participated in previously
conducted training programs, or has been coached by an experienced staff member or
sensory professional. If the panel leader lacks this knowledge, support/consultation can be
obtained to provide guidance and assistance to the panel leader in the training process.
The responsibilities of the Texture Profile panel leader have been described by
Muñoz et al.7 In addition, this author considers that the panel leader’s duties depend on
the position/assignments of the Texture Profile panel leader. In some organizations,
the panel leader has only panel responsibilities. In others, the panel leader is also a

BK-AST-MNL_13-200365-Chp02.indd 29 30/10/20 4:31 PM


30 Descriptive Analysis Testing for Sensory Evaluation: 2nd Edition

sensory scientist or project manager with additional duties. In general, the Texture
Profile panel leader’s responsibilities are as follows:
• Schedule assessors for all panel sessions and activities described below
• Screen candidates
• Design and conduct (or be involved in) the training and practice sessions
• Conduct or supervise the execution of panel sessions
• Supervise sample and reference preparation
• Monitor the panel
• Record, compile, and report panel results
In addition, the panel leader might be a project manager with additional responsibili-
ties such as designing projects, interpreting and reporting results, managing resources
and providing guidance to other staff in the administration and performance of the
daily operation of the Texture Profile program.

Language Development/Panel Training


The training of a Texture profile panel requires 40–100 hours of training depending on
the complexity of the products and the evaluation protocol, the desired level of train-
ing, and the number of attributes to be learned.7
Early Texture Profile publications do not discuss issues to consider in designing a
training program. This author considers that an assessment of all products’ and pro-
gram’s elements should be considered to design the best training program.25 As exam-
ples, important elements to consider when designing a Texture Profile training
program include the following:
• Number and type of product categories; longer programs can be expected with a
large number of product categories, complex products, and fatiguing products
(which limit the length of sessions)
• Type of evaluation protocols (i.e., simple or complex)
• Type of program (e.g., universal vs. product-specific, complete vs. simple
evaluation); product-specific programs and evaluations of simple or only a few
attributes require shorter programs
• Type of needed references
• Budget or resource limitations
The training approach described herein focuses on the process to establish a “univer-
sal” Texture Profile panel that incorporates many product categories. This approach
and philosophy were established by this method’s developers, and these concepts are
explained in their published work.1,4,5
While the development of universal Texture Profile panels continues being the
most common process followed by companies and organizations, other training ver-
sions of the comprehensive training described herein have been implemented by many
researchers, such as “product-specific” panels. Some of these other training

BK-AST-MNL_13-200365-Chp02.indd 30 30/10/20 4:31 PM


THE TEXTURE PROFILE METHOD 31

approaches are briefly described in several sections of this chapter (e.g., “Adaptations”).
For the philosophical and practical considerations of universal and product-specific
panels, the reader is directed to the discussion presented by Muñoz and Civille.26
The training of a universal Texture Profile panel initially involves a considerable
amount of learning because the panel learns all the Texture Profile technical concepts,
language, evaluation practices, the scaling approach, and intensity references. This
method’s universal training consists of orientation and practice sessions. During the
orientations, the panel learns the philosophy and elements of the Texture Profile
Method and completes product evaluations. During the practice sessions, the panel
practices all concepts learned in the orientations and completes product evaluations.
Early publications23 report that a training program to fully train the panel in the
Texture Profile Method would involve 2 weeks of daily orientation sessions (each last-
ing 2–3 hours), followed by about 6 months of hourly practice sessions (4–5 times per
week). As currently practiced, the duration of a universal Texture Profile training pro-
gram is determined by the number and type of product categories included and the
assessors’ schedules. In general, current Texture Profile panels might require 2 weeks of
orientation sessions, 3–4 days/each (the second one conducted a few weeks after the
first orientation session and weeks of practice), and 3–6 months of practice sessions
(3–4 times per week). The training of the Texture Profile panel covers the following:7
• Basic concepts of texture/rheology and texture perception
• Principles of the Texture Profile Method
• Use of intensity reference scales to demonstrate specific texture characteristics and
the procedure to quantify their intensities
• Evaluation of practice samples
• Expansion of the basic method’s procedures to specific products
Civille and Szczesniak23 presented an overview of the suggested training ele-
ments, such as basic concepts, the components of both the Flavor and Texture Profile
Methods, the definitions of texture attributes, the use of intensity reference scales,
practice in the use of scales, and products’ texture evaluations. A complete description
of the Texture Profile training approach is given in the next section. This approach is
typical for the training of a Texture Profile panel1,4,5 that is universal in nature, since
the training program includes a large variety of products. This approach is easily
adapted for training programs on nonfood texture that maintain the core technical
philosophy of the Texture Profile Method.

THE LEARNING OF THE BASIC PRINCIPLES OF THE TEXTURE


PROFILE METHOD AND TEXTURE PERCEPTION
In the first orientation, the following is usually covered:
• The basic concepts of sensory methodology and the principles of the Texture Profile
Method

BK-AST-MNL_13-200365-Chp02.indd 31 30/10/20 4:31 PM


32 Descriptive Analysis Testing for Sensory Evaluation: 2nd Edition

• The Profile methodology


The panel is taught the philosophy and methodology of the Profile methods
 A brief introduction to the Flavor Profile Method, as well as the principles that
were adapted by the Texture Profile Method
 The history and key characteristics of the Texture Profile Method
• Rheology and texture perception
The panel is familiarized with general rheology concepts and the perception of
sensory rheological/texture characteristics. The three categories/groups of texture
characteristics, as proposed by Szczesniak,4 and the fundamental classification of
texture sensory properties are introduced.

THE LEARNING OF THE CORE ELEMENTS OF THE TEXTURE


PROFILE METHOD: ATTRIBUTES, DEFINITIONS, AND SCALES
The panel is then introduced to the core elements of the Texture Profile Method. In a
universal program, assessors learn the main texture attributes that apply across many
product categories and the way they are evaluated. As suggested for the training of a
Texture Profile panel,1,12,23 the above is accomplished by presenting the intensity refer-
ence scales for as many texture attributes as appropriate for the training.
The original intensity reference scales5 were successfully used for many years in
Texture Profile training programs.10 Currently updated intensity reference scales are
used.12 Intensity scales have also been adapted or developed in other countries, projects,
or for other objectives.16–18
Civille and Szczesniak23 stressed the value of the reference scales as a learning tool
because the standards in intensity scales exhibit the particular characteristic being
taught as the outstanding property.
For each attribute, the texture procedure is taught to the panel followed by the
presentation of the intensity reference scale (i.e., individual references demonstrating
low to high intensities of the particular attribute).
This process is also followed in training programs for nonfood texture. The spe-
cific evaluation procedures, attribute definitions, and references corresponding to the
specific nonfood product are used.24
The procedure for introducing and teaching a texture attribute in foods is as
follows:
Part 1: Review of the evaluation procedure, the attribute definition, and the
direction of the scale
For example: cohesiveness of mass
• Place sample between molars, chew approximately 12–15 times, and evaluate
for:
Degree to which the sample/mass holds together.
• Rate attribute (cohesiveness of the mass)
Low (not a cohesive mass/loose particles) to high (very cohesive/compact mass).

BK-AST-MNL_13-200365-Chp02.indd 32 30/10/20 4:31 PM


THE TEXTURE PROFILE METHOD 33

Part 2: Review of the attribute intensity reference scale


For attributes being taught/demonstrated, the individual samples/reference scales
are presented to the panel. References are arranged from low to high intensity.
Several authors have described the way intensity reference scales are used in
training a Texture Profile panel.10,12,13,16 For the above example, after reviewing the
definition, the cohesiveness of mass scale’s references are shown, demonstrating
low to high intensities, as follows:
Reference scale: cohesiveness of mass (Adapted from Muñoz.12 Additional
information on products listed below can be found in Muñoz.12)

Intensity Reference Intensity


Licorice 0
Carrot 2.0
Mushroom 4.0
Frankfurter 7.5
Cheese 9.0
Brownie 13.0
Dough 15.0

The panel reviews each of the references using the protocol described in Part 1.
In reviewing each of the reference points, the panel learns the evaluation procedure
and the attribute by experiencing low, medium, and high intensities of that attribute.
The above procedure is followed for all the attributes chosen by the trainer. This initial
review gives the panel not only the core knowledge of the main texture attributes but
also, more importantly, of the required evaluation procedures and the direction of
each of the scales. The intensity scales are reviewed several times during the training.
In addition, these scales are periodically reviewed after the training is completed and
during project work, as needed.

REINFORCEMENT OF CONCEPTS AND PRACTICE SCORING USING


INTENSITY REFERENCE SCALES
This step is completed during the training phase or in practice sessions.
To reinforce the understanding of the texture attributes and evaluation proce-
dures, the panel completes the review of scales and other samples.1,23 As described by
Muñoz et al.,7 samples from each scale are prepared and presented together with one
or two other carefully prepared foods. The panel is instructed to rate the intensities of
each of the “unknowns.” In this exercise, the panel has the opportunity to review the
attribute, the evaluation procedure, and the intensity references and to practice scor-
ing. A panel discussion follows to review the intensity results and address any dis-
agreements. The same process to reinforce concepts and practice scales is followed
when working with nonfood texture panels. The corresponding nonfood products and
references are used.24

BK-AST-MNL_13-200365-Chp02.indd 33 13/11/20 6:00 PM


34 Descriptive Analysis Testing for Sensory Evaluation: 2nd Edition

DEVELOPMENT OF COMPLETE TEXTURE PROFILE EVALUATION


PROTOCOLS
Upon completion of the previous steps, the panel has learned the main concepts of the
Texture Profile Method’s approach. These concepts include the classification of texture
characteristics, the core texture attributes and their definitions and evaluation proce-
dures, and the main intensity reference scales. In addition, the panel has practiced
scoring the intensity of unknown samples relative to the intensity scales. With these
foundations, the panel is capable of initiating/practicing the development of complete
Texture Profile evaluation protocols.
As Brandt et al.1 recommended, several samples of a product category are used for
the panel to develop the evaluation procedures. These authors1 provided several exam-
ples of Texture Profile evaluation protocols for products such as rice, cereals, whipped
toppings, and biscuits. These evaluation protocols and examples have been the basis of
all Texture Profile evaluations and have provided guidance to practitioners involved in
texture training programs. Tables 1 and 2 present examples of complete Texture Profile
evaluation protocols (i.e., evaluation stages, attributes’ definitions, and evaluation
procedures).

PRACTICE OF COMPLETE PRODUCT EVALUATIONS


The Texture Profile product evaluation protocol developed by the panel is discussed,
refined, and used to practice and evaluate products. Two to three products are pro-
vided to the panel for evaluation following the established procedure. Upon comple-
tion, results are discussed by the panel, as well as any issues encountered in the
evaluation.

DEVELOPMENT OF TEXTURE PROTOCOLS FOR ALL CATEGORIES


OF INTEREST AND PRACTICE OF PRODUCT EVALUATIONS
For each additional product category of interest, the panel develops the evaluation
protocol and continues developing expertise in the key steps delineated above to estab-
lish the Texture Profile evaluation protocol:
• Selection of the product category’s texture attributes to be evaluated
• Development of definitions and evaluation procedures for each attribute
• Review of intensity references
• Product evaluation
Muñoz and Keane6 review the techniques followed in training assessors in the origi-
nal Texture and Flavor Profile Methods and contrast these techniques with those
used in derivative/modified Flavor and Texture Profile Methods currently used.
They also present a valuable example of how the original Texture Profile Method
training techniques can be modified to train a nonfood Texture Profile panel (e.g.,
a skinfeel panel).

BK-AST-MNL_13-200365-Chp02.indd 34 30/10/20 4:31 PM


THE TEXTURE PROFILE METHOD 35

Data Collection
CONSENSUS
The section “Background and Core Philosophy” presents the Flavor Profile Method’s
techniques that were adapted by the Texture Profile Method. A few of these techniques
included the use of the original 5-point (or the modified 7- or 14-point) Flavor Profile
scale and collecting data through consensus.
The process followed by a Texture Profile panel to reach consensus has been
described by Brandt et at.1 and Muñoz et al.7 Profilists score samples independently
and a discussion follows. Individual scores for each attribute are tallied and shown to
the group for the discussion. The panel reviews and discusses individual results, shar-
ing opinions and reaching consensus. In these discussions and in the case of disagree-
ments, the attributes, definitions, evaluation procedures, and references are reviewed.
When needed, samples in question are reevaluated in an attempt to reach consensus.
Texture Profile scores are reported as consensus values.7,23
Currently, most Texture Profile panels provide individual scores (see next
section). However, collecting data through consensus in the Texture Profile Method
and in other descriptive methods is still practiced. This approach is controversial.
While some researchers criticize the technique, others highlight its advantages.
Several researchers have assessed the value of consensus and/or have discussed their
advantages/disadvantages.6,27–29

INDIVIDUAL EVALUATIONS AND STATISTICAL ANALYSIS


Currently and in most Texture Profile derivative methods, the original method’s prac-
tices of data collection and analysis have been modified and improved.

Scales
Currently, Texture Profile data are collected using a variety of scales and treated as
individual responses.9,12,15,20,30–32 Assessors evaluate the products individually using
the protocol developed to score all products’ attributes. Different scales (e.g., line, 10,
15, 100-point scales) are currently used, and a few researchers have also used magni-
tude estimation in Texture Profile studies.33

Replications
While original Texture Profile Method practices did not address replication, most
current Texture Profile panels incorporate replications in products’ evaluation of foods
and nonfoods.

Data Analysis
Texture Profile data are currently analyzed using diverse statistical analyses. When
describing how the Texture Profile Method had evolved since its inception, Larmond9

BK-AST-MNL_13-200365-Chp02.indd 35 30/10/20 4:31 PM


36 Descriptive Analysis Testing for Sensory Evaluation: 2nd Edition

and Skinner10 discussed the use of statistical analyses of Texture Profile data. Both
authors also emphasized the use of multivariate statistical analysis approaches to ana-
lyze and present Texture Profile results and showed examples.
Researchers and practitioners who have trained and used Texture Profile panels
have used a variety of statistical techniques to analyze and present their results. These
techniques include the use of analysis of variance (ANOVA), correlation analysis, and
multivariate analysis. For example, Dransfield et al.20,32 described the way meat Texture
Profile results were analyzed through ANOVA, correlation, principal coordinate anal-
ysis, and principal component analysis (PCA). Other authors have reported the use of
similar statistical approaches (ANOVA, correlation analysis, and multivariate analy-
sis) for the analysis of Texture Profile data.15,22,30
In summary, interval Texture Profile data are currently treated as any other inter-
val descriptive data using univariate and the multitude of statistical techniques avail-
able to sensory practitioners. The specific analyses are chosen based on the test design
and study objectives.
Texture Profile data can be intrinsically variable; in foods because of assessors’
differences in mouth size/volume, teeth, saliva, dental status, etc., and in nonfoods
because of assessors’ differences in skin, hair type, etc. However, this method’s results’
variability due to scoring might be lowered through the use of intensity reference
scales. This practice, which aids in yielding lower variability in the data, provides addi-
tional discriminative ability to the statistical techniques used.

Pilot Test and Validation


In the early Texture Profile Method’s publications, no validation studies were recom-
mended. However, Skinner10 stressed the importance of practice exercises while a
Texture Profile panel is being trained and indicated that a database should be devel-
oped in this phase to check individual assessor and panel data repeatability. As the
method evolved, and as done currently, a validation study is usually conducted prior to
having a Texture Profile panel initiate project work. The process followed to validate
Flavor and Texture Profile panels is described by Muñoz and Keane,6 who highlighted
the following important steps to consider:

SAMPLE SELECTION
Samples to be evaluated should be carefully selected in a validation study. They should
be sufficiently but not extremely different. One or several sets of samples can be
included in the validation study. Romero del Castillo et al.34 covered the importance of
the samples’ characteristics in validation studies and described the selection and
preparation of samples to validate a texture panel. When samples with known vari-
ables are selected, it is possible to assess the panel performance and determine whether
the panel is ready for routine projects. A profile panel is expected to show differences
in the attributes known to be different.6

BK-AST-MNL_13-200365-Chp02.indd 36 30/10/20 4:31 PM


THE TEXTURE PROFILE METHOD 37

ANALYSIS AND ASSESSMENT/ANALYSIS


Results are assessed and analyzed based on the validation study’s test design. A few of
the procedures and parameters used to validate a Texture Profile panel are as follows:

Assessing results of new and trained panels


When the evaluations designed for the validation study can be conducted by
another trained and established/experienced Texture Profile panel, the comparison
of results from the recently trained Texture Profile panel with results from an
experienced/trained panel is a useful validation tool. A criterion/guideline for
validating a new panel comparing the results of the two panels is that the results
show the same direction of difference in at least 80% of the attributes.24

Panel performance measures


The validation study can be designed to study different panel performance
measures.25 The most important measures are described in the next section. For
example, Thybo and Martens13 presented the study design and statistical analyses
used to assess the discriminative ability and reproducibility of a Texture Profile panel.
Romero del Castillo et al.34 focused on checking the discriminatory capacity and the
reproducibility of a texture panel and on analyzing the validity of their judgments.

Comparison of panel data with consumer results


The comparison of results from the recently trained Texture Profile panel with
consumer attribute data for those attributes for which consumers have demonstrated
their understanding and the ability to differentiate among products24 is another
way to validate Texture Profile panels. If the Texture Profile panel data correlate
well with the previously validated consumer responses for similar attributes, the
recently trained panel is considered trained and reliable.

ADDITIONAL TRAINING AND CALIBRATION


Following any of the above approaches, results are analyzed/assessed to determine the
readiness of the Texture Profile panel for project evaluations or if more training is
necessary. Additional training or calibration exercises are designed to focus on the
issues unveiled by the validation study, such as review of specific attributes’ procedures
and definitions, improvement of intensity reference scales, improvement of ballots, etc.
Upon completion of the additional training/panel sessions, the data from these ses-
sions are assessed to determine whether the issues encountered in the validation study
have been addressed. If these additional training sessions have been successful, a new
validation study as described above is conducted.

Panel Performance Measures


Civille and Szczesniak23 recommended three factors for monitoring panel perfor-
mance: the reliability of the panel to duplicate its findings from one evaluation to the
other, the ability of an individual panel member to duplicate his or her finding from

BK-AST-MNL_13-200365-Chp02.indd 37 30/10/20 4:31 PM


38 Descriptive Analysis Testing for Sensory Evaluation: 2nd Edition

one evaluation to the other, and the ability of panel members to agree with one another.
Skinner10 mentioned the importance of checking individual and group data
repeatability.
Currently, because different scales are used to obtain attribute ratings, diverse
criteria can be studied to evaluate the Texture Profile panel performance. The most
important panel performance measures used to assess and monitor Texture Profile
individual assessors and panel are repeatability, scale usage, discrimination ability/
sensitivity, agreement, and reproducibility.35
Several of these performance measures have been used by other researchers to
assess texture profile panels. For example, Thybo and Martens13 published a valuable
approach for assessing the performance of a Texture Profile panel, such as its discrim-
ination ability and reproducibility, using several univariate and multivariate statistical
techniques, such as ANOVA and discriminant partial least-squares regression.
Rousset-Akrim et al.21 used a variety of statistical techniques such as correlation,
ANOVA, and stepwise discriminant analysis in their research work of selective tests
for discerning “an efficient assessor” in texture profiling.
The topic and techniques for panel performance have been discussed extensively
in other publications. Thus, the reader is encouraged to consult other publications that
have discussed the importance of these measures,35 provided recommendations, and
covered suggested statistical approaches to assess profile data and sensory profile pan-
els’ performance.36–38

Reporting of Results
Texture Profile results are summarized in reports or presentations.

CONTENTS OF REPORTS
Muñoz et al.7 recommended that a Texture Profile report include the objectives of the
study; identification and preparation of the samples; the ballot (showing the phases of
evaluation, terminology [attribute definitions and evaluation procedures], and inten-
sity scales used); summary of the techniques used; tables, charts, and graphs; and a
discussion of the results focused on differences and similarities among the test
samples.
Civille and Szczesniak23 also recommended that the Texture Profile report pres-
ents the frame of reference, i.e., how the panel was “standardized” prior to the evalua-
tion and the reference standards used to orient the panel. Additional important
characteristics that Muñoz and Keane6 recommend in a Flavor or Texture Profile
report are: executive summary, background, sensory test design, summary of data
analysis, conclusions, and recommendations.

PRESENTATION OF RESULTS
When consensus results are reported, charts or graphs are the most effective way of
displaying these results. Figure 1 shows the distinctive graphic representation of Texture

BK-AST-MNL_13-200365-Chp02.indd 38 30/10/20 4:31 PM


THE TEXTURE PROFILE METHOD 39

 raphic representation of texture profile results for caramels.7 (With


FIG. 1 G
permission).

Profile results, which was commonly used in the early years of the method and is
typical of the Texture Profile Method.7,10
Tables and a variety of graphs are currently used to depict Texture Profile results
depending on the study objectives and test design, and the corresponding statistical
analysis is completed. The section above “Data Collection” presented a few of the nota-
ble publications that have shown the use of assessors’ individual scores, the corre-
sponding statistical analysis, and the presentation of results of Texture Profile data.
As examples, below are a few of the presentation formats that have been used to
report Texture Profile results, which are specific to the studies completed and the data
analysis performed.

BK-AST-MNL_13-200365-Chp02.indd 39 30/10/20 4:31 PM


40 Descriptive Analysis Testing for Sensory Evaluation: 2nd Edition

Table 4 shows an example of tabular Texture Profile panel evaluation results. This
table shows selected information from the complete results presented by Dransfield et
al.20 on texture evaluation of different types of meats and preparation procedures.
Figures 2 and 3 are examples of current graphical representations of texture data.
Figure 2 depicts the Texture Profile panel results of herbal gels as a function of concen-
tration, studied and reported by Cui et al.16 These authors also showed the correlation
and data relationship analysis results and plots of the Texture Profile and instrumental
measures studied in their research.
Figure 3 is an example of a multivariate map of texture results. This format is cur-
rently commonly used to depict the diverse and sometimes complex texture data. The
application of multivariate statistical analyses presently used allow depicting texture
results in multivariate maps showing attribute relationships and other specific analysis
results. For example, figure 3 shows the PCA map (PC2 vs. PC1) of the texture and
appearance evaluation results of dry dog food products.39 Other presentations of the
statistical analysis results or Texture Profile and other texture evaluations include the
discriminant partial least-squares regression maps of texture research results on pota-
toes,13 preference mapping of the texture of dulce de leche,40 the texture and appear-
ance evaluation results of lip products,41 etc.
In general, the presentation format of Texture Profile results depends on the test
design, the data analyses conducted, and the individual researchers’ preferences.

TABLE 4 Texture profile results of different types of meats20

Attributes

Meat Type n IR RUB Fco CT J Res T Ch

Chicken
Roast 3 4.1 0.8 5.1 0.4 6.8 2.1 4.2 1.4
Casserole 3 5.1 1.1 5.4 0.4 6.1 2.5 4.9 1.6
Beef
Microwave 5 7.7 4.0 8.2 4.4 7.0 2.9 8.2 4.8
Grill 3 5.1 1.6 6.3 1.3 8.4 1.4 5.3 2.3
Lamb
Roast 4 5.6 2.0 6.2 1.4 7.6 1.7 5.8 2.8
Processed 1 2.9 0.7 5.4 1.1 6.6 2.5 3.1 0.7
Pork
Roast 11 6.0 1.8 6.2 0.8 6.3 2.5 6.0 2.3
Processed 1 4.3 1.4 4.8 1.0 6.5 1.5 4.4 0.8

Note: Only selected information from original source results are presented.

Values are the means of n samples of diverse types of meat/cooking for: IR = initial resistance; RUB =
rubberiness; Fco = fiber cohesiveness; CT = connective tissue; J = juiciness; Res = residue; T = toughness;
Ch = chewiness.

BK-AST-MNL_13-200365-Chp02.indd 40 09/11/20 11:50 AM


THE TEXTURE PROFILE METHOD 41

FIG. 2 T
 exture profile panel scores of herbal gel samples as a function of the
ingredients’ concentration.16 (With permission).

FIG. 3 P
 rincipal components analysis map (PC2 vs. PC1) of the texture and
appearance evaluation results of dry dog food products.39 (With permission).

BK-AST-MNL_13-200365-Chp02.indd 41 30/10/20 4:31 PM


42 Descriptive Analysis Testing for Sensory Evaluation: 2nd Edition

Adaptations
Several of the modifications and advances of the original Texture Profile Method
have been mentioned throughout this chapter. In this section, a summary of this
method’s adaptations is presented under two main categories: (a) evolution of the
Texture Profile Method and development of other texture evaluation methodology
(covering Texture Profile shortened/focused program, Derivative Profile Methods,
Temporal Texture Evaluation Methods and Descriptive Analysis Technical Texture
Evaluations) and (b) adaptation of the Texture Profile Method for the evaluation of
nonfoods (covering Skinfeel, Handfeel and other products [e.g., lip products, pet
food, etc.]).

EVOLUTION OF THE TEXTURE PROFILE METHOD AND DEVELOPMENT


OF OTHER METHODOLOGY

Adaptation of the Texture Profile Method’s Universal Approach: Evaluation of


Selected Products or Attributes (Shortened/Focused Program)
This chapter has described the characteristics of a comprehensive (universal) Texture
Profile capability, (i.e., a program that focuses on developing a panel capable of evalu-
ating many product categories to provide the most complete product evaluation [all
perceived texture attributes]). This approach is based on the original methodology.1,4,5
The approach was developed for R&D projects and to support research in the field of
rheology.4 Therefore, since its inception and currently, this comprehensive/universal
program is implemented to support R&D projects, which usually require the most
complete product evaluation. As described in this chapter, this process is involved and
time-consuming.
There are, however, many instances when a shortened program is implemented
because an R&D/complete Texture Profile panel is not needed or cannot be established
due to budget and other resource constraints. In these cases, the philosophy and nature
of the complete and shortened programs remain the same, but the implementation,
process, and results of the shortened program are different. The situations/applications
described below are implemented when a less comprehensive program/Texture Profile
panel is needed:
• Program for the evaluation of one or only a few product categories
Instead of covering many product categories, a focused Texture Profile program
includes only one or a few selected product categories. Several researchers have
reported the methodology used to train Texture Profile panels focused on specific
products. Examples of product-specific Texture Profile training approaches have
been reported on potatoes,13 herbal gels,16 meats,20,22 cheeses,30 etc.
• Complete product evaluation (appearance, texture, flavor or fragrance)
In general, when a panel is trained to evaluated all sensory dimensions (not only
texture), fewer texture attributes might be included in the training and thus
evaluated.

BK-AST-MNL_13-200365-Chp02.indd 42 30/10/20 4:31 PM


THE TEXTURE PROFILE METHOD 43

• Focused/attribute panels
For specific objectives (e.g., shelf-life studies, category-specific evaluations), attribute
Texture Profile panels might be trained to focus only on specific attributes of interest.
• QC/QA sensory programs
This application is another type of a focused program, in which only specific
attributes need to be evaluated. In this application, selected texture and possibly
other sensory attributes that have been identified as variable or with production
issues are included in the evaluation.42
• Others

Texture Profile Derivative Methods


Texture Profile-based/derivative descriptive methods very closely follow the principles
of the original Texture Profile Method philosophically and methodologically.6 This
means that they embrace and use the fundamental methodology elements of the
Texture Profile Method, such as rigorous training, development of technical/rheolog-
ically based texture terminology, development and use of detailed evaluation proce-
dures, and use of intensity references. Muñoz and Keane6 thoroughly describe how the
original Profile methodology has been followed or adapted by the different profile-
based/derivative methods.

Texture Attributes Evaluated by Temporal Methods


The Texture Profile Method intrinsically incorporates the temporal element and pro-
vides information on the changes of the product’s texture by the perceived attributes
evaluated in the different stages of manipulation. Additionally in foods, these tempo-
ral changes can be studied by evaluating the same attribute in the different stages of
mastication (e.g., stickiness) or by assessing the way attributes appear or disappear
throughout the chew/mastication (e.g., crispness disappears and cohesiveness of the
mass appears). Likewise, in the evaluation of many nonfoods the same attribute can be
evaluated in different stages of manipulation or in the afterfeel (or after application) at
specific time intervals (e.g., oily, waxy).24
The development of temporal methods and more sophisticated data-entry sys-
tems have allowed a more detailed evaluation of products’ texture temporal behavior.
The reader is referred to chapter 6 in this manual and the research and methodology
developments for the evaluation of texture attributes by temporal methods.43–46

Descriptive analysis technical texture evaluations


Texture attributes in foods and nonfoods can currently be evaluated by a gamut of
descriptive analysis methods. While the methodology might not be called Texture
Profile or a Texture Profile-based Method, this author considers that all technical tex-
ture evaluations are based on the Texture Profile Method because this descriptive
method provided the fundamental approach for sound sensory rheology-based

BK-AST-MNL_13-200365-Chp02.indd 43 30/10/20 4:31 PM


44 Descriptive Analysis Testing for Sensory Evaluation: 2nd Edition

texture evaluations, specifically the way (a) texture attributes should be clearly and
technically named and defined, (b) a detailed texture evaluation procedure or protocol
should be developed, (c) proper texture scale anchors should be established (e.g., soft to
firm, airy to dense, etc.), and (d) quantitative/intensity references can be used.
In summary, texture evaluations are technical in nature and methods or evalua-
tions that apply any of the above features adapt/use the fundamental characteristics of
the Texture Profile Method.

ADAPTATION OF THE TEXTURE PROFILE METHODOLOGY FOR THE


EVALUATION OF NONFOODS
The fundamental classification and attributes’ structure developed by Szczesniak4 and
overall the technically sound principles established by the Texture Profile Method1
served as a guide in the development of texture evaluation techniques, not only for all
foods but also for many other products, including nonfoods. Some of these adaptations
are described in the following sections.

Skinfeel Profile Descriptive Analysis


Schwartz11 is acknowledged for developing the Skinfeel Profile Descriptive Analysis
based on the Texture Profile Method’s procedures. Schwartz provided the terminology
for skinfeel evaluations, which as she indicated, represented the application of the
Texture Profile Method’s techniques to the evaluation of skincare products.
The following evaluation stages for skinfeel assessments were suggested
by Schwartz: pickup, rub-out, and after feel. The skincare product attributes and
definitions she suggested included thickness, consistency, spreadability, absorbency,
and residue (film, coating, flaky or powdery particles, dryness/moistness, oiliness,
irritation, etc.).
The contributions of Schwartz11 continued to be used and further refined. Aust
et al.31 developed and published the methodology for the evaluation of perceived attri-
butes of skincare products that was based on Schwartz’s work. In addition, the above
contributions were further advanced in the development of a standard guide for
descriptive skinfeel analysis of creams and lotions using either a technical/
profile-based or consumer-based approach.24

Handfeel Evaluations
The principles of the Texture Profile were also applied to develop the sensory and
descriptive methodology for the evaluation of fabrics and paper products. Parallel to
the developments of the Texture Profile Method, Brand47 developed and published the
fundamental concepts of what he called “aesthetics of fabrics,” incorporating appear-
ance and handle. These concepts provided the framework for assessing handfeel char-
acteristics of fabrics. Brand outlined the aesthetics “concepts” of fabric (defined as the
factors or elements whose relationship defines aesthetics) as body, cover, surface tex-
ture, resilience, drape, and style.

BK-AST-MNL_13-200365-Chp02.indd 44 30/10/20 4:31 PM


THE TEXTURE PROFILE METHOD 45

A comprehensive terminology was provided for each of these elements in Brand’s


publication. Examples of the texture terms provided for each of the above categories
included body (bulky, thin, light/heavy, soft/firm, etc.); cover (smooth, fuzzy, soft/
hard, dense, full, etc.); surface texture (cottony, greasy, non-smooth [grainy, rough],
etc.); resilience (rubbery, mushy, bounce, stiff, etc.); drape (stiff, crisp, clinging, full,
etc.); and style (pattern, coloration, etc.)
Through his published work, Brand47 established the foundations for the sensory
methodology used to assess handfeel priorities of fabrics and paper products. Based on
Brand’s work/contributions and the foundations/concepts of the Texture Profile
Method, the terminology to describe handfeel properties of paper and fabrics was pub-
lished by Civille and Dus.48 As indicated by Civille and Dus, they used the Texture
Profile Method as a model in the handfeel work they published, specifically using the
texture attribute classification outlined by Szczesniak,4 mechanical (e.g., force to
gather, stiffness), geometrical (e.g., grittiness, graininess), and moisture and thermal
properties.

Texture Evaluations in Other Nonfood Product Categories


Other noteworthy developments of texture-based evaluations in nonfoods include the
following:
• Donfrancesco et al.39 developed appearance, flavor, texture, and aroma descriptive
lexicons for dry dog food products. The suggested texture lexicon included
cohesiveness of mass, fibrous, firmness, fracturability, graininess, gritty, hardness,
initial crispness, mouthcoat, oily mouthfeel, powdery, and springiness.
• Dooley et al.41 developed a texture and appearance lexicon for the sensory analysis
of lip products. Texture profile-based terms recommended for the evaluation of lip
products included smoothness, spreadability, initial drag, product to product drag,
tackiness, wet, degree of absorption, and amount of residue. These attributes were
evaluated in the phases of evaluation chosen for this product category: initial
texture/application and after texture/residual.
• ASTM E2346, Standard Guide for the Sensory Evaluation of Household Hard
Surface-Cleaning Products with Emphasis on Spray Triggers,49 presents technical
texture attributes for this product category, including force to open/close trigger,
thickness, force to dispense, degree of product cling, degree of effort to remove soil,
feel of product while wiping, residual oiliness, residual grittiness, etc.

Summary
The Texture Profile Method represents a milestone in the evolution of sensory science
and descriptive analysis. This method is a notable contribution not only because it was
one of the first descriptive methods established, but also because it truly provided the
foundations for all sensory evaluations of the texture/rheological properties of
products.

BK-AST-MNL_13-200365-Chp02.indd 45 30/10/20 4:31 PM


46 Descriptive Analysis Testing for Sensory Evaluation: 2nd Edition

This method’s contributions are palpable when used as originally developed and
in all its adaptations in descriptive methods. Specifically, in terms of adaptations, the
Texture Profile Method’s techniques have been replicated by other descriptive meth-
ods or the techniques have been used as the basis for texture evaluations in all descrip-
tive methods.
In terms of contributions, the Texture Profile Method provided to the field of
sensory science the following key philosophical concepts and methodological
approaches, which continue being used as originally developed or in a modified form:
• The value of developing and using precise and analytical texture attributes,
definitions, and evaluation procedures.
• Together with the Flavor Profile Method, the use of consensus as a dynamic panel
practice to discuss results or to reach agreement, or both.
• The practice of using intensity/quantitative references to provide the benefits
described by the founders of the Texture Profile Method and by other professionals
who have adopted this approach.
• The pioneering and early methodology of temporal evaluations because the Texture
Profile Method captures temporal changes of the products’ texture throughout
their manipulation, from initial contact to complete use (i.e., complete chew-down
in foods, complete lotion absorption in skin, etc.).
This chapter’s main objectives were to summarize (a) the original method’s contribu-
tions, key characteristics, needs, and steps for establishing a texture profile descriptive
program and (b) the way many of the philosophical concepts and approaches devel-
oped by the original Texture Profile Method have been updated, as well as how they
were adapted and are being used in other descriptive methods.
The original Texture Profile Method was developed as a global/universal method
for evaluating the texture of many products, mainly foods. Over the years this meth-
od’s practices have been adapted in the following ways, demonstrating the impact that
this method has had in the evolution of descriptive analysis:
• Evolution of the Texture Profile Method and development of other methodology
(e.g., development of texture derivative profile methods, temporal methodology,
etc.)
• Adaptation of the Texture Profile Method for the evaluation of nonfoods (e.g.,
lotions, hair, paper and lip products, pet food, etc.)

References
1. M. A. Brandt, E. Z. Skinner, and J. A. Coleman, “The Texture Profile Method,” Journal of
Food Science 28 (1963): 404–409.

2. S. E. Cairncross and L. B. Sjöström, “Flavor Profiles—A New Approach to Flavor


Problems,” Food Technology 4 (1957): 308–311.

3. J. F. Caul, “The Profile Method of Flavor Analysis,” Advances in Food Research 7 (1950):
1–40.

BK-AST-MNL_13-200365-Chp02.indd 46 30/10/20 4:31 PM


THE TEXTURE PROFILE METHOD 47

4. A. S. Szczesniak, “Classification of Textural Characteristics,” Journal of Food Science


28 (1963): 385–389.

5. A. S. Szczesniak, M. A. Brandt, and H. H. Friedman, “Development of Standard Rating


Scales for Mechanical Parameters of Texture and Correlation between the Objective and
Sensory Methods of Texture Evaluation,” Journal of Food Science 28 (1963): 397–403.

6. A. M. Muñoz and P. A. Keane, “Original Flavor and Texture Profile and Modified/Derivative
Profile Descriptive Methods,” in Descriptive Analysis in Sensory Evaluation, ed. S. E. Kemp,
J. Hort, and T. Hollowood (Hoboken, NJ: John Wiley & Sons, 2018), 237–286.

7. A. M. Muñoz, A. S. Szczesniak, M. A. Einstein, and N. O. Schwartz, “The Texture Profile,” in


Manual on Descriptive Analysis Testing, ed. R. C. Hootman (West Conshohocken, PA:
ASTM International, 1992), 35–50.

8. A. S. Szczesniak, “Sensory Texture Evaluation Methodology,” Reciprocal Meat Conference


Proceedings of the American Meat Science Association 39 (1986): 86–96.

9. E. Larmond, “Beyond the Texture Profile,” in Food Structure: Its Creation and Evaluation,
ed. J. M. V. Blanshard and J. R. Mitchell (Oxford, UK: Butterworth-Heinemann, 1988),
449–464.

10. E. Z. Skinner, “The Texture Profile Method,” in Applied Sensory Analysis of Foods, ed.
H. Moskowitz (Boca Raton, FL: CRC Press, 1988), 89–110.

11. N. O. Schwartz, “Adaptation of the Sensory Texture Profile Method to Skin Care Products,”
Journal of Texture Studies 6 (1975): 33–42.

12. A. M. Muñoz, “Development and Application of Texture Reference Scales,” Journal of


Sensory Studies 1 (1986): 55–83.

13. A. K. Thybo and M. Martens, “Development of a Sensory Texture Profile of Cooked


Potatoes by Multivariate Data Analysis,” Journal of Texture Studies 29 (1998): 453–468.

14. S. Desobry-Banon and Z. Vickers, “Cohesiveness of Mass Evaluation by Time-Intensity


and Single Value Measurements,” Journal of Food Science 63 (1998): 174–176.

15. J. F. Meullenet, B. G. Lyon, J. A. Carpenter, and C. E. Lyon, “Relationship between Sensory


and Instrumental Texture Profile Attributes,” Journal of Sensory Studies 13 (1998): 77–93.

16. S. Cui, K. Yu, Z. Hu, and G. Wei, “Texture Profile of Herbal Gel Affected by Instrumental
Parameters and Correlations of Instrumental and Sensory Evaluation,” Journal of Texture
Studies 42 (2011): 349–358.

17. G. Hough, A. Contarini, and A. M. Muñoz, “Training a Texture Profile Panel and
Constructing Standard Rating Scales in Argentina,” Journal of Texture Studies 25 (1994):
45–57.

18. M. C. Bourne, A. M. R. Sandoval, M. C. Villalobos, and T. S. Buckle, “Training a Texture


Profile Panel and Development of Standard Rating Scales in Colombia,” Journal of
Texture Studies 6 (1975): 43–52.

19. B. Otegbayo, J. Aina, E. Sakyi-Dawson, M. Bokanga, and R. Asiedu, “Sensory Texture


Profiling and Development of Standard Rating Scales for Pounded Yam,” Journal of
Texture Studies 36 (2005): 478–488.

20. E. Dransfield, M. A. Francombe, and O. P. Whelehan, “Relationships between Sensory


Attributes in Cooked Meat,” Journal of Texture Studies 15 (1984): 33–48.

BK-AST-MNL_13-200365-Chp02.indd 47 30/10/20 4:31 PM


48 Descriptive Analysis Testing for Sensory Evaluation: 2nd Edition

21. S. Rousset-Akrim, J-F. Martin, C. Pilandon, and C. Touraille, “Research of Selective Tests
for Discerning ‘An Efficient Assessor’ in Texture Profiling,” Journal of Sensory Studies
10 (1995): 217–237.

22. B. W. Berry and G. V. Civille, “Development of a Texture Profile Panel for Evaluating
Restructured Beef Steaks Varying in Meat Particle Size,” Journal of Sensory Studies
1 (1986): 15–26.

23. G. V. Civille and A. S. Szczesniak, “Guidelines to Training a Texture Profile Panel,” Journal
of Texture Studies 4 (1973): 204–223.

24. Standard Guide for Two Sensory Descriptive Analysis Approaches for Skin Creams and
Lotions, ASTM E1490–19 (West Conshohocken, PA: ASTM International, approved
November 1, 2019), http://doi.org/10.1520/E1490-19

25. A. M. Muñoz, “A5daptive Profile Method®,” in Descriptive Analysis in Sensory Evaluation,


ed. S. E. Kemp, J. Hort, and T. Hollowood (Hoboken, NJ: John Wiley & Sons, 2018),
389–446.

26. A. M. Muñoz and G. V. Civille, “Universal, Product and Attribute Specific Scaling and the
Development of Common Lexicons in Descriptive Analysis,” Journal of Sensory Studies
13 (1998): 57–75.

27. E. Chambers IV, “Consensus Methods for Descriptive Analysis,” in Descriptive Analysis in
Sensory Evaluation, ed. S. E. Kemp, J. Hort, and T. Hollowood (Hoboken, NJ: John Wiley &
Sons, 2018), 213–236.

28. M. Rodriguez, M. Osornio, and G. Hough, “Comparison of Consensus Profiles Obtained at


the End of Product-Specific Training with Profiles Obtained by Individual Measurements
and Statistical Analysis,” Acta Alimentaria 43 (2014): 61–75.

29. H. Syarief, D. D. Hamann, F. G. Giesbrecht, C. T. Young, and R. J. Monroe, “Comparison of


Mean and Consensus Scores from Flavor and Texture Profile Analyses of Selected Food
Products,” Journal of Food Science 50 (1985): 647–650.

30. V. Moiny, J. F. Meullenet, and R. Xiong, “Uniaxial Compression of Cheddar Cheese at


Various Loading Rates and Its Correlation to Sensory Texture Profiles,” Journal of Texture
Studies 33 (2002): 231–254.

31. L. B. Aust, L. P. Oddo, J. E. Wild, and O. H. Mills, “The Descriptive Analysis of Skin Care
Products by a Trained Panel of Judges,” Journal of the Society of Cosmetic Chemists
38 (1987): 443–449.

32. E. Dransfield, R. C. D. Jones, and J. M. Robinson, “Development and Application of a


Texture Profile for U.K. Beef Burgers,” Journal of Texture Studies 15 (1984): 337–356.

33. A. V. Cardello, A. Matas, and J. Sweeney, “The Standard Scales of Texture: Rescaling by
Magnitude Estimation,” Journal of Food Science 47 (1982): 1738–1740.

34. R. Romero del Castillo, J. Valero, F. Casañas, and E. Costell, “Training, Validation and
Maintenance of a Panel to Evaluate the Texture of Dry Beans (Phaseolus Vulgaris L.),”
Journal of Sensory Studies 23 (2008): 303–319.

35. Standard Guide for Measuring and Tracking Performance of Assessors on a Descriptive
Sensory Panel, ASTM E3000-18 (West Conshohocken, PA: ASTM International, approved
April 1, 2018), http://doi.org/10.1520/E3000-18

BK-AST-MNL_13-200365-Chp02.indd 48 30/10/20 4:31 PM


THE TEXTURE PROFILE METHOD 49

36. J. A. McEwan, E. A. Hunter, L. J. Van Gemert, and P. Lea, “Proficiency Testing for Sensory
Profile Panels: Measuring Panel Performance,” Food Quality and Preference 13 (2002):
181–190.

37. E. A. Hunter and D. D. Muir, “A Comparison of Two Multivariate Methods for the Analysis
of Sensory Profile Data,” Journal of Sensory Studies 10 (1995): 89–104.

38. T. Næs and R. Solheim, “Detection and Interpretation of Variation within and between
Assessors in Sensory Profiling,” Journal of Sensory Studies 6 (1991): 159–177.

39. B. D. Donfrancesco, K. Koppel, and E. Chambers IV, “An Initial Lexicon for Sensory
Properties of Dry Dog Food,” Journal of Sensory Studies 27 (2012): 498–510.

40. G. Ares, A. Giménez, and A. Gámbaro, “Preference Mapping of Texture of Dulce de Leche,”
Journal of Sensory Studies 21 (2006): 553–571.

41. L. M. Dooley, K. Adhikari, and E. Chambers IV, “A General Lexicon for Sensory Analysis of
Texture and Appearance of Lip Products,” Journal of Sensory Studies 24 (2009):
581–600.

42. A. M. Muñoz, “Sensory Evaluation in Quality Control: An Overview, New Developments


and Future Opportunities,” Food Quality and Preference 13 (2002): 329–339.

43. E. J. Mayhew, S. J. Schmidt, P. Schlich, and S. Y. Lee, “Correlation of Consumer Perception


of Stickiness and Contributing Texture Attributes to Trained Panelist Temporal
Evaluations in a Caramel System,” Food Quality and Preference 65 (2018): 72–80.

44. S. C. Hutchings, K. D. Foster, D. I. Hedderley, and M. P. Morgenstern, “Differences between


Age Groups in the Use of the Temporal Dominance of Sensations Technique across a
Range of Food Textures,” Journal of Texture Studies 45 (2014): 206–219.

45. C. Kuesten, J. Bi, and Y. Feng, “Exploring Taffy Product Consumption Experiences Using a
Multi-Attribute Time–Intensity (MATI) Method,” Food Quality and Preference 30 (2013):
260–273.

46. L. M. Duizer, E. A. Gullett, and C. Findlay, “The Relationship between Sensory Time-
Intensity, Physiological Electromyography and Instrumental Texture Profile Analysis
Measurements of Beef Tenderness,” Meat Science 42 (1996): 215–224.

47. R. H. Brand, “Measurement of Fabric Aesthetics: Analysis of Aesthetic Components,”


Journal of Textile Research 34 (1964): 791–804.

48. G. V. Civille and C. A. Dus, “Development of Terminology to Describe the Handfeel


Properties of Paper and Fabrics,” Journal of Sensory Studies 5 (1990): 19–32.

49. Standard Guide for Sensory Evaluation of Household Hard Surface-Cleaning Products with
Emphasis on Spray Triggers, ASTM E2346 (West Conshohocken, PA: ASTM International,
approved November 1, 2015), http://doi.org/10.1520/E2346_E2346M-15

BK-AST-MNL_13-200365-Chp02.indd 49 30/10/20 4:31 PM


Chapter 3 | Quantitative Descriptive
Analysis
Rebecca N. Bleibaum,1,2 K. W. Clara Tao, 3 and Heather Thomas1

Introduction
Quantitative descriptive analysis (QDA) is a unique system for obtaining quantitative
sensory perceptual information about products. The system relies on a group of quali-
fied subjects who are trained to verbalize their sensory perceptions for attributes of
products. Individual responses are collected in replicate, yielding a quantitative data
set for further statistical analysis, a key component of the QDA method.
The QDA method can be used across the product life cycle, including, for exam-
ple, product development, ingredient substitution, monitoring competition, sensory
claims, and quality control. The QDA method is based on experimental and behav-
ioral psychology and relies on repeated measures and statistical analysis to determine
significant effects.
One value of a quantitative data set is to identify relationships between a QDA
panel’s description of those products, the technical language used by experts to
describe those products, and a larger population of consumer preferences to products.
This information provides a basis for relating specific ingredients and processing vari-
ables to consumer perception.
The QDA method was first published in 1974 by Stone et al.,1 introducing a proce-
dure for quantifying sensory descriptive perceptions. The method also introduced
several concepts new to descriptive testing, including individual data rather than con-
sensus opinion, an unstructured graphic rating scaling, replication, and specified sta-
tistical treatment of the resulting data. QDA is consumer-oriented in panel selection,
training, and attribute language and is useful for a wide variety of product develop-
ment and marketing activities, including product optimization research. This chapter
describes several product applications and more rapid approaches, including

1
 ragonfly SCI, Inc., 2360 Mendocino Ave., Ste. A2-375, Santa Rosa, CA 95403, USA
D
2
 niversity of California, Davis, Continuing and Professional Education, Sensory & Consumer Science Certificate
U
Program, Davis, CA, USA
3
Tragon Corp. (deceased)
DOI: 10.1520/MNL1320170005
Copyright © 2020 by ASTM International, 100 Barr Harbor Dr., PO Box C700, West Conshohocken, PA 19428-2959
ASTM International is not responsible, as a body, for the statements and opinions expressed in this chapter.
ASTM International does not endorse any products represented in this chapter.

BK-AST-MNL_13-200365-Chp03.indd 51 30/10/20 4:31 PM


52 Descriptive Analysis Testing for Sensory Evaluation: 2nd Edition

diagnostic descriptive analysis (DDA), which includes more subjects, selected attri-
butes, and fewer replications; and the master QDA panel approach, which utilizes
subjects who are professionals and have training in their area of expertise (e.g., chefs,
brewmasters, winemakers, cosmetologists). It also describes the procedure and ratio-
nale and experimental and behavioral sciences that have supported the methodology.

Background and Core Philosophy


The QDA method relies on measurement and quantification. The approach reflects the
science and experimental psychology backgrounds of its developers Dr. Herbert Stone
and Joel L. Sidel and their commitment to the scientific method. It was developed over
a period of years before, during, and after they worked together at the Stanford
Research Institute in Palo Alto, California, now known as SRI International.
The sensory language used by the panel is a common, everyday language that is
easily understood by the product development team, marketing and sales teams, and
consumers. Because the QDA method uses subjects who are likers and users of the
product category and selects those consumers with demonstrated sensory acuity
within the product category, the resulting language and subsequent quantitative
descriptive data can be combined with consumer-affective results to better understand
how the products’ sensory experience influences consumer behavior.
QDA has been described in several publications and books and in the writings of
Stone, Sidel, and their associates.1–3 From QDA’s inception, its developers compiled the
best available sensory practices and knowledge of human sensory measurement and
experimental psychology and expected the method to evolve as practices and the
knowledge base grew across multiple categories. The fundamental elements that define
QDA are as follows:
1. Provide a scientifically rigorous procedure with quantifiable results
2. Produce output directly generalizable to consumers
3. Employ a panel leader who is a discussion facilitator, with minimal influence
4. Involve all sensory modalities and their interaction
5. Be cost-effective
6. Screen and train subjects in less than 2 weeks
7. Provide multiple applications in product development, quality control, and marketing
A key challenge in sensory science is accounting for the inherent variability
within the research process—products vary, people vary, and methods vary, and the
QDA methodology embraces this inherent variability. QDA uses an iterative process to
create a comprehensive descriptive language that reflects the panel’s perceptions of the
product array. The QDA moderator then trains subjects on the use of the unstructured
graphic ratings scale to measure their perceptions, and then individual subject data are
collected using repeated measures (replication) to provide a robust data set for statisti-
cal analysis. The data obtained from QDA panels can be used with the same level of
confidence that one would expect from other analytical measures.

BK-AST-MNL_13-200365-Chp03.indd 52 30/10/20 4:31 PM


QUANTITATIVE DESCRIPTIVE ANALYSIS 53

Overview of Method
QDA is a consumer behavioral approach that uses trained descriptive panels to mea-
sure a product’s sensory characteristics. Panel members use their senses to perceive
product similarities and differences and articulate those perceptions in their own
words.
The QDA method involves measuring perceptions derived from products as the
source of the stimulation. The method embraces the uniqueness of our measuring
instrument (humans) based on physiological, psychological, and behavioral differ-
ences, along with experiential differences with a specific product or related products.
The responses reflect what is perceived in its entirety—how a product looks, smells,
feels (in the hand and mouth), and tastes; or, for nonfood items, the product experience
before, during, and after usage.
This matter of “perceived in its entirety” is an aspect of the sensory process that is
often disregarded or not fully appreciated by the descriptive analysis methods. When
planning a test and considering the evaluation process, there was a tendency to think
in narrow terms of flavor or texture; however, this is not the entire picture of what is
perceived. These sensory interrelationships must be considered regardless of the type
of test that is planned. If not accounted for, useful information may be lost, and the
overall value of the sensory resource may be compromised. The overall goal of the
QDA method is to provide a complete measure of the perceived sensory experience of
the products under study.
The QDA methodology has been applied to a wide variety of consumer goods,
including all types of foods and beverages, personal and household care, office sup-
plies, apparel, furniture, wearable electronics, and appliances. The application of the
QDA methodology is limited only by the creativity of the sensory scientist.
In the QDA method, qualitative references are used to clarify the definitions of
sensory attributes on an as-needed basis and to ensure that all subjects have a common
experience. References are complex, and the sensory scientist must take care to ensure
that they are used to help clarify the definition and to provide a common experience to
all subjects. The reference material may consist of a complex finished product (sour
lemon candy), less complex products (lemon juice), or individual ingredients such as
lemon peels, lemon oil, or lemon extract. However, each reference is carefully selected
when necessary, as deemed by the panel moderator, to aid panel discussions, provide a
common sensory experience, and promote a common language among the panel
members.

Subjects/Assessors Screening and Qualification


A high priority in the QDA methodology is the choice of subjects. All individuals (con-
sumers) interested in participating must be voluntary and qualified to participate on
the basis primarily of their demonstrated sensory skills in the product category of
interest. The following guidelines are a reminder that subjects must be qualified and

BK-AST-MNL_13-200365-Chp03.indd 53 30/10/20 4:31 PM


54 Descriptive Analysis Testing for Sensory Evaluation: 2nd Edition

provide directions as to how best to proceed with the qualifying process. Each product
category and each test type will have some unique requirements. Therefore, a basic set
of guidelines is far more useful for the sensory professional when developing a pool of
qualified subjects.
Subject guidelines to consider include the following:
• Participation should be voluntary and can be stopped at any time.
• All subjects’ personal information should be kept confidential.
• Any allergic or medical conditions that may impact testing should be determined.
• Subjects should like and be average or above-average users of the product categories
under study.
• Subjects should have demonstrated sensory skill within the product categories
under study.
People have a wide range of sensory skills in terms of their sensitivities as well as
in their abilities to articulate what they perceive. There is currently no way of knowing
in advance who these people are. However, it has been well established that consumers
who use a product frequently tend to be more skilled (than those who do not or use it
infrequently). Also, one’s ability to discriminate among products of the type being
tested is a very good indicator of sensory skill. The combination of these two criteria
serves as a reasonable basis for selecting those individuals best suited for language
development.
These procedures are designed to identify those individuals who are most likely to
perceive differences, articulate the basis for those differences, and provide intensity
judgments with a high degree of reliability. However, even with this screening, those
individuals who meet all criteria will not be equally sensitive and reliable. The screen-
ing process is intended to identify those individuals who are insensitive or unreliable,
or both, and eliminate them from further consideration as subjects, leaving a group of
people who are reasonably homogeneous in their sensory skills.
Potential subjects are prerecruited following a specific script (usually via tele-
phone or online) that begins with appropriate category usage qualification. The inter-
view includes an assessment of availability, verbal fluency, interest, and comfort with
participating in a group activity.
Once a potential set of qualified volunteers has been identified, the sensory staff
should expect to begin with some form of orientation and screening within a few
weeks. Failure to initiate screening within a few weeks will reduce interest and make it
more difficult to recruit others. The orientation includes a brief explanation (about 10
minutes) of sensory science followed by a booth-test demonstration. This first test in a
booth is planned to yield almost unanimous agreement among the subjects and allow
them to gain familarity with evaluating products, and this should have a direct and
positive impact on motivation.
Orientation and screening should be done with not more than about 10 to 12
individuals at a time. This ensures that each person is given adequate attention and

BK-AST-MNL_13-200365-Chp03.indd 54 30/10/20 4:31 PM


QUANTITATIVE DESCRIPTIVE ANALYSIS 55

that all of his or her questions are answered. It also makes for a more manageable effort
for the experimenter. The objective is to make each person feel that his or her responses
are important and a benefit to the program. At the same time, it also is important to
avoid creating the concept of an elitist group.
The actual screening will vary depending on each company’s products; how-
ever, the common point is that the screening tests use products of interest (i.e.,
products that will be evaluated in tests that follow). Sensory acuity screening is best
done using the discrimination model to determine whether subjects can find the
known product differences. While procedures such as ranking and scoring can be
used to supplement discrimination testing, they are of secondary importance in
assessing sensory skill.
The primary purposes of screening are to identify and eliminate those individuals
who cannot follow instructions, are insensitive to differences among products being
evaluated, exhibit little interest in the activity, or lack verbal fluency. Perhaps most
importantly, the screening should identify those individuals who can detect differ-
ences at better than chance among the products of interest. These are the ones identi-
fied as qualified based on sensory acuity. This can easily be achieved in about 5 to 6
hours (over several days and sessions), even starting with individuals with no prior
experience (i.e., naive consumers).
Once qualified in the preinterview, about 30 subjects (in groups of about 10)
report to a central location testing facility and are given a series of up to 20 discrimina-
tion tests over the course of several days. These tests are designed to cover the range of
products in the category (or categories) of interest and cover each modality that will be
evaluated by the panel, including differences before, during, and after usage. The dis-
crimination tests range from easy to moderate to difficult and are based on the product
set(s) of interest. Subjects scoring significantly above and beyond chance in the series
of screening tests (e.g., ≥70% correct duo-trio testing) qualify as having satisfactory
sensory acuity.
To establish an array of discrimination tests, the sensory professional must bench-
screen the category of products to be tested, identify observable differences, and pre-
pare a series of product pairs that include differences before, during, and after usage.
The number of pairs should be up to 20; with replication, this provides up to 30 or 40
trials/judgments, which is sufficient for subjects to demonstrate their abilities and for
the sensory professional to be able to classify individuals on the basis of their sensitiv-
ity and reliability. The effectiveness of this approach is based on the subsequent test
performance of those individuals. The degree of difficulty of the product pairs should
be considered. If more than a few product pairs are easy to differentiate the screening
process will not be effective, as most everyone will qualify. Only after this procedure
has been used once or twice will the sensory professional know whether pairings rep-
resent the desired easy, moderate, and difficult options. After screening, about 12 sub-
jects will be identified as potential candidates for language development on the basis of
their sensory acuity and their availability to complete the project.

BK-AST-MNL_13-200365-Chp03.indd 55 30/10/20 4:31 PM


56 Descriptive Analysis Testing for Sensory Evaluation: 2nd Edition

Qualifications of Panel Leader/Moderator


The QDA panel moderator performs the role of a true moderator and is not a panel
participant. The panel moderator facilitates the development of the language, ensures
that all subjects participate, explores individual differences among panelists, and
encourages open discussion to resolve disagreements.
Panel moderator responsibilities include the following:
• Understand research objectives
• Organize sessions
• Keep subjects focused
• Build confidence through positive reinforcement
• Help subjects define terms by reaching a consensus on language and procedures
• Encourage participation from all subjects
• Provide training materials and pacing of the sessions

Language Development/Panel Training


Developing a language that describes the sensory characteristics of a set of products is
the overall objective of the language development phase of the QDA method. The
method uses a consumer-based language, not technical words or a previously deter-
mined set of words. QDA subjects develop this language under the guidance of a panel
moderator, the person who is responsible for organizing and leading the language
development effort. The panel moderator plays a pivotal role in the success of the lan-
guage effort and in the success of all subsequent tests with that panel.
Language development is the most demanding component of the QDA method-
ology, not only in terms of the actual time required of the panel moderator but of the
subjects as well. For most products, five 90-minute sessions, held on consecutive days,
are usually sufficient. However, there can be more or fewer sessions depending on the
skill of the panel and the number and complexity of the products. Determining the
number of sessions required for language and overall panel readiness is the responsi-
bility of the panel moderator. For personal and homecare products, alternating days of
the group sessions, shorter in duration, and having the evaluations at home probably
will be necessary. More sessions also may be required because usage and scoring
during each session will be limited. However, the basic principles of the QDA method-
ology still apply, whether there are one, five, or more sessions or whether the product is
evaluated at home or in a facility designed specifically for the sensory evaluation of
products. It is essential for the panel moderator to listen to a request for assistance and
only after due consideration decide how and where the language development and
subsequent testing/data collection will be conducted.
The time required of the panel moderator will be more than that of the subjects.
For each session, about 90 to 120 minutes of preparation are required, outside of the

BK-AST-MNL_13-200365-Chp03.indd 56 30/10/20 4:31 PM


QUANTITATIVE DESCRIPTIVE ANALYSIS 57

session, to select products or product pairings, or both, and to determine the emphasis
that will be given to specific product characteristics (color, flavor, etc.). An additional
90 to 120 minutes of moderator time is required after each session to record the infor-
mation developed in the session, organize the paperwork for the next session, and so
forth. Panel moderators usually have an assistant to help with these activities, such as
creating and editing the scorecard and definitions of terms and preparing products.
The overall objectives of the language-development sessions are to develop a
scorecard with a list of attributes that reflect the subject’s perceptions in common,
everyday language. The panel creates a set of definitions for those attributes and an
evaluation procedure for a set of products. Because it is the subjects who will evaluate
the product, it is important to keep in mind that these activities reflect what they per-
ceive, not what you, the panel moderator, would like them to perceive. Within this
framework, there are a series of specific tasks that need to be accomplished:
• Develop a language (the attributes) that fully describes the products
• Develop clear definitions/explanations for each attribute
• Order the attributes within each modality (appearance, aroma, etc.)
• Develop and document the evaluation procedures
• Develop a frame of reference for scale use
• Practice scoring products
To begin the language development process, an individual and group orientation
is provided for the subjects. During the orientation, the panel moderator facilitates
introductions and introduces the general concepts of language development, describ-
ing their sensations and perceptions of the product category. This introduction and
orientation should take no longer than 20 minutes.
Immediately following the orientation, the language development process begins
as a group activity. Panel members are provided with an appropriate amount of the
product for evaluation—more than they will actually use. The first product given to
each subject should be the one most “typical” for the category (e.g., a benchmark,
gold-standard control, or one that defines the category), along with a category sheet.
Each panel member is asked to divide his or her perceptions into categories/modalities
such as before usage (visual, aroma, etc.), during usage (application, flavor, mouthfeel,
handfeel, etc.), and after usage. Once all panel members have written down their indi-
vidual perceptions, the panel moderator will call on all subjects to describe what they
have written, tracking each response on the board. This process is repeated for three or
four products that best represent the range of products in the research, when typically,
90% of the words needed to describe the product category will have been generated.
Products are thoughtfully selected by the panel moderator to ensure the range of
differences within the product array of interest has been provided to the panel. Each
panel session is about 90 minutes in length, and it may require more than one session
to describe three or four products because of the physical nature of the category. The

BK-AST-MNL_13-200365-Chp03.indd 57 30/10/20 4:31 PM


58 Descriptive Analysis Testing for Sensory Evaluation: 2nd Edition

language development process is iterative (i.e., the words generated in the first few ses-
sions to cover the product category are reviewed, discussed, and defined). The subjects
then practice scoring products using an unstructured graphic rating scale (fig. 1). They
develop a comprehensive list of words to describe the product array and the specific
procedures for their evaluation that is most typical for the category of interest. In addi-
tion, the subjects decide upon appropriate anchor words for each scale (such as weak to
strong or slightly to very).
The panel moderator creates definitions for each attribute scored based on input
from the panel so that the final definitions represent a true group consensus. The defi-
nitions are always present in the data-collection sessions so that subjects can reference
the meaning of each attribute. This panel method is designed to provide direct con-
sumer feedback to the technical developers and marketing teams on how these prod-
ucts are similar and different based on their sensory properties, absent of brand and
imagery.
During language development, subjects practice scoring and discuss their per-
ceptions on the scale. Scale context is established by exposing subjects to the actual
range represented by the stimuli (i.e., products) in the category or in the project.
Subjects establish their own scale location for products included in the range of, and
practice with, the training stimuli. It is not expected, or necessary, that all subjects
perceive or score the low-intensity product at the same exact scale location. Individual
subject differences due to scale location are accounted for, and extracted with, the
analysis of variance (ANOVA) model used to analyze the data.
In training sessions in which subject’s scores are distributed over an unusually
large-scale range, those scores are shown (e.g., whiteboard, flipchart, computer screen)
and the panel discusses whether that range is consistent with their perception. The
subjects decide for themselves whether their score represents what they perceive. As
indicated in the previous paragraph, differences in scale location are accounted for. All
that is required is that subjects remain reasonably consistent (i.e., normal expected
variation) in applying their internalized frame of reference and scale and through
repeated product exposure and practice with the scale; subjects will generally demon-
strate increased consistency in their ratings. The initial reaction from most individuals
is to look for “the correct location” on the scale or to see whether the person adjacent to
them is making a mark in the same place.

FIG. 1 QDA unstructured graphic line scale.

BK-AST-MNL_13-200365-Chp03.indd 58 23/11/20 10:43 AM


QUANTITATIVE DESCRIPTIVE ANALYSIS 59

In the third and subsequent sessions, qualitative reference materials may be used
to help subjects where necessary. Generally, references are presented without identifi-
cation, only with a letter or number and subjects are asked to write a word or two to
describe what they perceive. The panel is then told that these are presented to provide
a common experience and asked whether these references match any of the attributes
on their scorecard. Any material can be used; however, the raw materials used to for-
mulate the products may be most helpful in this regard. Subjects may also want to
bring materials from home if they believe it will help to explain a particular sensation.
This is perfectly acceptable, but here too, the panel moderator should exercise control
so that it does not become a mission (i.e., each subject feels he or she must bring some-
thing). Sometimes references encourage the subjects to talk with each other, which is
essential if the attributes are to be understood in a similar manner. References may be
useful in clarifying attribute meaning and in deciding which attributes are most useful
to the panel in describing the product array. Where references are helpful, they should
be documented so they are available for future applications. References also can be
helpful when training new subjects; however, it is not necessary, required, or recom-
mended to have a reference for every attribute. When references are used, they should
be identified and sampled in relation to the specific attribute(s) being discussed.
Some comments are warranted about the QDA language that is used to character-
ize the products. Because it is the basis for a descriptive test, questions are often raised
as to how subjects develop attributes, how one knows if the attributes are correct, and
how one knows when there are enough attributes.
First, the reader should keep in mind that the words used by the subjects are labels
to represent sensations. For example, if a panel uses the words “salty flavor” it does not
necessarily mean that there is a direct relationship with the amount of salt (or of
sodium ions) in a product’s formula. Likewise, attributes such as fruit flavor or artifi-
cial are most likely composites of several ingredients rather than a specific ingredient.
They represent a panel consensus as to what word (or group of words) to use to best
represent a particular sensation in the context of that set of products—hence, the
observation that the attributes should be viewed as labels and not as having a direct
relationship with a specific ingredient or having some greater meaning beyond the
product category being evaluated. Information about the effects of ingredients can be
developed through use of appropriate design studies. It is most important, at least in
the short term, for the panel to develop a list of attributes that enables them to describe
product similarities and differences in a reliable and valid manner.
Theoretically, there is no limit as to the number of attributes that a panel can
develop to fully describe a set of products. However, there are practical limits that
reflect the ability of humans to process information. In addition, time spent in discus-
sion also helps to minimize the duplication of attributes. Empirically it has been
observed that most product scorecards have about 40 attributes (heterogeneous prod-
ucts could have more attributes), and these attributes are grouped by modality, e.g.,
appearance, aroma, etc. While the total number of attributes can be quite large, the

BK-AST-MNL_13-200365-Chp03.indd 59 30/10/20 4:31 PM


60 Descriptive Analysis Testing for Sensory Evaluation: 2nd Edition

number associated with each modality will be much smaller. For example, there could
be 10 attributes to encompass product appearance, another 7 or 8 for aroma, 10 to 12
more for flavor, and so on. Because each modality is evaluated sequentially, the total
number of attributes is much less important. As previously stated, there is nothing
special about the number of attributes or the specific words the subjects use, within
certain constraints. It is better to have too many, rather than too few, attributes. Too
few attributes results in less sensitivity and less product differentiation. Too many
attributes adds some panel time, but it can result in improved product differentiation.
Because all subjects will not be equally sensitive to all attributes, having a few
additional attributes increases the likelihood that product differentiation will occur. It
reduces frustration for some of the subjects to include attributes that some of them
detect better than others. During training the subjects have numerous opportunities
to evaluate products and to use the attributes. It is the panel moderator’s responsibility
to help the subjects address the different “perceptions” and where possible to resolve
any disagreements that arise over the presence or absence of a particular attribute.
However, it is risky to decide, a priori, what attributes should or should not be retained
without first obtaining sufficient responses from each subject in a pilot test. Training is
a group effort, and it takes time for the subjects to develop a working relationship with
one another and to be able to describe their perceptions. It is relatively easy for an
inexperienced panel moderator to be misled into concluding too early in the training
process that the panel is or is not sensitive to a particular attribute or to be unduly
influenced by one or two subjects. Entirely different results may be obtained when the
subjects score the products in the booths. A critical element in this training process is
to be sure that the subjects have sufficient time to practice evaluating products and
discuss their perceptions as a panel. The discussions help to resolve the issue of whether
there is too much overlap or too many words to represent a particular sensation.
Therefore, it is best to leave some attributes for which the subjects are having difficulty
reaching a consensus, and only after a test will it be possible to assess how well those
attributes are being used. However, the panel leader must keep in mind that there are
limits as to how much information humans can process. If there are too many choices,
sensitivity is reduced, and variability increases substantially. Clearly, if there are 60 or
70 attributes for a homogeneous product, that is probably too many.
It has been suggested that one can reduce the number of attributes through the
use of factor analysis of responses obtained during training. Such action is not recom-
mended for numerous reasons, including the preliminary nature of the language, the
limited database, the psychological impact of the loss of attributes proposed by certain
subjects, and so forth.
The subjects are told they can use any words they want to describe their percep-
tions provided they do not use words that relate to product liking, connote quality, or
are technical in nature. For example, using “a product I would like,” “good for me,” “is
not my brand,” and so forth, are actively discouraged by the panel leader. Technical
jargon is especially problematic; the words will have very different meanings

BK-AST-MNL_13-200365-Chp03.indd 60 30/10/20 4:31 PM


QUANTITATIVE DESCRIPTIVE ANALYSIS 61

depending on the individual, even among a group of technologists. In addition, the


subjects introducing technical attributes often provide explanations that are them-
selves technical, and this adds to the confusion about those attributes.
Once a language (attributes, definitions, and product-handling procedures) has
been developed for a product category and used for data collection, changes are easily
made, usually requiring not more than a single 90-minute session. Changes might
include adding attributes because of a product not previously evaluated or dropping
some attributes because they are not used. As a panel leader, you will decide on the
extent of training required for a particular task on the basis of your experience with
these products and subject performance.

SUMMARY OF THE LANGUAGE DEVELOPMENT PROCESS


Successful use of the QDA method depends on the panel moderator’s preparation and
work with the subjects. As already noted, training is generally done in a series of five
90-minute sessions involving the entire panel. As a general guide, the emphasis in the
first two sessions is on developing attributes, familiarizing the subjects with the prod-
ucts and the evaluation process, and developing some explanations. The third and
fourth sessions emphasize the use of the scale, ordering of attributes, adding/subtract-
ing of attributes, developing specific definitions, and an evaluation protocol. The fifth
session emphasizes the completion of all these tasks.
Once these general tasks are accomplished, training is completed, and subjects
should proceed directly with data collection. At this stage, especially if the panel is
new, the product set is new, or the language is new, a pilot test may be recommended.
For example, if training were completed on Thursday or Friday, then Monday would be
the first test day. If for some reason the training were completed on Wednesday, testing
should begin on Thursday.

GENERAL CONSIDERATIONS IN QDA LANGUAGE DEVELOPMENT


It is important to realize that subjects will not be in complete agreement about every
attribute; nor will all subjects be in complete agreement as to which product is more
intense for every attribute. For some attributes, the subjects will be evenly divided,
reflecting small or no real differences between products. Such scoring disagreements
among subjects are not unusual and reflect individual differences in sensitivity among
subjects and individual product variation.
Subjects are volunteers from the consumer population and have been selected on
the basis of their sensory acuity in perceiving small differences among the products. If
each has the freedom to contribute fully to the development of the terminology, the
resulting scorecard will represent consumer perceptions of the products. Because the
subjects are not involved directly with the product development process and do not
know the variables among the products, they are expected to have no more bias than
the average consumer. One subject may be especially sensitive to one attribute but not
very sensitive to others; another may be the opposite. By using 10 to 12 people, the
range of sensitivities will be well balanced for all the products’ sensory properties.

BK-AST-MNL_13-200365-Chp03.indd 61 30/10/20 4:31 PM


62 Descriptive Analysis Testing for Sensory Evaluation: 2nd Edition

In every way, the scorecard is a product of the subjects. Care should be taken to
avoid forcing them to use attributes with which they are uncomfortable; the resultant
data will be biased or have considerably more variability. Also, if an attribute is to be
dropped (not used), the reason should be clear, and the subjects should discuss their
decision. Deleting an attribute that the panel considers important risks losing an
important attribute, undermines their self-confidence, and implies there are correct
answers to the test.
Although the panel leader should avoid providing attributes that the panel may
feel obligated to use (especially during the first two or three sessions), anything to help
the panel’s understanding should be used (e.g., if a panel member says something has
a musty/earthy odor, it is fine to ask whether it is somewhat like the odor of a dirt cellar,
raw mushrooms, or soil on potato skins to help describe it). As a general rule, subjects
should be told that if an attribute proposed by the panel leader is not helpful then they
can drop it by a simple vote. This procedure serves as a reminder that they (the sub-
jects) are in charge of the process.
If everything tried (discussion, reference, etc.) fails to bring about a consensus
agreement for an attribute, it may be better to allow the subjects to remove it from the
scorecard. If subjects feel frustrated, it will affect their responses to other attributes.
The data obtained from that attribute will most likely be uninformative. On the other
hand, even if only two or three subjects are responding consistently to an attribute that
others cannot perceive, it is better to retain that attribute. As always, the decision is the
subjects’ responsibility, although the leader may have to make the final decision if there
is disagreement.
There is no such thing as a problem subject—only a subject who has a problem. If
it cannot be solved, then you will likely have a frustrated/unhappy subject. The panel
leader develops the panel by keeping an open atmosphere, enhancing subject self-es-
teem by active listening, providing feedback as needed, and helping subjects to describe
a sensation when words will not come. Other subjects may follow the leader’s example
and aid each other in clarifying descriptions. Sometimes one encounters an individual
who has functioned effectively during screening but is so uncomfortable in the group
sessions that they ask to be excused. This is a reasonable request and should be granted.
Alternatively, there are individuals whose behavior is so disruptive that they should be
excused rather than allow them to destroy the panel’s effectiveness. These are not easy
decisions but reflect actual situations.
The initial goal is for subjects to describe products in their own words. When a
subject has difficulty with a description, the panel leader may make suggestions; how-
ever, subjects should first be encouraged to help one another. The panel leader must be
objective (i.e., without judgment of right/wrong, good/bad). Because of the high degree
of group dynamics involved, subjects are more sensitive to the responses of others than
they might otherwise be.
While the inexperienced panel leader will worry that subjects will not develop
many words, it is often the converse that occurs (i.e., the subjects will develop many

BK-AST-MNL_13-200365-Chp03.indd 62 30/10/20 4:31 PM


QUANTITATIVE DESCRIPTIVE ANALYSIS 63

words). The task for the panel leader in this case is to focus the panel on consolidation.
A second issue is the ability to control the group dynamics without appearing to con-
trol it. At some point, usually by the third session, the subjects develop sufficient
self-confidence, and there is a tendency for them to have simultaneous conversations.
While such discussions may appear to be useful, they are not, and the panel leader will
need to control this activity.
Once a scorecard (fig. 2) and attribute definitions (fig. 3) have been developed and
used in a test, the panel will be able to use it more effectively as more tests are initiated.
Prior to each test, the panel meets for a session, usually 90 minutes in duration, to
review the scorecard and make any changes based on the products in the new test.
Here the panel leader provides the scorecard, list of definitions, and a product. After
the second product, the panel leader will focus on selected groupings of attributes
reflecting the current products and results from the previous test. The panel leader will
need to review past results and identify specific attributes that warrant attention (e.g.,
product differentiation was lost due to interaction).

Data Collection
Replication and analyses (ANOVA) take into account differences in scale use; this
determines whether the mean is a realistic measure of the distribution of the scores.
Products are evaluated in individually controlled evaluation areas (e.g., booths) or in a
typical usage situation (at home, work, or play).
Strengths of perceptions are converted to numerical values through the use of a
line or graphic rating scale. The QDA scale developed by Stone et al. in 1974 is a 6-inch
(∼15 cm) unstructured line scale with two anchors, each 1/2 inch (1.25 cm) from the
end of the scale. These anchors are typically labeled with low/weak at the left to high/
strong on the right. Strength of a perception is recorded by placing a vertical mark on
the scale that best represents that individual’s sensitivity to that characteristic. The
effectiveness of the scale is based on the subject experiencing the range of products (the
stimuli) being tested.

Pilot Test and Validation


Once the sensory language has been developed for the product category of interest, a
pilot test is initiated. This test is designed to evaluate individual subject performance
and determine whether the panel, as a whole, is using the sensory attributes in the
same or similar ways. Before the pilot test, the panel moderator reviews the proce-
dures, definitions, and rating scales for each attribute. Generally, four products may be
selected from the category that represents a relatively broad sensory range. Next, these
products are evaluated by each subject, on an unbranded basis, in the individual test-
ing booths and, if necessary, in an extended usage situation.
For the pilot test, four replications are recommended. In this case only, the prod-
ucts are evaluated in the same order for each subject. The reasoning behind this is that

BK-AST-MNL_13-200365-Chp03.indd 63 30/10/20 4:31 PM


64 Descriptive Analysis Testing for Sensory Evaluation: 2nd Edition

FIG. 2 QDA scorecard (partial).

the pilot test data are analyzed to determine individual panel performance and attri-
bute agreement, not product differences. The data should be analyzed thoroughly with
one- and two-way ANOVA for each sensory attribute to determine whether the panel,
as a whole, scored products differently from one another. Multiple-comparison proce-
dures (such as Duncan’s new multiple range test) are calculated after the ANOVA to
identify statistically significant differences among products for each sensory attribute.

BK-AST-MNL_13-200365-Chp03.indd 64 30/10/20 4:31 PM


QUANTITATIVE DESCRIPTIVE ANALYSIS 65

FIG. 3 QDA definitions of terms (partial).

CHOCOLATE BEVERAGE QDA DEFINITIONS OF TERMS

APPEARANCE

Judge Instructions: Look at the cup of product, pick up, examine closely, then evaluate:
BROWN COLOR Overall impression of brown color of the product, ranging from a light
(light-dark) brown to a darker brown to a deep or very dark brown color.
THICKNESS (thin-thick) Overall impression of the thickness of the product, ranging from thin and
watery to thick and viscous.
FOAMY (slightly-very) Overall impression of the amount of foam on and in the product.

AROMA

Judge Instructions: Pick up the cup of product, hold close to nose, smell, and then evaluate:
CHOCOLATE Intensity of distinct chocolate aroma of any type; like chocolate chips or
(weak-strong) candy bars.
DAIRY/MILKY Intensity of milk or dairy aroma; reminiscent of the lactic tang with fresh
(weak-strong) dairy milk.
MALT (low-high) Intensity of malt or malted aroma.

FLAVOR

Judge Instructions: Take a few sips of the product, as necessary, to evaluate:


CHOCOLATE Intensity of distinct chocolate flavor of any type; like chocolate chips or
(weak-strong) candy bars.
DAIRY/ MILKY Intensity of milk or dairy flavor; reminiscent of the lactic tang with fresh
(weak-strong) dairy milk.
POWDERED MILK Intensity of powdered or dry milk powder flavor; like non or low-fat milk
(weak-strong) powder.
MEDICINAL (weak- Intensity of medicinal flavor; may include vitamin types of flavors.
strong)
MALT (weak-strong) Intensity of malt or malted flavor.
ALCOHOL/EXTRACT Intensity of alcohol or extract flavor; reminiscent of alcohols used in extracts
(weak-strong) of spices or flavorings.

Calculations for panel performance measures include six key measures: standard devi-
ation, crossover, scale range usage, discrimination, scale position usage, and subject
attribute decision influence. This panel performance analysis is conducted with the
pilot test data, and with each subsequent QDA study to monitor subject behavior.
More on this topic will be covered in the next section, “Panel Performance Measures.”
The pilot test data are only for the sensory scientist to make decisions on areas for
remedial panel training if necessary. Based on results of the pilot test, the panel is
reconvened to discuss attributes and products in which the panel leader seeks clarifi-
cation of the definitions or evaluation methods or both. After initial training and pilot
testing, remedial training sessions may be scheduled if necessary, or the panel may
begin their data-collection sessions.

BK-AST-MNL_13-200365-Chp03.indd 65 30/10/20 4:31 PM


66 Descriptive Analysis Testing for Sensory Evaluation: 2nd Edition

Panel Performance Measures


Each subject’s individual data are analyzed to determine whether that subject is satis-
factorily contributing to product discrimination. A one-way ANOVA for each subject
is computed for each attribute. This computation takes into account the subject’s rat-
ings for replications of each product. The subject’s one-way F-test is based on the mean
square for the between variable (i.e., products) divided by the mean square for error
based on the replication scores for each product. Subjects contributing worse than
chance (in this case defined as 0.50–0.99) are judged to exhibit poor discrimination of
product differences for that sensory attribute.
A two-way ANOVA is used to determine whether the overall panel result for
products on an attribute is significant. The source of variance for this model includes
product, subject, product × subject, and error. The main effects and interactions for
replications are typically collapsed into the error. Experience with numerous studies
have shown that this has little statistical impact on the error term, as thousands of
QDA panels have seldom exhibited significant error associated with replication.
A proprietary analysis included in the original Tragon statistical program and
now incorporated into several statistical programs computes a probability for each
subject’s contribution to the two-way product × subject interaction. The goal is to
examine data from subjects with significant scores for this interaction to determine
whether the significance is one of magnitude-scale differences or crossover-scale
differences compared with the panel. Converting scores to ranks and performing a
rank analysis will remedy magnitude differences. Crossover differences cannot be
remedied in the same way.
Crossover may relate to genetic or experiential sensory differences or may reflect
a need for additional training. The experienced analyst examines the data to determine
whether a subject’s interaction has seriously skewed the data for an attribute. If the data
are seriously skewed, that subject’s data may be omitted from the analysis for that attri-
bute as a last resort. This action will reduce the number of subjects and observations
for the data, which may consequently result in a loss of product significance. The rec-
ommended statistical calculations will also measure the degree of individual subject
crossover for attributes when the two-way ANOVA is significant. That calculation is
valuable for the final decision about attribute, subject, and product differences.

Reporting of Results
QDA analyses are designed to provide an in-depth understanding of these sensory
similarities and differences among products and include the following:
• One-way ANOVA for each sensory attribute to measure subject consistency and to
identify product differences
• Two-way ANOVA for each sensory attribute to determine whether the panel, as a
whole, scored the products as different from one another

BK-AST-MNL_13-200365-Chp03.indd 66 30/10/20 4:31 PM


QUANTITATIVE DESCRIPTIVE ANALYSIS 67

• Duncan’s new multiple range test to identify statistically significant differences


among products for each sensory attribute
• Spider graphs for each modality or another grouping (appearance, aroma, flavor,
texture, handfeel, and aftertaste; before, during, and after usage; significant
attributes only, etc.) to illustrate product differences
• Principal component analysis (PCA) to identify those attributes or groups of
attributes that may be related and that best differentiate the products
• Means matrixes and attribute correlations to understand the relationship and
potential redundancy of the array of attributes
Data are provided as means tables (table 1) and in spider plot format (figs. 4 and 5).

Practical Considerations
The product developer often needs to make changes quickly to accommodate project
objectives. In general, changes can be readily incorporated into the QDA process.
Several practical considerations and helpful suggestions are discussed below. The
panel moderator’s responsibility is to manage the language, help subjects clarify any
confusing words, and reduce the redundancy of words used in describing the product.
Most projects evolve, and descriptive analysis may be useful at various stages of devel-
opment. To best accommodate the research objectives, amendments to the original
test design or additional testing may be required.

Rapid Methods
If there are resource constraints or if there is less risk in decision making, consider
options for more rapid language development and data collection. Within the QDA
methodology, this is referred to as diagnostic descriptive analysis, or DDA. There
are a couple of approaches that are generally used. The first is to conduct a full QDA
approach and then select a subset of attributes based on various data-reduction
techniques such as PCA or with a more in-depth understanding of the important
sensory attributes that have the greatest impact on consumer behavior and pur-
chase interest. The research team could select key attributes on the basis of their
product knowledge and on previous results. Keep in mind that the subjects should
still have direct input to ensure that the reduced language is well defined and reflects
differences that they have observed.
Another common use of DDA involves screening previously trained subjects
qualified as likers and users (or potential likers and users) for a new product category.
One could screen potential subjects with as few as three to six discrimination tests to
qualify them, and then conduct only two to three language sessions to develop a short-
ened set of around 12 to 15 attributes. It is advised to use the full 12 to 15 subjects and
two replications of the product set may suffice.

BK-AST-MNL_13-200365-Chp03.indd 67 30/10/20 4:31 PM


TABLE 1 QDA means table 68

Means for DA Attributes

p value 0.05
Interaction as error value 0.10
Post hoc method Duncan’s
Brown Color Thickness Foamy

BK-AST-MNL_13-200365-Chp03.indd 68
Product 2 51.08 A Product 14 16.92 A Product 8 60.29 A
Product 9 49.75 AB Product 7 16.79 A Product 16 49.42 B
Product 13 42.29 ABC Product 15 16.12 AB Product 13 48.00 B
Product 12 40.54 BC Product 8 11.50 ABC Product 11 42.00 BC
Product 7 35.08 CD Product 13 11.13 ABC Product 15 41.75 BC
Product 14 33.92 CD Product 16 10.67 ABC Product 3 35.75 CD
Product 15 31.96 CDE Product 2 9.96 ABCD Product 2 34.79 CD
Product 16 28.58 DEF Product 4 8.79 BCD Product 4 34.46 CD
Product 8 27.79 DEF Product 5 7.38 CD Product 10 32.67 CDE
Product 4 24.37 DEFG Product 12 6.67 CD Product 12 32.33 CDE
Product 3 22.50 EFG Product 11 6.21 CD Product 9 30.63 DE
Product 5 20.17 FG Product 10 6.00 CD Product 5 27.50 DEF

Product 10 19.67 FG Product 6 5.12 CD Product 6 27.50 DEF


Descriptive Analysis Testing for Sensory Evaluation: 2nd Edition

Product 11 18.25 FG Product 9 4.67 CD Product 7 24.38 EF


Product 6 14.21 GH Product 3 4.29 CD Product 14 19.96 F
Product 1 8.04 H Product 1 2.79 D Product 1 19.08 F
Chocolate Aroma Dairy/Milky Aroma Malt Aroma
Product 13 44.58 A Product 3 33.71 A Product 13 14.88 A
Product 14 41.08 AB Product 8 33.25 AB Product 16 14.25 A
Product 16 31.50 BC Product 4 29.29 ABC Product 15 13.17 A
Product 12 29.92 CD Product 13 29.21 ABC Product 12 11.88 AB
Product 3 25.62 CDE Product 11 24.13 ABCD Product 14 10.29 ABC
Product 15 23.04 CDEF Product 10 23.50 BCD Product 9 6.88 BCD

30/10/20 4:31 PM
Product 11 19.50 DEFG Product 9 22.25 CDE Product 2 5.83 BCD
Product 9 18.96 DEFG Product 16 22.25 CDE Product 5 4.75 CD
Product 4 18.67 DEFG Product 15 19.67 CDEF Product 7 4.67 CD
Product 8 17.21 EFGH Product 1 17.71 DEFG Product 10 4.17 CD
Product 10 16.92 EFGH Product 6 13.42 EFGH Product 6 3.96 CD
Product 6 11.79 FGHI Product 14 11.50 FGH Product 4 3.92 CD

BK-AST-MNL_13-200365-Chp03.indd 69
Product 1 8.17 GHI Product 12 10.71 FGH Product 8 3.62 D
Product 7 5.87 HI Product 7 8.83 GH Product 3 3.50 D
Product 5 3.50 I Product 2 6.79 H Product 11 3.42 D
Product 2 3.29 I Product 5 6.04 H Product 1 2.92 D
Chocolate Flavor Dairy/Milky Flavor Powdered Milk Flavor
Product 14 44.71 A Product 8 27.25 A Product 15 18.08 A
Product 6 42.50 AB Product 3 26.25 AB Product 12 16.04 AB
Product 3 41.75 AB Product 11 25.38 AB Product 14 15.00 AB
Product 13 40.75 AB Product 2 24.46 ABC Product 13 14.42 AB
Product 5 40.54 AB Product 10 24.08 ABC Product 8 12.12 ABC
Product 11 40.25 AB Product 6 22.58 ABCD Product 9 12.08 ABC
Product 9 39.46 AB Product 13 22.21 ABCD Product 16 11.96 ABC
Product 16 39.42 AB Product 16 21.58 ABCDE Product 2 10.46 ABC
Product 10 39.00 AB Product 5 21.33 ABCDE Product 7 9.17 BC
Product 2 38.25 AB Product 9 20.21 BCDE Product 10 8.96 BC
Product 15 36.46 AB Product 15 19.71 BCDE Product 3 8.75 BC
Product 7 34.96 B Product 12 18.37 CDE Product 11 8.46 BC
Product 12 34.17 B Product 4 18.21 CDE Product 6 7.71 BC
Product 8 33.75 B Product 14 16.42 DE Product 5 5.58 C
Product 4 22.04 C Product 1 16.17 DE Product 1 5.21 C
Product 1 12.08 D Product 7 15.21 E Product 4 5.04 C
Product 9 39.46 AB Product 13 22.21 ABCD Product 16 11.96 ABC
QUANTITATIVE DESCRIPTIVE ANALYSIS

Product 4 22.04 C Product 1 16.17 DE Product 1 5.21 C


Product 1 12.08 D Product 7 15.21 E Product 4 5.04 C
69

30/10/20 4:31 PM
70 Descriptive Analysis Testing for Sensory Evaluation: 2nd Edition

FIG. 4 QDA spider plot—how to read.

These types of DDA programs are more common with nonflagship products such
as the lower-risk second- and third-tier products.

MASTER QDA PANELS


Another approach that is slightly different from traditional QDA is the master QDA
panel approach. The master QDA approach still uses a qualified sensory panel consist-
ing of screened, trained, and experienced panel members. The panel members are
individuals who have demonstrated a commitment to the product category through
work experience or personal hobbies (e.g., chefs, brewmasters, winemakers, cosmetol-
ogists). Although these panelists have a sensitivity to the products being evaluated and
do not represent the naive consumer population, their experience and commitment to
the products being evaluated are in fact an advantage in the language development and
testing protocol development process. Following in the QDA methodology, these mas-
ter QDA panel members are not taught about perception but are used to understand
perception from a more highly skilled viewpoint.4 As with any quality scientific test

BK-AST-MNL_13-200365-Chp03.indd 70 23/11/20 10:40 AM


QUANTITATIVE DESCRIPTIVE ANALYSIS 71

FIG. 5 QDA spider plot—data display.

methodology, repetitions are included, as well as timed rest intervals given between
evaluations, and data are collected in appropriate test environments following best
sensory practices.

Use in Sensory Claims Substantiation and Legal


Disputes
QDA is an important research tool that can be used to substantiate consumer-based
sensory claims or provide evidence of sensory similarities and differences in legal
disputes. Claims such as a smoother shave, softer feel, more flavor, longer lasting,
crunchier, crispier, and other sensory claims require that the perceptions are “con-
sumer-noticeable” and significant, generally at the 95% confidence interval. If the
research followed the QDA methodology, by definition of how the subjects/con-
sumers are recruited and screened and how the sensory language is developed, data
are collected and analyzed; the results are a direct reflection of the naive consumer

BK-AST-MNL_13-200365-Chp03.indd 71 23/11/20 10:41 AM


72 Descriptive Analysis Testing for Sensory Evaluation: 2nd Edition

perception. QDA is used successfully in claims substantiation and legal disputes


regarding sensory properties. For more information, see ASTM E1958, Standard
Guide for Sensory Claim Substantiation,5 which provides industry guidelines for
using sensory research tools to substantiate perceptual claims and is updated
at least every 5 years, and the Practical Guide for Comparative Advertising by
Corbin et al.6

Adaptations
The QDA methodology is readily adapted to any consumer product, as the overar-
ching objective is to describe and measure the product’s sensory experience. Each
product comes with its own set of challenges, and the sensory professional must
decide on testing protocols and methodology on the basis of the research objective.
The method has been adapted for a wide variety of products around the world,
including, but not limited to, food and confections, beverages (including wine and
spirits), tobacco, apparel, footwear, sporting goods, office supplies, writing instru-
ments, personal care, household care, art supplies, furniture, automotives, and pet
foods, among others.
To adapt the QDA methodology, the sensory scientist must understand how
product preparation and testing protocols can be designed to best reflect the end use
of that product. Research design choices may impact the selection of the test location
and tasks associated with language development and data collection. For some cate-
gories, it may be more appropriate to conduct an in-home evaluation or combine it
with a laboratory procedure and an away-from-the-lab or in-home procedure.
Generally, the overall objective is to work with the constraints of the product category
and develop a consumer-based language that best reflects the product’s sensory expe-
rience. Many consumer-usage situations, especially for nonfood products, cannot
fully be replicated in a laboratory environment. The sensory professional must deter-
mine how best to get as close as possible to typical behavior associated with consumer
usage. In this way, the QDA results can be readily correlated with consumer affective
behavior to determine how the products’ sensory experience affects consumer
perception.
For example, with running shoes, subjects could hold and bend the shoes, put
them on in the laboratory or a central location, and then walk around a room or build-
ing and describe their initial product sensory impressions. However, the running
experience cannot be fully understood without a variety of surfaces, distances, and
running times, along with considerations for indoor/outdoor and various weather
conditions. The sensory scientist must make key decisions during the recruitment of
subjects/runners, the distance, frequency, duration, and training level necessary to
meet the project objectives. These types of sensory procedural decisions must be dis-
cussed before initiating the panel to ensure that they meet the business objectives and

BK-AST-MNL_13-200365-Chp03.indd 72 30/10/20 4:31 PM


QUANTITATIVE DESCRIPTIVE ANALYSIS 73

that the results reflect the intended usability of the data. This is just one example of how
the QDA methodology may be adapted for any product of interest.
With widespread usage of direct-data entry and web-based systems, both labora-
tory and out-of-laboratory techniques will increase in usage and application. QDA is a
flexible methodology and is easily adapted to a wide variety of products for which the
laboratory may not be the best location for data collection.

Summary
As stated previously, the QDA methodology has been applied to a wide variety of con-
sumer goods, and the applications are limited only by the creativity of the sensory
science, development, and brand-building teams. The real power of the method comes
in the researcher’s approach to language development. It must start with a curious
mind and a panel of subjects with known sensory acuity who are likers and users of the
product. A key element of the moderator’s role is to understand what to measure and
how to measure it—not to teach the subjects anything but to learn from them. How do
individuals on the panel approach the category to begin with? How do they discuss
and talk about their observations and evaluate and make decisions?
For the sensory scientist, the QDA methodology is a dynamic and flexible system
but requires that the panel moderator make numerous decisions before, during, and
after the language development process. Key decisions must be made when organiz-
ing, recruiting, and screening the panel, observing behavior throughout the language
development process, when reviewing results and when making decisions on the find-
ings. All of these aspects are important and can affect the usefulness of the data. The
sensory scientist must decide on the appropriate experimental design, when and where
products will be evaluated, and, last but not least, analyzing data and reporting key
finding and conclusions. This is not unique to the QDA method but is an important
aspect of reviewing data and drawing appropriate conclusions from results. Without
sufficient knowledge about human behavior and the research/project objectives, incor-
rect product decisions could be reached to the detriment of our sensory science, to the
business, and to the product category.
From a business perspective, small data derived from QDA studies can provide
robust guidance to the development and marketing teams.7 QDA has been an essential
part of brand building and brand protection for well over 40 years, and it is continuing
to find new applications across a host of new categories. Consumer research is in a hard
data-driven era, and companies are looking for additional insight to understand prod-
ucts and how consumers view and interact with them to discover improvement and
new product opportunities. For example, the results can be analyzed with simple and
complex consumer data to understand relationships between sensory attributes and
consumer ratings with single correlations (fig. 6) or with multisegment data (table 2
and fig. 7).

BK-AST-MNL_13-200365-Chp03.indd 73 30/10/20 4:31 PM


74 Descriptive Analysis Testing for Sensory Evaluation: 2nd Edition

FIG. 6 XY graph relating consumer liking to QDA data with a single correlation.

Sensory data that quantifies unbiased consumer perceptions are essential to


understanding one aspect of marketplace behavior. Descriptive data can be used to
track any product changes and are used as a strategic, insightful, and essential tool in
many multinational companies. It has oftentimes been referred to as a secret weapon
to protect the brand by understanding the similarities and differences in the product
relative to key competition, and the greater power is when this small quantitative data
set is mathematically combined with larger-scale consumer assessment data to develop
predictive relationships between sensory parameters and consumer satisfaction.
Designed experiments to understand which areas to highlight in product development
versus which areas to avoid yield valuable information that can help establish corpo-
rate product strategy and priorities for innovation using data driven models of
consumer behavior.

BK-AST-MNL_13-200365-Chp03.indd 74 23/11/20 10:42 AM


QUANTITATIVE DESCRIPTIVE ANALYSIS 75

TABLE 2 C
 orrelation matrix with QDA attributes and CLT data with subgroups (e.g.,
brand user, preference segments)

Total Population Brand User A Brand User B Brand User C

Dairy/milky flavor 0.87 0.79 0.82 0.86


Dairy/milky aroma 0.86 0.77 0.86 0.84
Thickness texture 0.59 0.49 0.53 0.61
Chocolate flavor 0.57 0.43 0.57 0.56
Chocolate aftertaste 0.51 0.32 0.49 0.49
Powdered milk flavor −0.48 −0.54 −0.51 −0.43
Malt flavor −0.57 −0.47 −0.55 −0.59
Bitter taste −0.58 −0.42 −0.55 −0.55
Powdered milk aftertaste −0.66 −0.60 −0.63 −0.67
Chalky texture −0.66 −0.59 −0.71 −0.66
Astringent texture −0.69 −0.58 −0.66 −0.64
Foamy appearance −0.70 −0.52 −0.65 −0.73
Malt aroma −0.70 −0.57 −0.64 −0.71
Alcohol/extract flavor −0.71 −0.50 −0.68 −0.65
Medicinal flavor −0.75 −0.65 −0.77 −0.67
Powdered milk aroma −0.75 −0.77 −0.74 −0.72
Preference Segment 1 Preference Segment 2 Preference Segment 3

Dairy/milky flavor 0.78 −0.71 −0.64


Dairy/milky aroma 0.78 −0.72 −0.68
Thickness texture −0.71 0.71 0.77
Chocolate flavor −0.16 0.28 0.62
Chocolate aftertaste −0.24 0.40 0.66
Powdered milk flavor −0.57 0.61 0.39
Malt flavor −0.78 0.62 0.67
Bitter taste −0.04 −0.49 −0.25
Powdered milk aftertaste −0.51 0.71 0.08
Chalky texture −0.71 0.68 0.72
Astringent texture 0.08 0.61 0.64
Foamy appearance −0.67 0.38 0.61
Malt aroma −0.52 0.58 0.58
Alcohol/extract flavor −0.04 −0.49 −0.25
Medicinal flavor 0.15 −0.61 −0.32
Powdered milk aroma −0.59 0.70 0.05

BK-AST-MNL_13-200365-Chp03.indd 75 30/10/20 4:31 PM


76 Descriptive Analysis Testing for Sensory Evaluation: 2nd Edition

FIG. 7 QDA preference segment plot for key driver attributes.

References
1. H. Stone, J. Sidel, S. Oliver, A. Woolsey, and R. C. Singleton, “Sensory Evaluation by
Quantitative Descriptive Analysis,” Food Technology 28, no. 11 (1974): 23–34.
2. H. Stone, R. N. Bleibaum, and H. A. Thomas, Sensory Evaluation Practices, 4th ed. (San
Diego, CA: Academic Press, 2012).
3. J. L. Sidel, R. N. Bleibaum, and K. W. C. Tao, “Quantitative Descriptive Analysis (QDA),” in
Descriptive Analysis in Sensory Evaluation, ed. S. E. Kemp, J. Hort, and T. Hollowood
(London: Wiley-Blackwell, 2018), 287–318.
4. R. N. Bleibaum, M. J. Kern, and H. Thomas, “Contextual Product Testing for Small to
Medium Sized Enterprises (SMEs),” in Context: The Effects of Environment on Product
Design and Evaluation, ed. H. Meiselman (Duxford, UK: Woodhead Publishing, 2019),
501–520.

5. Standard Guide for Sensory Claim Substantiation, ASTM E1958-16a (West Conshohocken,
PA: ASTM International, approved October 1, 2016), http://doi.org/10.1520/E1958-16A
6. R. M. Corbin, R. N. Bleibaum, T. Jirgal, D. Mallen, and C. A. Van Dongen, Practical Guide to
Comparative Advertising: Dare to Compare (San Diego, CA: Academic Press, 2018).
7. M. Lindstrom, Small Data: The Tiny Clues That Uncover Huge Trends (New York: St. Martin’s
Press, 2016).

BK-AST-MNL_13-200365-Chp03.indd 76 30/10/20 4:31 PM


Chapter 4 | Spectrum Descriptive
Analysis
Gail Civille 1 and Katie Osdoba1

Introduction
The Spectrum™ Descriptive Analysis Method was developed by Gail Vance Civille
in the 1970s. The Spectrum Method expands on the rigorous training and structure of
the Flavor and Texture Profile Methods with the use of a more discriminating scale
(151 points) and statistical techniques to improve data interpretation. The Spectrum
Method has been used for a variety of consumer products beyond the world of food
and beverages, including personal care,1,2 fabrics, and paper.3

Objective Setting
Spectrum panel results allow for documentation of a product’s sensory characteristics
to guide product ideation, product maintenance, product development, quality con-
trol/assurance, and advertising claims. The rationale for the Spectrum Method is to
have technical terms and a universal scale that provide product guidance to product
developers at the bench and quality control engineers at the plant. The data are meant
to be used like other technical data to make decisions and gather critical product infor-
mation that is objective and consistent.

Background and Core Philosophy


The Spectrum Descriptive Analysis Method uses technical-based language for termi-
nology and intensity based on internal and published reference scales to define the
attributes and the intensity boundaries. It was developed for the purpose of providing
standardized, universal sensory profiles that completely and accurately describe the
sensory attributes of consumer products. Three foundational elements define the Spec-
trum Method: the use of highly trained panelists, the use of clear qualitative references
for attributes, and a universal intensity scale for quantitative ratings.

1
Sensory Spectrum Inc., 554 Central Ave., New Providence, NJ 07974, USA
G. C. https://orcid.org/0000-0002-5007-7546 , K. O. https://orcid.org/0000-0003-2298-535X
DOI: 10.1520/MNL1320150029
Copyright © 2020 by ASTM International, 100 Barr Harbor Dr., PO Box C700, West Conshohocken, PA 19428-2959
ASTM International is not responsible, as a body, for the statements and opinions expressed in this chapter.
ASTM International does not endorse any products represented in this chapter.

BK-AST-MNL_13-200365-Chp04.indd 77 30/10/20 5:04 PM


78 Descriptive Analysis Testing for Sensory Evaluation: 2nd Edition

With the universal scale, all attributes in all samples are rated using the same
scale, and samples may be compared over time, across panels, and across sessions.
Products can also be compared across categories because all attributes are rated rela-
tive to the universal scale and not to each other or within a particular category.
A key challenge in sensory science is overcoming the disconnect between how
people express what they perceive (usually consumers) compared with what is actually
perceived (panels). By training panelists to recognize sensory attributes and rate them
on a universal intensity scale, the Spectrum Method helps to bridge this gap. The cali-
brated data obtained from Spectrum panels can be used with the same level of confi-
dence that a researcher or product developer would expect from instrumental data.

Overview of Method
The Spectrum Method uses highly trained panelists to descriptively profile the sensory
attributes of consumer products. Depending on the product category, these profiles
may include appearance, aroma/fragrance, flavor (including aromatics, basic tastes,
and chemical feeling factors), texture, fabric/paper handfeel, skinfeel, and product
sound. The panelists are calibrated to the sensory attributes in each of these modalities
using qualitative and quantitative references and a universal intensity scale.
For ease of reading, many topics discussed in this chapter will use food flavor and
texture as examples. The beauty of the Spectrum Method, however, is that it can be uni-
versally used to evaluate the sensory properties of virtually any consumer product,
including fragrance, personal care, hair care, paper, fabric, and home care products. For
example, the principles that apply to flavor evaluation also apply to fragrance evaluation.

QUALITATIVE REFERENCES
Qualitative references are used to clarify the definition of each sensory attribute and to
ensure that all panelists have learned and internalized these definitions. A qualitative
reference has four features: a name that is specific to the attribute (e.g., distilled lemon), a
physical reference that demonstrates the attribute (e.g., lemon oil), examples of products
with the attributes (e.g., lemon-lime beverage, lemon candy), and a definition to explain
the attribute (e.g., the aromatics associated with distilled lemon oil).4 References can
include chemical odorants/flavorants (e.g., benzaldehyde), simple ingredients (e.g., white
vinegar), and controlled processes (e.g., coffee at different brew times or dilutions). Refer-
ences are also helpful for teasing apart nuanced differences in sensory character. For
example, the use of references can help separate the smoky flavor of bacon into ash, phe-
nolic, and wood smoke (these attributes would be grouped under “Smoke Complex”).
The more experienced the panelists, the more they can detect and describe these nuances.
Qualitative references can also be demonstrated by examples. Examples are pre-
dominating in one attribute but are not singular examples of that attribute (references,
on the other hand, are singular). Sweetened condensed milk would be an example of
the attribute “caramelized,” while caramelized sugar would be a reference. Examples
can help panelists see attributes in a broader context and more complex environment.

BK-AST-MNL_13-200365-Chp04.indd 78 30/10/20 5:04 PM


SPECTRUM DESCRIPTIVE ANALYSIS 79

Several examples demonstrating a single attribute may be shown to panelists so they


learn how the attribute manifests itself in various contexts.

THE UNIVERSAL SCALE


In the Spectrum Method, all sensory attributes are rated on the universal scale for
intensity. The boundaries of the universal scale cover the full range of product space
across hundreds of products (or “universe”). The use of this scale allows attributes to be
compared across products, product categories, panels, and time. The scale ranges from
0 to 15, with tenths allowed, for a total of 151 points of discrimination (the original
paper and pencil line scales were 15 cm in length, and thus the 15-point scale was
born). The universal scale is based on absolute zero intensity and ratio relationships
between points (a 10 rating is twice as intense as a 5 rating). Because of this use of ratio
scaling, the scale can be expanded beyond 15 points to capture extreme sensations
(e.g., the sourness of pure lemon juice or the aroma intensity of pure mint oil). The
Spectrum Method for personal care products uses a 0–100 point intensity scale.

QUANTITATIVE REFERENCES
A quantitative reference demonstrates an attribute at a specific intensity (e.g., Sour 2,
Sour 5, Sour 10, Sour 15). For a given attribute scale, references should be available for
several points on the scale. Well-chosen references help to reduce panel variability,
ensuring that data can be compared across time and products. Table 1 shows the Spec-
trum universal scale for aromatics, with commonly used quantitative references. Spec-
trum intensity references for basic tastes, texture, skinfeel, and handfeel attributes
have been published elsewhere.5 These references have been generated collectively over
several replicates and several panels.

Lexicon Development
The Spectrum lexicon development process has been covered extensively in other pub-
lications.5–7 Fig. 1 outlines the five primary steps in lexicon development (by a trained
panel). A shortened lexicon development procedure is demonstated for the flavor and
texture of macaroni and cheese in the sections that follow.

EXAMPLE: MACARONI AND CHEESE FLAVOR


1. Frame of reference: The process begins by selecting a broad range of samples that
cover the entire product sensory space. It is absolutely crucial to have the products

TABLE 1 Universal scale for aromatics

Intensity 2.0 5.0 7.5 10.0 12.0


Reference Vegetable Cooked apple Orange note in Grape note in Cinnamon
note in note in Mott’s Minute Maid frozen Welch’s note in Big
vegetable oil applesauce orange juice from Concord grape Red gum
concentrate juice

BK-AST-MNL_13-200365-Chp04.indd 79 30/10/20 5:04 PM


80 Descriptive Analysis Testing for Sensory Evaluation: 2nd Edition

FIG. 1 The five steps of lexicon development.

that cover the full frame of reference available when developing a new lexicon. Ulti-
mately, a set of 6 to 12 representative samples are chosen. Four samples were chosen
for the macaroni and cheese demonstration.
2. Term generation: The panel convenes to evaluate the frame of reference samples and
generate a list of terms. The panelists taste the samples and describe them in their
own words. Together, the panel discusses the samples and puts together an exhaus-
tive list of sensory attributes (possibly grouped into categories or complexes).
3. Use of references: A reference is found for each of the attributes listed in Step 2. Pan-
elists review the references and come to agreement on the appropriate definition for
each attribute. Macaroni and cheese references are shown in table 2.

TABLE 2 References used in the lexicon development for macaroni and cheese

Raw white wheat Diacetyl (odorant)


Cooked white wheat Butyric acid (odorant)
Cooked milk Isovaleric acid (odorant)
Nonfat dry milk Propionic acid (odorant)
Egg Lactic acid

BK-AST-MNL_13-200365-Chp04.indd 80 30/10/20 5:04 PM


SPECTRUM DESCRIPTIVE ANALYSIS 81

4. Examples: Examples that encompass various sensory attributes are shown to the
panel. An evaluation of these examples allows the panelists to see the attributes in
context to help them further refine the lexicon. Macaroni and cheese examples are
shown in table 3.
5. Lexicon refinement: The final step in lexicon development involves refining the
attribute list by eliminating redundant and integrated, consumer-type terms (e.g.,
“creamy”). The result is a working lexicon that is ready for validation. It is important to
remember that a lexicon can always be updated, changed, or modified. As new samples
are introduced, so may be new attributes. The “finished” macaroni and cheese flavor
lexicon is listed in table 4. After a new lexicon is developed, and before it is used for
product evaluations, it must be validated. Typically, two samples are evaluated using the
lexicon to demonstrate its ability to differentiate and profile products in the category.

EXAMPLE: MACARONI AND CHEESE TEXTURE


The development of a lexicon for texture evaluation follows a slightly different proce-
dure than that for flavor. The Spectrum Method uses a defined set of texture attributes
(based on the research and publications of Szczesniak et al.8,9 using rheological princi-
ples) that can be used across product categories. The texture attributes that are relevant
to the category in question (i.e., macaroni and cheese) are used in the working lexicon.

TABLE 3 Examples used in macaroni and cheese lexicon development sessions

Example Reference(s) Shown in Example

Cheddar cheese Butyric acid, cooked milk, bite/sharp


Swiss cheese Propionic acid, cooked milk, fruity, bitter
Asiago cheese Isovaleric acid, cooked milk, nasal pungency
American cheese Butyric acid, cooked milk, caramelized, soured dairy/cheesy

TABLE 4 Macaroni and cheese flavor lexicon

Chemical Feeling
Aromatics Aromatics (continued) Basic Tastes Factors

Total aromatics Dairy complex Sweet Bite/sharp


Wheat complex Milky Sour
Raw white wheat Non-fat dry milk Salt
Cooked white wheat Soured dairy/cheesy Bitter
Cheese acids complex Butter impression
Butyric acid Butterfat
Propionic acid Diacetyl
Caramelized Eggy

BK-AST-MNL_13-200365-Chp04.indd 81 30/10/20 5:04 PM


82 Descriptive Analysis Testing for Sensory Evaluation: 2nd Edition

New texture attributes occasionally arise during evaluation. New scales (with appro-
priate references) can be created to capture these new attributes.
In addition to attribute definitions, the techniques used to evaluate texture attri-
butes must be clearly defined. Texture evaluation techniques include identifying the
stage at which a particular attribute is experienced (i.e., first bite, chewdown, residual),
the amount of sample required for evaluation (i.e., an average bite), and any additional
manipulations needed (i.e., move the tongue around the outside of the mass). Table 5
shows an example of a texture attribute (hardness), its definition and technique, and its
references along the universal intensity scale. Table 6 shows the texture lexicon for
macaroni and cheese.

TABLE 5 Standard hardness scale

Scale Value Reference Brand/Type/Manufacturer Sample Size

1.0 Cream cheese Philadelphia Cream Cheese—Kraft Foods ½ in. cube


4.5 Cheese Yellow American pasteurized processed deli ½ in. cube
cheese—Land O’Lakes
7.0 Frankfurter Hebrew National, beef, large, cooked 5 min ½ in. slice
9.0 Peanuts Planters cocktail-type in vacuum tin 1 nut, whole
11.0 Almonds Planters or Blue Diamond, shelled 1 nut
14.5 Hard candy Lifesavers 3 pieces, 1 color
Technique: For solids, place food between the molars and bite down evenly, evaluating the force
required to compress the food. For semisolids, measure hardness by compressing the
food against the palate with the tongue once. When possible, the height for hardness
standards is ½ in.
Definition: The force to attain a given deformation, such as:
• Force to compress between molars, as above
• Force to compress between tongue and palate
• Force to bite through with incisors
[low force/soft ————————————————————————- high force/hard]

TABLE 6 Texture lexicon for macaroni and cheese

First Bite Chewdown Chewdown (continued) Residual

Hardness Mixes with saliva (sauce) Chalkiness of the matrix Toothpack/toothstick


Cohesiveness Cohesiveness of mass Graininess of the matrix Mouth coating (type)
Denseness Moistness of mass Stickiness in mouth/teeth/gums Loose particles
Uniformity of Bite Roughness of mass Rate of breakdown Mouth drying
Awareness of sauce

BK-AST-MNL_13-200365-Chp04.indd 82 30/10/20 5:04 PM


SPECTRUM DESCRIPTIVE ANALYSIS 83

EXAMPLE: PERSONAL CARE PRODUCTS


Standard lexicons and evaluation protocols are typically used for personal care prod-
ucts. For example, a standard lexicon would be used for a variety of hand and body
lotion projects, while a different lexicon would be used for color cosmetics. Additional
attributes are added as needed for project-specific objectives. Again, the products
themselves define what the lexicon will be. Like food texture, attributes are defined not
only by the actual feel of the products but also by attribute-specific evaluation tech-
niques. In the Spectrum Method, personal care products are evaluated on a 0–100
intensity scale. (This scale behaves in the same way as the flavor/texture universal scale
discussed earlier.)
Table 7 shows a typical evaluation protocol and lexicon for lotions and creams.
More detailed information can be found in ASTM E1490-2019, Standard Guide for
Two Sensory Descriptive Analysis Approaches for Skin Creams and Lotions.10

RECRUITING
Prospective panelists go through several stages of screening and qualification. Panel-
ists may be recruited internally or externally. Internally recruited panelists are conve-
nient because they are already located in the building and may have some knowledge
of the sensory evaluation process. However, people already employed by the company
must carve time out of their current work schedule to attend trainings and practice
sessions. Additionally, their prior knowledge of the company’s products may introduce
bias in evaluations. Externally recruited panelists may be more reliable and may have
more time to devote to the panel. Because they are more naive in regard to sensory
evaluation, however, they may require more extensive training. Also, because they will
likely be working only part time as panelists, their availability to come in for evalua-
tions on short notice may be limited.

PRESCREENING
Typically, 60 to 80 people are prescreened for a Spectrum panel, with the goal of 34 to
45 moving to the acuity screening step and finally 18 to 20 participating in the training
program (fig. 2). The final panel ideally has 12 to 15 members. A primary consideration
is availability for training, practice sessions, and afterward, routine panel work. An
accurate schedule should be available during this time. Potential panelists should be
available for a minimum of 80% of sessions during a specified time frame (typically
2–3 years). Prescreening for a food panel should also serve to eliminate those who
require special diets (medical or otherwise), take medication that affects taste percep-
tion, have food allergies, or have dentures/dental work that may affect texture percep-
tion. Potential skinfeel panelists must not be sensitive or allergic to ingredients
commonly found in fragrances, lotions, etc. Potential panelists should also express
interest in learning more about sensory evaluation of the product category. Prescreen-
ing is typically completed via online questionnaire, although it may also be completed
over the phone or in person.

BK-AST-MNL_13-200365-Chp04.indd 83 30/10/20 5:04 PM


84 Descriptive Analysis Testing for Sensory Evaluation: 2nd Edition

TABLE 7 Standard evaluation protocol and lexicon for lotions and creams

Product Appearance
In a petri dish, panel leader dispenses the product in a spiral shape using a nickel size circle, filling it in
from the edge to the center. Evaluate for:
Attribute Definition Qualitative Reference

Integrity of Shape Degree product holds shape Petrolatum


[flattens—retains shape]
Integrity of Shape Degree product holds shape Petrolatum
(10 seconds) [flattens—retains shape]
Tilt petri dish to see reflective properties and evaluate for:
Gloss Amount of reflected light from product Baby oil or light mineral
[dull/flat—shiny/glossy] oil

Product Pickup
Using automatic pipette, panel leader delivers 0.1 cc of product to the tip of the thumb or index finger.
Compress product slowly between index finger and thumb one time and then separate fingers. Evaluate
for:
Firmness Force required to fully compress product Petrolatum
between thumb and index finger
[no force—high force]
Stickiness Force required to separate finger tips Petrolatum
[no force/not sticky—high force/very sticky]
Compress and separate product between the index finger and thumb three times and evaluate for:
Cohesiveness Amount sample strings rather than breaks Petrolatum
when fingers are separated
[no strings—high strings]
Amount of Peaking Degree to which product makes stiff peaks on Petrolatum
finger tips
[flat/no peaks—stiff peaks]
Rub-Out
Using automatic pipette, panel leader delivers 0.05 cc of product to center of 5-cm-diameter circle on
volar forearm. Spread the measured amount of product within the circle using index or middle finger,
using a gentle circular motion—stroke at a rate of two strokes per second. After three rubs, evaluate for:
Wetness Amount of water perceived while rubbing Water
[none—high amount]
Spreadability Ease of moving product over the skin Baby oil or light mineral
[difficult/drag—easy/slip] oil

After 12 rubs, evaluate for:


Thickness Amount of product felt between fingertip and Petrolatum
skin
[thin/almost no product—thick/lots of product]
After 15–20 rubs, evaluate for:
Oil Amount of oil perceived in the product during Baby oil or light mineral
rub out oil
[none—extreme]

(continued)

BK-AST-MNL_13-200365-Chp04.indd 84 30/10/20 5:04 PM


SPECTRUM DESCRIPTIVE ANALYSIS 85

TABLE 7 Standard evaluation protocol and lexicon for lotions and creams (continued)

Wax Amount of wax perceived in the product during Surface of wax taper
rub out candle; cheese wax
[none—extreme]
Grease Amount of grease perceived in the product Petrolatum
during rub out
[none—extreme]
Continue rubbing and evaluate for:
Rubs to Absorbency The number of rubs at which the product loses Vaseline Total Moisture
wet, moist feel and a resistance to continue is Body Lotion; light
perceived mineral oil
[upper limit = 120 rubs]
Afterfeel (immediate—can be repeated at additional time points)
Visually analyze the forearm test site and evaluate for:
Gloss Amount or degree of light reflected off skin Baby oil or light mineral
[dull/matte—shiny] oil

Tap cleansed finger lightly over application site and evaluate for:
Stickiness Degree to which fingers adhere to residual Petrolatum
product
[not sticky—very sticky]
Stroke cleansed fingers (1–2 strokes) lightly across skin and evaluate for:
Slipperiness Ease of moving fingers across skin Untreated volar forearm
[difficult/drag—easy/slip] skin; light mineral oil

Thickness of Residue Amount of product residue felt between fingers Petrolatum


and skin
[thin—thick]
Amount of Residue Amount of product on skin Baby oil or light mineral
[none—large amount] oil

Type of Residue Oily, waxy, greasy, silicone, plastic/coated,


powdery (may be rated for intensity, for
proportion of signal, or for presence)

ACUITY SCREENING
Potential panelists who successfully complete prescreening are invited to participate in
a two-part screening session—sensory acuity testing and a personal interview. Mock
panel sessions may also be part of the screening process.

Sensory acuity testing


The goal of sensory acuity testing is to judge the potential panelist’s sensitivity to prod-
uct differences, their ability to descriptively describe sensory attributes, and their
understanding of scaling. Specific screening tests will vary depending on type of panel.
For food panels, tests include basic taste perception, identification of flavors/aromatics,

BK-AST-MNL_13-200365-Chp04.indd 85 30/10/20 5:04 PM


86 Descriptive Analysis Testing for Sensory Evaluation: 2nd Edition

FIG. 2 Number of potential panelists during each stage of recruitment.

and ability to use scales and references. Texture, skinfeel, and handfeel acuity can be
assessed by asking potential panelists to rank or rate the references for specific attri-
bute scales.10,11

Personal interview
The interview (by the panel leader or panel trainer) serves to gauge the potential pan-
elist’s interest in and commitment to the job, their ability to work in a team setting,
their general ability to learn the skills needed, and their ability to follow instructions.

Mock panel session


Panel leaders engage the panelist candidates in a mock panel session to observe the
group dynamic and gauge how well panelists will work together. Experienced sensory
professionals may also be present, playing the role of panel members to help make
these mock sessions more realistic.

Qualifications of Panel Leader


Spectrum panel leaders are highly involved in the development and maintenance of
their panels. Their duties include setting panel guidelines, leading group discussions,
clarifying attribute definitions, and ensuring that test controls, references, and sam-
ples are in place. They are also responsible for monitoring panel performance (both
individually and as a group), thereby maintaining panel integrity and ensuring
high-quality data output.

BK-AST-MNL_13-200365-Chp04.indd 86 30/10/20 5:04 PM


SPECTRUM DESCRIPTIVE ANALYSIS 87

Spectrum panel leaders have ideally been previously trained as panelists and so
are familiar with the methodology and have considerable skills of their own. Alterna-
tively, new panel leaders may observe an existing panel for several months. Spectrum
panel leaders should also have the demonstrated ability to manage, teach, and commu-
nicate within a group. An additional asset is a basic understanding of statistical analy-
sis, which is helpful for monitoring panel data and performance.

PANEL FUNCTION
The primary day-to-day function of the panel leader is to lead the panel discussions
during product evaluation. Spectrum evaluation sessions are very lively and everyone is
an active participant, especially when data are collected by consensus. The panel leader’s
job is to manage these group discussions to ensure that the panelists work together as a
team to come to a uniform, agreed upon intensity rating for each sensory attribute. The
panel leader’s guidance should be free from bias, giving equal voice to each panelist and
refraining from excessively positive or negative judgments of panelists’ comments. All
efforts should be made to avoid influencing the outcome of the evaluations.

PANEL LIFE
The panel leader facilitates the practical as well as cultural aspects of day-to-day panel
functioning. The practicalities include managing work schedules, pay, time off, and
individual performance monitoring. Culturally, the panel leader is responsible for
maintaining panel morale. The panel leader should be positive, prompt, professional,
and well prepared for each day’s session. The panel leader should also actively partici-
pate in all evaluation sessions.

Panel Training
LENGTH OF TRAINING
Spectrum panelists complete a minimum of 100 hours of training for each sensory
modality (flavor, texture, etc.) before beginning to evaluate products. Panels may be
trained to evaluate one or all modalities and one type or a variety of products. The 3- to
4-month training period includes orientation weeks, led by an experienced trainer,
interspersed with practice weeks, led by the panel leader (table 8).

TABLE 8 Training timeline (for each modality)

Training Stage Weeks Hours/Week Hours/Stage

Orientation 1 1 24 24
Practice 5–7 4–6 26–30
Orientation 2 1 24 24
Practice 5–7 4–6 26–30
Total 12–16 4–24 100+

BK-AST-MNL_13-200365-Chp04.indd 87 30/10/20 5:04 PM


88 Descriptive Analysis Testing for Sensory Evaluation: 2nd Edition

TRAINING OVERVIEW
The first orientation week covers basic sensory principles and Spectrum descriptive
analysis methodology. For food panels, these concepts include basic tastes, aromatics,
chemical feeling factors, texture, and appearance. The panelists also learn about objec-
tive versus subjective evaluation and the principles of scaling to help orient them to the
type of data they will be providing. Panelists start to develop their evaluation skills by
developing lexicons for a range of product categories using the standard lexicon devel-
opment procedure. References are used extensively throughout training. During lexi-
con development qualitative references are used to clarify terms. The panel trainer
then introduces the universal scale and the use of intensity references. In this way,
panelists learn the procedure while developing their evaluation skills and familiarity
with those products. Panelists also gain practice working as a group to define the
terms, evaluation procedures, and references needed so the entire panel works as
a unit.
Subsequent orientation weeks are tailored to the needs of the panel’s future work
(e.g., a wine panel will be trained on grapes and dark fruit). Practice sessions allow
panelists to hone their skills and gain exposure to more and more product categories
as well as attributes and their references.
During training, the role of the trainer or panel leader is to coach the panelists
through the exercises. Guidance will likely include additional references to clarify
terms and intensities, assistance in assigning terms when only integrated or consumer
terms are offered by the new panelists, and clarification around terms where two or
more panelists may be assigning different terms to the same attribute. For example,
Panelist 1 perceives orange, Panelist 2 perceives ginger, and Panelist 3 perceives cilan-
tro. All of these have a citrus component or character associated with them. It is the
panel leader’s responsibility to show this connection to the panel and work to assign
which citrus term they can all agree upon. Panel leaders also often point out instances
in which panelists’ newly learned sensory skills can be applied in other situations (to
other categories/modalities). For example, a woody note observed in coffee may also
appear in chocolate. Spectrum panel leaders do not provide the lexicon for the panel
but encourage exploration of nuances. If panelists describe an Oreo as vanilla, they get
to see several references for vanilla and vanillin and learn by evaluating that the cor-
rect descriptor for the sweet aromatic in an Oreo is vanillin.

TEXTURE
Texture training involves orienting panelists to the fundamental optical, rheological,
and physiological concepts of texture evaluation. They learn how to assess geometric
and mechanical properties of products, moistness-related attributes, and the sounds
products make during manipulation. Panelists then learn specific texture attribute
scales and the corresponding references and evaluation techniques. Panelists also
develop their skills by practicing texture evaluations of various products.

BK-AST-MNL_13-200365-Chp04.indd 88 30/10/20 5:04 PM


SPECTRUM DESCRIPTIVE ANALYSIS 89

PERSONAL CARE PRODUCTS


Panelists are first oriented to the different aspects of personal care product evaluation,
including pickup, appearance, rubout, and afterfeel and the attributes that appear in
each. Panelists learn the evaluation procedure and definition for each attribute and are
shown several references that span the intensity range (0–100). Panelists then practice
evaluating each attribute with real products. Panelists may also learn “reference prod-
ucts,” for which full profiles are known. Lotion is often used as a core platform/train-
ing tool because it exhibits a broad range of intensities of many different attributes, it is
familiar to panelists, and it is easily removed from the skin. Training may also include
exposure to a more diverse array of products, including gels, shampoos, and
cosmetics.

Panel Validation
Validation is an important step in ensuring the panel is acting as a calibrated mea-
surement instrument. Validation first occurs at the end of training and then annu-
ally or semiannually to monitor panel (group) and panelist (individual)
performance. An important part of validation is choosing appropriate decision
criteria and identifying what course of action will be taken for each outcome. Deci-
sion criteria may vary depending on product, modality, and the level of training of
the panel. Panelists who fail validation are given probationary status and undergo
retraining. Panelists may return to regular panel evaluations after the successful
completion of validation.

VALIDATION PROCEDURE
Validation assesses the precision, accuracy, and sensitivity of both the individual pan-
elists and the panel as a whole. These three areas can be addressed simultaneously by
implementing a well-designed validation study.
Panelists should be validated on products and attributes that are likely to occur in
their day-to-day evaluations. For example, a panel that evaluates potato chips should
be validated on not only potato and oil aromatics but also common off notes, such as
painty and cardboard. Four to six samples are typically used for validation. The sample
set as a whole should showcase differences (varying from small to large) in all or most
sensory attributes. Samples for which known product profiles exist (blind controls)
should also be included. Panelists should evaluate two to three replications of each
validation sample, in random order, each with a different three-digit blinding code.

VALIDATION OF INDIVIDUAL PANELISTS


The validation of individual panelists ensures that panelists are providing reproducible
and accurate data within themselves and that individually their data align with the
panel as a whole.

BK-AST-MNL_13-200365-Chp04.indd 89 30/10/20 5:04 PM


90 Descriptive Analysis Testing for Sensory Evaluation: 2nd Edition

Precision. Precision refers to the ability of panelists to repeatedly rate the same sample
in the same way. For individual panelists, standard deviation (across replicates of
the same sample) for one or more attributes serves as a good measure of precision.
Standard deviations greater than a specified decision criterion (dependent on the
product, modality, and panelist’s level of training) indicate that a panelist is not
sufficiently replicating his or her ratings for that attribute.
Accuracy. Accuracy is the degree to which a panelist’s ratings agree with a known sam-
ple profile (control). By-attribute accuracy may be assessed by subtracting the
expected rating for the control from the panelist’s average rating across replicates of
blind controls and converting the difference to a percentage of the total scale length.
(Alternatively, if a sample with a known profile is unavailable, panelists’ data may
be compared to the panel mean or median data.) Above a certain percentage (deci-
sion criterion), a panelist is considered insufficiently accurate.
Sensitivity. Sensitivity is the degree to which panelists can detect small differences,
measured as panelists’ ability to discriminate among samples. To determine dis-
crimination ability, each panelist’s data are subjected to an analysis of variance
(ANOVA). A p value for the sample effect that is less than a specified cutoff (deci-
sion criterion) indicates that the panelist is successfully discriminating among
samples for that attribute.
Overall precision, accuracy, and sensitivity (across attributes) may be examined by
calculating the frequency with which panelists “pass” based on the decision criteria.
Panelists’ abilities to both discriminate (sensitivity) and replicate (precision) can
be assessed visually by plotting their mean standard deviation versus their mean
p value (fig. 3). Panelists with low p values and low standard deviations have high sen-
sitivity and precision and are performing better than those with high p values and high
standard deviations.

Validation of Whole Panel


The validation of the panel as a whole ensures that the panel is providing accurate and
precise data (i.e., that the panel can be used as a calibrated measurement tool). Full-
panel assessments can be done with or without the data from panelists who did not
meet the individual validation criteria.
Precision and accuracy. Panel precision is evaluated similarly to individual panelist
precision but uses mean attribute ratings across panelists. Again, the test diagnos-
tics (standard deviation for precision and bias for accuracy) across replicates are
calculated and then averaged across samples.
Sensitivity. Panel sensitivity is evaluated by performing an ANOVA for each attribute
on the data from the full panel. A significant sample effect indicates that the panel
is adequately differentiating among samples.

BK-AST-MNL_13-200365-Chp04.indd 90 30/10/20 5:04 PM


SPECTRUM DESCRIPTIVE ANALYSIS 91

 lot of consistency (mean standard deviation) versus discriminability (mean p


FIG. 3 P
value) for visualization of panelist performance.

Examining the panelist × sample and panelist × replication interactions in an ANOVA


can also be helpful for monitoring panel performance. A plot of the panelist × sample
interaction shows how panelists’ ratings visually align with each other. If the panelists
are rating the same samples similarly, the effect will not be significant. The panelist ×
replication interaction is an indicator of panelist precision—if it is not significant, the
panelists’ ratings are more precise.

Panel Motivation and Recognition


Because humans are the measurement instruments of the Spectrum (and other) descrip-
tive analysis methods, a certain level of humanity is required for their maintenance and
motivation. Like analytical instruments, however, Spectrum panels must also be care-
fully calibrated and validated. All of these considerations may be addressed by investing
in a good panel leader. Routine feedback is key to keeping panelists motivated. Periodic
reviews and training exercises introduce a well-deserved break in routine panel work
while reinforcing skills at the same time. Panelists want to know that the work they are
doing is meaningful and contributes to real progress and concrete results. Project man-
agers should share the results of projects with the panel and help them understand the
contribution they made. A recognition or reward system can also be motivating. Short-
and long-term rewards such as certificates, small gifts, and occasional parties are easy

BK-AST-MNL_13-200365-Chp04.indd 91 30/10/20 5:04 PM


92 Descriptive Analysis Testing for Sensory Evaluation: 2nd Edition

ways to show the panelists that their work is valued. A descriptive panel should also
operate with a sense of fun. Although evaluations can be stressful at times, it is import-
ant to remember that it is not brain surgery. No one’s life hangs in the balance.

Testing Protocol/Protocol Development


During sample evaluations the panel should have clear protocols for sampling and
evaluating so that error and confusion is minimized. Evaluation protocols need to
consider all levels of test controls such as guidelines for the panelists, sample prepa-
ration, sample evaluation, reference review (if needed), and room environment. The
panel leader makes sure that panelists understand the instructions around all of
these guidelines and that they are followed. Total sample volume and evaluation
procedures should be designed and documented prior to evaluation sessions. It is
important that enough product be provided for panelists to accurately evaluate and
rate the entire lexicon. The volume of sample needed for evaluations may need to be
increased depending on the complexity of the sample (e.g., needs to be evaluated at
a set temperature or time point) and the length of the lexicon (such as a longer lex-
icon with both flavor and texture attributes). In food evaluations it is sometimes
helpful to have a separate setup for appearance so products can be compared side by
side. Personal care product evaluation needs three separate applications to cover
evaluation procedures—once for rub out and afterfeel and twice for appearance and
pickup. Sample evaluation replication may also be considered depending on project
design. Elements to consider for presentation design are the number of replications
(dependent on sample volume and time), randomized as a total sample set or by and
within sessions.

Data Collection
Spectrum descriptive analysis data may be collected individually or by consensus, and
there are benefits to each. When a study requires statistical analysis, individual data are
best. In this case, 6 to 12 panelists and 2 or more replications are required. Consensus
data are used when directional or nuanced information is needed. The process of con-
sensus data collection allows the panel to discuss attributes and intensities while the
products are being evaluated. The discussions that occur during evaluation ensure that
panelists are carving attributes similarly and rating them the same way. For example, a
product may have a browned dairy note that some panelists call “cooked milk” and
some call “caramelized.” Through these discussions, the panel as a whole can decide in
which attribute to put that particular flavor note. Attributes may also be added to the
lexicon during panel discussions. Consensus data are not simply an average of individ-
ual panelists’ data; it is an intensity rating that is agreed upon by all panelists. Statistical
analysis can be done with consensus data only when multiple replications are
completed.

BK-AST-MNL_13-200365-Chp04.indd 92 30/10/20 5:04 PM


SPECTRUM DESCRIPTIVE ANALYSIS 93

Analysis and Reporting of Results


DATA ANALYSIS
Data are clearly and easily summarized by calculating means and standard deviations.
A common analysis procedure for descriptive data is ANOVA followed by post hoc
multiple comparison tests. ANOVAs indicate whether statistically significant differ-
ences exist among samples for each attribute. The post hoc tests indicate which sam-
ples are statistically different from each other and how large the differences are. Other
types of useful data analysis procedures include nonparametric tests and multivariate
tests such as principal component analysis and cluster analysis. The type of data anal-
ysis used will always depend on specific project objectives, the experimental design,
and the type of data collected.

REPORTING
Descriptive analysis reports must clearly and concisely communicate the results of a
project to all interested parties. These parties may have differing degrees of sensory
knowledge, so results should be reported at varying levels. The use of graphs and visual
displays is highly encouraged. Reports should contain the test objectives, an executive
summary of the results, and key recommendations right up front. This information is
followed by more detailed sections on methodology, sample information, evaluation
procedures, data analysis, and results. Appendices containing additional information
(attribute definitions, prep instructions, etc.) may also be included.

PERCEPTUAL MAPPING
When many products must be compared, perceptual mapping is a useful tool. Percep-
tual mapping aids in summarizing descriptive data according to key attributes that
contribute most to the sensory variability within a specific product category. Spectrum
descriptive data are subjected to principal component analysis to group like attributes
and reduce the data to orthogonal (independent) factors. These factors are called key
sensory dimensions of the perceptual space. The dimensions may be mapped two at a
time to provide a visual representation of the space. Products may be added to the
maps to understand similarities and differences among products along the depicted
sensory dimensions. See fig. 4 for an example of a perceptual map of the orange juice
category. Attributes used in the evaluation are shown in table 9.

Practical Considerations
As a panel becomes more and more experienced, the Spectrum Method may be
adapted to allow for more efficient product evaluations. After a panel has had extensive
training on one product category, the lexicon developed for that category can be used
as a starting point for subsequent evaluations of similar products. Lexicons do not
need to be developed from scratch every time if the panel and project managers have a

BK-AST-MNL_13-200365-Chp04.indd 93 30/10/20 5:04 PM


94 Descriptive Analysis Testing for Sensory Evaluation: 2nd Edition

FIG. 4 Perceptual map of orange juice flavor.

TABLE 9 Orange juice lexicon

Chemical
Aromatics Aromatics Basic Feeling
Appearance Aromatics (continued) (continued) Tastes Factors Texture

Color intensity Orange complex Other citrus Hydrolyzed oil Sweet Astringent Viscosity
Chroma Raw orange Albedo Cardboard/ Sour Burn
oxidized
Opacity Cooked orange Other fruit Vinyl Bitter
Distilled orange Sweet Balance and
oil aromatic blend
Expressed orange Green
oil
Fruity/floral Fermented

BK-AST-MNL_13-200365-Chp04.indd 94 30/10/20 5:04 PM


SPECTRUM DESCRIPTIVE ANALYSIS 95

good idea of what sensory characteristics to expect. Additionally, a highly experienced


panelist can recall the sensory experience of a product simply by looking at its descrip-
tive profile. From there, the panelist can accurately describe how a new product differs
from that previously evaluated product.

Situational Adaptations
Descriptive analysis data generated using the Spectrum Method combines synergisti-
cally with other types of data to provide even greater guidance to researchers than any
of these methods alone.
Degree of difference. A Spectrum descriptive profile can easily be augmented with a
degree of difference (DOD) score. DOD scores indicate how different, perceptually,
two samples are from one another (e.g., how different a test sample is from a con-
trol). DOD is rated on a 10-point scale, with 0 = no difference and 10 = very large
difference. DOD scores can be helpful in both research and development and qual-
ity control applications.
Quality. Spectrum panels can use their highly developed descriptive vocabulary to
qualitatively describe sample differences. Product quality ratings are appropriate
when panelists have extensive experience with the product category. Panelists dis-
cuss and classify descriptive attributes according to their appropriateness for the
category. These attributes may be considered positive, as in expected “on notes,” or
negative, as in unexpected “off notes.” Quality ratings are generally given on a
10-point scale.
Product grouping/sorting/napping. Spectrum panels can use their clearly defined tech-
nical language in product sorting exercises, and the resulting data can be easily
translated to research needs.
Fidelity. Fidelity is an overall rating that takes into account the degree to which the
sample of interest matches an identified target or expectation.
Balance and blend. Balance and blend measures the degree to which the individual
attributes of a product blend together, making it difficult to identify each compo-
nent. In products with high balance and blend, nothing is “sticking out” (i.e., attri-
butes are balanced), and all attributes overlap so closely that pulling the profile
apart into discrete components is very difficult (i.e., attributes are blended).
Product-adapted universal scaling. For panels that primarily evaluate products within
the same or similar product categories, the universal scale may be adapted accord-
ingly. Intensities of specific attributes from a specific product category are rated
using the universal scale as a base and thus may be related back to the overall uni-
versal scale.
Flavor pairings. The breadth and depth of sensory expertise of a well-trained descrip-
tive panel may be leveraged in nontraditional ways. In a flavor pairings session,
creative problem-solving techniques (more commonly used in the consumer

BK-AST-MNL_13-200365-Chp04.indd 95 30/10/20 5:04 PM


96 Descriptive Analysis Testing for Sensory Evaluation: 2nd Edition

co-creation process) are used to find synchronous and asynchronous pairings of a


target ingredient with hundreds of other ingredients alone or in combination.12
Using a trained panel in this way accelerates the process of identifying successful
flavor/fragrance combinations and avoids lengthy trial-and-error cycles of tradi-
tional product development.

Advanced Applications
KEY DRIVERS AND PREDICTED LIKING
Spectrum descriptive data can, in combination with consumer data, help researchers
assess which sensory attributes drive consumer liking of a product category. Data from
perceptual mapping, such as the orange juice map shown in fig. 4, are combined with
consumer acceptance data using multiple linear regression. Statistical models gener-
ated from key-driver studies can also be used to predict liking of additional products
in a category and to generate sensory profiles of “ideal” products within the sensory
space.

SENSORY DRIVERS OF FUNCTIONAL AND EMOTIONAL BENEFITS


Likewise, Spectrum descriptive data may be combined with consumers’ perceptions of
emotional and functional benefits of products. In this case, consumer data, collected
via iterative creative consumer panels and online surveys, can be overlaid on descrip-
tive maps to better understand how sensory attributes cue each type of benefit. This
methodology has been demonstrated in categories such as strawberry yogurt13 and
ready-to-eat cereals14 and has been found to provide valuable information in new and
emerging categories.

References
1. N. O. Schwartz, “Adaptation of the Sensory Texture Profile Method to Skin Care
Products,” Journal of Texture Studies 6 (1975): 33–42.

2. G. V. Civille and C. A. Dus, “Evaluating Tactile Properties of Skincare Products: A


Descriptive Analysis Technique,” Cosmetics & Toiletries 106, no. 5 (1991): 83–88.

3. G. V. Civille and C. A. Dus, “Development of Terminology to Describe the Handfeel


Properties of Paper and Fabrics,” Journal of Sensory Studies 5 (1990): 19–32.

4. G. V. Civille and H. T. Lawless, “The Importance of Language in Describing Perceptions,”


Journal of Sensory Studies 1, no. 3/4 (1986): 203–215.

5. M. C. Meilgaard, G. V. Civille, and B. T. Carr, Sensory Evaluation Techniques (Boca Raton,


FL: CRC Press, 2007).

6. Lexicon for Sensory Evaluation: Aroma, Flavor, Texture, and Appearance, (West
Conshohocken, PA: ASTM International, 2011).

7. L. J. R. Lawless and G. V. Civille, “Developing Lexicons: A Review,” Journal of Sensory


Studies 28 (2013): 270–281.

BK-AST-MNL_13-200365-Chp04.indd 96 30/10/20 5:04 PM


SPECTRUM DESCRIPTIVE ANALYSIS 97

8. A. S. Szczesniak, “Classification of Textural Characteristics,” Journal of Food Science 28,


no. 4 (1963): 385–389.

9. A. S. Szczesniak, M. A. Brandt, and H. H. Friedman, “Development of Standard Rating


Scales for Mechanical Parameters of Texture and Correlation between the Objective and
the Sensory Methods of Texture Evaluation,” Journal of Food Science 28, no. 4 (1963):
397–403.

10. Standard Guide for Two Sensory Descriptive Analysis Approaches for Skin Creams and Lotions,
ASTM E1490-19 (West Conshohocken, PA: ASTM International, approved November 1, 2019),
http://doi.org/10.1520/E1490-19

11. Standard Guide for Descriptive Analysis of Shampoo Performance, ASTM E2082-12
(West Conshohocken, PA: ASTM International, approved October 15, 2012),
http://doi.org/10.1520/E2082-12

12. I. Koelliker, A. Retiveau-Krogmann, G. V. Civille, and S. Gibbons, “Flavor Synchrony—Using


Expert Panels to Identify and Describe Successful Flavor Synergies: A Powerful Tool to
Shorten the Development Process” (paper presentation, Society of Sensory Professionals
Conference, Atlanta, GA, October 26–28, 2016).

13. K. N. Oftedal, L. J. R. Lawless, I. Koelliker, and G. V. Civille, “Harmonizing Sensory


Attributes with Nutrition Claims and Emotional Benefits” (paper presentation, 10th
Pangborn Sensory Science Symposium, Rio de Janeiro, Brazil, August 11–13, 2013).

14. A. Retiveau-Krogmann, E. Toronto Doyle, G. V. Civille, and I. Koelliker, “Harmonizing


Sensory Attributes with Nutritional Claims and Emotional Benefits to Ready-to-Eat
Cereals” (paper presentation, 12th Pangborn Sensory Science Symposium, Providence, RI,
August 20–24, 2017).

BK-AST-MNL_13-200365-Chp04.indd 97 30/10/20 5:04 PM


Chapter 5 | Free Choice Profiling
Margaret J. Hinds1

Introduction
Free choice profiling (FCP), unlike conventional descriptive methods, is a descriptive
sensory technique in which each assessor uses his or her own objective terms for pro-
file analysis and is not required to explain or provide definitions for these terms.1 This
approach, in which attribute intensities of products are rated, is based on the premise
that assessors do not differ in their perceptions but merely in the way in which they
describe them.2
Conventional descriptive methods such as Quantitative Descriptive Analysis
(e.g., Tragon QDA developed in 1974 by Stone and Sidel3) and Spectrum™ Descriptive
Analysis4 have consensual steps (see respective chapters in this manual for details).
FCP also differs from conventional descriptive methods in that there is no consensus
among assessors on which terms should be used for profiling; thus, actual descriptor
terms vary in kind and quantity between assessors.
The training of panelists is another major factor that distinguishes conventional
descriptive methods from FCP. FCP is considered as a rapid alternative to conventional
descriptive analysis because it does not require extensive training of assessors.1 Train-
ing may take approximately 1 h4 and focuses on understanding the concept of subjec-
tive versus objective terms for vocabulary development, consistent use of a descriptor
term once the assessor has defined that term in his or her mind, and how to use scales
to score intensities. FCP also has no established training protocols; the rigor and
period of training typically depend on the study objective, complexity of the products,
and whether assessors are technical professionals or naive consumers. For example,
the training time for FCP may be limited to practicing scale usage by naive consumers
who are heavy product users5 or 10 half-hour sessions for the texture of complex prod-
ucts6 compared with 8–12 h for QDA and 100 h or more training time per modality for
the Spectrum method.3

1
Givaudan Flavors Corp., 1199 Edison Dr., Cincinnati, OH 45216, USA https://orcid.org/0000-0001-5447-9561

DOI: 10.1520/MNL1320150034
Copyright © 2020 by ASTM International, 100 Barr Harbor Dr., PO Box C700, West Conshohocken, PA 19428-2959
ASTM International is not responsible, as a body, for the statements and opinions expressed in this chapter.
ASTM International does not endorse any products represented in this chapter.

BK-AST-MNL_13-200365-Chp05.indd 99 18/11/20 3:45 PM


100 Descriptive Analysis Testing for Sensory Evaluation: 2nd Edition

Data from FCP cannot be analyzed by the typical univariate or multivariate


methods used for traditional descriptive sensory because of the lack of uniformity of
scoring attributes among assessors. The variation in terminology and scale usage by
the assessors, and consequent development of individual profiles, is typically
addressed by applying Generalized Procrustes Analysis (GPA) to the data. The GPA
procedure rationalizes the individual spatial configurations to an average configura-
tion called a consensus configuration, and thus provides information on sensory
spatial interrelationships between the samples. GPA results also present insights for
comparative assessor performance. However, GPA results do not provide statistically
significant differences between products, robust data that would be stable over time
and within a certain sensory space (consistent results from the same assessors at a
later time enabling products to repeatedly occupy the same relative positions to each
other), or facilitate the generation of numerically detailed profiles (e.g., spider or
radial graphs) typically associated with conventional descriptive analysis methods.

Core Philosophy
FCP is based on the concept that assessors do not differ in their perceptions of prod-
ucts but mainly in the way that they describe them1 because of their individual expe-
riences and familiarity with the products.7 FCP contains some principles of
conventional descriptive methods in that assessors must be able to reliably detect,
describe, and quantify attribute differences between products. However, unlike con-
ventional descriptive methods, FCP does not require agreement in the usage and
interpretation of the terms used to describe the products.5 Yet insights from FCP can
provide information on consumer perceptions of and the key features they use for
differentiating among products.8 Thus, the FCP approach benefits from terms used
uniquely by consumers, which is not typical with conventional descriptive
methods.9
FCP originated in food sensory science1 but has since been applied to explore
characteristics of nonfood items, including personal care products, and expressive
qualities of animals and landscapes. In this chapter, and throughout this manual, the
term product refers to any stimulus that may be described by sensory profiling (e.g.,
food, beverage, household and personal care products, animal behavior, and other
environmental stimuli that may be rated objectively).

Background and Scope


FCP was introduced1,10 as a method to evaluate flavor profiles in which assessors use
their individual objective terms to describe perceived attributes. The rationale behind
this approach1 was to limit the extensive training time typically spent on language
development and standardization used in conventional descriptive methods, espe-
cially profile methods. In the study conducted by Williams and Langron,1 assessors,
familiar with the product type, were instructed on the concept of FCP and introduced

BK-AST-MNL_13-200365-Chp05.indd 100 18/11/20 3:45 PM


FREE CHOICE PROFILING 101

to quantitative scaling prior to individual vocabulary development and sample profil-


ing.1 Multidimensional plots derived from GPA on the data showed fair agreement
between profiles from the assessors and provided information on the samples compa-
rable to previous knowledge.1 In subsequent studies,11,12 sensory insights obtained from
using FCP have compared favorably with those from conventional descriptive
methods.
Since its introduction, FCP has been applied to assess profiles of various types of
products across various modalities by technical and nontechnical assessors and for a
variety of study objectives. FCP has been used as follows:
• As a standalone method for profile analysis13–16
• In combination with conventional descriptive methods to better understand
consumers’ vocabulary and perceptions of products5,17
• By a traditional descriptive panel to develop terms to supplement generic
descriptors in a standard method for quality testing (e.g., terms to expand on
“other ripe fruit” to facilitate the profiling and blending of virgin olive oils to cater
to specific markets18)
• With a multi-block data-analysis method, Common Components and Specific
Weights Analysis (CCSWA), instead of GPA, as a rapid tool to characterize novel
food products19
• To compare insights from results from FCP panels with those from conventional
descriptive analysis panels (e.g., appearance differentiation between port wines,11
impact of drying and storage time on sensory characteristics of rosemary,20 and
aromas of whole-cone hops and dry-hopped beer12)
• To investigate sensory perceptions of agricultural produce varying in biochemical
compositions due to genetic variability21 and/or growth under different
microclimatic conditions22,23 for the subsequent improvement of breeding and
selection protocols
• As a precursor to other descriptive sensory methods24
• To profile characteristics of consumer products in tandem with hedonic tests to
better understand and predict acceptability25–27
• To roughly assess flavor quality during the temporal release of aromas of flavored
milks at two time points: first impression and aftertaste28
• To yield important insights into consumer differentiation of products5
• To investigate cross-cultural consumer perceptions of products with different
levels of flavor familiarity within the product set8,25
• To investigate human genetic ability, in terms of 6-n-propylthiouracil (PROP)
taster status, on creaminess perception29
• To examine and interpret behavioral expressions of animals (e.g., horses and
ponies to environmental challenges30 and pregnant ewes/sheep to feed-motivation
tests31)

BK-AST-MNL_13-200365-Chp05.indd 101 18/11/20 3:45 PM


102 Descriptive Analysis Testing for Sensory Evaluation: 2nd Edition

• To test the inter-observer reliability of assessing the emotional expression of


domestic dogs32
• To assess the effects of shelter stay on dogs33
• To understand expressive qualities of a range of landscapes34
• To understand consumer perceptions of brands from product photographs35
• With small groups of assessors specifically trained for FCP1,10,13
• With large groups of naive consumers who are heavy users of the product5
• With small groups of naive consumers who have never used the product and who
do replicate evaluations7

WHEN TO USE FREE CHOICE PROFILING


Free Choice Profiling (FCP) is appropriate only when a one-time evaluation of the
products can fulfill the overall goal of a study and consensus on the usage of descriptor
terms is not required. Depending on the sensory objectives, FCP may be conducted
when a panel needs to be set up quickly with minimal training or when insights from
naive consumers are preferred or required. FCP may be used for a general understand-
ing of or to compare profiles of products. FCP may be utilized as a stand-alone method
for profile analysis, as a precursor to conventional descriptive sensory methods (e.g., to
enhance the lexicon for a trained panel), to screen through large numbers of products
to provide preliminary insights, or concurrently with hedonic testing to predict con-
sumer acceptability.
FCP should not be used when robust, detailed, reproducible data that need to be
stable over time or within a certain sensory space are required. In these scenarios,
conventional descriptive methods (e.g., QDA or the Spectrum Descriptive Analysis
Method) are recommended. Examples of inappropriate uses of FCP include the
following:
• If a study objective is to determine statistically significant differences between
products or their specific attributes (e.g., by Analysis of Variance [ANOVA])
• For studies that desire insights from comparisons of or additions of data over time
points
• For shelf-life studies
• When consensus is desired on descriptor terms (e.g., as in the QDA method) and/
or product attributes (e.g., as in the Consensus Profile Method)
• When definitions are required for the descriptor terms used by assessors

Statistical Analysis of FCP Data


GENERALIZED PROCRUSTES ANALYSIS
GPA36 is typically used to analyze data collected from FCP because it caters to data
with different numbers of attributes per assessor as well as differences in the use of the

BK-AST-MNL_13-200365-Chp05.indd 102 18/11/20 3:45 PM


FREE CHOICE PROFILING 103

intensity scales. The data set from each assessor is a two-way table or rectangular array
of attribute intensities per product and can thus be treated as a (n × p) matrix of n
products described across p attributes.
During GPA, the matrices from all assessors undergo simultaneous transfor-
mations of translation, rotation, and scaling.36,37 Translation corrects for intensity
variations from using different parts of the scale by moving the centroids of each
configuration (a separate configuration is first developed from each assessor’s scores)
to a common origin. Rotation accounts for differences in the descriptive terms used
by the assessors. Scaling corrects for the range effects by generating weighted scaling
factors to compensate for individual scale usage differences (e.g., narrow scale ranges
versus wide scale ranges). The weighted scaling factors either stretch or shrink the
volumes of the individual configurations to make them as similar as possible. Sev-
eral iterations of these transformations are performed to decrease the distance
between the individual configurations (from each assessor’s matrix) and maximize
agreement between assessors on each product.37 Finally, GPA generates an optimum
consensus configuration of the samples for different principal axes combinations,
and thus provides information on the sensory spatial interrelationships between the
samples.
Results from GPA also include information on the efficacy of the transformations,
and performance of the assessors in terms of their agreement on each product and
comparison of their individual configurations with the consensus configuration.
Agreement between the shape and orientation of the individual configurations with
each other (and hence agreement with the consensus configuration) is more important
than agreement on scale usage.37 This is because similar shapes and orientations show
alignment on perceptions of product attributes by the assessors.37 Results from
optional permutation tests included with the GPA would indicate if a true consensus
was reached after the transformations, and suggest the appropriate number of factors
(principal axes) to retain.

GENERALIZED CANONICAL ANALYSIS


Generalized Canonical Analysis (GCA) is a less popular alternative to GPA for ana-
lyzing FCP data. It was proposed by van der Burg, de Leeuw, and Dijksterhuis38 and
has been applied to FCP data using the OVERALS procedure.39,40 OVERALS is a
technique for canonical correlation analysis that searches for commonalities between
two or more sets of variables measured on the same products.38 However, when using
GCA, the descriptive terms may need to be reorganized into fewer new categories to
minimize the existence of empty categories.40 Another disadvantage of using GCA
for FCP data is that GCA has a tendency of overfitting if there are many descriptive
terms and few products41 (i.e., generation of the consensus configuration might
include “noise” or unexplained/residual variation in the data, and thus the consensus
configuration would not truly represent the overall perception of the assessors).

BK-AST-MNL_13-200365-Chp05.indd 103 18/11/20 3:45 PM


104 Descriptive Analysis Testing for Sensory Evaluation: 2nd Edition

Conducting FCP
The main objective of FCP is to limit the time spent on training assessors to reach
consensus on the use of descriptor terms (lexicon), some of which might not be famil-
iar to them or applicable for a given study. Thus, the main distinguishing feature of
FCP from conventional descriptive methods is the development of individual descrip-
tive terms by each assessor. However, established sensory protocols/best practices
should be incorporated as much as possible to enable the collection of reliable data and
to obtain meaningful results.
The FCP method may be used with a small group of assessors trained specifi-
cally for FCP (referred to as FCP panel herein) or with naive consumers who are
familiar with the product. The sensory professional should ensure that all potential
assessors, irrespective of training experience, are articulate in critically examining
and describing important product attributes and are willing to participate in the
necessary tasks.

STAGES IN FCP
The following stages should be included in FCP, but approaches to Stages 1–6 would
vary depending on whether a small number of assessors are specifically screened and
validated for their ability to detect, describe, and quantify differences between prod-
ucts (FCP panel) or a large number of naive consumers are used.
1. Assessor screening and qualifications
2. Training and selection
3. Vocabulary development
4. Scaling
5. Validation of assessors’ performance
6. Sample evaluation and data collection
7. Data analysis and interpretation
8. Reporting results
Stages 1-6 for a FCP panel are more rigorous than those for naive consumers
because FCP panels have less numbers of respondents than the latter, and for a smaller
number of assessors, the reliability of the data is expected to increase with increased
training.

ASSESSOR SCREENING AND QUALIFICATIONS

FCP Panel
Potential assessors for a FCP panel are typically recruited from a pool of employees,
students, or professionals who have experience with the product. Recruitment criteria
should include good health to make reliable judgments of the product type and good
communication skills.42 Candidates should be screened for normal sensory acuity and
the ability to discriminate and reproduce results.4,42

BK-AST-MNL_13-200365-Chp05.indd 104 18/11/20 3:45 PM


FREE CHOICE PROFILING 105

At least two to three times the required number of qualified assessors (e.g., 40–50
candidates) should be screened to cater for failures and dropouts. A minimum of eight
trained panelists are recommended to achieve stable data in conventional descriptive
methods,43 thus suggesting the need for more than eight assessors for a FCP panel
because they undergo less rigorous training. For example, in a study with different biscuit
formulations, 15 persons were retained as assessors from the 48 that were screened.16
The types of products used for screening should be relevant to those that would be
in the actual study. Screening tests should be conducted across all modalities that
would be germane to the study objective. Types of tests may include matching, dis-
crimination, and intensity rating; and standard screening tests4,42 should be adapted
for the products to be tested. A series of discrimination tests (e.g., Triangle, Duo-Trio)
and intensity-ranking tests should always be conducted to facilitate the selection of
assessors who can reliably perceive and quantify differences between products. To
qualify for subsequent training, candidates should score at least 60% correct from a
series of discrimination tests with relevant products.4,42
Specific examples of screening tests that have been used with potential FCP can-
didates are as follows:
• For food and beverages:
¡ Biscuits containing 70% and 110% of sugar in a standard formulation and
predetermined to be significantly different (95% confidence) by a 30-member
triangle test panel were used. To qualify for training, each candidate was
required to pass three consecutive triangle tests with the two biscuit samples.16
• For personal care products:
¡ Ability to identify and describe fragrances/scents or textures and estimate
proportions of shaded areas in scaling exercises.4
¡ Ability to detect and describe sensations during and after the application of
sticks with different conditioning agents.44
• For animal-behavior studies:
¡ In reported studies, students of animal science or behavior, and livestock or pet
owners are typically recruited by e-mail or questionnaire, and no additional
screening tests are applied.32,33,45,46
• For environmental studies:
¡ Candidates are recruited from persons with the relevant technical expertise.
For example, in a study to assess landscape expressivity,34 graduate students in
holistic science whose academic program included landscape quality were
recruited and were not required to take additional screening tests.

Naive Consumer Panel


Potential candidates may be randomly recruited via the Internet or from an existing
database using a filter questionnaire (prescreener) specifically designed for the study
objectives. The recruitment protocol should facilitate the identification of consumers

BK-AST-MNL_13-200365-Chp05.indd 105 18/11/20 3:45 PM


106 Descriptive Analysis Testing for Sensory Evaluation: 2nd Edition

who are heavy users of the product but who have no formal training in sensory evalu-
ation and no affiliations to consumer/market research. The prescreener should contain
sections to assess respondents’ abilities to describe and quantify product characteris-
tics. For example, if the study objective is to evaluate the flavors of products, one would
include questions asking respondents to describe noticeable flavors of specific items,
along with scaling exercises to indicate the proportions of shaded areas.4 Depending
on the goal of the study, demographic quotas may need to be considered, and candi-
dates may be recruited from one geographical region or from multiple regions.
Candidates who pass the prescreener are invited on-site to receive additional
information about the study protocol, and interviewed to determine their potential
ability to participate in FCP. In some instances, additional screening tests are admin-
istered to naive consumers. For example, in a study in which flavor nuances between
black coffee products were being investigated, candidates who were regular coffee
drinkers had to score greater than 70% for odor recognition and 100% accuracy for
basic tastes recognition.21

VOCABULARY DEVELOPMENT
In this phase, assessors are required to generate their own terms to best describe the
products of interest. The main criteria are that objective terms only must be developed
and assessors must be consistent in the definition and use of each term throughout the
study. These criteria must be properly explained to and understood by the assessors
before commencing the exercise. Several approaches to facilitate vocabulary develop-
ment may be used. Typical approaches are outlined below.
• Individual Sample Description (ISD)
¡ Assessors may be presented with samples from all of the study products one at
a time and asked to provide spontaneous information/descriptors about the
modality/modalities of interest (e.g., listing the odor attributes of dried parsley
samples;15 appearance, odor, flavor, and texture of dry-cured ham;5 and aroma
attributes of whole cone hops and dry-hopped beer12).
¡ For animal-behavior studies, assessors may watch individual video footage of
each animal to capture the behavior of interest of a group of animals (e.g.,
pregnant ewes/sheep,31 domestic dogs,32 dogs in shelters,33 pigs,45 and lambs46).
Video clips are typically 2–4 min long, after which assessors would have
approximately 2–3 min to write down terms that best describe the behavioral
and/or emotional expressions of the animals.
¡ For landscape studies, assessors may be presented with monadic sequential
digital images of the actual landscapes in addition to site visits.34
• All Sample Description
¡ In some animal studies, multiple animals are observed in each video clip to
assess their group behavioral responses, such as social interaction of dairy

BK-AST-MNL_13-200365-Chp05.indd 106 18/11/20 3:45 PM


FREE CHOICE PROFILING 107

cows (e.g., social licking and head butting47) and environmental challenges to
horses and ponies.30
• Comparative Products Method
¡ A version of Kelly’s Repertory Grid Method (RGM) may be used in which
triads of products/samples with accompanying elicitations are presented to the
assessors.25,48,49 This is a systematic approach in which two of the products
within each triad are randomly grouped such that each product appears in at
least one triad, and one product from each triad is carried over to the next
triad. Assessors are asked to describe how the two grouped products are
similar to each other and different from the third one.49,50 The products may be
actual products in the study,25,40,51 or they may be related to the study
products.50 In either scenario, assessors are encouraged to list all of the
differences and similarities they can perceive, and these terms are used
subsequently to develop their vocabulary list. RGM is typically used with naive
consumers to trigger the generation of descriptors but is criticized for being
difficult and time-consuming to implement.52
¡ Samples of actual products may also be presented in pairs to assessors who are
asked to describe their similarities and differences.21,22
The study objective and product type should be considered when selecting a
method for vocabulary development. For example, in the ISD approach, assessors are
asked to describe products individually, whereas in RGM they are required to com-
pare samples. One study reported that ISD and RGM approaches for evaluating choc-
olate produced sample configurations that were very similar, as were interpretations
of the main perceptual dimensions.52 However, it has been proposed that ISD might
be more appropriate when the main objective is to discover the consumers’ vocabu-
lary, but RGM might be more useful when the goal is to discriminate among product
profiles.49
The sensory professional should review the terms developed by each assessor
and provide assistance, as needed, in finalizing their vocabulary. For example, the
sensory moderator should inspect the terms for synonyms and discuss their mean-
ings with the particular assessor to guide the selection of the most relevant term. If
there are an extensive number of terms for a given modality, the assessor and sensory
moderator should discuss their definitions to ensure that all are essential (e.g., no
redundancy) and that the assessor would not be overwhelmed by having all of them
on the ballot.

SCALING
Candidates are introduced to scaling with the actual type of scale for the study. There
is no standard scale for FCP. Scales may vary in length (e.g., 6 points, 6 inches, 15 cm)
and type (e.g., category, structured, unstructured/continuous with or without anchors).
However, scale characteristics4 should be carefully chosen by the sensory professional

BK-AST-MNL_13-200365-Chp05.indd 107 18/11/20 3:45 PM


108 Descriptive Analysis Testing for Sensory Evaluation: 2nd Edition

based on study objectives and product type. Any words, anchors, or graphics used with
the scale must be descriptive in nature and not related to hedonic sentiments.

FCP Panel
Examples of scales and their applications with FCP are listed below:
• 7-point continuous scale with anchors at 1 = null or very slight and 7 = very intense
for evaluating the odor, flavor, and textural attributes of ewes’ milk cheeses40
• 9-cm unstructured scale anchored at endpoints with intensity terms for
evaluating the appearance, aroma, flavor, and texture of biscuits with different
nutritional formulations16
• 10-cm unstructured line scale anchored at the ends with the terms not perceptible
and strongly perceptible for evaluating the effect of different drying protocols on
the odor of dried parsley15
• 125-mm-length visual analog scale anchored at 0 = minimum and 125 = maximum
for scoring the behavioral expressions of dogs32
• 12.5-cm unstructured line scale anchored, respectively, on the left and right sides
with 0 (descriptor not detected) and 100 (descriptor detected at the highest
intensity) for investigating the behavioral expression of lambs when exposed to
tactile interactions with humans46
• 125-mm-length visual analog scale ranging from minimum (quality absent) to
maximum (quality could not be more dominant) for assessing the expressive
qualities of a range of landscapes34

Naive Consumer Panel


Examples of scales and their applications with naive consumers are listed below:
• 100-mm unstructured line scale anchored with the words absence and intense at
the ends for evaluating hams that were dry-cured at different aging times5
• 100-mm unstructured line scale anchored with low to high for evaluating cereals25
• 10-cm line scale anchored with the words not at all and extremely at the ends for
evaluating different passion-fruit juice formulations7
• 8-cm structured line scale with 1 = not perceived and 9 = intense for evaluating
different formulations of chickpea gels51
• 16-point category scale with 0 = none, 1 = weak, and 15 = strong for evaluating
hot/spicy sauces by consumers of different cultures and experiences with spicy
foods8

ASSESSOR TRAINING
For training sessions, individual ballots or score cards are prepared by the sensory
professional for each assessor from their own list of terms (from the Vocabulary
Development stage). Assessors are shown how to use the ballots, after which they
would rate attribute intensities of one or more replicates of the products using the

BK-AST-MNL_13-200365-Chp05.indd 108 18/11/20 3:45 PM


FREE CHOICE PROFILING 109

prescribed scale. In situations in which the stimuli to be evaluated are not tangible
products (e.g., animal behavior or environmental factors), relevant or representative
graphics (e.g., photographs, video clips) or other items and/or site visits may be used
for training, and assessors would likewise score attribute intensities. Irrespective of
type, products to be evaluated should be blind-coded and randomized, and product
characteristics (e.g., spicy flavor, alcohol content) site location and potential fatigue
should be considered in determining the number of products to be evaluated in a
given session.42

FCP Panel
Training sessions are also opportunities for monitoring scale usage and for revising
anchors or ballots, or both, as necessary. After two replicates of the products have been
evaluated, the sensory professional should confirm the adequacy of scale and ballot
with each assessor. Considerations would include how the assessor is representing
slight versus moderate versus high versus very high intensities on the scale; how useful
the anchors are in their meanings and positions on the scale, and if any of them need
to be revised to home in on intensity ranges specific to the products; and whether there
are terms that definitely need to be deleted from or added to the ballot.

Naive Consumer Panel


Training for naive consumers is typically limited to one practice session with their
personalized ballots and some or all of the products to facilitate familiarity with the
product-evaluation protocols.

VALIDATION OF ASSESSORS’ PERFORMANCE

FCP Panel
Data collected from each assessor during training must be evaluated by the sensory
professional for consistency in terminology usage and scoring, ideally between repli-
cate product samples. The rigor of validation will depend on the study objective and
experience of the assessors with the products. In some instances, ANOVA on replicate
data from each assessor might be essential to monitor consistency in scoring relative
intensities (i.e., no significant difference between replicates for profiles of the same
product). The sensory professional would determine the number of training sessions
required to achieve this goal. Two training sessions are typically sufficient, but one
session might suffice when assessors have been prescreened for relevant technical
expertise. Data should also be inspected for terms that are used with very low fre-
quency, and these terms should be reviewed with the respective assessors to determine
whether they should be removed from the ballot or replaced by more suitable alterna-
tives. For example, in animal-behavior studies, descriptors with negative meanings
might need to be converted to the positive form (e.g., “unhappy” to “happy” ) to facili-
tate uniformity in the use of the scales.46

BK-AST-MNL_13-200365-Chp05.indd 109 18/11/20 3:45 PM


110 Descriptive Analysis Testing for Sensory Evaluation: 2nd Edition

Naive Consumer Panel


Validation of naive consumers’ performance would focus on consistency in terminol-
ogy usage and relative scoring of attributes with different intensities.

ASSESSOR SELECTION AND NUMBERS OF ASSESSORS

FCP Panel
At least 10 assessors who provided reliable and consistent data during the training are
selected to form the FCP panel. However, this number may be increased depending on
the nature of the product and whether the assessors have limited experience with the
product, FCP method, or other sensory practices. The sensory professional should
keep in mind that a minimum of 8 trained assessors are recommended to obtain stable
data when using conventional descriptive methods.43 The number of assessors on FCP
panels range from 10 to 18 in reported studies.

Naive Consumer Panel


From reported studies, the number of naive consumers participating in FCP range
from 10 or 14 doing triplicate product evaluations7,8 to 109 consumers spread over
three regions.5 Several studies use between 20 and 30 consumers.

SAMPLE EVALUATION AND DATA COLLECTION


Products to be evaluated should be blind-coded, presented in sequential monadic
form, and randomized between assessors as much as possible to balance first-order
effects and carryover (if applicable). Attribute ratings may be carried out at the same
time that the stimulus is being experienced (e.g., food, beverages) or after the stimulus
(e.g., scoring animal behavior after watching the relevant video clip32). Rest time and
sensory recuperation between consecutive product evaluations will depend on the
nature of the product/stimulus to be evaluated. The sensory professional should
develop an appropriate plan arising from observations and discussions during train-
ing sessions.

FCP Panel
From reported studies, assessors in FCP panels typically evaluate one or two repli-
cates of the product depending on their experience with the product and its degree of
complexity, but up to four replications13 may be necessary if a detailed characteriza-
tion of products is the objective. In comprehensive studies on animal-behavioral
responses, video clips are usually viewed and evaluated in duplicate, and the two sets
of scores are entered into one matrix defined by the number of animals observed and
the number of terms used by the individual assessor.33 At least duplicate evaluations
of all products should be carried out by FCP panels to enhance the stability and inter-
pretation of the data.

BK-AST-MNL_13-200365-Chp05.indd 110 18/11/20 3:45 PM


FREE CHOICE PROFILING 111

Naive Consumer Panel


Naive consumers typically evaluate one replicate of all the products. However, if there
is a small number of these assessors, replicate evaluations should be performed. For
example, in a study with 10 naive consumers, products were evaluated in triplicate.7

DATA ANALYSIS AND INTERPRETATION USING GPA


For this phase, the sensory professional should collaborate with a statistician if he or
she does not have prior training and experience using GPA. If many modalities of the
products are evaluated, the data from some modalities (e.g., aroma, taste, flavor, after-
taste) might need to be analyzed separately from others (e.g., appearance, texture/
mouthfeel) to facilitate interpretation of the results: this would depend on the com-
plexity of the product and the number of terms generated. In instances when assessors
conduct replicate evaluations of the products, the average data from each assessor
should be used for the GPA.

Data Preparation
Prior to statistical analysis, the data from each assessor must be formatted in a two-
way table with rows for the products and columns for the numerical scores of the
attributes or descriptor terms. The latter are then configured as a series of matrices,
where each matrix represents the intensities of n products across p attributes for one
assessor. Most software packages have options to cater for unequal numbers of col-
umns among the matrices arising from the diverse quantity of terms used by assessors.
However, if such options are unavailable, one would need to pad each matrix with
zero-intensity columns per product to achieve equal numbers of columns across all
matrices.32,37

Data Analysis
During data input, one would include information on the number of matrices (config-
urations) and indicate if their dimensions are equal or to be defined. The latter option
would require denoting the number of attribute columns per assessor. In FCP, asses-
sors typically evaluate all products, thereby allowing an algorithm based on Gower’s
GPA approach36 to be applicable. However, if one or more assessors did not evaluate all
products, some software packages offer an alternative Commandeur algorithm53 to
account for the missing data.
Several statistical packages support GPA, and thus care should be taken in select-
ing tests that are applicable to sensory data and the objectives of the particular study.
Basic steps to include in the GPA are translation, rotation, scaling, and Procrustes
Analysis of Variance (PANOVA). It is also advisable to incorporate a Consensus Test
(i.e., a permutation test to determine whether a true consensus is reached after the GPA
transformations). The consensus profile is defined in terms of its geometrical proper-
ties only.32,37 If available, a Dimensions Test might also be a helpful option to determine
the appropriate number of factors (dimensions) to keep in the consensus

BK-AST-MNL_13-200365-Chp05.indd 111 18/11/20 3:45 PM


112 Descriptive Analysis Testing for Sensory Evaluation: 2nd Edition

configuration. To develop an optimal graphical representation of the consensus con-


figuration and facilitate its interpretation, conducting Principal Component Analysis
(PCA) at the end of the GPA is recommended.32 PCA reduces the number of dimen-
sions of the consensus configuration, determines the principal axes, and reveals the
variability between the products that each of these axes explains.

Interpreting Efficiency of GPA Transformations


The PANOVA results indicate the efficiency of each of the transformations: the smaller
the p value from the Fisher’s F statistic the more efficient is a transformation in reduc-
ing the variability between the individual configurations. For example, a p value of
0.007 for rotation would denote that it was more efficient than scaling with a p value of
0.075 in constructing the consensus configuration.
Results from the consensus test would allow one to determine whether the pro-
portion of the original variance explained by the consensus configuration (Rc) is sig-
nificantly greater than that obtained from random permutations of the data.54
Positively significant results for Rc would indicate that the consensus configuration is
a true consensus space (not developed by chance).

Interpreting the Consensus Configuration and Product Descriptors


The final graphical output from the GPA shows the spatial arrangement of products
and attributes from the consensus configuration. One would use the following infor-
mation to interpret the relationships between the products: (a) eigenvalues to deter-
mine the appropriate number of principal components (axes or dimensions) to select,
and how much of the variability each axis captures, (b) descriptor term loadings on
and correlations with the axes to identify the most important attributes and how they
align with the axes, and (c) product loadings to discern similarities and differences
between products, and how the products relate to the axes.
Correlations between the (previously selected) dimensions and individual asses-
sor terms are examined. Terms displaying high (≥0.5) positive and negative correla-
tions with dimensions should be considered for describing the respective dimensions.
For example, the following terms (and correlations) on Dimension 1 for appearance of
unflavored yogurt products would indicate distinct separation of the products on
Dimension 1 based on their appearances: dull (−0.8), thick (−0.84), white (−0.85),
rough (−0.72), grainy (−0.69), smooth (+0.77), glossy (+0.8), shiny (+0.74), creamy
(+0.8), and silky (+0.64). One could deduce that one end of Dimension 1 is associated
with perceptions of products that look white, thick, dull, rough, and grainy, whereas
the opposite end of the axis is associated with products that look glossy, creamy,
smooth, and silky.
After descriptors for dimensions of interest are established, product loadings
from individual assessor configurations on those dimensions would facilitate descrip-
tion of product profiles. For example, if in the yogurt study above, loadings for Greek
and high-protein yogurts from individual assessors were on the negative side of

BK-AST-MNL_13-200365-Chp05.indd 112 18/11/20 3:45 PM


FREE CHOICE PROFILING 113

Dimension 1 (e.g., ranged from −2.63 to −4.59), and regular and light yogurts were on
the positive side of Dimension 1 (e.g., ranged from 1.98 to 5.72), one would conclude
that the Greek and high-protein yogurts were perceived as white, thick, dull, rough,
and grainy looking, whereas the regular and light yogurts had glossy, creamy, smooth,
and silky appearances.

Interpreting Assessor Performance


“Residuals by Object” results show the variation between assessors for each product,
and may be used to evaluate assessor performance in terms of their agreement on
product profiles. Smaller numerical values for these residuals indicate higher agree-
ment among assessors. For example, a residual value of 2 would indicate almost con-
sensus agreement among assessors for a particular product, whereas 19 would indicate
some disagreement.
“Residuals by Configuration” results show the variation between the individual
configurations (from each assessor) and the final consensus configuration (i.e., after all
transformation iterations were completed). This facilitates determination of which, if
any, assessors gave ratings that did not align with the consensus: the lower the residual
the closer is the fit of an individual configuration to the consensus configuration. For
example, from a panel of 12 assessors, 2 assessors with residuals ranging from 16 to 18
would be considered out of alignment with the consensus configuration compared
with the other 10 assessors who have residuals ranging from 1 to 7. Further inspection
of individual configurations from unaligned assessors would show for which products
there was disagreement.
Additional insights on agreement between individual configurations could be
garnered from a scrutiny (plot) of “assessor variance by configuration and by dimen-
sion.” This would enable one to observe whether all of the individual configurations
were aligned with the same dimensions or whether one or more individual configura-
tions aligned heavily with a different dimension (i.e., insights on how the variability
between the products was explained by each assessor and the extent to which they
agreed). For example, from a panel of 12 assessors, one assessor with 60% variance on
Dimension 3 and less than a 20% spread over Dimensions 1 and 2 would be considered
out of agreement with the other 11 assessors with more than 80% of variance spread
over Dimensions 1 and 2 and less than 20% aligned with Dimension 3.
The “Scaling Factors” indicate the relative scale ranges used by the assessors.
Assessor configurations with scaling factors close to 1 indicate a relatively similar
scale-range use. Scaling factors greater than 1 indicate the use of a relatively narrower
scale range compared with the majority of assessors, whereas those less than 1 indicate
the use of a relatively wider scale range.

REPORTING RESULTS
One would indicate the number of spatial dimensions that the products occupy based
on the number of axes that were important in explaining their variability. One would

BK-AST-MNL_13-200365-Chp05.indd 113 18/11/20 3:45 PM


114 Descriptive Analysis Testing for Sensory Evaluation: 2nd Edition

then discuss the descriptor terms and products that aligned with the important axes
and focus on the ones that loaded highly (typically ≥0.5 but definitely ≥0.3) with the
respective axes. This would facilitate describing the attributes and product properties
associated with each dimension as well as profile differences between the products (as
discussed/exemplified in the previous section). The sensory professional should be
mindful of project objectives when interpreting and reporting results (e.g., selecting
names for axes that provide meaningful and relevant insights to the project). Graphical
representations of the important dimensions of the consensus configurations, and
tables listing descriptor terms that correlated highly with these dimensions should be
included to enhance discussion of similarities and differences in product profiles.

Other Practical Considerations


SENSORY PROFESSIONALS
When FCP is selected as the descriptive method for a study, the sensory panel leader or
sensory moderator must be cognizant of the omission of consensual steps in FCP,
especially if he or she has experience leading conventional descriptive panels. Particu-
larly during the stages of vocabulary development and individual ballot preparation,
sensory panel leaders/moderators must be extremely cautious to avoid biasing the
assessors with their opinions on terms to describe characteristics of the products to be
evaluated.

USE OF FCP WITH EXPERT ASSESSOR PANELS


Expert assessor panels are made up of assessors who possess a high degree of sensory
acuity, are trained in sensory descriptive techniques, and have an established ability to
make consistent and repeatable sensory assessments.55 Although not the norm, expert
assessor panels might also use FCP to develop more comprehensive descriptors to sup-
plement standard generic terms18 or for projects requiring a fast turnaround. It has also
been proposed that the use of FCP with expert assessor panels would offer the dual
advantage of traditional descriptive techniques plus free eliciting techniques and thus
eliminate challenges linked to consensual procedures.56 These highly trained panelists
would simply need an introduction to the concept of FCP but would require no addi-
tional training with references peculiar to the study or calibration for scale usage—
steps that are customary in some conventional descriptive methods.

Pros and Cons of Using FCP


Conventional sensory descriptive methods can provide very detailed, robust, and
reproducible results that are stable over time and within specific sensory spaces and
thus cannot be replaced completely by FCP. Although the study objective should be the
main criterion when selecting a method, some pros and cons of using FCP are listed to
help determine when it might be appropriate to use FCP versus conventional descrip-
tive methods.

BK-AST-MNL_13-200365-Chp05.indd 114 18/11/20 3:45 PM


FREE CHOICE PROFILING 115

Advantages of FCP
• Is faster than conventional descriptive methods: short-term training allows for
quick responses and insights to business needs and market.
• Increases the number of potential descriptors for samples: can enrich sensory
descriptive vocabulary, especially when working with complex products or
scenarios (e.g., animal behavior, environment). In a study to evaluate aromas of
whole-cone hops and dry-hopped beer, FCP generated more than 180 valid aroma
descriptors versus only 14 developed by a conventional descriptive panel.12
• Individual assessor reproducibility is similar to conventional methods because it
allows assessors to use their own terminology.
• Facilitates profiling of samples because assessors feel more comfortable/can better
relate to using their own terms.1
• Affords less frustration to panel leaders because there is no need to force
agreement on the use of common terminology by assessors.
• Avoids potential challenges that might occur during consensual steps (e.g.,
different conceptions of the same stimulus,57 unbalanced group dynamics within
the panel, or a panel leader with a dominant personality58).
• Is cheaper because panel maintenance associated with some conventional
methods (e.g., Spectrum Descriptive Analysis, Consensus Profiling) is not
necessary: thus less time and resources are required.
• Requires only short-term commitment from assessors compared with long-term
commitment needed for some conventional methods (e.g., Spectrum Descriptive
Analysis, Consensus Profiling)
• Has potential for yielding insights on product perceptions from target consumers
versus technical descriptions only from trained panelists.

Disadvantages
• No formal or prescribed lexicon is provided to assessors, and assessors do not
define their descriptor terms; thus, the inconsistent use of terms by assessors,
especially naive consumers, is possible.
• Terms used by assessors might be too personal or difficult to interpret by the
sensory professional, or terms might be related to product benefits.56
• Ballot construction is time-consuming: the sensory professional must prepare a
separate ballot for each assessor containing their own descriptor terms
(vocabulary).
• Data are not robust: individual attribute means cannot be calculated, and thus
FCP does not facilitate the determination of significant flavor profile differences
between products (e.g., using ANOVA) or generation of numerical detailed
profiles (e.g., spider or radial graphs) typically associated with conventional
descriptive analysis methods.

BK-AST-MNL_13-200365-Chp05.indd 115 18/11/20 3:45 PM


116 Descriptive Analysis Testing for Sensory Evaluation: 2nd Edition

• Results are not reproducible, stable over time or in a specific sensory space: FCP
does not cater for the addition and analysis of supplementary data from similar
types of products collected at a different time point.

FCP versus Other Sensory Methods


FCP VERSUS CONVENTIONAL DESCRIPTIVE METHODS
A brief comparison between FCP and two main descriptive sensory methods, QDA3
and Spectrum Descriptive Analysis,4 will be made here. The reader should consult the
respective chapters in this manual for details on these methods.

Quantitative Descriptive Analysis


In QDA, assessors are selected from consumers who are likers and users of the product
category, have demonstrated sensory acuity, and have good descriptive capabilities.
Descriptor terms are developed by assessors and finalized over a series of at least five
2-h training sessions with guidance from a panel leader. Assessors screen through
samples from the product set and provide terms that describe the characteristics they
perceive. After every two samples, discussions are held to define the terms and to check
for redundancies. There is consensus among the panel on the final ballot terms and
their definitions, and the panel leader provides the sequence in which they appear on
the common ballot. There is a prescribed scale for product evaluation (6-inch-long
unstructured line scale anchored with weak/low and strong/high at half-inch positions
from the ends) which is typically carried out in triplicate. Data are robust enough to
allow detailed univariate and multivariate statistical operations, and the tracking of
data over time if best QDA practices are followed.

Spectrum Descriptive Analysis


For Spectrum Descriptive Analysis, persons who previously passed discrimination
tests and demonstrated good communication skills and group dynamics qualify to
be assessors after several (six or more) months of training. Protocols for sample eval-
uation (e.g., quantity to be placed in mouth, time for manipulation), definitions of
terms, and references to aid in lexicon development and intensity scoring are well
defined by the method or the panel leader. There is consensus among the assessors as
to the final terms for the ballot. There is a prescribed 15-point scale for product eval-
uation which is typically carried out in triplicate. Data are robust enough to allow
detailed univariate and multivariate statistical operations and the tracking of data
over time if Spectrum best practices (including panel recalibration with references)
are followed.

FCP VERSUS OTHER RAPID SENSORY METHODS


In addition to FCP, other rapid sensory methods, as potential alternatives to conven-
tional descriptive methods, have been developed. A brief comparison of FCP and some
of these methods is provided in the sections that follow.

BK-AST-MNL_13-200365-Chp05.indd 116 18/11/20 3:45 PM


FREE CHOICE PROFILING 117

Flash Profile
In the Flash Profile (FP) method, which was derived from FCP, assessors also develop
their own descriptor terms; and like FCP, FP has been carried out with trained panels,
semi-trained assessors, and naive consumers.2 However, in FP, there is a simultaneous
presentation of the whole sample set, and products are ranked for each attribute.59
Typically, FP is carried out in two sessions or in one session with two steps. In the first
step, assessors taste all the products comparatively to generate terms to best discrimi-
nate between the products. In the second step, they rank the products from low to high
for each descriptor.59 For data analysis, numbers are assigned to the relative ranks of
the products for each term, and (n × p) matrices are developed for GPA.2 The use of
ranking in FP has been criticized as time-consuming and more difficult than FCP
because assessors have to re-taste samples, but re-tasting might facilitate greater dis-
crimination between the products.60 From studies in which the two methods have
been directly compared, recommendations favor using FCP for large numbers of prod-
ucts and FP when there is a small product set.60

Check-All-That-Apply
Both FCP and Check-All-That-Apply (CATA) require no consensus on descriptor
terms selected by assessors for product evaluation. However, the CATA method uses a
ballot consisting of product descriptor terms that are preselected by the sensory profes-
sional. The selection of descriptor terms is one of the main ways in which CATA differs
from FCP. The CATA terms may originate from trained assessor panels or previous
qualitative or quantitative consumer studies, or both,2 and unlike FCP, assessors have
no input in vocabulary development. Another major difference between FCP and
CATA is that the latter requires assessors to simply select (check/tick) all of the terms
they consider as descriptors of the product, but there is no attribute intensity rating. A
disadvantage of CATA is that terms might not be selected because assessors were neu-
tral or undecided about them or because they did not pay attention to them.2 The
CATA method requires minimal instruction, no training, is relatively easy to perform,
and is completed quickly.61 Frequency scores from the binary data collected from
CATA can be subjected to Cochran’s Q test to determine whether there are significant
differences between the products for their respective attributes, a feat that is not possi-
ble with FCP data. However, insights on spatial sensory interrelationships of products
may be obtained by both methods: using Correspondence Analysis on CATA data and
GPA on FCP data.

Sorting/Multidimensional Sorting
Sorting is a technique in which products are grouped in two dimensions by assessors
according to their overall perceptions or perceptions of specific attributes of the prod-
ucts. It requires little or no assessor training, and is based on the concept that the more
frequently two products are grouped together the more similar they are. Major differ-
ences from FCP are that in Sorting no language or descriptor terms are developed or

BK-AST-MNL_13-200365-Chp05.indd 117 18/11/20 3:45 PM


118 Descriptive Analysis Testing for Sensory Evaluation: 2nd Edition

used by the assessors, and there is no quantitative rating of attributes. For data analy-
sis, a matrix developed by summing the number of times pairs of products are sorted
into the same group (across all assessors) is subjected to multiple dimensional scaling
(MDS). Compared with the sensory dimensions developed from GPA on FCP data,
those from MDS on Sorting data might illustrate product relationships from a more
integrated perspective than from separable attributes.62 However, unlike FCP, the
dimensions created by MDS from Sorting data are not readily interpretable unless the
sensory professional has additional knowledge of the stimuli.

Case Study Example


DATA ANALYSIS AND INTERPRETATION OF BEER APPEARANCE
USING GPA
This is a generic case study that focuses on concepts germane to the analysis and inter-
pretation of GPA on FCP data. Explanation of basic and multivariate statistical princi-
ples is not within the scope of this case study: the sensory professional should be
grounded in these principles before conducting GPA or should consult the expertise of
a statistician. The sensory professional should also be mindful of his or her own project
objectives when conducting GPA and interpreting and reporting results, and either
omit or expand on the approaches and procedures presented in this case study.

Data Source
The data used for this case study were part of a beer data set provided by Hal MacFie
(Hal MacFie Sensory Training Limited, Keynsham, United Kingdom). Vocabulary for
the original beer data was developed using the RGM. Eight assessors with technical
expertise or consumption experience with beers, or both, evaluated 12 freshly-poured
draft beers of different brands using 9-point intensity scales.

Data Preparation
This case study focused on data from one modality only—appearance of the beers—to
facilitate visualization of the accompanying tables and graphics. The number of
descriptor terms (attributes) developed by the assessors for the appearance of the beers
ranged from 3 (Assessor P4) to 10 (Assessor P6) and are listed in table 1.
Data were prepared in an appropriate format to use XLSTAT Premium version
2019.3.2 for GPA analysis. The data set from each assessor was formatted in a two-way
table or matrix with rows for the 12 products (A–M) and columns for the numerical
scores of the attributes. Table 2 illustrates the matrix for Assessor P1: the attribute
intensity scores for each product are in columns 2–9.
Data from the eight assessors were configured as a series of matrices, where each
matrix represented intensities of the 12 products across the particular number of
attributes per assessor. Table 3 shows the matrices derived from Assessors P1 and P2:
data from Assessor P1 are shown in columns 1–8, and data from Assessor P2 are

BK-AST-MNL_13-200365-Chp05.indd 118 18/11/20 3:45 PM


FREE CHOICE PROFILING 119

TABLE 1 D
 escriptor terms developed by eight assessors for the appearance of 12
beers

P1 P2 P3 P4

Dark appearance Light appearance Dark appearance Dark color


Amount of head Amount of color Amount of head Light color
Light appearance Constant head Thick head Foamy appearance
Carbonated appearance Thin, watery head Foamy head
Amount of color Long-lasting head Light appearance
Long-lasting head Frothy head
Hearty appearance Golden color
Fizzy appearance
P5 P6 P7 P8

Light color Pale appearance Dark color Thick appearance


Creamy head Dark color/appearance Amount of head Coats glass
Amount of head Light appearance Caramel-color Holds head for a long
brown time
Dark color Carbonated appearance Yellow color Light color
Golden color Golden color/appearance Depth of shade Dark color
Duration of head Amount of head Amount of head
Fluffy appearance
Golden-brown appearance
Flat appearance
Bubbly appearance

Note: P1–P8 denotes the individual assessors.

shown in columns 9–15. Data sets from the other six panelists (not shown here) were
merged in a similar format to develop the complete series of matrices. Information on
the number of columns designated for attributes from each assessor was tabulated
separately for data-entry purposes (e.g., 8 columns for Assessor P1 and 7 columns for
Assessor P2).
During data input, information on the number of products, assessors, and attri-
bute columns per assessor were defined. (XLSTAT has options to account for unequal
numbers of attribute columns from assessors, so there was no need to add columns
with zeroes to each matrix.)

Data Analysis
Obtaining meaningful insights from FCP data relies heavily on properly conducting
GPA. Thus, the sensory professional should have prior training in using suitable statis-
tical software, including the selection of appropriate options; otherwise, a statistician
should be consulted. The data-analysis procedures and options used in this case study

BK-AST-MNL_13-200365-Chp05.indd 119 18/11/20 3:45 PM


120

TABLE 2 Matrix developed from attribute intensity scores of Assessor P1

BK-AST-MNL_13-200365-Chp05.indd 120
Product Dark Appearance Amount of Head Light Appearance Carbonated Appearance Amount of Color Long-Lasting Head Hearty Appearance Fizzy Appearance

A 2 3 6 5 3 5 2 5
B 3 5 2 3 5 5 5 5
C 2 5 7 6 2 6 2 6
D 2 5 4 5 4 5 4 5
E 1 5 5 5 4 5 5 6
G 2 3 3 4 3 3 3 5
H 5 4 2 2 6 5 6 3
I 2 4 4 3 4 3 3 3
J 2 3 3 5 4 3 3 5
Descriptive Analysis Testing for Sensory Evaluation: 2nd Edition

K 2 3 5 5 5 4 4 5
L 6 5 1 4 6 4 5 5
M 6 5 1 2 7 3 6 2

18/11/20 3:45 PM
TABLE 3 Matrices derived from attribute intensity scores from Assessors P1 and P2

BK-AST-MNL_13-200365-Chp05.indd 121
Long- Thin, Long-
Dark Amount Light Carbonated Amount Lasting Hearty Fizzy Light Amount Constant Watery Lasting Frothy Golden
Appearance of Head Appearance Appearance of Color Head Appearance Appearance Appearance of Color Head Head Head Head Color

2 3 6 5 3 5 2 5 6 2 4 5 3 3 3
3 5 2 3 5 5 5 5 4 4 2 6 2 2 5
2 5 7 6 2 6 2 6 6 2 5 2 6 6 2
2 5 4 5 4 5 4 5 5 1 7 1 1 6 2
1 5 5 5 4 5 5 6 6 3 2 6 5 2 3
2 3 3 4 3 3 3 5 2 5 5 3 5 5 6
5 4 2 2 6 5 6 3 4 5 6 1 6 6 6
2 4 4 3 4 3 3 3 4 4 6 2 6 6 5
2 3 3 5 4 3 3 5 2 6 5 2 5 5 6
2 3 5 5 5 4 4 5 2 5 5 3 5 4 6
6 5 1 4 6 4 5 5 1 6 5 1 5 5 3
6 5 1 2 7 3 6 2 1 7 6 1 6 6 2

Note: Scores from Assessor P1 are shown in columns 1–8; scores from Assessor P2 are shown in columns 9–15.
121

18/11/20 3:45 PM
122 Descriptive Analysis Testing for Sensory Evaluation: 2nd Edition

are provided below but may have to be altered depending on the study objective and
capabilities of the statistical software.
• The GPA was set to perform a maximum of 100 iterations: this is a typical number
of iterations to request to allow the GPA algorithm to develop and test a sufficient
quantity of consensus spaces from the data to get the best possible fit for the
consensus configuration.
• Gower’s GPA approach36 was selected because the 12 products were evaluated by
all 8 assessors.
• Steps included in the GPA were translation, rotation, scaling, PANOVA, a
Consensus Test, and a Dimensions Test.
• For the Consensus and Dimension tests, 300 permutations each were requested:
typically a minimum of 200 is used. These permutation tests are necessary
because the amount of variance explained “on its own” cannot give an indication
of the true fit of the consensus configuration from the data (because assessors
used different descriptors to score attribute intensities of the products) and should
be tested against the variance from random permutations to ensure that results
from the original data did not occur by chance. During the Consensus and
Dimension tests, respective results of the consensus variance and number of
dimensions from the original data were tested (at 95% confidence) against results
from the 300 random permutations of the various attribute intensity scores for
each product.
• Finally, a PCA was conducted to obtain an optimal graphical representation of the
consensus configuration and facilitate its interpretation.

Interpreting Efficiency of GPA Transformations


The PANOVA results (table 4 in which the column “Pr > F” shows the probability of
significance of the Fisher’s F statistic) indicated that the three transformations were
significantly effective (Pr > F was less than 0.05). However, the scaling transformation
(p = 0.02) was less efficient than rotation and translation (p < 0.0001) in reducing the

TABLE 4 PANOVA results

Source df Sum of Squares Mean Squares F Pr > F

Residuals after scaling 448 457.462 1.021


Scaling 7 17.127 2.447 2.396 0.020
Residuals after rotation 455 474.588 1.043
Rotation 315 683.922 2.171 2.126 <0.0001
Residuals after translation 770 1,158.510 1.505
Translation 70 1,556.365 22.234 21.774 <0.0001
Corrected total 840 2,714.875 3.232

BK-AST-MNL_13-200365-Chp05.indd 122 18/11/20 3:45 PM


FREE CHOICE PROFILING 123

FIG. 1 C
 onsensus Test results: variance explained by original data that generated the
consensus configuration versus variance explained by 300 random
permutations of the data.

variability between the volume that each of the individual configurations occupied in
the sensory space.
Results from the Consensus Test (95% confidence) are shown in figure 1, in which
the horizontal axis depicts the variance explained by the original data (that generated
the Consensus Configuration) and by the 300 random permutations of the data. The
histogram indicates the distribution of the variance from 300 random permutations of
the data: the proportion of variance explained from the majority of random permuta-
tions is less than 0.25, with a few permutations explaining almost 0.28 of the variance
(fig. 1). The Rc value of 0.319, which corresponds to the proportion of the original vari-
ance explained by the consensus configuration, is significantly greater than the results
obtained from the random permutations of the data, thus indicating that the consen-
sus configuration was not developed by chance.

Interpreting the Consensus Configuration and Product Descriptors


Results from the Dimensions Test (fig. 2) were used to determine the number of
dimensions on which the products were significantly separated in the Consensus

BK-AST-MNL_13-200365-Chp05.indd 123 23/11/20 10:53 AM


124 Descriptive Analysis Testing for Sensory Evaluation: 2nd Edition

Configuration. The horizontal axes in figure 2A and 2B show the values for F, where F
represents the following ratio: (variance between products)/(variance between
assessors).
For each dimension in the consensus configuration, the F value from the original
data was tested against the F value obtained from 300 random permutations of the
data in that dimension: products were considered significantly separated in a given
dimension if the F value from the original data was larger than the highest F value
from the random permutations. Figure 2A shows that the F value (34.292) from the
original data in Dimension F1 was significantly greater than the distribution (histo-
gram) of F values from the 300 random permutations of the data in that dimension.
However, the F value (9.989) from the original data in Dimension F2 (fig. 2B) was not
significantly higher than the F values from the 300 random permutations (some ran-
dom permutations had F values as high as 10.3). Therefore, one would conclude that in
the consensus configuration, the products were significantly separated (95% confi-
dence) on Dimension F1 but not on Dimension F2.

FIG. 2 D
 imensions Test results: The significant number (95% confidence) of
dimensions in which the products were separated in the consensus
configuration. (A) Test in Dimension F1: Original data significantly different
from 300 random permutations of the data; (B) Test in Dimension F2: Original
data not significantly different from 300 random permutations of the data.

BK-AST-MNL_13-200365-Chp05.indd 124 23/11/20 10:55 AM


FREE CHOICE PROFILING 125

FIG. 2 (continued)

Eigenvalues from the PCA (table 5) were also used to determine the appropriate
number of dimensions or axes to retain for discussion and display. The eigenvalues
(table 5) showed that after PCA transformation of the Consensus Configuration, fac-
tors/axes F1 and F2 represented 66.9% and 15.5%, respectively, of the total variability
compared with less than 7% variability from the other factors. These results (table 5)
suggested the retention of two axes for display, and further indicated that the

TABLE 5 Eigenvalues after PCA transformation of the consensus configuration

F1 F2 F3 F4 F5 F6

Eigenvalue 6.57 1.52 0.63 0.45 0.32 0.17


Variability, % 66.87 15.49 6.46 4.59 3.21 1.75
Cumulative % 66.87 82.35 88.82 93.40 96.61 98.37

Note: F1–F6 denotes the first six axes (factors) from the PCA.

BK-AST-MNL_13-200365-Chp05.indd 125 23/11/20 10:55 AM


126 Descriptive Analysis Testing for Sensory Evaluation: 2nd Edition

variability in appearance attributes of the products was captured mainly along the
first axis (F1) and to a much lesser extent along the second axis (F2), agreeing with the
previous Dimensions Test on the consensus configuration.
To understand how the attributes aligned with the two main axes, correlations
between descriptor terms and Dimensions F1 and F2 from the individual assessors
(table 6) were inspected. Terms displaying high (≥0.5) positive and negative correlations
with the dimensions were considered important for this case study example. Depend-
ing on a study objective, terms with correlations greater than ±0.3 may also be exam-
ined by the sensory professional to obtain more details on product profile perceptions.
Table 6 shows that the terms dark appearance or dark color were used by all asses-
sors except Assessor P2 and showed a high positive correlation (greater than +0.75)

TABLE 6 C
 orrelation between assessor attributes and dimensions in the consensus
configuration

Assessor Descriptor Terms F1 F2 F3 F4

P1 Dark appearance 0.86 0.27 −0.02 −0.31


Amount of head 0.16 0.46 −0.31 −0.19
Light appearance −0.76 0.20 0.04 0.25
Carbonated appearance −0.56 −0.12 −0.03 0.39
Amount of color 0.78 −0.08 −0.37 −0.28
Long-lasting head −0.50 0.37 −0.19 −0.29
Hearty appearance 0.64 −0.07 −0.45 −0.28
Fizzy appearance −0.58 −0.16 −0.08 0.41
P2 Light appearance −0.79 0.30 −0.08 −0.23
Amount of color 0.79 −0.21 0.09 0.13
Constant head 0.42 −0.05 0.40 −0.28
Thin, watery head −0.63 0.03 −0.44 0.25
Long-lasting head 0.46 0.19 0.44 0.29
Frothy head 0.45 0.10 0.58 −0.24
Golden color −0.03 −0.63 0.38 0.07
P3 Dark appearance 0.94 0.23 0.05 −0.11
Amount of head 0.31 0.84 −0.11 0.01
Thick head 0.57 0.55 0.07 −0.18
Foamy head −0.16 0.34 0.53 −0.15
Light appearance 0.79 0.25 −0.13 0.19
P4 Dark color 0.95 0.12 −0.08 0.16
Light color −0.26 −0.74 0.02 −0.37
Foamy appearance 0.39 −0.55 0.20 −0.45

(continued)

BK-AST-MNL_13-200365-Chp05.indd 126 18/11/20 3:45 PM


FREE CHOICE PROFILING 127

TABLE 6 C
 orrelation between assessor attributes and dimensions in the consensus
configuration (continued)

Assessor Descriptor Terms F1 F2 F3 F4

P5 Light color −0.79 0.16 −0.06 0.02


Creamy head 0.46 0.36 0.15 0.62
Amount of head 0.29 0.32 0.40 0.62
Dark color 0.76 0.18 −0.06 −0.31
Golden color 0.58 −0.67 0.18 0.11
Duration of head 0.21 0.16 0.38 0.73
P6 Pale appearance −0.74 0.46 0.26 −0.35
Dark-color appearance 0.93 −0.03 −0.06 0.22
Light appearance −0.90 0.18 0.01 0.23
Carbonated appearance −0.29 0.05 0.74 −0.19
Golden-color appearance 0.12 −0.65 0.48 0.45
Amount of head 0.40 0.48 0.53 0.14
Fluffy appearance 0.28 0.45 0.57 0.13
Golden-brown appearance 0.83 0.35 −0.24 0.05
Flat appearance −0.38 −0.50 −0.67 −0.14
Bubbly appearance −0.07 0.09 0.83 −0.30
P7 Dark color 0.86 −0.37 −0.17 0.05
Amount of head 0.37 0.19 −0.17 0.11
Caramel-color brown 0.88 −0.22 −0.28 0.07
Yellow color −0.95 −0.14 0.01 0.19
Depth of shade 0.86 −0.34 −0.11 0.31
P8 Thick appearance 0.66 −0.36 0.03 −0.25
Coats glass 0.09 0.01 0.06 −0.01
Holds head for a long time 0.42 0.46 −0.13 −0.24
Light color 0.01 0.36 −0.59 0.08
Dark color 0.87 −0.34 0.24 −0.02
Amount of head 0.06 0.60 −0.34 −0.31

Note: Numerical values in F1–F4 denote the correlation between the assessors’ descriptor terms and the
first four axes (dimensions).

with Dimension F1. The terms light color or light appearance were used by six asses-
sors and correlated negatively (more negatively than −0.75) with Dimension F1 from
four assessors (P1, P2, P5, and P6), negatively (−0.74) with Dimension F2 from Asses-
sor P4, but positively (+0.79) with Dimension F1 from Assessor P3 (table 6). Terms
used by one or two assessors only were amount of color, caramel-color brown, depth of
shade, golden-brown appearance, golden color, thick appearance, hearty appearance,
and thick head that correlated positively with Dimension F1; and yellow color, fizzy

BK-AST-MNL_13-200365-Chp05.indd 127 18/11/20 3:45 PM


128 Descriptive Analysis Testing for Sensory Evaluation: 2nd Edition

appearance, carbonated appearance, and thin-watery head that correlated negatively


with Dimension F1 (table 6). With reference to Dimension F2 (table 6), there was a
positive correlation with the term amount of head from four assessors (P1, P3, P6, and
P8) and a negative correlation with golden color/appearance from three assessors (P2,
P5, and P6). Thus, one could conclude that the most important attributes of the prod-
ucts that aligned with the first two dimensions (F1 and F2) were related to color (light
vs. dark on F1), thick/hearty appearance (on F1), amount of head (on F2), and golden
color (on F2).
If a study objective included gaining a deeper understanding of product profiles
or capturing as many differences between products as possible, one would inspect
additional dimensions beyond F1 and F2 for possible correlations with the descriptor
terms. For example, descriptor terms that correlated positively with F3 were frothy
head, foamy head, carbonated appearance, fluffy appearance, and bubbly appearance,
whereas flat appearance correlated negatively with F3 (table 6).
A plot that projects the individual assessor descriptor terms onto the consensus
configuration may be generated. One might inspect this type of plot for an overview of
the details that would be provided in a table such as table 6. Figure 3 shows a projection
of the individual assessor terms for light color/appearance versus dark color/appear-
ance that correlated with Dimension F1. Figure 3 indicates that Dimension F1 defi-
nitely separated the products according to perceptions of their color (other terms that
correlated with F1 and F2 were deleted from this figure to facilitate visualization).
The PCA results showing the product coordinates on axes F1 and F2 (table 7) and
the relative positions of the 12 products (A to M) in the sensory space (fig. 4) facilitated
the discernment of product alignment with the axes and of their similarities and dif-
ferences in appearance. From table 7 and figure 4, Products A, B, and C occupied the
highest negative positions on F1 (i.e., −3.01, −2.66, and −2.51, respectively), whereas
Products M and L had the highest positive positions on F1 (i.e., 5.05 and 4.50, respec-
tively). Previous knowledge about the correlation of the descriptor terms with the
dimensions (table 6) combined with information from table 7 and figure 4 indicated
that there was a continuum from light color and appearance to dark color and hearty/
thick appearance as one moved from negative (left) to positive (right) of the horizontal
axis or Dimension F1 (fig. 4). Therefore, Products A, B, and C were the lightest in color
and appearance, whereas M and L were dark beers with a hearty/thick appearance.
Product H was brown (probably caramel color) but not as dark and hearty as M and L
(table 7). Using a similar rationale for the vertical axis or Dimension F2, it was deduced
that Products C and M had the largest head (2.31 and 1.67, respectively, on F2), whereas
Products J and K were the most golden in color (−2.50 and −1.20, respectively, on F2).
Products E, I, D, and G were probably yellow-brown but not as distinctly golden as J
and K (table 7, fig. 4). The terms foamy, carbonated, and bubbly from some assessors
correlated positively with Dimension F3 (table 6), and this correlation suggested that
Products G, I, and C were perceived as more foamy/carbonated/bubbly than the other
beers because of their positive coordinates on Dimension F3 (table 7).

BK-AST-MNL_13-200365-Chp05.indd 128 18/11/20 3:45 PM


FREE CHOICE PROFILING 129

FIG. 3 P
 rojection of individual assessor terms for color on Dimension F1 of the
consensus configuration.

TABLE 7 PCA results showing the product coordinates on the first four axes

Product F1 F2 F3 F4

A −3.01 1.12 −0.40 −0.29


B −2.66 0.21 −0.62 −0.44
C −2.51 2.31 0.97 0.19
D −1.27 −0.87 −0.57 −0.76
E −1.46 −0.32 −1.22 0.67
G 0.62 −0.40 1.02 1.04
H 2.20 −0.32 0.70 −1.32
I −1.22 0.19 1.02 0.17
J −0.11 −2.50 0.57 −0.37
K −0.13 −1.20 −0.72 0.87
L 4.50 0.10 0.19 0.50
M 5.05 1.67 −0.94 −0.25

BK-AST-MNL_13-200365-Chp05.indd 129 23/11/20 10:56 AM


130 Descriptive Analysis Testing for Sensory Evaluation: 2nd Edition

FIG. 4 R
 elative positions of the products (A–M) on Dimensions F1 and F2 after the
PCA.

Interpreting Assessor Performance


The results denoting residual variance after the GPA were used to determine the
extent of agreement between the assessors as a group and between individual asses-
sors and the consensus configuration. Table 8 shows the results of “Residuals by
Object,” which represented the residual variance from each product after the GPA.
The smaller numerical values for the residuals from Products E (28.66), A (30.19),
and M (30.57) suggested that there was a higher agreement between the assessors
on the profiles of Products E, A, and M than on the profiles of Products D (49.45)
and J (49.39).
An inspection of product (i.e., object) coordinates on the main dimensions/axes
of the individual configurations generated from each assessor’s scores (table 9) pro-
vided additional insights on the differences in residual variance observed in table 8.
In table 9, each row represents the coordinates of a product on Dimensions F1 and F2
in the configuration developed from the scores of one assessor (e.g., J/P1 states that in
the configuration derived from Assessor P1’s scores, Product J had respective coordi-
nates of −0.69 on the first dimension [F1] and −1.81 on the second dimension [F2]).
Table 9 shows that Product M had positive coordinates on both F1 and F2 axes from
all eight assessors (high agreement on M’s coordinates on the first two dimensions/
axes in the individual configurations—therefore less residual variance after the GPA
when M is positioned in the consensus configuration). On the other hand, the

BK-AST-MNL_13-200365-Chp05.indd 130 23/11/20 11:01 AM


FREE CHOICE PROFILING 131

TABLE 8 Residuals by object: the residual variance from each product after the GPA

Product Residual Variance

A 30.19
B 45.52
C 39.42
D 49.45
E 28.66
G 35.21
H 32.59
I 35.45
J 49.39
K 36.79
L 44.22
M 30.57

TABLE 9 C
 oordinates for Products J and M on the first two dimensions from
individual assessors

Product Coordinates from Each


Assessor F1 F2

J/P1 −0.69 −1.81


J/P2 1.81 −1.57
J/P3 −1.95 −2.82
J/P4 −1.72 −4.66
J/P5 −1.60 −2.75
J/P6 0.79 −3.34
J/P7 0.34 −1.00
J/P8 2.15 −2.02
M/P1 6.01 0.74
M/P2 4.03 1.14
M/P3 5.40 2.75
M/P4 5.16 1.66
M/P5 4.09 2.33
M/P6 5.88 1.01
M/P7 6.30 0.29
M/P8 3.56 3.47

Note: F1 and F2 denote the coordinates of Products J and M on the first two dimensions of the individual
configurations developed from the scores of each assessor (P1-P8).

BK-AST-MNL_13-200365-Chp05.indd 131 18/11/20 3:45 PM


132 Descriptive Analysis Testing for Sensory Evaluation: 2nd Edition

coordinates for Product J on F1 ranged from −1.95 to 2.15 (disagreement on its posi-
tion on F1)—hence, the larger numerical residual for Product J (vs. Product M) pre-
viously observed in table 8.
The “Residuals by Configuration” results (table 10) show the residual variance
from each assessor after the consensus configuration was generated by the GPA and
indicate the variation between the individual configurations from each assessor
and the final consensus configuration. The configuration from Assessor P6 had the
best fit to the consensus configuration (smallest residual variance of 41.82) com-
pared with less goodness of fit (residual variance of 64.66) from the configurations
of Assessors P2, P5, and P8 (table 10). A revisit of table 6 provided further explana-
tion for the differences in the residual variance between these assessors: the descrip-
tors developed by Assessor P6 correlated with Dimensions F1, F2, and F3 in a
similar manner to those in the consensus configuration, whereas descriptors for
head from Assessor P5 correlated very highly with Dimension F4 but not F2, and
light color from Assessor P8 correlated negatively with Dimension F3 instead of
Dimension F1.
The “Scaling Factors” (table 11) indicate the relative scale ranges used by the
assessors and how they were transformed during the GPA. Assessors P1, P5, P6,
and P7 had scaling factors close to 1, indicating a relatively similar scale-range use
(table 11). For those assessors who used a relatively narrower scale range (e.g.,
Assessor P3, whose actual intensity scores were mainly between 2 and 5) com-
pared with the other assessors, the GPA algorithm compensated with a high scale
factor (1.28 for Assessor P3; table 11) to get a better fit (expansion) of the volume of
P3’s configuration to the consensus configuration. On the other hand, Assessor P2
used a relatively wider scale range (P2’s actual scores from the raw data were
between 1 and 7), and thus the GPA algorithm used a (reduction) scale factor of
0.81 on the volume of P2’s configuration to improve its fit to the volume of the
consensus configuration (table 11).

TABLE 10 R
 esiduals by configuration: residual variance from each assessor’s
configuration after the consensus configuration was developed by the GPA

Assessor Residual Variance

P1 57.64
P2 64.36
P3 54.46
P4 61.90
P5 64.29
P6 41.82
P7 48.34
P8 64.66

BK-AST-MNL_13-200365-Chp05.indd 132 18/11/20 3:45 PM


FREE CHOICE PROFILING 133

TABLE 11 S
 caling factors used on the individual configurations from each assessor

Assessor Scaling Factor Used by GPA

P1 0.99
P2 0.81
P3 1.28
P4 1.23
P5 0.98
P6 0.92
P7 0.93
P8 1.13

Acknowledgments
The author thanks Hal MacFie for providing the data that were used for statistical
analysis and development of the case study presented in this chapter.

References
1. A. A. Williams and S. P. Langron, “The Use of Free-Choice Profiling for the Evaluation of
Commercial Ports,” Journal of the Science of Food and Agriculture 35, no. 5 (1984):
558–568, https://doi.org/10.1002/jsfa.2740350513

2. P. Varela and G. Ares, “Sensory Profiling, the Blurred Line between Sensory and Consumer
Science. A Review of Novel Methods for Product Characterization,” Food Research
International 48, no. 2 (2012): 893–908, https://doi.org/10.1016/j.foodres.2012.06.037

3. H. Stone, R. Bleibaum, and H. Thomas, “Descriptive Analysis,” in Sensory Evaluation


Practices, 4th ed. (San Diego, CA: Academic Press, 2012), 233–290.

4. C. M. Meilgaard, G. V. Civille, and B. T. Carr, “Selection and Training of Panel Members,” in


Sensory Evaluation Techniques, 5th ed. (Boca Raton, FL: CRC Press, 2016), 165–200.

5. M. D. Guàrdia, A. P. S. Aguiar, A. Claret, J. Arnau, and L. Guerrero, “Sensory


Characterization of Dry-Cured Ham Using Free-Choice Profiling,” Food Quality and
Preference 21, no. 1 (2010): 148–155, https://doi.org/10.1016/j.foodqual.2009.08.014

6. R. J. Marshall and S. P. J. Kirby, “Sensory Measurement of Food Texture by Free-Choice


Profiling,” Journal of Sensory Studies 3, no. 1 (1988): 63–80.

7. R. Deliza, H. MacFie, and D. Hedderley, “The Consumer Perception of a Passion-Fruit Juice


Using Free-Choice Profiling,” Journal of Sensory Studies 20 (2005): 17–27.

8. S. M. Lee, S. E. Kim, J. X. Guinard, and K. O. Kim, “Exploration of Flavor Familiarity Effect


in Korean and US Consumers’ Hot Sauces Perceptions,” Food Science and Biotechnology
25, no. 3 (2016): 745–756, https://doi.org/10.1007/s10068-016-0128-1

9. J. R. Piggott, M. R. Sheen, and S. G. Apostolidou, “Consumers’ Perceptions of Whiskies


and Other Alcoholic Beverages,” Food Quality and Preference 2, no. 3 (1990): 177–185,
https://doi.org/10.1016/0950-3293(90)90022-M

BK-AST-MNL_13-200365-Chp05.indd 133 18/11/20 3:45 PM


134 Descriptive Analysis Testing for Sensory Evaluation: 2nd Edition

10. A. A. Williams and G. M. Arnold, “A Comparison of the Aromas of Six Coffees


Characterised by Conventional Profiling, Free-Choice Profiling and Similarity Scaling
Methods,” Journal of the Science of Food and Agriculture 36, no. 3 (1985): 204–214,
https://doi.org/10.1002/jsfa.2740360311

11. E. Cristovam, A. Paterson, and J. R. Piggott, “Differentiation of Port Wines by Appearance


Using a Sensory Panel: Comparing Free Choice and Conventional Profiling,” European
Food Research and Technology 211, no. 1 (2000): 65–71, https://doi.org/10.1007/
s002170050590

12. B. A. Donaldson, C. W. Bamforth, and H. Heymann, “Sensory Descriptive Analysis and


Free-Choice Profiling of Thirteen Hop Varieties as Whole Cones and after Dry Hopping of
Beer,” Journal of the American Society of Brewing Chemists 70, no. 3 (2012): 176–181,
https://doi.org/10.1094/ASBCJ-2012-0710-01

13. S. M. Rubico and M. R. McDaniel, “Sensory Evaluation of Acids by Free-Choice


Profiling,” Chemical Senses 17, no. 3 (1992): 273–289, https://doi.org/10.1093/
chemse/17.3.273

14. M. Morzel, E. M. Sheehan, C. M. Delahunty, and E. K. Arendt, “Sensory Evaluation of


Lightly Preserved Salmon Using Free-Choice Profiling,” International Journal of Food
Science and Technology 34, no. 2 (1999): 115–123, https://doi.org/10.1046/j.1365-2621.
1999.00239.x

15. M. C. Díaz-Maroto, M. Consuelo, M. A. González Viñas, and M. D. Cabezudo, “Evaluation of


the Effect of Drying on Aroma of Parsley by Free Choice Profiling,” European Food
Research and Technology 216, no. 3 (2003): 227–232, https://doi.org/10.1007/
s00217-002-0643-6

16. D. A. M. dos Santos, J. Lobo, L. M. Araujo, and P. S. Marcellini, “Free Choice Profiling,
Acceptance and Purchase Intention in the Evaluation of Different Biscuit Formulations,”
Ciencia e Agrotecnologia 39, no. 6 (2015): 613–623.

17. C. Gómez, F. Fiorenza, L. Izquierdo, and E. Costell, “Perception of Mealiness in Apples: A


Comparison of Consumers and Trained Assessors,” European Food Research and
Technology 207, no. 4 (1998): 304–310.

18. L. Guerrero, A. Romero, and J. Tous, “Importance of Generalised Procrustes Analysis in


Sensory Characterisation of Virgin Olive Oil,” Food Quality and Preference 12 (2001):
515–520.

19. M. M. L. Tormena, L. T. de Medeiros, P. C. de Lima, G. Possebon, R. H. B. Fuchs, and


E. Bona, “Application of Multi-Block Analysis and Mixture Design with Process Variable
for Development of Chocolate Cake Containing Yacon (Smallanthus sonchifolius) and
Maca (Lepidium meyenii),” Journal of the Science of Food and Agriculture 97, no. 11 (2017):
3559–3567, https://doi.org/10.1002/jsfa.8211

20. M. C. Díaz-Maroto, M. S. Perez-Coello, E. Sanchez-Paloma, and M. A. and Gonzalez Vinas,


“Impact of Drying and Storage Time on Sensory Characteristics of Rosemary
(Rosmarinus Officinalis L.),” Journal of Sensory Studies 22 (2007): 34–48.

21. C. S. G. Kitzberger, M. B. dos Santos Scholz, J. B. D. G. da Silva, M. Benassi, and L. F. P.


Pereira, “Free Choice Profiling Sensory Analysis to Discriminate Coffees,” AIMS
Agriculture and Food 1, no. 4 (2016): 455–469.

BK-AST-MNL_13-200365-Chp05.indd 134 18/11/20 3:45 PM


FREE CHOICE PROFILING 135

22. C. S. G. Kitzberger, C. M. da Silva, M. B. dos Santos Scholz, M. I. F. Ferreira, I. M.


Bauchrowitz, J. B. Eilert, and J. dos Santos Neto, “Physicochemical and Sensory
Characteristics of Plums Accesses (Prunus salicina),” AIMS Agriculture and Food 2, no. 1
(2017): 101–112, https://doi.org/10.3934/agrfood.2017.1.101

23. M. B. dos Santos Scholz, C. S. G. Kitzberger, N. Durand, and M. Rakocevic, “From the Field
to Coffee Cup: Impact of Planting Design on Chlorogenic Acid Isomers and Other
Compounds in Coffee Beans and Sensory Attributes of Coffee Beverage,” European Food
Research and Technology 244, no. 10 (2018): 1793–1802, https://doi.org/10.1007/
s00217-018-3091-7

24. Y. Xia, F. Zhong, Y. Chang, and Y. Li, “An Aromatic Lexicon Development for Soymilks,”
International Journal of Food Properties 18, no. 1 (2015): 125–136, https://doi.org/
10.1080/10942912.2013.780255

25. B. Meudic and D. N. Cox, “Understanding Malaysian Consumers’ Perceptions of Breakfast


Cereals Using Free Choice Profiling,” Food Australia 53, no. 7 (2001): 303–307.

26. L. Torri, M. Piochi, R. Marchiani, G. Zeppa, C. Dinnella, and E. Monteleone, “A Sensory- and
Consumer-Based Approach to Optimize Cheese Enrichment with Grape Skin Powders,”
Journal of Dairy Science 99, no. 1 (2016): 194–204, https://doi.org/10.3168/jds.2015-9922

27. M. E. Kustyawati, O. Nawansih, and S. Nurdjanah, “Profile of Aroma Compounds and


Acceptability of Modified Tempeh,” International Food Research Journal 24, no. 2 (2017):
734–740.

28. S. M. Miettinen, L. Hyvönen, R. S. T. Linforth, A. J. Taylor, and H. Tuorila, “Temporal Aroma


Delivery from Milk Containing 0–5% Added Fat, Observed by Free Choice Profiling, Time
Intensity, and Atmospheric Pressure Chemical Ionization-Mass Spectrometry Techniques,”
Journal of Agricultural and Food Chemistry 52, no. 26 (2004): 8111–8118, https://doi.org/
10.1021/jf040218v

29. S. V. Kirkmeyer and B. J. Tepper, “Understanding Creaminess Perception of Dairy


Products Using Free-Choice Profiling and Genetic Responsivity to 6-n-Propylthiouracil,”
Chemical Senses 28, no. 6 (2003): 527–536, https://doi.org/10.1093/chemse/28.6.527

30. F. Napolitano, G. De Rosa, A. Braghieri, F. Grasso, and A. Bordi, “The Qualitative


Assessment of Responsiveness to Environmental Challenge in Horses and Ponies,”
Applied Animal Behaviour Science 109 (2008): 342–354, https://doi.org/10.1016/j.
applanim.2007.03.009

31. C. A. Stockman, T. Collins, A. L. Barnes, D. Miller, S. L. Wickham, E. Verbeek, L. Matthews,


F. Wemelsfelder, and P. A. Fleming, “Qualitative Behavioural Assessment of the Motivation
for Feed in Sheep in Response to Altered Body Condition Score,” Animal Production
Science 54, no. 7 (2013): 922–929, https://doi.org/https://doi.org/10.1071/AN13020

32. J. Walker, A. Dale, N. Waran, N. Clarke, M. Farnworth, and F. Wemelsfelder, “The


Assessment of Emotional Expression in Dogs Using a Free Choice Profiling Methodology,”
Animal Welfare 19, no. 1 (2010): 75–84.

33. J. K. Walker, A. R. Dale, R. B. D’Eath, and F. Wemelsfelder, “Qualitative Behaviour


Assessment of Dogs in the Shelter and Home Environment and Relationship with
Quantitative Behaviour Assessment and Physiological Responses,” Applied Animal
Behaviour Science 184 (2016): 97–108, https://doi.org/10.1016/j.applanim.2016.08.012

BK-AST-MNL_13-200365-Chp05.indd 135 18/11/20 3:45 PM


136 Descriptive Analysis Testing for Sensory Evaluation: 2nd Edition

34. S. P. Harding, S. E. Burch, and F. Wemelsfelder, “The Assessment of Landscape


Expressivity: A Free Choice Profiling Approach,” PLoS ONE 12, no. 1 (2017): 1–18, https://
doi.org/10.1371/journal.pone.0169507

35. J. R. Elmore and H. Heymann, “Perceptual Maps of Photographs of Carbonated


Beverages Created by Traditional and Free-Choice Profiling,” Food Quality and Preference
10, no. 3 (1999): 219–227, https://doi.org/10.1016/s0950-3293(99)00028-2

36. J. C. Gower, “Generalized Procrustes Analysis,” Psychometrika 40, no. 1 (1975): 33–50.

37. G. B. Dijksterhuis and J. C. Gower, “The Interpretation of Generalized Procrustes Analysis


and Allied Methods,” Food Quality and Preference 3 (1992): 67–87.

38. E. van der Burg, J. de Leeuw, and G. Dijksterhuis, “OVERALS: Nonlinear Canonical
Correlation with k Sets of Variables,” Computational Statistics and Data Analysis 18, no. 1
(1994): 141–163, https://doi.org/10.1016/0167-9473(94)90136-8

39. E. van der Burg and G. Dijksterhuis, “Generalised Canonical Analysis of Individual Sensory
Profiles and Instrumental Data,” Data Handling in Science and Technology 16 (1996):
221–258.

40. P. Bárcenas, F. J. Pérez Elortondo, and M. Albisu, “Comparison of Free Choice Profiling,
Direct Similarity Measurements and Hedonic Data for Ewes’ Milk Cheeses Sensory
Evaluation,” International Dairy Journal 13, no. 1 (2003): 67–77, https://doi.org/10.1016/
S0958-6946(02)00139-5

41. T. Næs, B. Brockhoff, and O. Tomic, “Miscellaneous Methodologies” in Statistics for


Sensory and Consumer Science (Chichester, UK: Wiley, 2010), 243–256.

42. Guidelines for the Selection and Training of Sensory Panel Members (West Conshohocken,
PA: ASTM International, 1981), https://doi.org/10.1520/STP758-EB

43. H. Heymann, B. Machado, L. Torri, and A. L. Robinson, “How Many Judges Should One
Use for Sensory Descriptive Analysis?” Journal of Sensory Studies 27 (2012): 111–122.

44. A. Gambaro, M. E. Parente, and A. Gimenez, “Free-Choice-Profile Descriptive Analysis


of Sticks with Conditioning Agents,” Journal of Cosmetic Science 57, no. 6 (2006):
455–463.

45. O. Wemelsfelder, T. E. A. Hunter, M. T. Mendl, and A. B Lawrence, “Assessing the


‘Whole Animal’: A Free Choice Profiling Approach,” Animal Behaviour 62 (2001): 209–220,
https://doi.org/10.1006/anbe.2001.1741

46. M. Serrapica, X. Boivin, M. Coulon, A. Braghieri, and F. Napolitano, “Positive Perception of


Human Stroking by Lambs: Qualitative Behaviour Assessment Confirms Previous
Interpretation of Quantitative Data,” Applied Animal Behaviour Science 187 (2017): 31–37,
https://doi.org/10.1016/j.applanim.2016.11.007

47. T. Rousing and F. Wemelsfelder, “Qualitative Assessment of Social Behaviour of Dairy


Cows Housed in Loose Housing Systems,” Applied Animal Behaviour Science 101 (2006):
40–53, https://doi.org/10.1016/j.applanim.2005.12.009

48. D. M. H. Thomson and J. A. McEwan, “An Application of the Repertory Grid Method
to Investigate Consumer Perceptions of Foods,” Appetite 10, no. 3 (1988): 181–193,
https://doi.org/10.1016/0195-6663(88)90011-6

BK-AST-MNL_13-200365-Chp05.indd 136 18/11/20 3:45 PM


FREE CHOICE PROFILING 137

49. S. Fiszman, N. Salgado, C. E. Orrego, and G. Ares, “Comparison of Methods for


Generating Sensory Vocabulary with Consumers: A Case Study with Two Types of
Satiating Foods,” Food Quality and Preference 44 (2015): 111–118, https://doi.org/10.1016/
j.foodqual.2015.04.005

50. P. Tarancón, S. M. Fiszman, A. Salvador, and A. Tárrega, “Formulating Biscuits with


Healthier Fats. Consumer Profiling of Textural and Flavour Sensations during
Consumption,” Food Research International 53, no. 1 (2013): 134–140, https://doi.org/
10.1016/j.foodres.2013.03.053

51. M. J. Jiménez, A. Tárrega, R. Fuentes, W. Canet, and M. D. Álvarez, “Consumer


Perceptions, Descriptive Profile, and Mechanical Properties of a Novel Product with
Chickpea Flour: Effect of Ingredients,” Food Science and Technology International 22,
no. 6 (2016): 547–562, https://doi.org/10.1177/1082013216639452

52. J. A. McEwan, J. S. Colwill, and D. M. H. Thomson, “The Application of Two Free-Choice


Profile Methods to Investigate the Sensory Characteristics of Chocolate,” Journal of
Sensory Studies 3, no. 4 (1989): 271–286, https://doi.org/10.1111/j.1745-459X.1989.
tb00450.x

53. J. J. F. Commandeur, “Generalized Procrustes Analysis” in Matching Configurations


(Leiden, The Netherlands: DSWO Press, 1991), 13–64.

54. I. N. Wakeling, M. M. Raats, and H. J. H. MacFie, “A New Significance Test for Consensus in
Generalized Procrustes Analysis,” Journal of Sensory Studies 7, no. 2 (1992): 91–96,
https://doi.org/10.1111/j.1745-459X.1992.tb00526.x

55. Standard Terminology Relating to Sensory Evaluation of Materials and Products,


ASTM E253-19 (West Conshohocken, PA: ASTM International, approved October 15,
2019), https://doi.org/https://doi.org/10.1520/E0253-19

56. O. Lazo, A. Claret, and L. Guerrero, “A Comparison of Two Methods for Generating
Descriptive Attributes with Trained Assessors: Check-All-That-Apply (CATA) vs. Free
Choice Profiling (FCP),” Journal of Sensory Studies 31 (2016): 163–176, https://doi.org/
10.1111/joss.12202

57. H. R. Moskowitz, “Language Development in Descriptive Analysis and Formation of


Sensory Concepts,” in Viewpoint and Controversies in Sensory Science and Consumer
Product Testing, ed. H. R. Moskowitz, A. M. Muñoz, and M. C. Gacula Jr. (Trumbull, CT:
Food & Nutrition Press, 2003), 313–336.

58. H. T. Lawless and H. Heymann, “Descriptive Analysis” in Sensory Evaluation of Food:


Principles and Practices (New York: Springer, 1999), 341–378.

59. V. Dairou and J. M. Sieffermann, “A Comparison of 14 Jams Characterized by


Conventional Profile and a Quick Original Method, the Flash Profile,” Journal of Food
Science 67, no. 2 (2002): 826–834, https://doi.org/10.1111/j.1365-2621.2002.tb10685.x

60. J. Liu, W. L. P. Bredie, E. Sherman, J. F. Harbertson, and H. Heymann, “Comparison of


Rapid Descriptive Sensory Methodologies: Free-Choice Profiling, Flash Profile and
Modified Flash Profile,” Food Research International 106 (2018): 892–900, https://doi.
org/10.1016/j.foodres.2018.01.062

BK-AST-MNL_13-200365-Chp05.indd 137 18/11/20 3:45 PM


138 Descriptive Analysis Testing for Sensory Evaluation: 2nd Edition

61. L. Dooley, Y.-S. Lee, and J.-F. Meullenet, “The Application of Check-All-That-Apply (CATA)
Consumer Profiling to Preference Mapping of Vanilla Ice Cream and Its Comparison to
Classical External Preference Mapping,” Food Quality and Preference 21, no. 4 (2010):
394–401, https://doi.org/10.1016/j.foodqual.2009.10.002

62. A. Saint-Eve, E. Kora, and N. Martin, “Impact of the Olfactory Quality and Chemical
Complexity of the Flavouring Agent on the Texture of Low Fat Stirred Yogurts Assessed
by Three Different Sensory Methodologies,” Food Quality and Preference 15 (2004):
655–668, https://doi.org/10.1016/j.foodqual.2003.09.002

BK-AST-MNL_13-200365-Chp05.indd 138 18/11/20 3:45 PM


Chapter 6 | Summary of Temporal
Methods
Suzanne Pecore 1 and Chris Findlay2

Introduction
Temporal methods encompass a broad range of techniques to gather information
about the time course of sensory sensations. Methodological developments have
focused on measuring temporality within descriptive analysis work, although research
has studied temporal liking as well. This chapter focuses solely on the analytical mea-
sure of temporal events.
One of the earliest examples of evaluating temporal characteristics was in Flavor
Profiling,1 in which the order that flavor attributes appeared were listed, as well as the
quantification of aftertastes. Larson-Powers and Pangborn2 detailed the time-intensity
method, which focused on the exclusive measurement of intensity changes of a single
taste sensation from onset through aftertaste. Since that time there have been many
methods developed to expand the measurement of sensations relative to time, some
focused on attribute intensity changes and others focused on attribute perceptual
changes. Examples of these methods are classified in table 1.

Background and Core Philosophy


The intent of temporal methods is to go beyond attribute intensities to capture another
aspect of the consumption or usage experience with a product. Where traditional
descriptive methods focus on measuring intensity at a single time point, temporal
methods measure changes in attributes with time over the course of the product expe-
rience. In some cases, such temporal changes are more relevant to product differences
than the attribute intensities alone. Thus, temporal methods provide a useful adjunct
to traditional descriptive analysis in measuring the full product experience.
Temporal methods are concerned with attribute changes within the immediate
product experience (e.g., eating a full serving of a food product or applying a lotion to

1
P&D Consulting LLC, 57 N St. NW, Unit 134, Washington, DC 20001, USA
2
Compusense Inc., 255 Speedvale Ave. West, Guelph, ON Canada N1H 1C5
https://orcid.org/0000-0002-2999-2061
DOI: 10.1520/MNL1320160016
Copyright © 2020 by ASTM International, 100 Barr Harbor Dr., PO Box C700, West Conshohocken, PA 19428-2959
ASTM International is not responsible, as a body, for the statements and opinions expressed in this chapter.
ASTM International does not endorse any products represented in this chapter.

BK-AST-MNL_13-200365-Chp06.indd 139 30/10/20 5:03 PM


140 Descriptive Analysis Testing for Sensory Evaluation: 2nd Edition

TABLE 1 Classification of temporal sensory methods (not an exhaustive list)

Acronym

Focus on attribute intensity changes with time


Aftertastes or residual textural attributes DA

Traditional time intensity2 TI


3 DATI
Dual-attribute time intensity
4,5 PP or SP
Progressive or sequential profiling

Multiple-attribute time intensity6 MATI

Focus on attribute perceptual changes with time

Order of occurrence (within flavor profiling)1 FP


7 TOS
Temporal order of sensations

Temporal dominance of sensations8 TDS


9 TCATA
Temporal check-all-that-apply

Note: Superscript numbers correspond to references in the reference list at the end of the chapter.

the skin). Changes that might occur beyond a single exposure period are generally not
evaluated using the temporal methods described here but are more in the realm of
longer-term methods (e.g., shelf-life studies or extended-use testing).

Overview of Method
Temporal methods measure the changes in attribute appearance or attribute intensi-
ties over time and are typically performed by a trained panel of assessors. They provide
additional information about the full product experience that cannot be captured
using traditional descriptive analysis, which is focused on maximal intensities at one
point in time. Temporal methods are concerned with attribute changes within the
immediate product experience and do not extend beyond the single exposure period.

Setting the Objective


Temporal methods are useful when changes in sensory attributes are experienced
while using the product. This will often become evident when the project team first
experiences the test product(s), especially when comparing the experience with a con-
trol product or a target competitive product. In such cases, the test product will show
sensory differences from the control not solely related to attribute intensities. In order
to understand the product experience more fully, the use of a temporal method as well
as traditional descriptive analysis can be useful in interpreting the data.

Subjects/Assessors
Generally temporal evaluations are conducted by trained panelists. As traditional
descriptive analysis will typically accompany temporal work as a means of

BK-AST-MNL_13-200365-Chp06.indd 140 30/10/20 5:03 PM


SUMMARY OF TEMPORAL METHODS 141

understanding the product experience more fully, it is efficient to use the same panel-
ists to conduct both evaluations. The advantage of using current descriptive panelists
is a reduction in training time, as they will already be familiar with the product, the
terminology to describe the product, and the data-entry system.
Given the similarity to descriptive panel work, the same numbers, background,
and expertise is expected of temporal method panelists. However, given the difficult
cognitive and physical tasks involved in some methods, not all existing panelists may
be successful in capturing their temporal responses. Thus, recruiting and training
additional panelists may have to be considered.

Subject/Assessor Screening and Qualification


Pre-recruiting and screening will be similar to other descriptive analysis methods. In
addition, some assessment of the ability to use the data-entry system and to success-
fully track sensations with time will need to be included in the screening process.
Screening for temporal method abilities should include practice with the data-en-
try system on samples of known and wide differences. It will be difficult to ensure that
absolute differences are accurate given the variability of responses to temporal changes
(temporal data are invariably noisy), but some assessment of reproducibility can be
made after several replicate sessions. A useful screening exercise is to interview the
potential panelists, as they often are the best judge of their comfort level with the cog-
nitive and physical tasks involved in temporal work.

Qualifications of Panel Leader/Moderator


The panel leader qualifications will be similar to other descriptive analysis methods. In
addition, the panel leader should be experienced with the data-entry system and the
temporal method to be applied.

Language Development/Panel Training


All of the training activities described in the other chapters in this manual can apply
to temporal methods, particularly if an existing descriptive panel will evaluate the
same products for temporal characteristics. If this is the case, an additional 3 to 5 days
of training should suffice to orient the panelists to the temporal protocol and data-
entry system.

Data Collection
The type and number of attributes collected are as listed in table 2.

Pilot Test and Validation


A similar test as described for screening temporal panelists (see earlier section) can be
used to confirm that panelists are sufficiently trained for temporal work. General con-
sensus among panelists on gross differences among samples may be the limit to any

BK-AST-MNL_13-200365-Chp06.indd 141 30/10/20 5:03 PM


142 Descriptive Analysis Testing for Sensory Evaluation: 2nd Edition

TABLE 2 Type of data and number of attributes listed by temporal method

Temporal Method Type of Data Typical Number of Attributes

Time intensity Intensity 1


Dual-attribute time intensity Intensity 2
Progressive or sequential profiling Intensity 4–6
Multiple-attribute time intensity Intensity 4–6
Temporal order of sensations Perceptiona 4–6
Temporal dominance of sensations Dominance a
4–7
Temporal check-all-that-apply Perception a
4–10

Nominal scale data.


a

validation, as temporality of sensations varies extremely among individuals. The


rapidity of some of the sensations to be recorded also makes it difficult for panelists to
accurately and reproducibly capture their perceptions.
In time intensity, there is a phenomena described as “signature curves,” wherein
individuals will repeatedly generate a specific curve shape. For example, some panel-
ists may show a very short duration to a taste, while others show an extended aftertaste.
This may be due to physical differences such as salivary flow or mouth temperature
and not to any deficiency in training. This will prevent the creation of similar curves
across all panelists but should not limit the panel leader from continuing to collect
such data. It is simply a known source of variability to be managed.

Panel Performance Measures


As described in the prior section, data from temporal methods can be extremely vari-
able, both within and across panelists. In general, using grossly different samples may
provide the best measure of performance, as relative differences should remain con-
stant across such different samples. For example, a comparison of the sweet taste of
sugar to a potent sweetener should consistently show the sugar to have less sweet after-
taste in a time-intensity experiment and could be used as a performance measure for
panelists.

Reporting of Results
Time will always be a key data point in any temporal method, whether a specific time
unit or an ordering of the attributes. The data for analysis will vary depending on the
method, as shown in table 3.

Practical Considerations
Choosing a temporal method depends on the attribute change being measured. In
selecting a method, consider the temporal aspect being affected by the product change,
as well as the capabilities and resources of the researcher.

BK-AST-MNL_13-200365-Chp06.indd 142 30/10/20 5:03 PM


SUMMARY OF TEMPORAL METHODS 143

TABLE 3 Key parameters and analysis appropriate for each temporal method

Temporal Method Key Parameters Summary Statistics

Time intensity Maximum intensity; time to Mean ratings/ANOVA by


maximum intensity; onset time; parameter; summary curve
duration time; AUC
Dual-attribute time intensity Maximum intensity; time to Mean ratings/ANOVA by
maximum intensity; onset time; parameter; summary curve
duration time; AUC
Progressive or sequential Intensity ratings/event ANOVA; summary curve
profiling
Multiple-attribute time intensity Maximum intensity; time to ANOVA; summary curve
maximum intensity; onset time;
duration time; AUC
Temporal order of sensations Proportion appearing first; Mean weighted order of
weighted order of occurrence occurrence/ANOVA; plots of
proportions and weighted order
of occurrence
Temporal dominance of Proportion of each attribute Summary curve of relative
sensations selected, by time point, above dominance proportions for all
chance attributes across panelists; plot
difference between product
curves
Temporal check-all-that-apply Proportion of each attribute Summary curve of relative
selected at each time point dominance proportions for all
attributes across panelists; plot
difference between product
curves

Note: Details on each method are provided in the cited literature references and in the ASTM guides on
time intensity and on other temporal methods. ANOVA = analysis of variance; AUC = area under the curve.

Situational Adaptations
Temporal methods should be used very judiciously, as they typically require highly
trained panelists, a specific data-entry system, and an experienced researcher to exe-
cute the study well. In some situations of multi-attribute evaluation, focusing on only
one or a few attributes of the temporal experience may be appropriate when a full study
cannot be conducted due to time and resource constraints. For example, sometimes
the early onset of a key attribute is critical to acceptability, and simply asking an asses-
sor to check the first appearing attribute on a list will suffice to understand this. In
another example, simply asking about any sensations remaining at a time point after
product application can determine whether the product is meeting specifications for
efficacy.

References
1. S. E. Cairncross and L. B. Sjöström, “Flavor Profiles—A New Approach to Flavor
Problems,” Food Technology (1950): 308–311.

BK-AST-MNL_13-200365-Chp06.indd 143 23/11/20 9:43 PM


144 Descriptive Analysis Testing for Sensory Evaluation: 2nd Edition

2. N. Larson-Powers and R. M. Pangborn, “Paired Comparison and Time-Intensity


Measurements of the Sensory Properties of Beverages and Gelatins Containing Sucrose
or Synthetic Sweeteners,” Journal of Food Science 43 (1978): 41–46.
3. L. M. Duizer, K. Bloom, and C. J. Findlay, “Dual-Attribute Time-Intensity Sensory
Evaluation: A New Method for Temporal Measurement of Sensory Perceptions,” Food
Quality and Preference 8, no. 4 (1997): 261–269.
4. F. R. Jack, J. R. Piggott, and A. Paterson, “Analysis of Textural Changes in Hard Cheese
during Mastication by Progressive Profiling,” Journal of Food Science 59, no. 3 (1994):
539–543.
5. L. Methven, K. Rahelu, N. Economou, L. Kinneavy, L. Ladbrooke-Davis, O. B. Kennedy,
D. S. Mottram, and M. A. Gosney, “The Effect of Consumption Volume on Profile and
Liking of Oral Nutritional Supplements of Varied Sweetness: Sequential Profiling and
Boredom Tests,” Food Quality and Preference 21, no. 8 (2010): 948–955.
6. C. Kuesten, J. Bi, and Y-H. Feng, “Exploring Taffy Product Consumption Experiences
Using a Multi-Attribute Time-Intensity (MATI) Method,” Food Quality and Preference 30,
no. 2 (2013): 260–273.
7. S. Pecore, C. Rathjen-Nowak, and T. Tamminen, “Temporal Order of Sensations” (paper
presentation, 9th Pangborn Sensory Science Symposium, Toronto, Canada, September
4–8, 2011).
8. N. Pineau, P. Schlich, S. Cordelle, C. Mathonnière, S. Issanchou, A. Imbert, M. Rogeaux,
P. Etiévant, and E. Köster, “Temporal Dominance of Sensations: Construction of the TDS
Curves and Comparison with Time-Intensity,” Food Quality and Preference 20, no. 6
(2009): 450–455.
9. J. C. Castura, L. Antúnez, A. Giménez, and G. Ares, “Temporal Check-All-That-Apply
(TCATA): A Novel Temporal Sensory Method for Characterizing Products,” Food Quality
and Preference 47A (2016): 79–90.

BK-AST-MNL_13-200365-Chp06.indd 144 23/11/20 9:43 PM


Index

A ANOVA techniques, 4, 36, 38


abstract concepts, 28 analysis with, 58
accuracy, 90 data collection and, 63
acuity, 28–29, 85–86 with FCP, 102
adaptations, 20, 42–46, 72–73, 95–96, for product discrimination, 66
143 for sensitivity ranges, 90
advanced applications, 96 appearance lexicons, 45
all sample descriptions, 106–107 aromatics, 79 (table)
analysis. See also specific analyses arrangement test, 6
with ANOVA techniques, 58 articulation, 54
assessment and, 37 assessors
CCSWA, 101 assessment, 37
data, 35–36, 93, 111–113, 120, 122 background of, 27–29
DDA, 52, 67, 70 of beer, 118–119, 119 (table), 120
descriptive analysis technical (table), 121 (table), 122
texture evaluations, 43–44 characteristics of, 27–29
of Flavor Profile Method, 14 consensus between, 13–14
GCA, 103 in Consensus Profiling Methods, 4–5
GPA, 100–103, 111–112 discussion with, 3, 3 (figure)
maps, 40, 41 (figure) evaluation for, 11
PANOVA, 111–112, 122–123, 122 expert, 114
(table) FCP for, 99
PCA, 67 Flavor Profile Method for, 11
proprietary, 66 individual assessor terms,
of results, 73 129 (figure)
Skinfeel Profile Descriptive performance, 109–110, 113, 130,
Analysis, 44 131 (table), 132, 132 (table),
Standard Guide for Two Sensory 133 (table)
Descriptive Analysis plots for, 128
Approaches, 83 of products, 1–2
statistical, 35–36, 102–103 of QDA, 53–55
for temporal methods, 143 (table) qualifications of, 5–7
animal-behavior studies, 105 scaling for, 15
anosmias, 15 scores of, 130, 131 (table), 132

BK-AST-MNL_13-200365-Index.indd 145 19/11/20 8:22 PM


146 Descriptive Analysis Testing for Sensory Evaluation: 2nd Edition

screening for, 5–7, 29, 104–105 behavior, 62, 65, 105


statistical treatment and, 15 behavioral approaches, 53
subjects and, 140–141 beverage industry, 16, 105
in Texture Profile Method, 27–30 biscuits, 105
training, 108–109 blank-sheet method, 11
validation of, 109–110 blank spots, 15
attitude, 28 blend, 95
attributes brand protection, 74
attribute panels, 43 breakdown, 24 (table)
correlation matrix with, 75 (table) budgeting suggestions, 42–44
data and, 142
definitions of, 21, 32–33, 63, 65 C
(figure) calculations, 65
for descriptions, 59–60 calibration, 26–27, 37
details of, 11–12 caramels, 24 (table), 38–39, 39 (figure)
development of, 46 case studies. See temporal methods
evaluation of, 22 categories, 10–11, 30, 34, 63
examples of, 81, 81 (table) characteristics, of assessors, 27–29
intensity of, 120 (table), 121 (table) check-all-that-apply (CATA) method,
key driver, 76 (figure), 96 117
of products, 102 chewdown, 24 (table)
psychology of, 60 Civille, Gail Vance, 77
quantitative references for, 79 classifications, 20, 139, 140 (table)
of samples, 4 CLT data, 75 (table)
scores for, 14, 16 Common Components and Specific
selection of, 29 Weights Analysis (CCSWA), 101
technical, 61 comparisons, 8, 10, 12–13, 37, 107
in training, 33 complex consumer data, 73
understanding of, 62 comprehensive texture evaluation,
availability, 27 22, 23 (table), 24, 24 (table), 25–26
(table), 26
B conducting, 104–114
balance, 61–62, 95 confidence, 9–10
basic concepts, 31 consensus, 1–2, 9, 13–14, 112–113
basic instruction course, 8 agreement, 62
basic perpetual test, 6 in data collection, 92
basic principles, 31–32 data collection and, 35
beer. See also free choice profiling interpretation of, 123–128,
assessors of, 118–119, 119 (table), 124–125 (figure), 125 (table),
120 (table), 121 (table), 122 126–127 (table), 129 (figure),
PANOVA results for, 122–123, 122 129 (table), 130 (figure)
(table) test results, 123 (figure)

BK-AST-MNL_13-200365-Index.indd 146 19/11/20 8:22 PM


INDEX 147

Consensus Profiling Methods direct-data entry, 73


assessors in, 4–5 FCP, 102–103
data collection for, 10–12 GCA for, 103
Flavor Profile Method and, 1–2, 10 GPA for, 100, 102–103
overall method of, 2–3, 3 (figure) as means tables, 67, 68–69 (table)
panels in, 5–7 preparation, 118–119, 119 (table),
panel training for, 7–10 120 (table), 121 (table)
practical considerations in, 13–15 QDA, 74 (figure)
situational adaptations for, 15–16 for sensory analysts, 12
consistency, 6, 58, 91 (figure) sources, 118
consumers in spider plots, 67, 70 (figure), 71
complex consumer data, 73 (figure)
consumer-based language, 56 texture, 40, 41 (figure)
consumer goods, 53 decision-making, 14, 72–73
consumer liking, 74 (figure) definitions, 21, 32–33, 57, 63, 65
consumer-oriented perceptions, (figure), 77
51–52 degree of difference (DOD), 95
consumer ratings, 73–74, 74 derivative methods, 43
(figure), 75 (table), 76 (figure) descriptions. See also specific topics
consumer results, 37 all sample, 106–107
naive consumer panels, 105–106, articulation in, 54
108–111 attributes for, 59–60
products for, 72, 77 comments and, 59
subjects and, 71–72 conventional descriptive methods,
conventional descriptive methods, 116 116
core characteristics, 20 DDA, 52, 67, 70
core elements, 32–33 descriptive analysis technical
core philosophy. See philosophy texture evaluations, 43–44
correlation matrix, 75 (table) descriptive data, 74
criticism, 15 descriptive lexicons, 45
descriptive methods, 19
D descriptor terms, 126–127 (table),
data 126–128, 129 (figure)
analysis, 35–36, 93, 111–113, 120, details in, 8
122 ISD, 106
from ANOVA techniques, 4 by panels, 51
attributes and, 142 product descriptors, 112–113
CLT, 75 (table) profile-based descriptive
collection, 10–12, 35–36, 63, 65, approach, 22, 23 (table)
92, 141, 142 (table) Standard Guide for Two Sensory
complex consumer data, 73 Descriptive Analysis Approaches,
descriptive, 74 83

BK-AST-MNL_13-200365-Index.indd 147 19/11/20 8:22 PM


148 Descriptive Analysis Testing for Sensory Evaluation: 2nd Edition

technical descriptive method, comprehensive texture evaluation,


21–22 22, 23 (table), 24, 24 (table),
texture terms, 25–26 (table) 25–26 (table), 26
details, 8–9, 11–12 descriptive analysis technical
development. See also language texture evaluations, 43–44
development error in, 92
of attributes, 46 handfeel evaluations, 44–45
of evaluation protocols, 34 individual, 35–36
lexicon, 79–83, 79 (table), of nonfoods, 44–45
80 (figure), 80 (table), 81 for panels, 42
(table), 82 (table), 84–85 (table), philosophy in, 31
86, 86 (figure) procedure, 32–33
of methods, 42–44 of products, 8
product developers, 67 protocols, 34
protocol, 92 replications in, 63–64
of Texture Profile Method, 19–21 sample evaluations, 92, 110–111
vocabulary, 106–107 standard evaluation protocol, 83,
diagnostic descriptive analysis (DDA), 84–85 (table)
52, 67, 70 for subjects, 60
differentiation, 60, 95 unconditional positive regard in,
dimensions test results, 124–125 8–9
(figure), 124–128, 125 (table), experience, 93, 95
126–127 (table), 129 (figure), experimental psychology, 52
129 (table), 130 (figure) expert assessor panels, 114
direct-data entry, 73
disagreements, 56 F
discrimination tests, 55, 90–91, feedback, 62, 91–92
91 (figure) fidelity, 95
discussion, 3, 3 (figure), 11–12, 14, 60 first bite phase, 23 (table)
distractions, 14 first chew, 24 (table)
dry dog food products, 41 (figure) five steps of lexicon development, 79,
80 (figure)
E flash profile (FP) method, 117
emotional benefits, 96 Flavor Profile Method, 1–2, 4–7,
end users, 10–11 10–11, 14–15, 19
environmental studies, 105 flavors. See specific topics
error, 92 focus, of subjects, 62–63
evaluation formats, of scores, 12
for assessors, 11 formula, for products, 59
of attributes, 22 free choice profiling (FCP)
complete product evaluations, advantages of, 115
34 for assessors, 99

BK-AST-MNL_13-200365-Index.indd 148 19/11/20 8:22 PM


INDEX 149

background of, 100–102 information, 13, 15–16, 21, 60


case study of, 118–119, 119 instructions, 64 (figure)
(table), 120–121 (table), intensity
122–128, 122 (table), 123 (figure), of attributes, 120 (table),
124–125 (figure), 125 (table), 121 (table)
126–127 (table), 129 (figure), comparisons of, 8
129 (table), 130, 130 (figure), of core characteristics, 20
131 (table), 132, 132 (table), of flavor, 6
133 (table) references, 26–27, 33
conducting, 104–114 interpretation, 10
data, 102–103 of assessor performance, 130,
disadvantages of, 115–116 131 (table), 132, 132 (table),
panels for, 104–105, 108–110 133 (table)
philosophy of, 100 of consensus, 123–128,
practical considerations for, 114 124–125 (figure), 125 (table),
sensory methods compared to, 126–127 (table), 129 (figure),
116–118 129 (table), 130 (figure)
fundamental principles, 20 interviews, 5, 7, 55, 86
introductions, 57
G
General Foods researchers, 19–20 J
Generalized Canonical Analysis jargon, 60–61
(GCA), 103
Generalized Procrustes Analysis K
(GPA), 100–103, 111–112 Kelly’s Repertory Grid Method (RGM),
graphic representations, 38–39, 107
39 (figure), 40, 41 (figure), 58, key driver attributes, 76 (figure), 96
58 (figure) key parameters, for temporal methods,
graphs, 67, 74 (figure), 130 (figure) 143 (table)
grouping, 95
L
H laboratory procedures, 72
handfeel evaluations, 44–45 language development
health, 28 panel training and, 141
herbal gel samples, 41 (figure) process, 73
histograms, 124–125 (figure) for QDA, 56–63, 58 (figure)
sensory language, 52
I terminology, 38
individual assessor terms, 129 (figure) for Texture Profile Method, 30–34
individual evaluation, 35–36 leaders, 86–87, 141. See panels
individual panelists, 89–90 learning, 32–33
individual sample description (ISD), 106 legal disputes, 71–72

BK-AST-MNL_13-200365-Index.indd 149 19/11/20 8:22 PM


150 Descriptive Analysis Testing for Sensory Evaluation: 2nd Edition

lexicon development. See development N


life, of panels, 87 naive consumer panels, 105–106,
line scale, 58, 58 (figure) 108–111
liquid pharmaceutical industry, 16 napping, 95
nonfoods, 21, 26, 44–45
M
macaroni and cheese, 79–82, O
79 (table), 80 (figure), 80 (table), objectives, 11
81 (table), 82 (table) objective settings, 77, 140
maps, 40, 41 (figure), 93, 94 (figure), observations, 12
94 (table) oiliness, 13
marketing, 73 opinions, 61–62
masticatory phase, 23 (table) organization, 56–58
matrix, 120 (table), 121 (table) orientation, 54–55, 57, 88
means tables, 67, 68–69 (table) OVERALS procedure, 103
measurement, 1, 13, 37–38, 65–66, 142
meats, 40, 40 (table) P
methods. See also specific methods pairings, 57, 95–96
adaptations of, 46 panels
blank-sheet method, 11 attribute, 43
CATA, 117 categories for, 10–11
comparative products method, in Consensus Profiling Methods,
107 5–7
consensus and, 9 consumer results and, 37
conventional descriptive, 116 descriptions by, 51
derivative, 43 discrimination tests for, 90–91,
descriptive, 19 91 (figure)
development of, 42–44 discussion in, 11–12, 60
FP, 117 evaluation for, 42
modifications of, 22 experience for, 93, 95
practice exercises with, 36 expert assessor, 114
products and, 7–8 for FCP, 104–105, 108–110
profile methodology, 32 leaders of, 10, 56, 62–63, 86–87,
for QDA, 53 141
rapid, 67, 68–69 (table), 70–71, location of, 16
70 (figure), 71 (figure), mock panel sessions, 86
116–118 motivation for, 91–92
mock panel sessions, 86 naive consumer, 105–106, 108–111
moderators, 29–30, 56, 73, 141 objective settings for, 77
modifications, 22 panelist confidence, 9–10
motivation, 91–92 panel performance measures, 13,
multidimensional sorting, 117–118 37–38, 65–66, 142

BK-AST-MNL_13-200365-Index.indd 150 19/11/20 8:22 PM


INDEX 151

personal care products for, 89 pilot tests, 12, 36–37, 63–65, 64


qualitative references for, 78–79 (figure), 65 (figure), 141–142
recognition for, 91–92 plot of consistency, 91 (figure)
recruitment of, 3, 3 (figure), 83, practical considerations
86 (figure) in Consensus Profiling Methods,
target panel size, 27 13–15
team decisions by, 14 for FCP, 114
training of, 4, 7–10, 13–14, 30–34, in QDA, 67
56–63, 58 (figure), 86 (figure), for spectrum descriptive analysis,
87–89, 87 (table), 141 93, 94 (figure), 94 (table), 95
validation of, 89–91, 91 (figure) for temporal methods, 142
paperwork, 57 Practical Guide for Comparative
partial consensus profiling, 1–2 Advertising, 72
participation, 54 practice, 10, 31, 33–34, 36
PCA maps, 40, 41 (figure) precision, 90
perceptions, 6–8, 51–53 predicted liking, 96
perceptual mapping, 93, preference segment plot, 76 (figure)
94 (figure), 94 (table) preinterviews, 55
texture perception, 28, 31–32 preparation, of data, 118–119, 119
performance (table), 120 (table), 121 (table)
assessor, 109–110, 113, 130, pre-recruiting, 141
131 (table), 132, 132 (table), prescreening, 83, 86 (figure)
133 (table) presentation, of profile results, 38–40,
consistency and, 91 (figure) 39 (figure), 40 (table), 41 (figure)
measures, 37 principal component analysis (PCA),
panel performance measures, 13, 67, 112
37–38, 65–66, 142 for FCP, 125–126, 125 (table)
trials, 3, 3 (figure) maps, 40, 41 (figure)
personal care products, 83, 84–85 results, 128, 129 (table)
(table), 85–86, 86 (figure), 89, 105 procedure, 23, 32–33, 72–73, 89, 103
personal interviews, 7, 86 process. See specific topics
philosophy Procrustes Analysis of Variance
of Consensus Profiling Method, (PANOVA), 111–112, 122–123, 122
2–3 (table)
in evaluation, 31 products
of FCP, 100 assessors of, 1–2
of QDA, 52 attributes of, 102
of spectrum descriptive analysis, brand protection for, 74
77–78 budgeting suggestions for, 42–44
of temporal methods, 139–140 categories of, 30, 63
of Texture Profile Method, comparative products method,
19–22 107

BK-AST-MNL_13-200365-Index.indd 151 19/11/20 8:22 PM


152 Descriptive Analysis Testing for Sensory Evaluation: 2nd Edition

comparisons of, 10 of panel leaders, 10, 56, 86–87


complete product evaluations, 34 screening for, 53–55
complexity of, 56 for temporal methods, 141
for consumers, 72, 77 qualitative references, 59, 78–79
descriptors, 112–113 quality, 95
developers of, 67 quantitative descriptive analysis
differentiation of, 60 (QDA), 99, 116
dry dog food, 41 (figure) adaptations for, 72–73
evaluation of, 8 language development for, 56–63,
formula for, 59 58 (figure)
in graphs, 130 (figure) practical considerations in, 67
grouping of, 95 for products, 51–52, 73–74, 74
marketing of, 73 (figure), 75 (table), 76 (figure)
methods and, 7–8 rapid methods for, 67, 68–69
oiliness in, 13 (table), 70–71, 70 (figure), 71
pairings of, 57 (figure)
personal care, 83, 84–85 (table), research from, 71–72
85–86, 86 (figure) results reporting for, 66–67
product-adapted universal scaling, subjects for, 53–55
95 testing with, 63–65, 64 (figure),
product discrimination, 66 65 (figure)
QDA for, 51–52, 73–74, 74 quantitative references, 79
(figure), 75 (table), 76 (figure) questions, in orientation, 54–55
screening and, 55
profiling, 9, 12, 15–16, 44. See also R
specific topics ranking tests, 6
for partial consensus profiling, 1–2 rapid approaches, 51–52
profile-based descriptive rapid methods, 67, 68–69 (table),
approach, 22, 23 (table) 70–71, 70 (figure), 71 (figure),
profile results, 38–40, 39 (figure), 116–118
40 (table), 41 (figure) ratings, 58, 73–74, 74 (figure), 75
tabular Texture Profile, 40, (table), 76 (figure)
40 (table) R&D projects, 43
proprietary analysis, 66 recognition, for panels, 91–92
protocol, 34, 83, 84–85 (table), 92 recruitment, 3, 3 (figure), 83, 104, 141
psychology, 5, 52, 60 references
acuity and, 28–29
Q intensity, 26–27, 33
QC/QA sensory programs, 43 qualitative references, 59, 78–79
qualifications use of, 80, 80 (table)
of assessors, 5–7 refinement, of lexicon, 81, 81 (table)
in FCP, 104–105 Repertory Grid Method (RGM), 107

BK-AST-MNL_13-200365-Index.indd 152 19/11/20 8:22 PM


INDEX 153

replications, 35, 63–64 for FCP, 107–108


reporting line scale, 58, 58 (figure)
of results, 13, 66–67, 93, 113–114, product-adapted universal scaling,
142, 143 (table) 95
for spectrum descriptive analysis, 93 scaling factors, 113, 132, 133 (table)
for Texture Profile Method, universal scales, 78–79, 79 (table),
38–40, 39 (figure), 40 (table), 95
41 (figure) unstructured graphic rating scale,
research, 36, 52, 71–72 58, 58 (figure)
residual phase, 23 (table), 24 (table) science, 14, 29–30, 54, 70–71, 72–73
residuals by configuration, 132, scope, of FCP, 100–102
132 (table) scores, 7–8, 11–12, 33, 61–63, 64
resource constraints, 67 (figure)
responses, 53 of assessors, 130, 131 (table), 132
responsibility, 5 for attributes, 14, 16
results matrix for, 120 (table), 121 (table)
analysis of, 73 screening
consensus test, 123 (figure) acuity, 85–86
consumer, 37 for assessors, 5–7, 29, 104–105
dimensions test, 124–125 (figure), of candidates, 30
124–128, 125 (table), DDA and, 67, 70
126–127 (table), 129 (figure), prescreening, 83, 86 (figure)
129 (table), 130 (figure) products and, 55
from panels, 37 for qualifications, 53–55
PANOVA, 122–123, 122 (table) in temporal methods, 141
PCA, 128, 129 (table) tools, 5
reporting of, 13, 66–67, 93, sensations, 59
113–114, 142, 143 (table) sensitivity ranges, 61, 90
for spectrum descriptive sensory acuity testing, 85–86
analysis, 93 sensory analysts, 12
for Texture Profile Method, sensory claims substantiation, 71–72
38–40, 39 (figure), 40 (table), sensory drivers, 96
41 (figure) sensory language, 52
review, 32–33 sensory perceptions, 7–8
rheology, 19–22, 31–32 sensory procedural decisions, 72–73
routine feedback, 91–92 sensory professionals, 114
sensory science, 54, 72–73
S sensory skills, 54
samples, 4, 36, 92, 106–107, 110–111 sensory tests, 15
scales, 1, 9, 15, 32–33, 35 sensory texture evaluations, 25–26
classifications and, 20 (table)
definitions and, 77 Sidel, Joel L., 52

BK-AST-MNL_13-200365-Index.indd 153 19/11/20 8:22 PM


154 Descriptive Analysis Testing for Sensory Evaluation: 2nd Edition

signature curves, 142 comments by, 59


single correlation graph, 74 (figure) consumers and, 71–72
situational adaptations, 15–16, 95–96, evaluation for, 60
143 focus of, 62–63
skills, 54, 70 jargon for, 60–61
Skinfeel Profile Descriptive Analysis, 44 problems for, 62
snacks, 23 (table) for QDA, 53–55
sorting, 95, 117–118 training for, 66
spectrum descriptive analysis, 116 surface properties phase, 23 (table)
advanced applications for, 96
lexicon development in, 79–83, 79 T
(table), 80 (figure), 80 (table), tabular Texture Profile, 40, 40 (table)
81 (table), 82 (table), 84–85 target panel size, 27
(table), 85–86, 86 (figure) tasks, 61
for panel leaders, 86–87 tasting, 6
panel recognition in, 91–92 teaching values, 8–9, 26–27
panel training for, 87–89, 87 (table) team decisions, 14
panel validation in, 89–91, technical attributes, 61
91 (figure) technical descriptive method, 21–22
philosophy of, 77–78 temporal methods, 43, 139–143,
practical considerations for, 93, 140 (table), 142 (table), 143 (table)
94 (figure), 94 (table), 95 terminology, 38, 45, 80,
results reporting for, 93 126–127 (table), 126–128,
situational adaptations for, 95–96 129 (figure)
spider graphs, 67 testing
spider plots, 67, 70 (figure), 71 (figure) aptitude for, 5
standard evaluation protocol, 83, arrangement test, 6
84–85 (table) basic perpetual test, 6
Standard Guide for Sensory Claim consensus test results, 123 (figure)
Substantiation (ASTM), 72 dimensions test results,
Standard Guide for Two Sensory 124–125 (figure), 124–128,
Descriptive Analysis Approaches 125 (table), 126–127 (table),
(ASTM), 83 129 (figure), 129 (table),
standard hardness scale, 82, 82 (table) 130 (figure)
Stanford Research Institute (SRI), 52 discrimination tests, 55, 90–91,
statistics, 15, 35–36, 102–103 91 (figure)
Stone, Herbert, 52 for partial consensus profiling,
structured training, 8–10 1–2
studies, 14, 101–102, 105 perceptual recognition series, 6–7
subjects personal interviews for, 7
assessors and, 140–141 pilot tests, 12, 36–37, 63–65, 64
behavior of, 62, 65 (figure), 65 (figure)

BK-AST-MNL_13-200365-Index.indd 154 19/11/20 8:22 PM


INDEX 155

protocol, 92 calibration and, 37


with QDA, 63–65, 64 (figure), in Flavor Profile Method, 4–5
65 (figure) length of, 87, 87 (table)
ranking tests, 6 observations in, 12
in science, 70–71 of panels, 4, 7–10, 13–14, 30–34,
scores from, 7 56–63, 58 (figure), 86 (figure),
sensory acuity, 85–86 87–89, 87 (table), 141
sensory tests, 15 responsibility and, 5
time for, 63 structured, 8–10
validation and, 141–142 for subjects, 66
texture, 81–82, 81 (table), 82 (table), 88 tasks and, 61
Texture Profile Method, 1 texture, 88
adaptations in, 42–45
assessors in, 27–30 U
core philosophy of, 19–22 unconditional positive regard, 8–9
data collection in, 35–36 universal scales, 78–79, 79 (table),
language development for, 30–34 95
overview of, 22, 23 (table), 24, unstructured graphic rating scale, 58,
24 (table), 25–26 (table), 26–27 58 (figure)
panel performance measures for,
37–38 V
pilot tests for, 36–37 validation
results reporting for, 38–40, 39 of assessors, 109–110
(figure), 40 (table), 41 (figure) of panels, 89–91, 91 (figure)
summary of, 45–46 pilot tests and, 12, 36–37,
thresholds, 15 63–65, 64 (figure),
time, 56–58, 63. See also temporal 65 (figure)
methods testing and, 141–142
tools, 5, 71–72 viewpoints, 70
Traditional Flavor Profiles, 13 vocabulary development, 106–107
training, 86 (figure) voluntary participation, 54
assessors, 108–109
X Y graph, 74 (figure)
attributes in, 33

BK-AST-MNL_13-200365-Index.indd 155 19/11/20 8:22 PM


BK-AST-MNL_13-200365-Index.indd 156 19/11/20 8:22 PM

You might also like