Calculus PDF
Calculus PDF
                                                                  f ðxÞ ¼ 0
Mass balances of stirred tank reactors may lead
to ordinary differential equations:
                                                                  it may be more difficult to find the x satisfying
dy
dt
   ¼ f ½yðtÞ                                                     this equation. These problems are compounded
                                                                  when there are more unknowns, leading to
Radiative heat transfer may lead to integral                      simultaneous equations. If the unknowns
equations:                                                        appear in a linear fashion, then an important
                  Z1                                              consideration is the structure of the matrix
yðxÞ ¼ gðxÞ þ l        Kðx; sÞf ðsÞd s                            representing the equations; special methods
                  0
                                                                  are presented here for special structures.
   Even when the model is a differential equa-                    They are useful because they increase the speed
tion or integral equation, the most basic step in                 of solution. If the unknowns appear in a
                                                       Mathematics in Chemical Engineering              3
nonlinear fashion, the problem is much more            determinant are all zero, the value of the deter-
difficult. Iterative techniques must be used (i.e.,    minant is zero. If the elements of one row or
make a guess of the solution and try to improve        column are multiplied by the same constant, the
the guess). An important question then is              determinant is the previous value times that
whether such an iterative scheme converges.            constant. If two adjacent rows (or columns)
Other important types of equations are linear          are interchanged, the value of the new determi-
difference equations and eigenvalue problems,          nant is the negative of the value of the original
which are also discussed.                              determinant. If two rows (or columns) are
                                                       identical, the determinant is zero. The value
                                                       of a determinant is not changed if one row (or
1.1. Matrix Properties                                 column) is multiplied by a constant and added
                                                       to another row (or column).
A matrix is a set of real or complex numbers              A matrix is symmetric if
arranged in a rectangular array.
     2                                        3        aij ¼ aji
      a11          a12       ...     a1n
    6 a21          a22       ...     a2n      7
    6                                         7
A ¼ 6 ..            ..        ..      ..      7        and it is positive definite if
    4 .              .         .       .      5
         am1      am2        . . . amn
                                                                    n X
                                                                    X n
                                                       xT Ax ¼                   aij xi xj  0
The numbers aij are the elements of the matrix                      i¼1 j¼1
A0ij ¼ ð1Þiþj M ij
                                                       If AT ¼ A1 the matrix is orthogonal.
The value of jAj is given by                              Matrices are added and subtracted element
                                                       by element.
        X
        n                    X
                             n
jAj ¼          aij A0ij or         aij A0ij            A þ B is aij þ bij
         j¼1                 i¼1
A tridiagonal matrix is zero except for elements         may be the size of a disturbance, and the output is
along the diagonal and one element to the right          the gain [1]. If the rank is less than n, not all the
and left of the diagonal.                                variables are independent and they cannot all be
                                                         controlled. Furthermore, if the singular values
        8
        < 0 if j < i  1                                 are widely separated, the process is sensitive to
aij ¼     a otherwise                                    small changes in the elements of the matrix and
        : ij
          0 if j > i þ 1
                                                         the process will be difficult to control.
Block diagonal and pentadiagonal matrices also
arise, especially when solving partial differen-
tial equations in two- and three-dimensions.             1.2. Linear Algebraic Equations
               
A ¼ USV                                                  The U is upper triangular; it has zero elements
                                                         below the main diagonal and possibly nonzero
where S is a k  k diagonal matrix with diagonal         values along the main diagonal and above it
elements dii ¼ s i > 0 for 1  i  k. The eigen-        (see Fig. 1). The L is lower triangular. It has
values of S S are s 2i . The vectors ui for k þ 1  i    the value 1 in each element of the main
 m and vi for k þ 1  i  n are eigenvectors            diagonal, nonzero values below the diagonal,
associated with the eigenvalue zero; the eigen-          and zero values above the diagonal (see
values for 1  i  k are s 2i . The values of s i are    Fig. 1). The original problem can be solved
called the singular values of the matrix A. If A is      in two steps:
real, then U and Vare real, and hence orthogonal
matrices. The value of the singular value decom-         Ly ¼ f; Ux ¼ y solves Ax ¼ LUx ¼ f
position comes when a process is represented
by a linear transformation and the elements of A,           Each of these steps is straightforward
aij , are the contribution to an output i for a          because the matrices are upper triangular or
particular variable as input variable j. The input       lower triangular.
                                                   Mathematics in Chemical Engineering               5
kðAÞ ¼k A k k A1 k
   The determinant is given by the product of      Generally, the inverse is not used in this way
the diagonal elements of U. This should be         because it requires three times more operations
calculated as the LU decomposition is per-         than solving with an LU decomposition. How-
formed. If the value of the determinant is a       ever, if an inverse is desired, it is calculated
very large or very small number, it can be         most efficiently by using the LU decomposition
divided or multiplied by 10 to retain accuracy     and then solving
in the computer; the scale factor is then accu-
mulated separately. The condition number k          AxðiÞ ¼ bðiÞ
                                                           
can be defined in terms of the singular value        ðiÞ
                                                    bj ¼
                                                            0 j 6¼ i
decomposition as the ratio of the largest dii to            1 j¼i
6               Mathematics in Chemical Engineering
If
requires multiplying a matrix by a vector, which    dimensionless groups govern that phenomenon.
can be done very efficiently on parallel com-       In chemical reaction engineering the chemical
puters: For sparse matrices this is a viable        reaction stoichiometry can be written as
method. The original method was devised by
HESTENES and STIEFEL [8]; however, more recent      X
                                                    n
                                                           aij Ci ¼ 0; j ¼ 1; 2; . . . ; m
implementations use a preconditioned conju-          i¼1
gate gradient method because it converges
faster, provided a good “preconditioner” can        where there are n species and m reactions. Then
be found. The system of n linear equations          if a matrix is formed from the coefficients aij,
                                                    which is an n  m matrix, and the rank of the
Ax ¼ f
                                                    matrix is r, there are r independent chemical
                                                    reactions. The other nr reactions can be
where A is symmetric and positive definite, is to   deduced from those r reactions.
be solved. A preconditioning matrix M is
defined in such a way that the problem
is easy to solve exactly (M might be diagonal,      Consider a single nonlinear equation in one
for example). Then the preconditioned conju-        unknown,
gate gradient method is
                                                    f ðxÞ ¼ 0
Guess x0
Calculate r0 ¼ f  Ax0
Solve M t0 ¼ r0 ; and set p0 ¼ t0
                                                    In Microsoft Excel, roots are found by using
for k ¼ 1; n ðor until convergenceÞ                 Goal Seek or Solver. Assign one cell to be x, put
        r T tk                                      the equation for f(x) in another cell, and let Goal
ak ¼ Tk
      pk Apk
xkþ1 ¼ xk þ ak pk                                   Seek or Solver find the value of x making the
rkþ1 ¼ rk  ak Apk                                  equation cell zero. In MATLAB, the process is
Solve Mtkþ1 ¼ rkþ1                                  similar except that a function (m-file) is defined
      rT tkþ1
bk ¼ kþ1T                                           and the command fzero(‘f’, x0) provides the
         rk tk
pkþ1 ¼ tkþ1 þ bk pk                                 solution x, starting from the initial guess x0.
test for convergence                                   Iterative methods applied to single equations
enddo
                                                    include the successive substitution method
Note that the entire algorithm involves only        xkþ1 ¼ xk þ bf ðxk Þ 	 gðxk Þ
matrix multiplications. The generalized mini-
mal residual method (GMRES) is an iterative         and the Newton–Raphson method.
method that can be used for nonsymmetric
systems and is based on a modified Gram–                               f ðxk Þ
                                                    xkþ1 ¼ xk 
Schmidt orthonormalization. Additional infor-                       df =dxðxk Þ
mation, including software for a variety of
methods, is available [9–13].                       The former method converges if the derivative
   In dimensional analysis if the dimensions of     of g(x) is bounded [3]. The latter method
each physical variable Pj (there are n of them)
                                                        
are expressed in terms of fundamental measure-      dg 
                                                     ðxÞ  m for jx  aj < h
                                                    dx 
ment units mj (such as time, length, mass; there
are m of them):
                                                    is based on a Taylor series of the equation about
           a    a
½Pj  ¼ m1 1j m2 2j    mammj                     the k-th iterate:
The second and higher-order terms are neglec-          that is fast and guaranteed to converge, if the
ted and f (xkþ1) ¼ 0 to obtain the method.             root can be bracketed initially [15, p. 251].
                                       2 
                                                           In the method of bisection, if a root lies
df                      0            
  > 0; x1  x0  ¼  f ðx Þ   b; and d f   c   between x1 and x2 because f(x1) < 0 and f(x2)
dx 0                df =dxðx0 Þ 0     dx2 
     x                              x                  > 0, then the function is evaluated at the center,
                                                       xc ¼ 0.5 (x1 þ x2). If f (xc) > 0, the root lies
Convergence of the Newton–Raphson method               between x1 and xc. If f (xc) < 0, the root lies
depends on the properties of the first and second      between xc and x2. The process is then repeated.
derivative of f(x) [3, 14]. In practice the method     If f (xc) ¼ 0, the root is xc. If f (x1) > 0 and f (x2)
may not converge unless the initial guess is           > 0, more than one root may exist between x1
good, or it may converge for some parameters           and x2 (or no roots).
and not others. Unfortunately, when the method             For systems of equations the Newton–Raph-
is nonconvergent the results look as though a          son method is widely used, especially for equa-
mistake occurred in the computer program-              tions arising from the solution of differential
ming; distinguishing between these situations          equations.
is difficult, so careful programming and testing
are required. If the method converges the dif-         f i ðfxj gÞ ¼ 0; 1  i; j  n; where fxj g ¼ ðx1 ; x2 ; . . . ; xn Þ ¼ x
ference between successive iterates is some-
thing like 0.1, 0.01, 0.0001, 108. The error          Then, an expansion in several variables
(when it is known) goes the same way; the              occurs:
method is said to be quadratically convergent
when it converges. If the derivative is difficult                                 X
                                                                                  n
                                                                                    @f i
to calculate a numerical approximation may be          fi ðxkþ1 Þ ¼ f i ðxk Þ þ               jxk ðxkþ1  xkj Þ þ   
                                                                                  j¼1
                                                                                        @xj         j
used.
                                                      The Jacobian matrix is defined as
df     f ðxk þ eÞ  f ðxk Þ
       ¼
dx xk            e                                             
                                                               @f 
                                                       Jkij ¼  i 
                                                                @x j xk
In the secant method the same formula is used
as for the Newton–Raphson method, except that
the derivative is approximated by using the            and the Newton–Raphson method is
values from the last two iterates:                     X
                                                       n
                                                             Jkij ðxkþ1  xk Þ ¼ f i ðxk Þ
    
df     f ðxk Þ  f ðxk1 Þ                           j¼1
       ¼
dx  k
    x        xk  xk1
                                                       For convergence, the norm of the inverse of the
This is equivalent to drawing a straight line          Jacobian must be bounded, the norm of
through the last two iterate values on a plot of f     the function evaluated at the initial guess
(x) versus x. The Newton–Raphson method is             must be bounded, and the second derivative
equivalent to drawing a straight line tangent to       must be bounded [14, p. 115], [3, p. 12].
the curve at the last x. In the method of false           A review of the usefulness of solution meth-
position (or regula falsi), the secant method is       ods for nonlinear equations is available [16].
used to obtain xkþ1, but the previous value is         This review concludes that the Newton–Raph-
taken as either xk1 or xk. The choice is made so      son method may not be the most efficient.
that the function evaluated for that choice has        Broyden’s method approximates the inverse
the opposite sign to f (xkþ1). This method is          to the Jacobian and is a good all-purpose
slower than the secant method, but it is more          method, but a good initial approximation to
robust and keeps the root between two points at        the Jacobian matrix is required. Furthermore,
all times. In all these methods, appropriate           the rate of convergence deteriorates for large
strategies are required for bounds on the func-        problems, for which the Newton–Raphson
tion or when df/dx ¼ 0. Brent’s method com-            method is better. Brown’s method [16] is
bines bracketing, bisection, and an inverse            very attractive, whereas Brent’s is not worth
quadratic interpolation to provide a method            the extra storage and computation. For large
                                                     Mathematics in Chemical Engineering             9
systems of equations, efficient software is avail-   is used as the initial guess and the homotopy
able [11–13].                                        equation is solved for x1.
                 dx                                              R          1
x1;0 ¼ x0 þ Dt                                       ynþ1;i ¼       xn;i þ     x0;i
                 dt                                             Rþ1        Rþ1
10                  Mathematics in Chemical Engineering
and the equilibrium equation gives                        where A and B are constants that must be
                                                          specified by boundary conditions of some kind.
yn;i ¼ K n;i xn;i                                             When the equation is nonhomogeneous, the
                                                          solution is represented by the sum of a particu-
If these are combined,                                    lar solution and a general solution to the homo-
                                                          geneous equation.
                     R          1
K nþ1;i xnþ1;i ¼        xn;i þ     x0;i
                    Rþ1        Rþ1
                                                          xn ¼ xn;P þ xn;H
is obtained, which is a linear difference equa-           The general solution is the one found
tion. This particular problem is quite compli-            for the homogeneous equation, and the partic-
cated, and the interested reader is referred to           ular solution is any solution to the non-
[19, Chap. 6]. However, the form of the differ-           homogeneous difference equation. This can
ence equation is clear. Several examples are              be found by methods analogous to those used
given here for solving difference equations.              to solve differential equations: The method of
More complete information is available in [20].           undetermined coefficients and the method of
   An equation in the form                                variation of parameters. The last method
                                                          applies to equations with variable coefficients,
xnþ1  xn ¼ f nþ1                                         too. For a problem such as
xn ¼ f n xnþ1 f n xn ¼ gn
Hamilton–Cayley theorem [19, p. 127] states            2. Experimental data must be fit with a math-
that the matrix A satisfies its own characteristic        ematical model. The data have experimental
equation.                                                 error, so some uncertainty exists. The
                                                          parameters in the model as well as the
Pn ðAÞ ¼ ðAÞn þ a1 ðAÞn1 þ a2 ðAÞn2                  uncertainty in the determination of those
         þ    þ an1 ðAÞ þ an I ¼ 0                   parameters is desired.
    A laborious way to find the eigenvalues of a          These problems are addressed in this chap-
matrix is to solve the n-th order polynomial for      ter. Section 2.2 gives the properties of polyno-
the li — far too time consuming. Instead the          mials defined over the whole domain and
matrix is transformed into another form whose         Section 2.3 of polynomials defined on segments
eigenvalues are easier to find. In the Givens         of the domain. In Section 2.4, quadrature meth-
method and the Housholder method the matrix           ods are given for evaluating an integral. Least-
is transformed into the tridiagonal form; then,       squares methods for parameter estimation for
in a fixed number of calculations the eigenval-       both linear and nonlinear models are given in
ues can be found [15]. The Givens method              Sections 2.5. Fourier transforms to represent
requires 4 n3/3 operations to transform a real        discrete data are described in Section 2.7. The
symmetric matrix to tridiagonal form, whereas         chapter closes with extensions to two-dimen-
the Householder method requires half that num-        sional representations.
ber [14]. Once the tridiagonal form is found, a
Sturm sequence is applied to determine the
eigenvalues. These methods are especially use-
ful when only a few eigenvalues of the matrix         2.2. Global Polynomial Approximation
are desired.
    If all the eigenvalues are needed, the QR         A global polynomial Pm (x) is defined over the
algorithm is preferred [21].                          entire region of space
    The eigenvalues of a certain tridiagonal
                                                                  X
                                                                  m
matrix can be found analytically. If A is a           Pm ðxÞ ¼          cj xj
tridiagonal matrix with                                           j¼0
1. A function is known exactly at a set of points     Note that each coefficient of yj is a polynomial
   and an interpolating function is desired. The      of degree m that vanishes at the points {xj}
   interpolant may be exact at the set of points,     (except for one value of j) and takes the value of
   or it may be a “best fit” in some sense.           1.0 at that point, i.e.,
   Alternatively it may be desired to represent
   a function in some other way.                      Pm ðxj Þ ¼ yj j ¼ 1; 2; . . . ; m þ 1
12               Mathematics in Chemical Engineering
If the function f (x) is known, the error in the             to within multiplicative constants, which can be
approximation is [23]                                        set either by requiring the leading coefficient to
                                                             be one or by requiring the norm to be one.
                jxmþ1  x1 jmþ1
jerrorðxÞj 
                   ðm þ 1Þ!                                  Zb
                          ðmþ1Þ                                      WðxÞP2m ðxÞdx ¼ 1
       maxx1 xxmþ1 jf           ðxÞj
                                                             a
The evaluation of Pm (x) at a point other than               The polynomial Pm (x) has m roots in the closed
the defining points can be made with Neville’s               interval a to b.
algorithm [15]. Let P1 be the value at x of                     The polynomial
the unique function passing through the point
(x1, y1); i.e., P1¼ y1. Let P12 be the value at x of         pðxÞ ¼ c0 P0 ðxÞ þ c1 P1 ðxÞ þ    cm Pm ðxÞ
the unique polynomial passing through the
points x1 and x2. Likewise, Pijk . . . r is the              minimizes
unique polynomial passing through the points
xi, xj, xk, . . . , xr. The following scheme is used:                 Zb
                                                             I¼                WðxÞ½f ðxÞ  pðxÞ2 dx
                                                                      a
when
                                                                          Zb
                                                                               WðxÞ f ðxÞPj ðxÞdx
These entries are defined by using                           cj ¼         a
                                                                                                          ;
                                                                                           Wj
Piðiþ1ÞðiþmÞ ¼
                                                                               Zb
ðx  xiþm ÞPiðiþ1Þðiþm1Þ þ ðxi  xÞPðiþ1Þðiþ2ÞðiþmÞ       Wj ¼                  WðxÞP2j ðxÞdx
                         xi  xiþm                                                 a
Consider P1234: the terms on the right-hand side             Note that each cj is independent of m, the
of the equation involve P123 and P234. The                   number of terms retained in the series. The
“parents,” P123 and P234, already agree at points            minimum value of I is
2 and 3. Here i ¼ 1, m ¼ 3; thus, the parents
agree at xiþ1, . . . , xiþm1 already. The formula                         Zb                             X
                                                                                                          m
makes Pi (iþ1) . . . (iþm) agree with the function           I min ¼                   WðxÞ f 2 ðxÞdx              W j c2j
                                                                                                              j¼0
at the additional points xiþm and xi. Thus, Pi                                 a
                                                             Z1
The orthogonality includes a nonnegative                             Wðx2 ÞPk ðx2 ÞPm ðx2 Þxa1 dx ¼ 0
weight function, W (x)  0 for all a  x  b.                    0
This procedure specifies the set of polynomials              k ¼ 0; 1; . . .; m  1
                                                                               Mathematics in Chemical Engineering                            13
1                  1                pffiffiffiffiffiffiffi
                                       1 ffi
                                                              Chebyshev                                 Tiþ1¼2xTiTi1
                                      1x2
                                         q1    pq
0                   1                x (1x)                  Jacobi (p, q)
                                         2
1                  1                ex                      Hermite                                   Hiþ1¼2xHi2 i Hi–1
                                       c x
0                   1                xe                       Laguerre (c)                              (iþ1) Liþ1c¼(xþ2 iþcþ1) L ci (iþc) Li1c
0                   1                1                        shifted Legendre
0                   1                1                        shifted Legendre, function of x2
where a ¼ 1 is for planar, a ¼ 2 for cylindrical,                                 Rational polynomials are useful for approxi-
and a ¼ 3 for spherical geometry. These func-                                     mating functions with poles and singularities,
tions are useful if the solution can be proved to                                 which occur in Laplace transforms (see
be an even function of x.                                                         Section 4.2).
                                                                                     Fourier series are discussed in Section 4.1.
Rational Polynomials. Rational polynomials                                        Representation by sums of exponentials is also
are ratios of polynomials. A rational polynomial                                  possible [24].
Ri(iþ1) . . . (iþm) passing through m þ 1 points                                     In summary, for discrete data, Legendre
                                                                                  polynomials and rational polynomials are
                                                                                  used. For continuous data a variety of orthogo-
yi ¼ f ðxi Þ; i ¼ 1; . . . ; m þ 1
                                                                                  nal polynomials and rational polynomials are
                                                                                  used. When the number of conditions (discrete
is
                                                                                  data points) exceeds the number of parameters,
                    Pm ðxÞ p0 þ p1 x þ    þ pm xm
                                                                                  then see Section 2.5.
Riðiþ1ÞðiþmÞ ¼          ¼                              ;
                    Qn ðxÞ   q 0 þ q 1 x þ    þ q n xn
mþ1¼ mþnþ1                                                                        2.3. Piecewise Approximation
An alternative condition is to make the rational                                  Piecewise approximations can be developed
polynomial agree with the first m þ 1 terms in                                    from difference formulas [3]. Consider a case
the power series, giving a Pade approximation, i.                                in which the data points are equally spaced
e.,
                                                                                   xnþ1  xn ¼ Dx
Keeping only the first two terms gives a straight           Thus, the approximation is
line through (x0, y0)  (x1, y1); keeping the first
three terms gives a quadratic function of posi-                      X
                                                                     NT                   X
                                                                                          NT
                                                            yðxÞ ¼         ci N i ðxÞ ¼         yðxi ÞN i ðxÞ
tion going through those points plus (x2, y2).                       i¼1                  i¼1
The value a ¼ 0 gives x ¼ x0; a ¼ 1 gives
x ¼ x1, etc.                                                where ci ¼ y (xi). For convenience, the trial
   Backward differences are defined by                      functions are defined within an element by
                                                            using new coordinates:
ryn ¼ yn  yn1
                                                                 x  xi
r2 yn ¼ ryn  ryn1 ¼ yn  2yn1 þ yn2                     u¼
                                                                  Dxi
The interpolation polynomial of order n                     The Dxi need not be the same from element to
through the points y0, y1, y2, . . . is                   element. The trial functions are defined as Ni (x)
                                                            (Fig. 3 A) in the global coordinate system and
                   aða þ 1Þ 2
ya ¼ y0 þ ary0 þ
                      2!
                           r y0 þ   þ                    NI (u) (Fig. 3 B) in the local coordinate system
aða þ 1Þ    ða þ n  1Þ n
                                                            (which also requires specification of the
            n!
                          r y0                              element). For xi < x < xiþ1
                                                                     X
                                                                     2
                                                            yðxÞ ¼         ceI N I ðuÞ                                ð2Þ
                                                                     I¼1
with ceI ¼ yðxi Þ; i ¼ ðe  1Þ2 þ I                               The formulas for the cubic spline are derived
                                                              as follows for one region. Since the function is a
in the e  th element
                                                              cubic function the third derivative is constant
                                                              and the second derivative is linear in x. This is
                                                              written as
   Hermite cubic polynomials can also be used;
these are continuous and have continuous first                                                                         x  xi
                                                              C 00i ðxÞ ¼ C00i ðxi Þ þ ½C 00i ðxiþ1 Þ  C 00i ðxi Þ
derivatives at the element boundaries [3].                                                                              Dxi
Splines. Splines are functions that match                     and integrated once to give
given values at the points x1, . . . , xNT, shown
in Figure 5, and have continuous derivatives up                C0i ðxÞ ¼ C 0i ðxi Þ þ C 00i ðxi Þðx  xi Þþ
to some order at the knots, or the points                                                     ðx  xi Þ2
                                                               ½C 00i ðxiþ1 Þ  C00i ðxi Þ
x2, . . . , xNT1. Cubic splines are most com-                                                  2Dxi
mon. In this case the function is represented by
a cubic polynomial within each interval and has               and once more to give
continuous first and second derivatives at the
knots.                                                         Ci ðxÞ ¼ C i ðxi Þ þ C 0i ðxi Þðx  xi Þ þ C 00i ðxi Þ
   Consider the points shown in Figure 5A. The                 ðx  xi Þ2                                 ðx  xi Þ3
                                                                          þ ½C 00i ðxiþ1 Þ  C 00i ðxi Þ
notation for each interval is shown in Figure 5B.                  2                                        6Dxi
Within each interval the function is represented
as a cubic polynomial.                                        Now
C i ðxÞ ¼ a0i þ a1i x þ a2i x2 þ a3i x3                       yi ¼ Ci ðxi Þ; y0i ¼ C 0i ðxi Þ; y00i ¼ C00i ðxi Þ
16              Mathematics in Chemical Engineering
xZ
 0 þh             Z1                                                            X2     Z1                          
                                                                                                           1      1
       yðxÞdx ¼           ya hda                                      ¼ Dxi         ceI N I ðuÞdu ¼ Dxi ce1 þ ce2
                                                                                I¼1
                                                                                                           2      2
 x0                   0                                                                  0
 h             1                                                        Dxi
¼ ðy0 þ y1 Þ  h3 y000 ðjÞ; x0  j  x0 þ h                           ¼     ðy þ yiþ1 Þ
 2            12                                                         2 i
This corresponds to passing a straight line                          Since ce1 ¼ yi and ce2 ¼ yiþ1 , the result is the
through the points (x0, y0), (x1, y1) and integrat-                  same as the trapezoid rule. These formulas can
ing under the interpolant. For equally spaced                        be added together to give linear elements:
points at a ¼ x0, a þ Dx ¼ x1, a þ 2 Dx ¼
x2, . . . , a þ N Dx ¼ xN, a þ (N þ 1) Dx ¼ b ¼                      Zb                  X Dxe
xnþ1, the trapezoid rule is obtained.                                       yðxÞdx ¼                   ðye1 þ ye2 Þ
                                                                                             e
                                                                                                   2
                                                                      a
Trapezoid Rule.
                                                                     If the quadratic expansion is used (Eq. 3), the
Zb                                                                   endpoints of the element are xi and xiþ2, and
                 h
      yðxÞdx ¼      y þ 2y1 þ 2y2 þ    þ 2yN þ yNþ1 Þ þ Oðh3 Þ    xiþ1 is the midpoint, here assumed to be equally
                 2 0
 a
                                                                     spaced between the ends of the element:
x0Zþ2h                Z2                                                        X
                                                                                3    Z1
                                   h
           yðxÞdx ¼        ya hda ¼ ðy0 þ 4y1 þ y2 Þ                  ¼ Dxi       ceI N I ðuÞdu
                                   3                                             I¼1
  x0                  0                                                                  0
                                                                                                            
       5
     h ðIVÞ                                                                             1     2      1
      y ðjÞ; x0  j  x0 þ 2h                                        ¼ Dxe          ce1 þ ce2 þ ce3
     90 0                                                                               6     3      6
18                    Mathematics in Chemical Engineering
For many elements, with different Dxe, qua-                 Table 2. Gaussian quadrature points and weights*
dratic elements:                                            N                         xi                          Wi
Zb                                                          1                         0.5000000000                0.6666666667
                X Dxe
       yðxÞ ¼                 ðye1 þ 4ye2 þ ye3 Þ           2                         0.2113248654                0.5000000000
                 e
                          6                                                           0.7886751346                0.5000000000
a
                                                            3                         0.1127016654                0.2777777778
                                                                                      0.5000000000                0.4444444445
If the element sizes are all the same this gives                                      0.8872983346                0.2777777778
Simpson’s rule.                                             4                         0.0694318442                0.1739274226
    For cubic splines the quadrature rule within                                      0.3300094783                0.3260725774
                                                                                      0.6699905218                0.3260725774
one element is                                                                        0.9305681558                0.1739274226
                                                            5                         0.0469100771                0.1184634425
xZiþ1                                                                                 0.2307653450                0.2393143353
                  1
        Ci ðxÞdx ¼ Dxi ðyi þ yiþ1 Þ                                                   0.5000000000                0.2844444444
                  2
 xi                                                                                   0.7692346551                0.2393143353
                                                                                      0.9530899230                0.1184634425
              1
               Dx3 ðy00 þ y00iþ1 Þ
              24 i i                                        *
                                                              For a given N the quadrature points x2, x3, . . . , xNP1 are given
                                                            above. x1 ¼ 0, xNP ¼ 1. For N ¼ 1, W1 ¼ W3 ¼ 1/6 and for N  2,
For the entire interval the quadrature formula              W1 ¼ WNP ¼ 0.
is
xZNT
                       X
                     1 NT1
       yðxÞdx ¼            Dxi ðyi þ yiþ1 Þ                 polynomials give the quadrature formula
                     2 i¼1
 x1
                 X
              1 NT1                                        Z1
                   Dx3 ðy00 þ y00iþ1 Þ                                         X
                                                                                n
              24 i¼1 i i                                         ex yðxÞdx ¼         W i yðxi Þ
                                                                                i¼1
                                                            0
The quadrature is exact when y is a polynomial              (points and weights are available in mathemat-
of degree 2 m  1 in x. The m weights and m                 ical tables) [23].
Gauss points result in 2 m parameters, chosen
to exactly represent a polynomial of degree                 Romberg’s method uses extrapolation tech-
2 m  1, which has 2 m parameters. The Gauss                niques to improve the answer [15]. If I1 is
points and weights are given in Table 2. The                the value of the integral obtained by using
weights can be defined with W (x) in the inte-              interval size h ¼ Dx, I2 the value of I obtained
grand as well.                                              by using interval size h/2, and I0 the true value
   For orthogonal collocation on finite ele-                of I, then the error in a method is approximately
ments the quadrature formula is                             hm, or
Z1                   X           X
                                 NP
       yðxÞdx ¼            Dxe          W j yðxeJ Þ         I 1  I 0 þ chm
                      e          j¼1
0
                                                                           m
   Each special polynomial has its own quad-                I2  I0 þ c
                                                                           h
rature formula. For example, Gauss–Legendre                                2
                                                       Mathematics in Chemical Engineering                        19
Replacing the  by an equality (an approxima-           sum of squares of the deviation between the
tion) and solving for c and I0 give                     experimental data and the theoretical equation
       2m I 2  I 1                                              XN                                    
I0 ¼                                                                  yi  yðxi ; a1 ; a2 ; . . . ; aM Þ 2
        2m  1                                          x2 ¼
                                                                 i¼1
                                                                                   s  i
I1                         I2                I3   I4             XN
                                                                     ½yi  yðxi ; a1 ; a2 ; . . . ; aM Þ2
                           J1                J2   J3    s2 ¼
                                                                 i¼1
                                                                                     N
                                             K1   K2
                                                  L1
likelihood.
   In a least-squares parameter estimation, it is       The parameters are easily determined using
desired to find parameters that minimize the            computer software. In Microsoft Excel, the
20                  Mathematics in Chemical Engineering
                                                          yk ¼ yðtk Þ; tk ¼ kD; k ¼ 0; 1; 2; . . . ; N  1
Nonlinear Regression. In nonlinear regres-
sion, the same procedure is followed except
that an optimization routine must be used to              and the sampling rate is D; with only N values
find the minimum x2 (see Chap. 10).                       {yk} the complete Fourier transform Y (v)
                                                          cannot be determined. Calculate the value Y
                                                          (vn) at the discrete points
2.6. Fourier Transforms of Discrete
Data [15]                                                 vn ¼
                                                                  2pn        N
                                                                      ; n ¼  ; . . . ; 0; . . . ;
                                                                                                   N
                                                                  ND         2                     2
Suppose a signal y (t) is sampled at equal                        X
                                                                  N1
                                                          Yn ¼          yk e2pikn=N
intervals                                                         k¼0
                                                          Yðvn Þ ¼ DY n
yn ¼ yðnDÞ; n ¼ . . . ;  2;  1; 0; 1; 2; . . .
D ¼ sampling rate
                                                          The discrete inverse Fourier transform is
ðe:g:; number of samples per secondÞ
          Z1
YðvÞ ¼         yðtÞeivt dt
                                                             The fast Fourier transform (FFT) is used to
          1
                                                          calculate the Fourier transform as well as the
               Z1
                                                          inverse Fourier transform. A discrete Fourier
          1                                               transform of length N can be written as the sum
yðtÞ ¼              YðvÞeivt dv
         2p
               1                                         of two discrete Fourier transforms, each of
                                                          length N/2, and each of these transforms is
(For definition of i, see Chap. 3.) The                   separated into two halves, each half as long.
Nyquist critical frequency or critical angular            This continues until only one component is
                                                    Mathematics in Chemical Engineering              21
left. For this reason, N is taken as a power of      each of the grid points:
2, N ¼ 2p.
   The vector {yj} is filled with zeroes, if need
                                                     dY      1 X N
be, to make N ¼ 2p for some p. For the               dx
                                                        jn ¼       y0 e2ikpxn =L
                                                             L k¼N k
computer program, see [15, p. 381]. The stan-
dard Fourier transform takes N2 operations to
calculate, whereas the fast Fourier transform        Any nonlinear term can be treated in the same
takes only N log2 N. For large N the difference      way: Evaluate it in real space at N points and
is significant; at N ¼ 100 it is a factor of 15,     take the Fourier transform. After processing
but for N ¼ 1000 it is a factor of 100.              using this transform to get the transform of a
   The discrete Fourier transform can also be        new function, take the inverse transform to
used for differentiating a function; this is used    obtain the new function at N points. This is
in the spectral method for solving differential      what is done in direct numerical simulation of
equations. Consider a grid of equidistant            turbulence (DNS).
points:
                                                L
xn ¼ nDx; n ¼ 0; 1; 2; . . . ; 2 N  1; Dx ¼         2.7. Two-Dimensional Interpolation
                                               2N
                                                     and Quadrature
the solution is known at each of these grid
points {Y(xn)}. First, the Fourier transform is      Bicubic splines can be used to interpolate a set
taken:                                               of values on a regular array, f (xi, yj). Suppose
                                                     NX points occur in the x direction and NY points
        1 2XN1                                      occur in the y direction. PRESS et al. [15] suggest
yk ¼            Yðxn Þe2ikpxn =L
       2 N n¼0                                       computing NY different cubic splines of size NX
                                                     along lines of constant y, for example, and
The inverse transformation is                        storing the derivative information. To obtain
                                                     the value of f at some point x, y, evaluate each
           1 X N                                     of these splines for that x. Then do one spline of
YðxÞ ¼           y e2ikpx=L
           L k¼N k                                  size NY in the y direction, doing both the deter-
                                                     mination and the evaluation.
which is differentiated to obtain                        Multidimensional integrals can also be bro-
                                                     ken down into one-dimensional integrals. For
dY 1 X N
           2pik 2ikpx=L
                                                     example,
  ¼      y     e
dx L k¼N k L
                                                      Zb f Z2 ðxÞ                   Zb
                                                                    zðx; yÞdxdy ¼        GðxÞdx;
Thus, at the grid points                               a f 1 ðxÞ                    a
                                                                      fZ
                                                                       2 ðxÞ
dY      1 X N
                2pik 2ikpxn =L
   jn ¼       y     e                                  GðxÞ ¼               zðx; yÞdx
dx      L k¼N k L
                                                                      f 1 ðxÞ
The variable i is the imaginary unit which has       Since the arctangent repeats itself in multiples
the property                                         of p rather than 2 p, the argument must be
                                                     defined carefully. For example, the u given
i2 ¼ 1                                              above could also be the argument of  (x þi y).
                                                     The function z ¼ cos u þ i sin u obeys |z| ¼ |cos u
The real and imaginary parts of a complex            þ i sin u| ¼ 1.
number are often referred to:                           The rules of equality, addition, and multipli-
                                                     cation are
     Re    ðzÞ ¼ x;        Im    ðzÞ ¼ y
                                                     z1 ¼ x1 þ i y1 ; z2 ¼ x2 þ i y2
A complex number can also be represented             Equality:
graphically in the complex plane, where the
                                                     z1 ¼ z2 if and only if x1 ¼ x2 and y1 ¼ y2
real part of the complex number is the abscissa
and the imaginary part of the complex number         Addition:
is the ordinate (see Fig. 7).                        z1 þ z2 ¼ ðx1 þ x2 Þ þ iðy1 þ y2 Þ
    Another representation of a complex number       Multiplication:
is the polar form, where r is the magnitude and
                                                     z1 z2 ¼ ðx1 x2  y1 y2 Þ þ iðx1 y2 þ x2 y1 Þ
u is the argument.
                  pffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi                   The last rule can be remembered by
r ¼ jx þ i yj ¼    x2 þ y2 ; u ¼ argðx þ i yÞ
                                                     using the standard rules for multiplication,
Write                                                keeping the imaginary parts separate, and
                                                     using i2 ¼  1. In the complex plane, addition
z ¼ x þ i y ¼ rðcosu þ i sinuÞ
                                                     is illustrated in Figure 8. In polar form, multi-
                                                     plication is
so that
                                                     z1 z2 ¼ r1 r2 ½cosðu1 þ u2 Þ þ i sinðu1 þ u2 Þ
x ¼ r cos u; y ¼ r sin u
                                                     The magnitude of z1 þ z2 is bounded by
and
                                                     jz1 
 z2 j  jz1 j þ jz2 j and jz1 j  jz2 j  jz1 
 z2 j
           y
u ¼ arctan
           x
                                                     as can be seen in Figure 8. The magnitude and
                                                     arguments in multiplication obey
Roots of a complex number are complicated by          The second equation follows from the defini-
careful accounting of the argument                    tions
                                                                  ez þ ez            ez  ez
                                                      cosh z 	             ; sinh z 	
z ¼ w1=n with w ¼ Rðcos Q þ i sin QÞ; 0  Q  2p                      2                   2
sinðz1 þ z2 Þ ¼ sin z1 cos z2 þ cos z1 sin z2         Let f (z) be a single-valued continuous function
                                                      of z in a domain D. The function f (z) is
                                                      differentiable at the point z0 in D if
The same is true of hyperbolic functions. The
absolute boundaries of sin z and cos z are not              f ðz0 þ hÞ  f ðz0 Þ
                                                      lim
bounded for all z.                                    h!0            h
   Trigonometric identities can be defined by
using                                                 exists as a finite (complex) number and is
                                                      independent of the direction in which h tends
eiu ¼ cos u þ i sin u                                 to zero. The limit is called the derivative, f 0 (z0).
                                                      The derivative now can be calculated with h
                                                      approaching zero in the complex plane, i.e.,
For example,                                          anywhere in a circular region about z0. The
                                                      function f(z) is differentiable in D if it is dif-
eiðaþbÞ ¼ cosða þ bÞ þ i sinða þ bÞ                   ferentiable at all points of D; then f(z) is said to
     ¼ eia eib ¼ ðcos a þ i sin aÞ                    be an analytic function of z in D. Also, f(z) is
     ðcos b þ i sin bÞ
                                                      analytic at z0 if it is analytic in some e neigh-
                                                      borhood of z0. The word analytic is sometimes
     ¼ cos a cos b  sin a sin b                      replaced by holomorphic or regular.
     þ i ðcos a sin b þ cos b sin aÞ                     The Cauchy–Riemann equations can be used
                                                      to decide if a function is analytic. Set
Equating real and imaginary parts gives
                                                      f ðzÞ ¼ f ðx þ i yÞ ¼ uðx; yÞ þ i vðx; yÞ
sinða þ bÞ ¼ cos a sin b þ cos b sin a                Theorem [30, p. 51]. Suppose that f (z) is
                                                      defined and continuous in some neighborhood
     The logarithm is defined as                      of z ¼ z0. A necessary condition for the exis-
                                                      tence of f 0 (z0) is that u (x, y) and v (x, y) have
ln z ¼ lnjzj þ i arg z                                first-order partials and that the Cauchy–Rie-
                                                      mann conditions (see below) hold.
and the various determinations differ by multi-       @u @v     @u    @v
                                                        ¼   and    ¼    at z0
ples of 2 pi. Then,                                   @x @y     @y    @x
eln z ¼ z
                                                      Theorem [30, p. 61]. The function f (z) is
lnðez Þ  z 	 0ðmod 2 piÞ                             analytic in a domain D if and only if u and v
                                                      are continuously differentiable and satisfy the
Also,                                                 Cauchy–Riemann conditions there.
                                                         If f1(z) and f2(z) are analytic in domain D,
lnðz1 z2 Þ  ln z1  ln z2 	 0ðmod 2 piÞ              then a1f1(z) þ a2f2(z) is analytic in D for any
                                                            Mathematics in Chemical Engineering               25
Maximum Principle. If f (z) is analytic in a               Also if s (t) is the arc length on C and l (C) is
domain D and continuous in the set consisting              the length of C
of D and its boundary C, and if | f (z)|  M on C,
                                                                   
then | f (z)| < M in D unless f (z) is a constant          Z
                                                           
                                                                    
                                                                    
                                                            f ðzÞdz  max jf ðzÞjlðCÞ
[30, p. 134].                                                      
                                                                      z2C
                                                               C
C : z ¼ zðtÞ; 0  t  1
                                                           Cauchy’s Theorem [25, 30, p. 111]. Suppose f
where z (t) is a continuous function of bounded            (z) is an analytic function in a domain D and C
variation; C is oriented such that z1 ¼ z (t1)             is a simple, closed, rectifiable curve in D such
precedes the point z2 ¼ z (t2) on C if and only if         that f (z) is analytic inside and on C. Then
t1 < t2. Define                                            I
                                                                   f ðzÞdz ¼ 0                                ð4Þ
                                                           C
Z                   Z1
        f ðzÞdz ¼            f ½z ðtÞ dz ðtÞ
 C                   0                                     If D is simply connected, then Equation (4)
                                                           holds for every simple, closed, rectifiable curve
                                                           C in D. If D is simply connected and if a and b
The integral is linear with respect to the inte-           are any two points in D, then
grand:
                                                           Zb
R
 C ½a1 f 1 ðzÞ þ a2 f 2 ðzÞdz                                     f ðzÞdz
      R                   R                                 a
¼ a1 C f 1 ðzÞdz þ a2 C f 2 ðzÞdz
Z                   Z1
        f ðzÞdz ¼        f ½zðtÞz0 ðtÞdt                  Power Series. If f (z) is analytic interior to a
 C                  0                                      circle |z  z0| < r0, then at each point inside the
                                                                 Mathematics in Chemical Engineering                27
n ¼ 0; 
 1; 
 2; . . . ;
                                                                  all m < n, then z0 is called a pole of order m. It is
                                                                  a simple pole if m ¼ 1. In such a case,
and C is a closed curve counterclockwise in R.
                                                                             A1    X
                                                                                    1
                                                                  f ðzÞ ¼         þ    An ðz  z0 Þn
                                                                            z  z0 n¼0
Singular Points and Residues [33, p. 159,
30, p. 180]. If a function in analytic in every
                                                                  If a function is not analytic at z0 but can be made
neighborhood of z0, but not at z0 itself, then z0 is
                                                                  so by assigning a suitable value, then z0 is a
called an isolated singular point of the function.
                                                                  removable singular point.
About an isolated singular point, the function
                                                                      When f (z) has a pole of order m at z0,
can be represented by a Laurent series.
        pðz0 Þ
A1 ¼
        q0 ðz0 Þ
             1
f 1 ðzÞ ¼
            1z                                           3.5. Other Results
is the analytic continuation onto the entire z            Theorem [32, p. 84]. Let P (z) be a polynomial
plane except for z ¼ 1.                                   of degree n having the zeroes z1, z2, . . . , zn
                                                      Mathematics in Chemical Engineering                              29
and let p be the least convex polygon contain-         4. Integral Transforms [34–39]
ing the zeroes. Then P0 (z) cannot vanish any-
where in the exterior of p.                            4.1. Fourier Transforms
   If a polynomial has real coefficients, the
roots are either real or form pairs of complex         Fourier Series [40]. Let f (x) be a function that
conjugate numbers.                                     is periodic on  p < x < p. It can be expanded
   The radius of convergence R of the Taylor           in a Fourier series
series of f (z) about z0 is equal to the distance
from z0 to the nearest singularity of f (z).                       a0 X 1
                                                       f ðxÞ ¼       þ     ðan cos n x þ bn sin n xÞ
                                                                   2   n¼1
It is desired that                                                 Zp
                                                               1
                                                        bn ¼            f ðxÞsin n x dx
                                                               p
U ðj; hÞ ¼ u½^xðj; hÞ; ^yðj; hÞ                                   p
be a harmonic function of j and h.                     The values {an} and {bn} are called the finite
                                                       cosine and sine transform of f, respectively.
Theorem [30, p. 284]. The transformation               Because
z ¼ f ðzÞ                                       ð6Þ              1
                                                        cos n x ¼ ðeinx þ einx Þ
                                                                 2
                                                                              1 inx
takes all harmonic functions of x and y into            and sin n x ¼
                                                                              2i
                                                                                 ðe  einx Þ
harmonic functions of j and h if and only if
either f (z) or f  (z) is an analytic function of
z ¼ j þi h.                                            the Fourier series can be written as
   Equation (6) is a restriction on the transfor-
mation which ensures that                                              X
                                                                       1
                                                       f ðxÞ ¼               cn einx
                                                                   n¼1
     @2 u @2 u          @2 U @2 U
if       þ     ¼ 0 then     þ 2 ¼0
     @x2 @y2            @z2  @h                        where
                                                               8
Such a mapping with f (z) analytic and f 0 (z) 6¼ 0            >
                                                               >
                                                                 1
                                                               < 2 ðan þ i bn Þ            for n  0
is a conformal mapping.                                cn ¼
                                                               >
                                                               >
    If Laplace’s equation is to be solved in the               : 1 ðan  i bn Þ          for n < 0
                                                                 2
region exterior to a closed curve, then the point
at infinity is in the domain D. For flow in a long
channel (governed by the Laplace equation) the         and
inlet and outlet are at infinity. In both cases the
transformation                                                         Zp
                                                                1
                                                       cn ¼                 f ðxÞeinx dx
                                                               2p
      azþb                                                             p
z¼
      z  z0
                                                       If f is real
takes z0 into infinity and hence maps D into a                     
bounded domain D .                                    cn ¼ cn :
30                     Mathematics in Chemical Engineering
@T @ 2 T
                                                             If f (x) is continuous and piecewise continu-
   ¼ 2                                                       ously differentiable,
@t  @x
T ðx; 0Þ ¼ f ðxÞ
                                                             Z1
T ðp; tÞ ¼ T ðp; tÞ                                                 f ðxÞeivx dx
                                                             1
Let
                                                             converges for each v, and
      X
      1
T¼          cn ðtÞeinx                                       lim f ðxÞ ¼ 0
      1                                                     x!
1
Then,                                                        then
                                                                       
X
1                          X
                           1                                         df
  dcn           inx                    2      inx          F            ¼  ivF ½f 
            e          ¼        cn ðtÞðn Þe                         dx
1
      dt                   1
                                                             Z1
Fourier Transform [40]. When the function                            jf ðxÞj2 dx
f(x) is defined on the entire real line, the Fourier         1
                                                                                    Mathematics in Chemical Engineering                           31
has a finite value. If f (x) and g (x) are square                                    This is also the total power in a signal, which
integrable, the product f (x) g (x) is absolutely                                    can be computed in either the time or the
integrable and satisfies the Schwarz inequality:                                     frequency domain. Also
1             2                                                                     Z1                             Z1
Z                                                                                        ^f ðvÞ^g ðvÞdv ¼ 2 p
                                                                                                                        f ðxÞg ðxÞdx
 f ðxÞ g ðxÞdx
              
                                                                                   1                              1
 1
Z1                             Z1                                                               Z1
                                                                                            1
      j^f ðvÞj2 dv ¼ 2 p            jf ðxÞj2 dx                                       ¼              eivx0 ^f ðvÞ^h ðvÞdv
                                                                                           2p
1                             1                                                               1
32                  Mathematics in Chemical Engineering
Theorem. The product                                           The finite Fourier sine and cosine transforms
                                                            are
^f ðvÞ^h ðvÞ
                                                                                          Zp
                                                                                      2
                                                            f s ðnÞ ¼ Fns ½f  ¼               f ðxÞsin n xdx;
                                                                                      p
                                                                                          0
is the Fourier transform of the convolution
product f  h. The convolution permits finding              n ¼ 1; 2; . . . ;
                                                            n ¼ 0; 1; 2; . . .
@T @ 2 T
   ¼ 2                                                                    X
                                                                          1
@t  @x                                                      f ðxÞ ¼              f s ðnÞ sin n x;
T ðx; 0Þ ¼ f ðxÞ;  1 < x < 1                                              n¼1
                                                                          1         X1
T bounded                                                   f ðxÞ ¼         f ð0Þ þ     f c ðnÞ cos n x
                                                                          2 c       n¼1
The solution is
                                                            f, f 0 are continuous, f 00 is piecewise continuous
^    tÞ ¼ ^f ðvÞev t                                       on 0  x  p.
                   2
Tðv;
                    Z1
                1
                         eivx ^f ðvÞev t dv
                                         2
T ðx; tÞ ¼
               2p
                    1
4pt
                                                    1                                               ðpxÞ
                                                    n                                                 p
                                                    ð1Þnþ1                                         x
                                                       n                                            p
Also,                                               1ð1Þn
                                                                                                    1
                                                       n
                                                                                                    
                                                                                                       ðp  cÞ x ðx  cÞ
     df
                                                    p
                                                    n2   sin n cð0 < c < pÞ
Fns       ¼ nFnc ½f                                                                                    c ðp  xÞ ðx  cÞ
     dx                                                                                             
                                                                                                          x         ðx < cÞ
                                                  p
                                                    n cos n cð0         < c < pÞ
     df                2             2                                                                   px           ðxÞ
Fnc       ¼ nFns ½f   f ð0Þ þ ð1Þn f ðpÞ
     dx                p             p              p2 ð1Þn1                      n
                                                                                        
                                                         n 2½1ð1Þ
                                                                 n3                                 x2
                                                                  2
                                                                     
                                                    pð1Þn n63  pn                                 x3
When two functions F (x) and G (x) are defined
on the interval  2 p < x < 2 p, the function                 ½1  ð1Þn ec p                      ecx
                                                       n
                                                    n2 þc2
                                                       n                                            sinh cðpxÞ
                                                    n2 þc2
                    Zp                                                                                sinh c p
          
F ðxÞ G ðxÞ ¼            f ðx  yÞ g ðyÞdy             n
                                                              ðjkj 6¼ 0; 1; 2; . . .Þ               sin kðpxÞ
                                                    n2 k2                                            sin k p
                    p
                                                          n cx
                                                    ð1Þ e 1
              Z1                                      n2 þc2
                                                                                                          1 cx
                                                                                                          ce
Fvs ½f  	         f ðxÞsin v x dx;
                                                       1                                                  cosh cðpxÞ
              0                                     n2 þc2                                                 c sinh c p
                                                          it must be that
Thus, the techniques described herein can be
                                                          F 2 ¼ F 1 þ NðtÞ
applied only to linear problems. Generally, the
assumptions made below are that F (t) is at least                    ZT         
                                                          where             N ðtÞ dt ¼ 0 for every T
piecewise continuous, that it is continuous in
each finite interval within 0 < t < 1, and that it
                                                                     0
may take a jump between intervals. It is also             Laplace transforms can be inverted by using
of exponential order, meaning eat|F (t)| is              Table 5, but knowledge of several rules is
bounded for all t > T, for some finite T.                 helpful.
   The unit step function is
                                                          Substitution.
              0    0t<k
Sk ðtÞ ¼ f
              1           t>k                             f ðs  aÞ ¼ L½eat FðtÞ
and its Laplace transform is This can be used with polynomials. Suppose
                  eks                                                1   1    2sþ3
L½Sk ðtÞ ¼                                               f ðsÞ ¼       þ    ¼
                   s                                                  s s þ 3 s ðs þ 3Þ
          1                                                           1
L½1 ¼                                                    L½1 ¼
          s                                                           s
                                                                            Mathematics in Chemical Engineering          35
L [F]                                            F (t)
1
s                                                1
1
s2                                               t
1                                                 tn1
sn                                               ðn1Þ!
p1ffi                                              p1ffiffiffiffiffi
  s                                                pt
                                                  pffiffiffiffiffiffiffi
s3/2                                            2 t=p
                                                                             Figure 10. Square wave function
G ðkÞ
 sk ðk       > 0Þ                                tk1
 1
sa                                              ea t
                                                                             has the Laplace transform
   1
ðsaÞn
         ðn ¼ 1; 2; . . .Þ                         1
                                                 ðn1Þ! t
                                                          n1 a t
                                                             e
                                                                                            1    1
 G ðkÞ                                                                       L½S ðtÞ ¼
ðsaÞk
         ðk > 0Þ                                 tk1ea t                                   s 1  ehs
    1                                             1     at
                                                 ab ðe        eb t Þ
ðsaÞðsbÞ
                                                                                The Dirac delta function d (t  t0) has the
     s
ðsaÞðsbÞ
                                                  1
                                                 ab ðae
                                                         at
                                                                 beb t Þ    property
   1                                             1
s2 þa2                                           a sin a t                   Z1
                                                                                  dðt  t0 Þ F ðtÞ dt ¼ Fðt0 Þ
   s
s2 þa2                                           cos a t
                                                                              0
   1                                             1
s2 a2                                           a sinh a t
                                                                             Its Laplace transform is
   s
s2 a2                                           cosh a t
                                                                             L½d ðt  t0 Þ ¼ est0 ; t0  0; s > 0
     s                                            t
ðs2 þa2 Þ2                                       2 a sin a t
  s2 a2
                                                 t cos a t                   The square wave function illustrated in
ðs2 þa2 Þ2
                                                                             Figure 10 has Laplace transform
    1                                            1 at
ðsaÞ2 þb2                                       be        sin b t
                                                                                         1     cs
   sa
                                                 ea t cos b t                L½F c ðtÞ ¼ tanh
ðsaÞ2 þb2                                                                               s     2
More generally, translation gives the following.                             Other Laplace transforms are listed in Table 5.
Translation.
                                  t 
                      b        1
f ða s  bÞ ¼ f a s      ¼ L ebt=a F        ;a>0
                      a        a       a
       8                     1
       >
       >            0t<
       >0
       >
       >
       >                     h
       >
       <
                    1     2
S ðtÞ ¼ 1             t<
       >
       >            h     h
       >
       >
       >
       >
       >
       :2           2     3
                      t<
                    h     h                                                  Figure 11. Triangular wave function
36                    Mathematics in Chemical Engineering
Convolution properties are also satisfied:                  If the factor (s  a) is repeated m times, then
                      Zt                                              p ðsÞ    Am       Am1
F ðtÞ  G ðtÞ ¼            F ðtÞ G ðt  tÞdt                f ðsÞ ¼         ¼       þ           þ
                                                                      q ðsÞ ðs  aÞm ðs  aÞm1
                      0
                                                                           A1
                                                                      þ       þ h ðsÞ
                                                                          sa
and
                                                            where
f ðsÞgðsÞ ¼ L½FðtÞ  GðtÞ
                                                                      ðs  aÞm p ðsÞ
                                                            f ðsÞ 	
                                                                          q ðsÞ
Derivatives of Laplace Transforms.The Laplace
integrals L [F (t)], L [t F (t)], L [t2 F (t)], . . . are                               1    dmk f ðsÞ
                                                            Am ¼ f ðaÞ; Ak ¼                            ja ;
                                                                                     ðm  kÞ! dsmk
uniformly convergent for s1  a and
                                                            k ¼ 1; . . . ; m  1
                                                                       1                   1
All the factors are linear, none are repeated,              f ðsÞ ¼       ; f0 ðsÞ ¼ 
                                                                      s2              ðs  2Þ2
and the an are all distinct. If p (s) has a smaller
                                                            f ð1Þ ¼ 1; f0 ð1Þ ¼ 1
degree than q (s), the Heaviside expansion can
be used to evaluate the inverse transformation:
                                                            The inverse Laplace transform is then
                  X m
    1       p ðsÞ       p ðai Þ ai t
L                  ¼              e
             q ðsÞ   i¼1
                         q0 ðai Þ                           FðtÞ ¼ e2t þ ½1  t et
                                                                    Mathematics in Chemical Engineering                37
Quadratic Factors. Let p (s) and q (s) have real                          To solve an integral equation:
coefficients, and q (s) have the factor
                                                                                        Zt
                                                                     Y ðtÞ ¼ a þ 2              Y ðtÞcos ðt  tÞdt
ðs  aÞ þ b ; b > 0
        2       2
                                                                                        0
            p ðsÞ     f ðsÞ
f ðsÞ ¼           ¼
            q ðsÞ ðs  aÞ2 þ b2                                      Then the Laplace transform is used to obtain
              AsþB
      ¼                   þ h ðsÞ                                               a             s
          ðs  aÞ2 þ b2                                               y ðsÞ ¼     þ 2 y ðsÞ 2
                                                                                s          s þ1
                                                                                     a ðs2 þ 1Þ
Let f1 and f2 be the real and imaginary parts of                      or y ðsÞ ¼
                                                                                     s ðs  1Þ2
the complex number f (a þi b).
                           1
ðs2  2 s þ 1Þ y ðsÞ ¼                                               The functions | f(s)| and |x f(s)| are bounded
                          s2
                                                                     in the half-plane x  x1 > x0 and f (s) ! 0 as
                    1
y ðsÞ ¼                                                              |y| ! 1 for each fixed x. Thus,
            ðs  2Þðs  1Þ2
                                                                      jf ðx þ i yÞj < M; jx f ðx þ i yÞj < M;
lead to                                                                         x  x1 > x 0
                                                                       lim f ðx þ i yÞ ¼ 0; x > x0
YðtÞ ¼ e2t  ð1 þ tÞ et                                               y!
1
38              Mathematics in Chemical Engineering
If F (t) is continuous, F0 (t) is piecewise contin-     Also F (t) and its n derivatives are continuous
uous, and both functions are O [exp (x0t)], then        functions of t of order O [exp (x0t)] and they
| f (s)| is O (1/s) in each half-plane x  x1 > x0.     vanish at t ¼ 0.
If F (t) and F0 (t) are continuous, F00 (t) is piece-   Series of Residues [41]. Let f (s) be an analytic
wise continuous, and all three functions are O          function except for a set of isolated singular
[exp (x0t)], then                                       points. An isolated singular point is one for
                                                        which f (z) is analytic for 0 < |z  z0| < % but z0
js2 f ðsÞ  s Fð0Þj < M; x  x1 > x0                    is a singularity of f (z). An isolated singular
                                                        point is either a pole, a removable singularity, or
                                                        an essential singularity. If f (z) is not defined in
The additional constraint F (0) ¼ 0 is necessary        the neighborhood of z0 but can be made analytic
and sufficient for | f (s)| to be O (1/s2).             at z0 simply by defining it at some additional
                                                        points, then z0 is a removable singularity. The
Inversion Integral [41]. Cauchy’s integral for-         function f (z) has a pole of order k  1 at z0 if
mula for f (s) analytic and O (sk) in a half-plane     (z  z0)k f (z) has a removable singularity at z0
x  y, k > 0, is                                        whereas (z  z0)k1f (z) has an unremovable
                                                        isolated singularity at z0. Any isolated singu-
                       Z
                       gþib                             larity that is not a pole or a removable singu-
           1                   f ðzÞ
f ðsÞ ¼         lim                  dz; Re ðsÞ > g     larity is an essential singularity.
          2 pi b!1             sz
                      gib
                                                           Let the function f (z) be analytic except for
                                                        the isolated singular point s1, s2, . . . , sn. Let
                                                        %n (t) be the residue of ezt f (z) at z ¼ sn (for
Applying the inverse Laplace transformation on
                                                        definition of residue, see Section 3.4). Then
either side of this equation gives
                                                                   X
                                                                   1
                       Z
                       gþib
                                                        F ðtÞ ¼          %n ðtÞ
         1                       zt
F ðtÞ ¼       lim              e f ðzÞdz                           n¼1
        2 pi b!1
                      gib
                       Z
                       gþib                             If sn is a removable pole of f (s), of order m,
dn F    1
     ¼       lim              ezt zn f ðzÞdz;           then
dtn    2 pi b!1
                      gib
n ¼ 1; 2; . . . ; m                                     fn ðzÞ ¼ ðz  sn Þm f ðzÞ
                                                       Mathematics in Chemical Engineering                                 39
                                                        The solution is
    Problem 1. Infinite domain, on  1 < x < 1.
                                                                                                  pffiffiffiffiffiffiffiffi
Tðx; 0Þ ¼ f ðxÞ; initial conditions                      Tðx ; tÞ ¼ T 0 þ ½T 1  T 0 ½1  erf ðx= 4a tÞ
                                                                                                 pffiffiffiffiffiffiffiffi
Tðx; tÞ bounded                                          or Tðx; tÞ ¼ T 0 þ ðT 1  T 0 Þ erfc ðx= 4a tÞ
40                  Mathematics in Chemical Engineering
   Problem 3. Semi-infinite domain, boundary                                          The solution for T1 can also be obtained by
condition of the first kind, on 0  x < 1                                           Laplace transforms.
Tðx; 0Þ ¼ f ðxÞ
                                                                                    t1 ¼ L½T 1 
Tð0; tÞ ¼ g ðtÞ
where
                                                                                    and solving gives
T 1 ðx; 0Þ ¼ f ðxÞ; T 2 ðx; 0Þ ¼ 0
Thus,
                                                                                       Problem 4. Semi-infinite domain, boun-
                                    2                                               dary conditions of the second kind, on
U 1 ðv; tÞ ¼ F vs ½f  ev              at
                                                                                    0  x < 1.
and [40, p. 322]                                                                    Tðx; 0Þ ¼ 0
                    Z1
               2                              2
                                                                                         @T
T 1 ðx; tÞ ¼                F vs ½f ev          at
                                                       sin vx dv                    k      ð0; tÞ ¼ q0 ¼ constant
               p                                                                         @x
                    0
Solve for T2 by taking the sine transform Take the Laplace transform
@U 2                                                                                           @2 t
     ¼  v2 a U 2 þ a gðtÞ v                                                        st ¼ a
 @t                                                                                            @x2
U 2 ðv; 0Þ ¼ 0                                                                           @t q0
                                                                                    k      ¼
                                                                                         @x   s
Thus,
                                                                                    The solution is
               Zt                                                                                    pffiffiffi
                              2                                                                    q0 a xpffiffiffiffiffi
U 2 ðv; tÞ ¼            ev       aðttÞ
                                             av g ðtÞdt                             t ðx; sÞ ¼            e s=a
                                                                                                   k s3=2
                0
                                                                                tð0; sÞ ¼ tðL; sÞ ¼ 0
    @T
k      ð0; tÞ ¼ h T ð0; tÞ
    @x
                                                                                The solution is
Take the Laplace transform                                                                              pffiffiffiffi          pffiffi
                                                                                                T 0 sinh as x T 0 sinh as ðL  xÞ T 0
                                                                                t ðx; sÞ ¼             pffiffis            pffiffi     þ
                                                                                                 s sinh a L    s     sinh as L     s
                      @2 t
s t  f ðxÞ ¼ a
                      @x2
    @t                                                                          The inverse transformation is [41, p. 220], [44,
k      ð0; sÞ ¼ h t ð0; sÞ                                                      p. 96]
    @x
                                                                                 Tð0; tÞ ¼ 0
                  Z1                                      Z1                     TðL; 0Þ ¼ T 0 ¼ constant
             2               2
T ðx; tÞ ¼             eb       at
                                      cos ½b x  m ðbÞ        f ðjÞ
             p
                  0                                       0
                                                                                Take the Laplace transform
                             cos ½b j  m ðbÞ djdb
   Problem 6. Finite domain, boundary condi-                                    and the inverse transformation is [41, p. 201],
tion of the first kind                                                          [44, p. 313]
                                                                                                 "                                                   #
Tðx; 0Þ ¼ T 0 ¼ constant                                                                           x 2X 1
                                                                                                          ð1Þn n2 p2 at=L2 n p x
                                                                                T ðx; tÞ ¼ T 0      þ          e            sin
                                                                                                   L p n¼1 n                    L
Tð0; tÞ ¼ T ðL; tÞ ¼ 0
42                  Mathematics in Chemical Engineering
with magnitude
                                                       The transpose of A is
        pffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi
jrj ¼    x2 þ y 2 þ z 2                                ATij ¼ Aji
but
                                                                              Because the dyadics may not be symmetric, the
u v 6¼ v u                                                                    order of indices and which indices are summed
                                                                              are important. The order is made clearer when
The Kronecker delta is defined as                                             the dyadics are made from vectors.
u v ¼ jujjvjcos u; 0 u p c ¼ u v ¼ a j u j j v j sin u; 0 u p
where u is the angle between u and v. The scalar                              where a is a unit vector in the direction of uv.
product of two vectors is a scalar, not a vector. It                          The direction of c is perpendicular to the plane
is the magnitude of u multiplied by the projec-                               of u and v such that u, v, and c form a right-
tion of v on u, or vice versa. The scalar product                             handed system. If u ¼ v, or u is parallel to v,
of u with itself is just the square of the magni-                             then u ¼ 0 and uv ¼ 0. The following laws are
tude of u.                                                                    valid for cross products.
u  u ¼ ju2 j ¼ u2
                                                                              uv¼vu                                        Commutative law fails for vector
The following laws are valid for scalar products                                                                                 product
                                                                              u(vw) 6¼ (uv)w                              Associative law fails for vector
                                                                                                                                 product
u v ¼ v u                            Commutative law for scalar             u(v þ w) ¼ uv þ uw                           Distributive law for vector
                                           products                                                                               product
u (v þ w) ¼ u v þ u                 Distributive law for scalar products   ex ex¼ ey ey¼ ez ez ¼ 0
     w
                                                                              ex ey¼ ez, ey ez¼ ex, ez
ex ex ¼ ey ey¼ ez ez ¼ 1
                                                                                 ex¼ ey
ex ey¼ ex ez¼ ey ez ¼ 0
                                                                                                2                         3
                                                                                           ex                  ey    ez
If the two vectors u and v are written in com-                                           6
                                                                                         6
                                                                                                                        7
                                                                                                                        7
ponent notation, the scalar product is                                        u  v ¼ det6 ux                  uy    uz 7
                                                                                         4                              5
                                                                                                    vx         vy    vz
u  v ¼ u x vx þ u y vy þ u z vz
                                                                              ¼ ex ðuy vz  vy uz Þ þ ey ðuz vz  ux vz Þ
If u  v ¼ 0 and u and v are not null vectors, then
                                                                              þez ðux vy  uy vx Þ
u and v are perpendicular to each other and u ¼
p/2.
                                                                              This can also be written as
    The single dot product of two dyadics is
            XX                 X                                                       XX
AB¼                    ei ej ð  Aik Bkj Þ                                    uv¼                           ekij ui vj ek
              i     j         k                                                             i       j
                                                                    Mathematics in Chemical Engineering                           45
Thus e123 ¼ 1, e132 ¼1, e312 ¼ 1, e112 ¼ 0, for                     can be formed where l is an eigenvalue. This
example.                                                             expression is
   The magnitude of uv is the same as the
area of a parallelogram with sides u and v. If
                                                                     l3  I 1 l2 þ I 2 l  I 3 ¼ 0
uv ¼ 0 and u and v are not null vectors, then u
and v are parallel. Certain triple products are
useful.                                                              An important theorem of HAMILTON and CAYLEY
                                                                     [47] is that a second-order dyadic satisfies its
ðu  vÞw 6¼ uðv  wÞ                                                 own characteristic equation.
u  ðv  wÞ ¼ v  ðw  uÞ ¼ w  ðu  vÞ
                                                                 Z t rffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi
                                                                       dr dr 
                                                      s ðtÞ ¼                        dt
                                                                      dt dt
                                                                  a
                                                      This gives
Figure 15. Vector differentiation
                                                       2         2  2  2
                                                       ds   dr dr  dx   dy   dz
                                                           ¼  ¼      þ    þ
                                                       dt   dt dt  dt   dt   dt
is the velocity. The derivative operation obeys
the following laws.
                                                      Because
d            du dv
   ðu þ vÞ ¼   þ
dt           dt dt                                    dr ¼ dxex þ dyey þ dzez
d            du       dv
   ðu  vÞ ¼    vþu
dt           dt       dt                              then
d            du       dv
   ðu  vÞ ¼    vþu
dt           dt       dt                              ds2 ¼ dr  dr ¼ dx2 þ dy2 þ dz2
d        du     du
   ðuuÞ ¼ u þ u
dt       dt     dt
                                                      The derivative dr/dt is tangent to the curve in
If the vector u depends on more than one              the direction of motion
variable, such as x, y and z, partial derivatives
are defined in the usual way. For example,                   dr
                                                      u ¼  dt 
                                                           dr 
if uðx; y; zÞ; then                                         dt 
@u       uðx þ Dx; y; zÞ  uðx; y; zÞ
   ¼ lim
@x Dt!0              Dx                               Also,
Rules for differentiation of scalar and vector              dr
                                                      u¼
products are                                                ds
@            @u       @v
   ðu  vÞ ¼    vþu                                 Differential Operators. The vector differential
@x           @x       @x
@            @u       @v                              operator (del operator) b is defined in cartesian
   ðu  vÞ ¼    vþu                                 coordinates by
@x           @x       @x
dðu  vÞ ¼ du  v þ u  dv                                            @f      @f      @f
                                                      rf ¼ ex            þ ey    þ ez
     @u     @u   @u                                                   @x      @y      @z
du ¼    dx þ dy þ dz
     @x     @y   @z
                                                      and is a vector. If f is height or elevation, the
                                                      gradient is a vector pointing in the uphill direc-
   If a curve is given by r(t), the length of the     tion. The steeper the hill, the larger is the
curve is [43]                                         magnitude of the gradient.
                                                         The divergence of a vector is defined by
     Z b rffiffiffiffiffiffiffiffiffiffiffiffi
          dr dr
L¼              dt                                                   @ux @uy @uz
          dt dt                                       r u¼              þ    þ
     a                                                                @x   @y   @z
                                                               Mathematics in Chemical Engineering                          47
and is a scalar. For a volume element DV, the net               Useful formulas are [49]
outflow of a vector u over the surface of the                    r  ðfuÞ ¼ rf  u þ fr  u
element is
                                                                 r  ðfuÞ ¼ rf  u þ fr  u
Z
                                                                 r  ðu  vÞ ¼ v  ðr  uÞ  u  ðr  vÞ
     u  ndS
DS                                                               r  ðu  vÞ ¼ v  ru  vðr  uÞ  u  rv þ uðr  vÞ
                                                                 r  ðr  uÞ ¼ rðr  uÞ  r2 u
This is related to the divergence by [48, p. 411]
                                                                                                @2 f @2 f @2 f
                                                                 r  ðrfÞ ¼ r2 f ¼                  þ    þ     ; where r2
                         Z                                                                      @x2 @y2 @z2
                     1
r  u ¼ lim                      u  ndS
          DV!0 DV                                               is called the Laplacian operator. r(r
                                                                                                     r w) ¼ 0.
                         DS
                                                                The curl of the gradient of f is zero.
Thus, the divergence is the net outflow per unit                r  ðr  uÞ ¼ 0
volume.
   The curl of a vector is defined by                           The divergence of the curl of u is zero. Formu-
                                                                las useful in fluid mechanics are
                               
               @      @      @                                   r  ðrvÞT ¼ rðr  vÞ
ru¼             þ ey þ ez
                ex                 ðex ux þ ey uy þ ez uz Þ
              @x     @y      @z
                                                             r  ðt  vÞ ¼ v  ðr  tÞ þ t : rv
       @uz @uy           @ux @uz
¼ ex              þ ey                                                 1
       @y     @z          @z    @x                               v  rv ¼ rðv  vÞ  v  ðr  vÞ
                                                                       2
        @uy @ux
 þez                                                           If a coordinate system is transformed by a rota-
        @x     @y
                                                                tion and translation, the coordinates in the new
and is a vector. It is related to the integral                  system (denoted by primes) are given by
                                                                 0   1               0                           1
                                                                  x0                   l11           l12   l13
Z               Z                                                B 0C                B                         C
                                                                 By C ¼              B l21           l22   l23 C
     u  ds ¼        us ds                                       @ A                 @                         A
C               C                                                 z0                      l31        l32   l33
                                                                 0 1                 0          1
                                                                  x                       a1
which is called the circulation of u around path                 B C
                                                                 ByC
                                                                                     B C
                                                                                     B a2 C
                                                                 @ A         þ       @ A
C. This integral depends on the vector and the
contour C, in general. If the circulation does not                   z                    a3
depend on the contour C, the vector is said to be               Any function that has the same value in all
irrotational; if it does, it is rotational. The                 coordinate systems is an invariant. The gradient
relationship with the curl is [48, p. 419]                      of an invariant scalar field is invariant; the same
                                  Z                             is true for the divergence and curl of invariant
                             1
n  ðr  uÞ ¼ lim
                     D!0 DS
                                       u  ds                   vectors fields.
                                   C                                The gradient of a vector field is required in
                                                                fluid mechanics because the velocity gradient is
Thus, the normal component of the curl equals                   used. It is defined as
the net circulation per unit area enclosed by the                        XX                    @vj
                                                                 rv ¼                ei ej         and
contour C.                                                               i       j
                                                                                               @xi
   The gradient, divergence, and curl obey a                                 XX                      @vi
distributive law but not a commutative or asso-                  ðrvÞT ¼                   ei ej
                                                                             i        j
                                                                                                     @xj
ciative law.
                                                                The divergence of dyadics is defined
rðf þ cÞ ¼ rf þ rc                                                                                     !
                                                                         P                X @t ji
                                                                 rt ¼       i ei                          and
r  ðu þ vÞ ¼ r  u þ r  v                                                                j
                                                                                                @xj
                                                                                                "                     #
r  ðu þ vÞ ¼ r  u þ r  v                                                          X              X @
                                                                 r  ðfu vÞ ¼              ei             ðfuj vi Þ
r  f 6¼ fr                                                                           i             j
                                                                                                      @xj
r  ðftÞ ¼ rf  t þ fr  t
                                                         Vector Integration [48, pp. 206–212]. If u is a
nt : t ¼ ttn ¼ t : nt                                  vector, then its integral is also a vector.
posed into
                                                         Theorems about this line integral can be written
                              @                          in various forms.
v ¼ vII þ nvn ; r ¼ rII þ n
                              @n
                                                         Theorem [43]. If the functions appearing in the
The velocity gradient can be decomposed into             line integral are continuous in a domain D, then
                                                         the line integral is independent of the path C if
                                         @vII      @vn
rv ¼ rII vII þ ðrII nÞvn þ nrII vn þ n        þ nn       and only if the line integral is zero on every
                                         @n        @n
                                                         simple closed path in D.
The surface gradient of the normal is the nega-
tive of the curvature dyadic of the surface.             Theorem [43]. If u ¼ rf where f is single-
                                                         valued and has continuous derivatives in D,
rII n ¼ B                                               then the line integral is independent of the
                                                         path C and the line integral is zero for any
The surface divergence is then                           closed curve in D.
rII  v ¼ dII : rv ¼ rII  vII  2 Hvn                   Theorem [43]. If f, g, and h are continuous
                                                         functions of x, y, and z, and have continuous
where H is the mean curvature.                           first derivatives in a simply connected domain
                                                         D, then the line integral
   1                                                     Z
H ¼ dII : B
   2                                                         ðf dx þ gdy þ hdzÞ
                                                         C
or if f, g, and h are regarded as the x, y, and z                 Then the divergence theorem in component
components of a vector v:                                         form is
rv¼0                                                              Z                           Z
                                                                         @ux @uy @uz
                                                                            þ    þ      dxdydz ¼   ½ux cos ðx; nÞ
                                                                         @x   @y   @z
Consequently, the line integral is independent                     V                                             S
of the path (and the value is zero for a closed                            þuy cos ðy; nÞ þ uz cos ðz; nÞdS
contour) if the three components in it are
regarded as the three components of a vector                      If the divergence theorem is written for an
and the vector is derivable from a potential (or                  incremental volume
zero curl). The conditions for a vector to be
                                                                                      Z
derivable from a potential are just those in the                  r  u ¼ lim
                                                                                  1
                                                                                           un dS
                                                                           DV!0 DV
third theorem. In two dimensions this reduces                                         DS
to the more usual theorem.
                                                                  the divergence of a vector can be called the
Theorem [48, p. 207]. If M and N are continu-
                                                                  integral of that quantity over the area of a closed
ous functions of x and y that have continuous
                                                                  volume, divided by the volume. If the vector
first partial derivatives in a simply connected
                                                                  represents the flow of energy and the diver-
domain D, then the necessary and sufficient
                                                                  gence is positive at a point P, then either a
condition for the line integral
                                                                  source of energy is present at P or energy is
Z
                                                                  leaving the region around P so that its temper-
    ðMdx þ NdyÞ
C
                                                                  ature is decreasing. If the vector represents the
                                                                  flow of mass and the divergence is positive at a
to be zero around every closed curve C in D is                    point P, then either a source of mass exists at P
                                                                  or the density is decreasing at the point P. For an
@M @N                                                             incompressible fluid the divergence is zero and
   ¼
@y   @x
                                                                  the rate at which fluid is introduced into a
   If a vector is integrated over a surface with                  volume must equal the rate at which it is
incremental area d S and normal to the surface                    removed.
n, then the surface integral can be written as                        Various theorems follow from the diver-
                                                                  gence theorem.
ZZ                 ZZ
        u  dS ¼        u  ndS                                   Theorem. If f is a solution to Laplace’s equa-
    S               S                                             tion
Also if f satisfies the conditions of the theorem                   Theorem [48, p. 423]. The necessary and suf-
and is zero on S then f is zero throughout D. If                    ficient condition that the divergence of a vector
two functions f and c both satisfy the Laplace                      vanish identically is that the vector is the curl of
equation in domain D, and both take the same                        some other vector.
values on the bounding curve C, then f ¼ c;
i.e., the solution to the Laplace equation is                       Leibniz Formula. In fluid mechanics and
unique.                                                             transport phenomena, an important result is
    The divergence theorem for dyadics is                           the derivative of an integral whose limits of
                                                                    integration are moving. Suppose the region V
Z                    Z                                              (t) is moving with velocity vs. Then Leibniz’s
     r  tdV ¼              n  t dS
                                                                    rule holds:
V                       S
                                                                         ZZZ             ZZZ                 ZZ
                                                                    d                            @f
                                                                                 fdV ¼              dV   þ        fvs  ndS
Stokes Theorem [48, 49]. Stokes theorem says                        dt                           @t
                                                                          VðtÞ            VðtÞ                S
that if S is a surface bounded by a closed,
nonintersecting curve C, and if u has continu-                      Curvilinear Coordinates. Many of the rela-
ous derivatives then                                                tions given above are proved most easily by
I               ZZ                             ZZ                   using tensor analysis rather than dyadics. Once
     u  dr ¼           ðr  uÞ  ndS ¼             ðr  uÞ  dS    proven, however, the relations are perfectly
C                   S                           S                   general in any coordinate system. Displayed
                                                                    here are the specific results for cylindrical and
The integral around the curve is followed in the                    spherical geometries. Results are available for a
counterclockwise direction. In component                            few other geometries: Parabolic cylindrical,
notation, this is                                                   paraboloidal, elliptic cylindrical, prolate sphe-
                                                                    roidal, oblate spheroidal, ellipsoidal, and
H
    C ½ux cos ðx; sÞ þ uy cos ðy; sÞ þ uz cos ðz; sÞds             bipolar coordinates [45, 50].
        ZZ                                                        For cylindrical coordinates, the geometry is
               @uz @uy                     @ux @uz
¼                         cosðx; nÞ þ                 cosðy; nÞ   shown in Figure 16. The coordinates are related
                @y    @z                   @z      @x
         S                                                       to cartesian coordinates by
                 @uy @ux
             þ              cosðz; nÞ dS
                 @x    @y
               @f eu @f       @f
rf ¼ er           þ      þ ez
               @r   r @u      @z
                                                                                         For spherical coordinates, the geometry is
                
    21@       @f     1 @2 f @2 f                                                         shown in Figure 17. The coordinates are related
r f¼        r      þ 2 2þ 2
     r @r     @r    r @u    @z                                                           to cartesian coordinates by
    @ 2 f 1 @f 1 @ 2 f @ 2 f
¼        þ    þ       þ
    @r2 r @r r2 @u2 @z2                                                                                                                      pffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi
                                                                                         x ¼ r sin u cosf                                r ¼ x2 þ p     y2ffiffiffiffiffiffiffiffiffiffiffiffiffiffi
                                                                                                                                                             þ z2 ffi
              1@             1 @vu @vz                                                   y ¼ r sin u sinf                                u ¼ arctanð
 
 x2 þ y2 =zÞ
rv¼               ðr vr Þ þ      þ                                                      z ¼ r cos u                                     f ¼ arctan yx
              r @r           r @u   @z
                                           
                1 @vz @vu             @vr @vz
r  v ¼ er                    þ eu      
                r @u     @z           @z   @r                                            The unit vectors are related by
                           
       1@             1 @vr
þ ez        ðr vu Þ                                                                      er ¼ sin u cos fex þ sin u sin fey þ cos uez
       r @r           r @u
           @c      1 @c         1 @c
rc ¼ er
           @r
              þ eu
                   r @u
                        þ ef
                             r sin u @f
                                                                           6. Ordinary Differential Equations
                                                                           as Initial Value Problems
                                            
         1 @       @c       1    @          @c
r2 c ¼          r2      þ 2           sin u      þ
         r @r
          2        @r    r sin u @u         @u                             A differential equation for a function that
           1    @2c
                                                                           depends on only one variable (often time) is
       þ 2 2                                                               called an ordinary differential equation. The
        r sin u @f2
                                                                           general solution to the differential equation
         1 @ 2              1 @                                            includes many possibilities; the boundary or
rv¼           ðr vr Þ þ            ðvu sin uÞþ
         r2 @r           r sin u @u                                        initial conditions are required to specify which
              1 @vf                                                        of those are desired. If all conditions are at one
       þ
           r sin u @f                                                      point, the problem is an initial value problem
                                                                           and can be integrated from that point on. If
                                                  
                  1 @                      1 @vu                           some of the conditions are available at one point
r  v ¼ er                ðvf sin uÞ                þ
              r sin u @u                r sin u @f                         and others at another point, the ordinary
                                           
                    1 @vr 1 @                                              differential equations become two-point bound-
           þeu                      ðr vf Þ þ
                 r sin u @f r @r                                           ary value problems, which are treated in
                                                                         Chapter 7. Initial value problems as ordinary
                 1@             1 @vr
           þef        ðr vu Þ 
                 r @r           r @u                                       differential equations arise in control of
                                                                          lumped-parameter models, transient models
             1 @ 2                 1 @                                     of stirred tank reactors, polymerization
r  t ¼ er         ðr t rr Þ þ             ðt ur sin uÞ
             r2 @r             r sin u @u
                                                                          reactions and plug-flow reactors, and generally
              1 @tfr t uu þ t ff
         þ
           r sin u @f
                          
                                  r
                                          þ                                in models where no spatial gradients occur in
                                                                          the unknowns.
               1 @                   1 @                        1 @tfu
         þeu 3 ðr3 t ru Þ þ                  ðt uu sin uÞ þ
               r @r              r sin u @u                 r   sin u @f
                                    
           tur  t ru  t ff cot u
         þ                            þ
             
                      r                                                    6.1. Solution by Quadrature
                1 @                  1 @                         1 @tff
         þef 3 ðr 3 t rf Þ þ                 ðt uf sin uÞ þ
               r @r              r sin u @u                 r   sin u @f
                                                                          When only one equation exists, even if it is
           tfr  t rf þ t fu cot u
         þ
                      r
                                                                           nonlinear, solving it by quadrature may be
                                                                           possible. For
           @vr           @vu           @vf
rv ¼ er er       þ er eu      þ er ef      þ
            @r             @r          @r                                  dy
                                                                          ¼ f ðyÞ
              1 @vr vu                  1 @vu vr                           dt
     þeu er                  þ eu eu          þ      þ
              r @u       r              r @u      r                        y ð0Þ ¼ y0
                                               
             1 @vf                1 @vr vf
     þeu ef         þ ef er                      þ
             r @u              r sinu @f      r                            the problem can be separated
                                      
                 1 @vu vf
     þef eu                  cot u þ
              r sinu @f        r                                            dy
                                                                                 ¼ dt
                                                                         f ðyÞ
                  1 @vf vr vu
     þef ef                 þ þ cot u
               r sinu @f       r     r
                                                                      and integrated:
            @ 1 @ 2                      1     @          @vr
r2 v ¼ er               ðr vr Þ þ 2                 sin u
           @r r2 @r                   r sin u @u          @u
                                                                          Zy               Zt
              1     @ 2 vr       2     @                   2 @vf                 dy0
      þ 2 2                              ðvu sin uÞ  2                                ¼        dt ¼ t
          r sin u @f2 r2 sin u @u                      r sin u @f               f ðy0 Þ
                                                                           y0               0
                                                     
            1 @         @vu      1 @    1 @
       þeu         r2          þ 2             ðvu sin uÞ
           r @r
             2          @r       r @u sin u @u                             If the quadrature can be performed analytically
                                                
            1    @ 2 vu 2 @vr 2 cot u @vf                                  then the exact solution has been found.
       þ 2 2             þ          2
        r sin u @f    2     r @u r sin u @f
                             2
                                                                          For example, consider the kinetics problem
            1 @         @vf       1 @    1 @
       þef 2        r2         þ 2             ðvf sin uÞ                  with a second-order reaction.
            r @r         @r      r @u sin u @u
                                                     
            1    @ 2 vf        2 @vr 2 cot u @vu
       þ 2 2             þ             þ 2                                 dc
        r sin u @f    2     r sin u @f r sin u @f
                             2
                                                                              ¼  kc2 ; cð0Þ ¼ c0
                                                                           dt
                                                   Mathematics in Chemical Engineering                       53
To find the function of the concentration versus    by using Frechet differentials [51]. In this case,
time, the variables can be separated and inte-
                                                                                      
grated.                                             exp
                                                              Ft        dc F
                                                                          þ c ¼
                                                                                d
                                                                                     exp
                                                                                          Ft
                                                                                             c
                                                              V         dt V    dt        V
dc
   ¼  kdt;
c2                                                  Thus, the differential equation can be written as
 1
 ¼  kt þ D                                                                 
 c                                                  d         Ft          Ft F
                                                         exp     c ¼ exp        cin
                                                    dt        V           V   V
Application of the initial conditions gives the
solution:                                           This can be integrated once to give
1        1                                                                     Zt      0
  ¼ kt þ                                                      Ft               F         Ft
c        c0                                         exp            c ¼ c ð0Þ þ      exp      cin ðt0 Þ dt0
                                                              V                V         V
                                                                                      0
can be converted into a set of first-order equa-        Euler’s method is first order
tions. By using
                                                        ynþ1 ¼ yn þ Dtf ðyn Þ
                   ði1Þ
                  d     y d ði2Þ dyi1
yi 	 yði1Þ ¼             ¼ y    ¼
                  dtði1Þ  dt      dt
                                                        Adams–Bashforth Methods. The second-order
                                                        Adams–Bashforth method is
the higher order equation can be written as a set
of first-order equations:                                             Dt
                                                        ynþ1 ¼ yn þ      ½3 f ðyn Þ  f ðyn1 Þ
                                                                      2
dy1
    ¼ y2
 dt
                                                        The fourth-order Adams–Bashforth method is
dy2
    ¼ y3
 dt
                                                                      Dt
dy3                                                     ynþ1 ¼ yn þ      ½55 f ðyn Þ  59 f ðyn1 Þ þ 37 f ðyn2 Þ 9 f ðyn3 Þ
    ¼ y4                                                              24
 dt
...
                                                        Notice that the higher order explicit methods
dyn
    ¼  Fðyn1 ; yn2 ; . . . ; y2; y1 Þ                require knowing the solution (or the right-hand
 dt
                                                        side) evaluated at times in the past. Because
The initial conditions would have to be speci-          these were calculated to get to the current time,
fied for variables y1(0), . . . , yn (0), or equiv-     this presents no problem except for starting the
alently y (0), . . . , y(n1)(0). The set of            evaluation. Then, Euler’s method may have to
equations is then written as                            be used with a very small step size for several
                                                        steps to generate starting values at a succession
dy                                                      of time points. The error terms, order of the
   ¼ fðy; tÞ
dt                                                      method, function evaluations per step, and
                                                        stability limitations are listed in Table 6.
All the methods in this chapter are described for       The advantage of the fourth-order Adams–
a single equation; the methods apply to the             Bashforth method is that it uses only one
multiple equations as well. Taking the single           function evaluation per step and yet achieves
equation in the form                                    high-order accuracy. The disadvantage is the
                                                        necessity of using another method to start.
dy
dt
   ¼ f ðyÞ                                              Runge–Kutta Methods. Runge–Kutta methods
                                                        are explicit methods that use several function
                                                        evaluations for each time step. The general
multiplying by dt, and integrating once yields          form of the methods is
Ztnþ1                                                               X
                                                                      v
        dy 0 X
                 tnþ1
            dt ¼      f y ðt0 Þ dt0                     ynþ1 ¼ yn þ         wi k i
        dt0       tn                                                  i¼1
tn
                                                        with
This is
                                                                                  X
                                                                                   i1        
                                                        ki ¼ Dt f tn þ ci Dt; yn þ     aij kj
                  Ztnþ1                                                               j¼1
                          dy 0
ynþ1 ¼ yn þ                   dt
                          dt0
                  tn
                                                        Runge–Kutta methods traditionally have been
                                                        writen for f (t, y) and that is done here, too. If
The last substitution gives a basis for the vari-       these equations are expanded and compared
ous methods. Different interpolation schemes            with a Taylor series, restrictions can be placed
for y (t) provide different integration schemes;        on the parameters of the method to make it first
using low-order interpolation gives low-order           order, second order, etc. Even so, additional
integration schemes [3].                                parameters can be chosen. A second-order
                                                                         Mathematics in Chemical Engineering                             55
Method Error term Order Function evaluations per step Stability limit, l Dt
Explicit methods
                                               h2 00
Euler                                          2 y               1               1                                     2.0
                                               5 3 000
Second-order Adams–Bashforth                   12 h y            2               1
                                               251 5 ð5Þ
Fourth-order Adams–Bashforth                   720 h y           4               1                                     0.3
Second-order Runge–Kutta
  (midpoint)                                                     2               2                                     2.0
Runge–Kutta–Gill                                                 4               4                                     2.8
                                                   nþ1   nþ1
Runge–Kutta–Feldberg                           y       z        5               6                                     3.0
Predictor–corrector methods
Second-order Runge–Kutta                                         2               2                                     2.0
Adams, fourth-order                                              2               2                                     1.3
Kutta–Gill method is popular because it is high        For example, Euler’s method gives
order and does not require a starting method (as
does the fourth-order Adams–Bashforth                    ynþ1 ¼ yn  l Dtyn or ynþ1 ¼ ð1  l DtÞyn
method). However, it requires four function            or rmn ¼ 1  l Dt
evaluations per time step, or four times as
many as the Adams–Bashforth method. For                The stability limit is then
problems in which the function evaluations
are a significant portion of the calculation           l Dt  2
time this might be important. Given the speed
of computers and the widespread availability of        The Euler method will not oscillate provided
desktop computers, the efficiency of a method
is most important only for very large problems         l Dt  1
that are going to be solved many times. For
other problems the most important criterion for        The stability limits listed in Table 6 are
choosing a method is probably the time the user        obtained in this fashion. The limit for the Euler
spends setting up the problem.                         method is 2.0; for the Runge–Kutta–Gill
   The stability of an integration method is best      method it is 2.785; for the Runge–Kutta–
estimated by determining the rational polyno-          Feldberg method it is 3.020. The rational poly-
mial corresponding to the method. Apply this           nomials for the various explicit methods are
method to the equation                                 illustrated in Figure 19. As can be seen, the
                                                       methods approximate the exact solution well as
dy
   ¼  ly; yð0Þ ¼ 1                                    l Dt approaches zero, and the higher order
dt
                                                       methods give a better approximation at high
                                                       values of l Dt.
and determine the formula for rmn:                         In solving sets of equations
                                                       dy
ykþ1 ¼ rmn ðl DtÞyk                                       ¼ A y þ f; yð0Þ ¼ y0
                                                       dt
The rational polynomial is defined as                  all the eigenvalues of the matrix A must be
                                                       examined. FINLAYSON [3] and AMUNDSON [54,
            pn ðzÞ                                     p. 197–199] both show how to transform these
rmn ðzÞ ¼           ez
            qm ðzÞ
jr mn ðzÞj 1
equations into an orthogonal form so that each                   error estimate, then Dt is reduced to meet that
equation becomes one equation in one                             criterion. Also, Dt is increased to as large a
unknown, for which single equation analysis                      value as possible, because this shortens com-
applies. For linear problems the eigenvalues do                  putation time. If the local truncation error has
not change, so the stability and oscillation                     been achieved (and estimated) by using a step
limits must be satisfied for every eigenvalue                    size Dt1
of the matrix A. When solving nonlinear prob-
lems the equations are linearized about the                      LTE ¼ c Dtp1
solution at the local time, and the analysis
applies for small changes in time, after which                   and the desired error is e, to be achieved using a
a new analysis about the new solution must be                    step size Dt2
made. Thus, for nonlinear problems the eigen-
values keep changing.
   Richardson extrapolation can be used to                       e ¼ c D tp2
differential equations it is called the Crank–                                difference formulas. The formulas of various
Nicolson method. Adams methods exist as                                       orders are [57]:
well, and the fourth-order Adams–Moulton
method is                                                                     1 : ynþ1 ¼ yn þ Dt f ðynþ1 Þ
                                                                                           4 n 1 n1 2
                                                                              2 : ynþ1 ¼     y  y  þ Dt f ðynþ1 Þ
                    Dt                                                                     3    3    3
ynþ1 ¼ yn þ            ½9 f ðynþ1 Þ þ 19 f ðyn Þ  5 f ðyn1 Þ þ f ðyn2 Þ
                    24
                                                                                           18 n    9 n1    2 n2
                                                                              3 : ynþ1 ¼      y     y   þ    y
                                                                                           11     11       11
The properties of these methods are given in                                       6
                                                                              þ      Dt f ðynþ1 Þ
Table 6. The implicit methods are stable for any                                  11
step size but do require the solution of a set of                                          48 n 36 n1 16 n2    3 n3
                                                                              4 : ynþ1 ¼      y     y þ    y     y
nonlinear equations, which must be solved                                                  25     25     25     25
iteratively. An application to dynamic distilla-                                  12
tion problems is given in [56].                                               þ      Dt f ðynþ1 Þ
                                                                                  25
    All these methods can be written in the
                                                                                           300 n 300 n1 200 n2
form                                                                          5 : ynþ1 ¼       y      y þ     y
                                                                                           137     137     137
                                                                                   75 n3    12 n4
              X
              k                           X
                                          k                                          y   þ     y
                                                                                  137       137
y   nþ1
          ¼         ai y   nþ1i
                                   þ Dt         bi f ðy   nþ1i
                                                                  Þ
              i¼1                         i¼0
                                                                                   60
                                                                              þ       Dt f ðynþ1 Þ
                                                                                  137
       maxi j Re ðli Þj
SR ¼                                                  dc1
       mini j Re ðli Þj                                   ¼ f ðc1 ; c2 Þ
                                                       dt
                                                      dc2
SR ¼ 20 is not stiff, SR ¼ 103 is stiff, and SR        dt
                                                          ¼ k1 c1  k2 c22
¼ 106 is very stiff. If the problem is nonlinear,
the solution is expanded about the current
                                                      The first equation could be the equation for a
state:
                                                      stirred tank reactor, for example. Suppose both
                     X
                                                      k1 and k2 are large. The problem is then stiff,
dyi                   n
                         @f i
    ¼ f i ½yðtn Þ þ          ½y  yj ðtn Þ          but the second equation could be taken at
dt                   j¼1
                         @yj j
                                                      equilibrium. If
    The question of stiffness then depends on the
                                                      c1 @2c2
eigenvalue of the Jacobian at the current time.
Consequently, for nonlinear problems the prob-
lem can be stiff during one time period and not       The equilibrium condition is then
stiff during another. Packages have been devel-
oped for problems such as these. Although the         c22 k1
                                                         ¼   	 K
                                                      c1 k 2
chemical engineer may not actually calculate
the eigenvalues, knowing that they determine
the stability and accuracy of the numerical           Under these conditions the problem becomes
scheme, as well as the step size employed, is
useful.                                               dc1
                                                          ¼ f ðc1 ; c2 Þ
                                                       dt
                                                      0 ¼ kl c1  k2 c22
6.5. Differential–Algebraic Systems
                                                      Thus, a differential-algebraic system of equa-
Sometimes models involve ordinary differen-           tions is obtained. In this case, the second equa-
tial equations subject to some algebraic con-         tion can be solved and substituted into the first
straints. For example, the equations governing        to obtain differential equations, but in the gen-
one equilibrium stage (as in a distillation           eral case that is not possible.
                                                     Mathematics in Chemical Engineering               61
satisfy the tolerance with each of these methods       Netlib web site, http://www.netlib.org/, choose
can be determined. Then the method and step            “ode” to find packages which can be
size for the next step that achieves the biggest       downloaded. Using Microsoft Excel to solve
step can be chosen, with appropriate adjustments       ordinary differential equations is cumbersome,
due to the different work required for each order.     except for the simplest problems.
The package generally starts with a very small
step size and a first-order method — the back-
ward Euler method. Then it integrates along,           6.7. Stability, Bifurcations, Limit
adjusting the order up (and later down) depend-        Cycles
ing on the error estimates. The user is thus
assured that the local truncation error meets          In this section, bifurcation theory is discussed
the tolerance. A further difficulty arises because     in a general way. Some aspects of this subject
the set of nonlinear equations must be solved.         involve the solution of nonlinear equations;
Usually a good guess of the solution is available,     other aspects involve the integration of ordinary
because the solution is evolving in time and           differential equations; applications include
past history can be extrapolated. Thus, the            chaos and fractals as well as unusual operation
Newton–Raphson method will usually converge.           of some chemical engineering equipment. An
The package protects itself, though, by only           excellent introduction to the subject and details
doing a few (i.e., three) iterations. If convergence   needed to apply the methods are given in [65].
is not reached within these iterations, the step       For more details of the algorithms described
size is reduced and the calculation is redone for      below and a concise survey with some chemical
that time step. The convergence theorem for the        engineering examples, see [66] and [67]. Bifur-
Newton–Raphson method (Chap. 1) indicates              cation results are closely connected with stabil-
that the method will converge if the step size         ity of the steady states, which is essentially a
is small enough. Thus, the method is guaranteed        transient phenomenon.
to work. Further economies are possible. The              Consider the problem
Jacobian needed in the Newton–Raphson
method can be fixed over several time steps.           @u
                                                          ¼ F ðu; lÞ
Then if the iteration does not converge, the           @t
Jacobian can be reevaluated at the current
time step. If the iteration still does not converge,   The variable u can be a vector, which makes F a
then the step size is reduced and a new Jacobian       vector, too. Here, F represents a set of equations
is evaluated. The successive substitution method       that can be solved for the steady state:
can also be used — which is even faster, except
that it may not converge. However, it too will         F ðu; lÞ ¼ 0
converge if the time step is small enough.
    The Runge–Kutta methods give extremely             If the Newton–Raphson method is applied,
good accuracy, especially when the step size is
                                                       F su d us ¼ F ðus ; lÞ
kept small for stability reasons. If the problem
is stiff, though, backward difference implicit         usþ1 ¼ us þ dus
methods must be used. Many chemical reactor
problems are stiff, necessitating the use of           is obtained, where
implicit methods. In the MATLAB suite of
ODE solvers, ode45 uses a revision of the                       @F s
                                                       F su ¼      ðu Þ
RKF45 program, while the ode15s program                         @u
uses an improved backward difference method.
Ref. [64] gives details of the programs in MAT-        is the Jacobian. Look at some property of the
LAB. Fortunately, many packages are available.         solution, perhaps the value at a certain point or
On the NIST web page, http://gams.nist.gov/            the maximum value or an integral of the solu-
choose “problem decision tree”, and then               tion. This property is plotted versus the param-
“differential and integral equations” to find          eter l; typical plots are shown in Figure 22. At
packages which can be downloaded. On the               the point shown in Figure 22 A, the determinant
                                                              Mathematics in Chemical Engineering                63
det F u ¼ 0                                                      sest X ¼ F ss st
                                                                            ue X
For the limit point, The exponential term can be factored out and
@F                                                               ðF ss
                                                                    u  s dÞX ¼ 0
   6¼ 0
@l
                                                                 A solution exists for X if and only if
whereas for the bifurcation-limit point
                                                                 det jF ss
                                                                        u  s dj ¼ 0
@F
   ¼0
@l
                                                                    The s are the eigenvalues of the Jacobian.
   The stability of the steady solutions is also of              Now clearly if Re (s) > 0 then u0 grows with
interest. Suppose a steady solution uss; the                     time, and the steady solution uss is said to be
function u is written as the sum of the known                    unstable to small disturbances. If Im (s) ¼ 0 it
steady state and a perturbation u0 :                             is called stationary instability, and the distur-
                                                                 bance would grow monotonically, as indicated
u ¼ uss þ u0                                                     in Figure 23A. If Im (s) 6¼ 0 then the distur-
                                                                 bance grows in an oscillatory fashion, as shown
This expression is substituted into the original                 in Figure 23B, and is called oscillatory
equation and linearized about the steady-state                   instability. The case in which Re (s) ¼ 0 is
value:                                                           the dividing point between stability and
                                                                 instability. If Re (s) ¼ 0 and Im (s) ¼ 0 —
@uss @u0
    þ    ¼ F ðuss þ u0 ; lÞ                                      the point governing the onset of stationary
 @t   @t
                                                                 instability — then s ¼ 0. However, this means
                       @F
    F ðuss ; lÞ þ        ju u0 þ                             that s ¼ 0 is an eigenvalue of the Jacobian, and
                       @u ss
                                                                 the determinant of the Jacobian is zero. Thus,
The result is                                                    the points at which the determinant of the
                                                                 Jacobian is zero (for limit points and bifurca-
@u0         0
                                                                 tion-limit points) are the points governing the
    ¼ F ss
        u u
@t                                                               onset of stationary instability. When Re (s) ¼ 0
                                                                 but Im (s) 6¼ 0, which is the onset of oscillatory
A solution of the form                                           instability, an even number of eigenvalues pass
                                                                 from the left-hand complex plane to the right-
u0 ðx; tÞ ¼ est XðxÞ                                             hand complex plane. The eigenvalues are
64              Mathematics in Chemical Engineering
complex conjugates of each other (a result of the       and apply Newton–Raphson with this initial
original equations being real, with no complex          guess and the new value of l. This will be a
numbers), and this is called a Hopf bifurcation.        much better guess of the new solution than just
Numerical methods to study Hopf bifurcation             u0 by itself.
are very computationally intensive and are not             Even this method has difficulties, however.
discussed here [65].                                    Near a limit point the determinant of the
   To return to the problem of solving for the          Jacobian may be zero and the Newton method
steady-state solution: near the limit point or          may fail. Perhaps no solutions exist at all for the
bifurcation-limit point two solutions exist that        chosen parameter l near a limit point. Also, the
are very close to each other. In solving sets of        ability to switch from one solution path to
equations with thousands of unknowns, the               another at a bifurcation-limit point is necessary.
difficulties in convergence are obvious. For            Thus, other methods are needed as well:
some dependent variables the approximation              arc-length continuation and pseudo-arc-length
may be converging to one solution, whereas              continuation [66]. These are described in
for another set of dependent variables it may be        Chapter 1.
converging to the other solution; or the two
solutions may all be mixed up. Thus, solution is
difficult near a bifurcation point, and special         6.8. Sensitivity Analysis
methods are required. These methods are dis-
cussed in [66].                                         Often, when solving differential equations, the
   The first approach is to use natural contin-         solution as well as the sensitivity of the solution
uation (also known as Euler–Newton contin-              to the value of a parameter must be known. Such
uation). Suppose a solution exists for some             information is useful in doing parameter estima-
parameter l. Call the value of the parameter l0         tion (to find the best set of parameters for a
and the corresponding solution u0. Then                 model) and in deciding whether a parameter
                                                        needs to be measured accurately. The differential
F ðu0 ; l0 Þ ¼ 0                                        equation for y (t, a) where a is a parameter, is
Exchanging the order of differentiation in the             that depends upon the location of all the parti-
first term leads to the ordinary differential equa-        cles (but not their velocities). Since the major
tion                                                       part of the calculation is in the evaluation of the
                                                           forces, or potentials, a method must be used that
   
d @y    @ f @y @ f                                         minimizes the number of times the forces are
      ¼       þ
dt @a   @y @a @a                                           calculated to move from one time to another
                                                           time. Rewrite this equation in the form of an
The initial conditions on @y/@a are obtained by            acceleration.
differentiating the initial conditions
                                                           d 2 ri  1
                                                                  ¼ Fi ðfrgÞ 	 ai
                                                           dt2     mi
@                       @y
   ½y ð0; aÞ ¼ y0 ; or    ð0Þ ¼ 0
@a                      @a
                                                           In the Verlot method, this equation is written
Next, let                                                  using central finite differences (Eq. 12). Note
                                                           that the accelerations do not depend upon the
               @y                                          velocities.
y1 ¼ y; y2 ¼
               @a
                                                           ri ðt þ DtÞ ¼ 2ri ðtÞ  ri ðt  DtÞ þ ai ðtÞDt2
and solve the set of ordinary differential equa-
tions                                                      The calculations are straightforward, and no
                                                           explicit velocity is needed. The storage require-
dy1
    ¼ f ðy1 ; aÞ                 y1 ð0Þ ¼ y0               ment is modest, and the precision is modest (it
 dt
                                                           is a second-order method). Note that one must
dy2 @ f              @f
    ¼   ðy ; aÞ y2 þ             y2 ð0Þ ¼ 0                start the calculation with values of {r} at time t
 dt   @y 1           @a
                                                           and tDt.
Thus, the solution y (t, a) and the derivative with            In the Velocity Verlot method, an equation is
respect to a are obtained. To project the impact           written for the velocity, too.
of a, the solution for a ¼ a1 can be used:
                                                           dvi
                                                               ¼ ai
                                                           dt
                           @y
y ðt; aÞ ¼ y1 ðt; a1 Þ þ      ðt; a1 Þða  a1 Þ þ   
                           @a
                                                           The trapezoid rule (see 2.4.) is applied to obtain
  ¼ y1 ðt; a1 Þ þ y2 ðt; a1 Þða  a1 Þ þ   
                                                                                 1
This is a convenient way to determine the                  vi ðt þ DtÞ ¼ vi ðtÞ þ ½ai ðtÞ þ ai ðt þ DtÞDt
                                                                                 2
sensitivity of the solution to parameters in the
problem.                                                   The position of the particles is expanded in a
                                                           Taylor series.
     d 2 ri
                                                           7. Ordinary Differential Equations
mi          ¼ Fi ðfrgÞ; Fi ðfrgÞ ¼ rV
     dt2                                                   as Boundary Value Problems
where mi is the mass of the i-th particle, ri is the       Diffusion problems in one dimension lead to
position of the i-th particle, Fi is the force acting      boundary value problems. The boundary con-
on the i-th particle, and V is the potential energy        ditions are applied at two different spatial
66                    Mathematics in Chemical Engineering
and integrated                                                         Dp r 2
                                                            hv ¼            þ c1 lnr þ c2
                                                                       L 4
Zy               Zt
      dy0
             ¼        dt ¼ t                                Now the two unknowns must be specified from
     f ðy0 Þ
y0               0                                          the boundary conditions. This problem is a
                                                            two-point boundary value problem because
If the quadrature can be performed analytically,            one of the conditions is usually specified at
the exact solution has been found.                          r ¼ 0 and the other at r ¼ R, the tube radius.
    As an example, consider the flow of a non-              However, the technique of separating variables
Newtonian fluid in a pipe, as illustrated                   and integrating works quite well.
in Figure 24. The governing differential                       When the fluid is non-Newtonian, it
                                                            may not be possible to do the second step
                                                            analytically. For example, for the Bird–
                                                            Carreau fluid [74, p. 171], stress and velocity
                                                            are related by
                                                                            h0
                                                            t¼h           
dv
2 ið1nÞ=2
                                                                  1þl      dr
   Putting this value into the equation for stress        The model for a chemical reactor with axial
as a function of r gives                              diffusion is
            h0              Dp r c1                    1 dc2 dc
h         
dv
2 ið1nÞ=2 ¼  L 2 þ r                          ¼ DaRðcÞ
    1þl                                                Pe dz2 dz
           dr
                                                           1 dc                     dc
                                                                ð0Þ þ cð0Þ ¼ cin ;    ð1Þ ¼ 0
This equation cannot be solved analytically for            Pe dz                    dz
dv/dr, except for special values of n. For prob-
lems such as this, numerical methods must be          where Pe is the Peclet number and Da the
used.                                                 Damk€ohler number.
                                                          The boundary conditions are due to
7.2. Initial Value Methods                            DANCKWERTS [75] and to WEHNER and WILHELM
                                                      [76]. This problem can be treated by using
An initial value method is one that utilizes the      initial value methods also, but the method is
techniques for initial value problems but allows      highly sensitive to the choice of the parameter s,
for an iterative calculation to satisfy all the       as outlined above. Starting at z ¼ 0 and making
boundary conditions. Suppose the nonlinear            small changes in s will cause large changes in
boundary value problem                                the solution at the exit, and the boundary con-
                                                      dition at the exit may be impossible to satisfy.
                                                    By starting at z ¼ 1, however, and integrating
d2 y           dy
     ¼ f x; y;
dx 2           dx                                     backwards, the process works and an iterative
                                                      scheme converges in many cases [77]. How-
with the boundary conditions                          ever, if the problem is extremely nonlinear the
                                                      iterations may not converge. In such cases, the
a0 yð0Þ  a1
                dy
                   ð0Þ ¼ a; ai  0                    methods for boundary value problems
                dx                                    described below must be used.
                dy                                        Packages to solve boundary value problems
b0 yð1Þ  b1       ð1Þ ¼ b; bi  0
                dx
                                                      are available on the internet. On the NIST web
                                                      page, http://gams.nist.gov/ choose “problem
Convert this second-order equation into two
                                                      decision tree”, and then “differential and integral
first-order equations along with the boundary
                                                      equations”, then “ordinary differential equa-
conditions written to include a parameter s.
                                                      tions”, “multipoint boundary value problems”.
du
                                                      On the Netlib web site, http://www.netlib.org/,
   ¼v                                                 search on “boundary value problem”. Any
dx
dv                                                    spreadsheet that has an iteration capability can
   ¼ f ðx; u; vÞ
dx                                                    be used with the finite difference method. Some
uð0Þ ¼ a1 s  c1 a                                    packages for partial differential equations also
vð0Þ ¼ a0 s  c0 a
                                                      have a capability for solving one-dimensional
                                                      boundary value problems [e.g., Comsol
The parameters c0 and c1 are specified by the         Multiphysics (formerly Femlab)].
analyst such that
the transport coefficients at the last value and                     particularly for chemical reaction engineering.
then solve                                                           In the collocation method [3], the dependent
                                                                     variable is expanded in a series.
Dðckiþ1=2 Þðckþ1    kþ1
             iþ1  ci   Þ  Dðcki1=2 Þðcikþ1  ckþ1
                                                 i1 Þ
                                                         ¼0
                            Dx2                                                 X
                                                                                Nþ2
                                                                     yðxÞ ¼               ai yi ðxÞ                             ð18Þ
                                                                                i¼1
The advantage of this approach is that it is
easier to program than a full Newton–Raphson
method. If the transport coefficients do not vary                    Suppose the differential equation is
radically, the method converges. If the method
does not converge, use of the full Newton–                           N½y ¼ 0
Raphson method may be necessary.
    Three ways are commonly used to evaluate                         Then the expansion is put into the differential
the transport coefficient at the midpoint. The                       equation to form the residual:
first one employs the transport coefficient eval-
uated at the average value of the solutions on                                                X
                                                                                              Nþ2
                                                                     Residual ¼ N½                    ai yi ðxÞ
either side:                                                                                  i¼1
                                  
                     1
Dðciþ1=2 Þ  D         ðciþ1 þ ci Þ
                     2                                               In the collocation method, the residual is set to
                                                                     zero at a set of points called collocation
The second approach uses the average of the                          points:
transport coefficients on either side:
                                                                       XN þ2
            1                                                        N½      ai yi ðxj Þ ¼ 0; j ¼ 2; . . . ; N þ 1
Dðciþ1=2 Þ  ½Dðciþ1 Þ þ Dðci Þ                              ð17Þ      i¼1
            2
The orthogonal collocation method has found                          Furthermore, if Equation (18) is differentiated
widespread application in chemical engineering,                      once and evaluated at all collocation points, the
                                                                            Mathematics in Chemical Engineering             71
first derivative can be written in terms of the                              Note that 1 is the first collocation point (x ¼ 0)
values at the collocation points:                                            and N þ 2 is the last one (x ¼ 1). To apply the
                                                                             method, the matrices Aij and Bij must be found
dy         X
           Nþ2
                 dy                                                          and the set of algebraic equations solved, per-
   ðxj Þ ¼     ai i ðxj Þ; j ¼ 1; . . . ; N þ 2
dx         i¼1
                 dx                                                          haps with the Newton–Raphson method. If
                                                                             orthogonal polynomials are used and the col-
                                                                             location points are the roots to one of the
This can be expressed as                                                     orthogonal polynomials, the orthogonal collo-
                                                                             cation method results.
dy         XN þ2
                                    dy                                          In the orthogonal collocation method the
   ðxj Þ ¼       ½yi ðxk Þ1 yðxk Þ i ðxj Þ; j ¼ 1; . . . ; N þ 2
dx         i;k¼1
                                    dx                                       solution is expanded in a series involving
                                                                             orthogonal polynomials, where the polynomi-
                                                                             als Pi1(x) are defined in Section 2.2.
or shortened to
           X
           Nþ2                                                                                     X
                                                                                                   N
dy                                                                            y ¼ a þ bx þ xð1  xÞ ai Pi1 ðxÞ
   ðxj Þ ¼     Ajk yðxk Þ;
dx         k¼1                                                                                            i¼1
                                                                                                                           ð22Þ
          X
          N þ2
                                dyi                                               X
                                                                                  Nþ2
Ajk ¼            ½yi ðxk Þ1       ðxj Þ                                     ¼         bi Pi1 ðxÞ
          i¼1
                                dx                                                i¼1
             X
             N þ2                                                            y¼           di xi1
d2 y
     ðxj Þ ¼      Bjk yðxk Þ;                                                      i¼1
dx2          k¼1
          X
          Nþ2
                                d2 yi
Bjk ¼            ½yi ðxk Þ1         ðxj Þ                                  The collocation points are shown in Figure 26.
          i¼1
                                dx2
                                                                             There are N interior points plus one at each end,
   This method is next applied to the differ-                                and the domain is always transformed to lie on
ential equation for reaction in a tubular                                    0 to 1. To define the matrices Aij and Bij this
reactor, after the equation has been made                                    expression is evaluated at the collocation point-
nondimensional so that the dimensionless                                     s; it is also differentiated and the result is
length is 1.0.                                                               evaluated at the collocation points.
                                                                                          X
                                                                                          Nþ2
1 d2 c dc                                                                     yðxj Þ ¼          di xi1
       ¼ Da RðcÞ;                                                                                  j
Pe dx2 dx                                                                                 i¼1
                                                                     ð19Þ
    dc                        dc
      ð0Þ ¼ Pe½cð0Þ  cin ;    ð1Þ ¼ 0                                      dy         X
                                                                                         Nþ2
    dx                        dx                                                 ðxj Þ ¼     di ði  1Þxji2
                                                                              dx         i¼1
1 XNþ2              X
                    Nþ2
       Bjk cðxk Þ      Ajk cðxk Þ ¼ Da Rðcj Þ                       ð20Þ
Pe k¼1              k¼1
 X
 Nþ2
 Alk cðxk Þ ¼ Peðc1  cin Þ;
    k¼1
                                                                     ð21Þ
X
N þ2
     ANþ2;k cðxk Þ ¼ 0
k¼1                                                                          Figure 26. Orthogonal collocation points
72             Mathematics in Chemical Engineering
These formulas are put in matrix notation,            found; and once d is known, the solution for
where Q, C, and D are N þ 2 by N þ 2 matrices.        any x can be found.
                                                          To use the orthogonal collocation method,
y ¼ Qd;
          dy      d2 y
             ¼ Cd; 2 ¼ Dd
                                                      the matrices are required. They can be calcu-
          dx      dx                                  lated as shown above for small N (N < 8) and by
Qji ¼ xji1 ; Cji ¼ ði  1Þxji2 ;                    using more rigorous techniques, for higher N
Dji ¼ ði  1Þði  2Þxi3
                                                      (see Chap. 7). However, having the matrices
                     j
                                                      listed explicitly for N ¼ 1 and 2 is useful; this is
In solving the first equation for d, the first and    shown in Table 7.
                                                          For some reaction diffusion problems, the
second derivatives can be written as
                                                      solution can be an even function of x. For
             dy
                                                      example, for the problem
d ¼ Q1 y;      ¼ CQ1 y ¼ Ay;
             dx
                                               ð23Þ
d2 y                                                  d2 c       dc
     ¼ DQ1 y ¼ By                                         ¼ kc;    ð0Þ ¼ 0; cð1Þ ¼ 1                      ð24Þ
dx2                                                   dx2        dx
Thus the derivative at any collocation point          the solution can be proved to involve only even
can be determined in terms of the solution at         powers of x. In such cases, an orthogonal
the collocation points. The same property is          collocation method, which takes this feature
enjoyed by the finite difference method (and          into account, is convenient. This can easily
the finite element method described below),           be done by using expansions that only involve
and this property accounts for some of                even powers of x. Thus, the expansion
the popularity of the orthogonal collocation
                                                                                    X
                                                                                    N
method. In applying the method to                     yðx2 Þ ¼ yð1Þ þ ð1  x2 Þ            ai Pi1 ðx2 Þ
Equation (19), the same result is obtained;                                          i¼1
The polynomials are defined to be orthogonal         orthogonal collocation is applied at the interior
with the weighting function W(x2).                   points
Z1                                                   X
                                                     N þ1
                                                            Bji ci ¼ f2 Rðcj Þ; j ¼ 1; . . . ; N
     Wðx2 ÞPk ðx2 ÞPm ðx2 Þxa1 dx ¼ 0                i¼1
                                             ð25Þ
0
k m1
                                                     and the boundary condition solved for is
                                                     cNþ1 ¼ 1
where the power on xa1 defines the geometry as
planar or Cartesian (a ¼ 1), cylindrical (a ¼ 2),
and spherical (a ¼ 3). An analogous develop-         The boundary condition at x ¼ 0 is satisfied
ment is used to obtain the (N þ 1)(N þ 1)           automatically by the trial function. After the
matrices                                             solution has been obtained, the effectiveness
                                                     factor h is obtained by calculating
           X
           Nþ1
yðxj Þ ¼         di x2i2                                   R1                           P
                                                                                         Nþ1
                     j                                           R½cðxÞxa1 dx                W j Rðcj Þ
           i¼1                                              0                            i¼1
                                                     h	                            ¼
                                                            R1                            P
                                                                                         Nþ1
dy         X
           N þ1                                                  R½cð1Þxa1 dx                W j Rð1Þ
   ðxj Þ ¼      di ð2i  2Þx2i3
                            j                               0                            i¼1
dx         i¼1
                 X
                 Nþ1                                 Note that the effectiveness factor is the average
r2 yðxi Þ ¼          d i r2 ðx2i2 Þjxj              reaction rate divided by the reaction rate eval-
                 i¼1
                                                     uated at the external conditions. Error bounds
           dy                                        have been given for linear problems [83, p.
y ¼ Qd;       ¼ Cd; r2 y ¼ Dd
           dx
                                                     356]. For planar geometry the error is
Qji ¼ x2i2
       j    ; Cji ¼ ð2i  2Þx2i3
                             j    ;
                                                                            w2ð2Nþ1Þ
Dji ¼ r2 ðx2i2 Þjxj                                 Error in h ¼
                                                                       ð2N þ 1Þ!ð2N þ 2Þ!
              dy
d ¼ Q1 y;       ¼ CQ1 y ¼ Ay;                      This method is very accurate for small N (and
              dx
                                                     small f2); note that for finite difference meth-
r2 y ¼ DQ1 y ¼ By
                                                     ods the error goes as 1/N2, which does not
                                                     decrease as rapidly with N. If the solution is
In addition, the quadrature formula is               desired at the center (a frequent situation
                                                     because the center concentration can be the
WQ ¼ f; W ¼ fQ1
                                                     most extreme one), it is given by
                                                                    X
                                                                    N þ1
                                                     cð0Þ ¼ d1             ½Q1 1i yi
where                                                                i¼1
Z1                      X
                        Nþ1                             The collocation points are listed in Table 8.
     x2i2 xa1 dx ¼           W j x2i2
                                    j                For small N the results are usually more
                         j¼1
0
                                                     accurate when the weighting function in
¼
         1
                	 fi                                 Equation (25) is 1  x2. The matrices for
     2i  2 þ a
                                                     N ¼ 1 and N ¼ 2 are given in Table 9 for
                                                     the three geometries. Computer programs to
      As an example, for the problem                 generate matrices and a program to solve
                                                     reaction diffusion problems, OCRXN, are
                                                   available [3, p. 325, p. 331].
  1 d          dc
          xa1      ¼ f2 RðcÞ
xa1 dx        dx                                       Orthogonal collocation can be applied to
dc                                                   distillation problems. STEWART et al. [84, 85]
   ð0Þ ¼ 0; cð1Þ ¼ 1
dx                                                   developed a method using Hahn polynomials
74               Mathematics in Chemical Engineering
Table 8. Collocation points for orthogonal collocation with        The boundary conditions are typically
symmetric polynomials and W ¼ 1
                                                                   dc           dc
                                     Geometry                         ð0Þ ¼ 0;  ð1Þ ¼ Bim ½cð1Þ  cB 
                                                                   dx           dx
N           Planar                 Cylindrical      Spherical
1           0.5773502692           0.7071067812     0.7745966692   where Bim is the Biot number for mass transfer.
2           0.3399810436           0.4597008434     0.5384693101   These become
            0.8611363116           0.8880738340     0.9061793459
3           0.2386191861           0.3357106870     0.4058451514    1 dc
            0.6612093865           0.7071067812     0.7415311856          ðu ¼ 0Þ ¼ 0;
                                                                   Dx1 du
            0.9324695142           0.9419651451     0.9491079123
4           0.1834346425           0.2634992300     0.3242534234
            0.5255324099           0.5744645143     0.6133714327   in the first element;
            0.7966664774           0.8185294874     0.8360311073
            0.9602898565           0.9646596062     0.9681602395
5           0.1488743390           0.2165873427     0.2695431560        1 dc
                                                                              ðu ¼ 1Þ ¼ Bim ½cðu ¼ 1Þ  cB ;
            0.4333953941           0.4803804169     0.5190961292       DxNE du
            0.6794095683           0.7071067812     0.7301520056
            0.8650633667           0.8770602346     0.8870625998
            0.9739065285           0.9762632447     0.9782286581   in the last element. The orthogonal collocation
                                                                   method is applied at each interior collocation
                                                                   point.
                                                                    1 XNP
The reaction–diffusion equation is written as                      Dxk J¼1
                                                                           A1;J cJ ¼ 0;
                   
    1   d        dc    d2 c a  1 dc
            xa1      ¼ 2þ           ¼ f2 RðcÞ                     in the first element;
xa1    dx       dx    dx     x dx
Table 9. Matrices for orthogonal collocation with symmetric polynomials and W¼1x2
76             Mathematics in Chemical Engineering
  These equations can be assembled into an                      the unknown solution is expanded in a series of
overall matrix problem                                          known functions {bi(x)}, with unknown coef-
                                                                ficients {ai}.
A Ac ¼ f
                                                                              X
                                                                              NT
                                                                cðxÞ ¼              ai bi ðxÞ
The form of these equations is special and is                                 i¼1
The error at the collocation points is more                     This process makes the method a Galerkin
accurate, giving what is known as superconver-                  method. The basis for the orthogonality condi-
gence.                                                          tion is that a function that is made orthogonal to
 i            
d                                                             each member of a complete set is then zero. The
 ðy  yexact Þ                     constantjDxj4
dxi           
                 collocation points                             residual is being made orthogonal, and if the
                                                                basis functions are complete, and an infinite
                                                                number of them are used, then the residual is
7.6. Galerkin Finite Element Method                             zero. Once the residual is zero the problem is
                                                                solved. It is necessary also to allow for the
In the finite element method the domain is                      boundary conditions. This is done by integrat-
divided into elements and an expansion is                       ing the first term of Equation (26) by parts and
made for the solution on each finite element.                   then inserting the boundary conditions:
In the Galerkin finite element method an addi-                  Z1                               
                                                                                1  d      a1 dbi
tional idea is introduced: the Galerkin method                       bj ðxÞ
                                                                              xa1 dx
                                                                                        x
                                                                                              dx
                                                                                                    xa1 dx ¼
is used to solve the equation. The Galerkin                     0
Combining this with Equation (26) gives                 in the e-th element. Then the Galerkin method
                                                        is
  NTZ
            1
 X    dbj dbi a1
                                                                    N PZ
                                                                                     1
            x dxai                                           X 1 X      dN J dN I
  i¼1
      dx dx                                                                       ðxe þ uDxe Þa1 duceI
        0
                                                              e
                                                                Dxe I¼1   du du
                                                                                 0
                  "                        #
                   XNT                                                                "                         #
Bim bj ð1Þ               ai bi ð1Þ  cB                          X                       X
                                                                                          NP
                                                         Bim          N J ð1Þ               ceI N I ð1Þ  c1
                    i¼1                         ð28Þ
                                                                   e                      I¼1                                         ð29Þ
       Z1            "              #                                                            "                    #
                      XNT
                                                                 X           Z1                   XNP
¼ f2        bj ðxÞ          ai bi ðxÞ xa1 dx            ¼ f2          Dxe            N J ðuÞR          ceI N I ðuÞ
                      i¼1                                          e                              I¼1
       0                                                                         0
This equation defines the Galerkin method,              The element integrals are defined as
and a solution that satisfies this equation (for
all j ¼ 1, . . . , 1) is called a weak solution.                             Z1
For an approximate solution the equation is                             1             dN J dN I
                                                         BeJI ¼                                ðxe þ uDxe Þa1 du;
                                                                       Dxe             du du
written once for each member of the trial                                    0
                                                                            a1
cð1Þ ¼ cB                                                ðxe þ uDxe Þ                du
then the boundary condition is used (instead of         whereas the boundary element integrals are
Eq. 29) for j ¼ NT,
                                                         B BeJI ¼ Bim N J ð1ÞN I ð1Þ;
X
NT
                                                         F F eJ ¼ Bim N J ð1Þc1
      ai bi ð1Þ ¼ cB
i¼1
                     a1
                                                            CHRISTIANSEN [94]. For a second-order differen-
ðxe þ uk Dxe Þ                                              tial equation and cubic trial functions on finite
                                                            elements, the error in the i-th element is given by
For an application of the finite element method
                                                            k Error ki ¼ cDx4i k uð4Þ ki
in fluid mechanics, see ! Fluid Mechanics.
                                                            Because cubic elements do not have a nonzero
7.7. Cubic B-Splines                                        fourth derivative, the third derivative in adjacent
                                                            elements is used [3, p. 166]:
Cubic B-splines have cubic approximations
                                                                    1 d 3 ci          1 d3 ciþ1
within each element, but first and second deriv-            ai ¼             ; aiþ1 ¼ 3
                                                                   Dx3i du3          Dxiþ1 du3
atives continuous between elements. The func-                               2                               3
tions are the same ones discussed in Chapter 2,                          1 6 ai  ai1           aiþ1  ai 7
                                                            k uð4Þ   ki  4                  þ               5
and they can be used to solve differential                               2 1                   1
                                                                              ðxiþ1  xi1 Þ     ðxiþ2  xi Þ
equations [92].                                                             2                  2
                                                  Mathematics in Chemical Engineering                   79
method of solving such problems is to refine the       The Blasius equation becomes
mesh near the singularity, by relying on the
                                                                              
better approximation due to a smaller Dx.                   d3 f      d2 f
                                                       2 z3 3  3z2 2  z
                                                                            df
Another approach is to incorporate the singular             dz        dz    dz
trial function into the approximation. Thus, if                      
                                                            d2     df
                                                       þf z2 2 þ z      ¼ 0 for 0  z  1:
the solution approaches f(x) as x goes to zero,             dz     dz
and f(x) becomes infinite, an approximation
may be taken as                                        The differential equation now has variable
                                                       coefficients, but these are no more difficult to
                 X
                 N
                                                       handle than the original nonlinearities.
yðxÞ ¼ f ðxÞ þ         ai yi ðxÞ
                 i¼1                                      Another approach is to use a variable mesh,
                                                       perhaps with the same transformation. For
This function is substituted into the differential     example, use z ¼ eh and a constant mesh
equation, which is solved for ai. Essentially, a       size in z. Then with 101 points distributed
new differential equation is being solved for a        uniformly from z ¼ 0 to z ¼ 1, the following
new variable:                                          are the nodal points:
     d3    d2
2       þf 2 ¼0
    dh3   dh
                                                       8. Partial Differential Equations
    df
f ¼    ¼ 0 at h ¼ 0
    dh
                                                       Partial differential equations are differential
df
   ¼ 1 at h ! 1
                                                       equations in which the dependent variable is
dh                                                     a function of two or more independent varia-
                                                       bles. These can be time and one space dimen-
Because one boundary is at infinity using a            sion, or time and two or more space dimensions,
mesh with a constant size is difficult! One            or two or more space dimensions alone. Prob-
approach is to transform the domain. For exam-         lems involving time are generally either hyper-
ple, let                                               bolic or parabolic, whereas those involving
                                                       spatial dimensions only are often elliptic.
z ¼ eh                                                Because the methods applied to each type of
                                                       equation are very different, the equation must
                                                       first be classified as to its type. Then the special
Then h ¼ 0 becomes z ¼ 1 and h ¼ 1 becomes             methods applicable to each type of equation are
z ¼ 0. The derivatives are                             described. For a discussion of all methods, see
                                                       [100–103]; for a discussion oriented more
dz              d2 z                                   toward chemical engineering applications, see
   ¼ eh ¼ z;      ¼ eh ¼ z
dh              dh2
                                                       [104]. Examples of hyperbolic and parabolic
df    df dz      df                                    equations include chemical reactors with radial
    ¼       ¼ z
dh dz dh         dz
                                                       dispersion, pressure-swing adsorption, disper-
        2  2
 2
d f
     ¼
       d f dz
                  þ
                    df d2 z     d2 f
                            ¼ z2 2 þ z
                                       df              sion of an effluent, and boundary value prob-
dh2 dz2 dh          dz dh2      dz     dz              lems with transient terms added (heat transfer,
                                                               Mathematics in Chemical Engineering              81
mass transfer, chemical reaction). Examples of                  type. The surface is defined by this equation
elliptic problems include heat transfer and mass                plus a normalization condition:
transfer in two and three spatial dimensions and
steady fluid flow.                                              X
                                                                n
                                                                       s 2k ¼ 1
                                                                 k¼0
                                                                Thus,
where
                                                                 s 21 þ s 22 ¼ 1 ðnormalizationÞ
                                             X
                                             n
a ¼ ða0 ; a1 ; . . . ; an Þ; jaj ¼                 ai            s 21 þ s 22 ¼ 0 ðequationÞ
                                             i¼0
                @ jaj
@a ¼
         @ta0 @xa1 1 . . . @xan n                               These cannot both be satisfied so the problem is
                                                                elliptic. When the equation is
the characteristic equation for P is defined as
 X                                                              @2 u @2 u
                                                                         ¼0
        aa s a ¼ 0; s ¼ ðs 0 ; s 1 ; . . . ; s n Þ              @t2 @x2
jaj¼m                                                   ð30Þ
s a ¼ s a0 0 s a1 1 . . . s an n
                                                                then
where s represents coordinates. Thus only the
highest derivatives are used to determine the                   j20 þ j21 ¼ 0
82                    Mathematics in Chemical Engineering
Now real j0 can be solved and the equation is                  First-order quasi-linear problems are written
hyperbolic                                                  in the form
s 20 þ s 21 ¼ 1 ðnormalizationÞ                             Xn
                                                                   @u
                                                                A1     ¼ f; x ¼ ðt; x1 . . . ; xn Þ
s 20    þ   s 21   ¼ 0 ðequationÞ                          l¼0
                                                                   @x1                                          ð31Þ
                                                            u ¼ ðu1 ; u2 ; . . . ; uk Þ
When the equation is
       2       
@c     @ c @2c                                              The matrix entries A1 is a kk matrix whose
   ¼D       þ 2
@t     @x 2  @y                                             entries depend on u but not on derivatives of u.
                                                            Equation (31) is hyperbolic if
then
                                                            A ¼ Am
s 20 þ s 21 þ s 22 ¼ 1 ðnormalizationÞ
bs 20 þ s 21 ¼ 0 n ¼ 1; A0 ¼ 1; A1 ¼ FðuÞ
is required. Combining these gives                          Thus, the roots are real and the equation is
                                                            hyperbolic.
1  ð1 þ bÞs 20 ¼ 0                                            The final example is the heat conduction
                                                            problem written as
The roots are real and the equation is hyper-
bolic. When b ¼ 0                                           %C p
                                                                   @T    @q
                                                                      ¼  ; q ¼ k
                                                                                   @T
                                                                   @t    @x        @x
j21 ¼ 0
                                                            In this formulation the constitutive equation for
and the equation is parabolic.                              heat flux is separated out; the resulting set of
                                                    Mathematics in Chemical Engineering             83
and for Equation (35):                                 The concentration profile is steeper for the
    dci       ci  ci1          df dci                MacCormack method than for the upstream
f       þ fui           þ ð1  fÞ ji    ¼0             derivatives, but oscillations can still be present.
    dt           Dx              dc dt
                                                       The flux-corrected transport method can be
If the flow were from right to left, then the          added to the MacCormack method. A solution
formula would be                                       is obtained both with the upstream algorithm
dci Fðciþ1 Þ  Fðci Þ    ciþ1  2ci þ ci1
                                                       and the MacCormack method; then they are
   þ                  ¼D                               combined to add just enough diffusion to elim-
dt         Dx                  Dx2
                                                       inate the oscillations without smoothing the
If the flow could be in either direction, a local      solution too much. The algorithm is compli-
determination must be made at each node i and          cated and lengthy but well worth the effort
the appropriate formula used. The effect of            [107–109].
using upstream derivatives is to add artificial           If finite element methods are used, an
or numerical diffusion to the model. This can be       explicit Taylor–Galerkin method is appropriate.
ascertained by taking the finite difference form       For the convective diffusion equation the
of the convective diffusion equation                   method is
dci    ci  ci1    ciþ1  2ci þ ci1                  1 nþ1            2                  1
    þu           ¼D                                     ðc    cniþ1 Þ þ ðcnþ1    cni Þ þ ðcnþ1  cni1 Þ
dt        Dx              Dx2                          6 iþ1            3 i                6 i1
                                                                                               
                                                           uDt n                   DtD u2 Dt2
and rearranging                                        ¼      ðciþ1  cni1 Þ þ         þ        ðcniþ1  2cni þ cni1 Þ
                                                           2Dx                     Dx2     2Dx2
                                                                           ac
                                                               f ðcÞ ¼
                                                                         1 þ Kc
Otherwise, if x is changed one side changes,          The boundary conditions for u are
but the other cannot because it depends on t.
Call the constant  l and write the separate          uð0; tÞ ¼ 0
equations                                             uðL; tÞ ¼ 0
TðtÞ ¼ Tð0ÞelDt
                                                      Separation of variables is now applied to this
                                                      equation by writing
and the second equation is written in the form
                                                      uðx; tÞ ¼ TðtÞXðxÞ
d2 X
     þ lX ¼ 0
dx2
                                                      The same equation for T (t) and X (x) is
   Next consider the boundary conditions. If          obtained, but with X (0) ¼ X (L) ¼ 0.
they are written as
                                                      d2 X
                                                           þ lX ¼ 0
                                                      dx2
cðL; tÞ ¼ 1 ¼ TðtÞXðLÞ
                                                      Xð0Þ ¼ XðLÞ ¼ 0
cð0; tÞ ¼ 0 ¼ TðtÞXð0Þ
the boundary conditions are difficult to satisfy          Next X (x) is solved for. The equation
because they are not homogeneous, i.e. with a         is an eigenvalue problem. The general solution
zero right-hand side. Thus, the problem must          is obtained by using emx pffiffiffi and finding that
be transformed to make the boundary condi-            m2þ l ¼ 0; thus m ¼ 
i l. The exponential
tions homogeneous. The solution is written as         term
the sum of two functions, one of which                   pffiffi
                                                      e
i lx
satisfies the nonhomogeneous boundary con-
ditions, whereas the other satisfies the homo-
geneous boundary conditions.                          is written in terms of sines and cosines, so that
                                                      the general solution is
cðx; tÞ ¼ f ðxÞ þ uðx; tÞ
                                                              pffiffiffi      pffiffiffi
uð0; tÞ ¼ 0                                           X ¼ Bcos lx þ Esin lx
uðL; tÞ ¼ 0
                                                      The boundary conditions are
Thus, f (0) ¼ 1 and f (L) ¼ 0 are necessary. Now                 pffiffiffi      pffiffiffi
the combined function satisfies the boundary          XðLÞ ¼ Bcos lL þ Esin lL ¼ 0
conditions. In this case the function f (x) can be    Xð0Þ ¼ B ¼ 0
taken as
                                                      If B ¼ 0, then E 6¼ 0 is required to have any
f ðxÞ ¼ L  x                                         solution at all. Thus, l must satisfy
                                                         pffiffiffi
The equation for u is found by substituting for c     sin lL ¼ 0
in the original equation and noting that the f (x)
drops out for this case; it need not disappear in     This is true for certain values of l, called
the general case:                                     eigenvalues or characteristic values. Here,
                                                      they are
@u   @2u
   ¼D 2
@t   @x                                               ln ¼ n2 p2 =L2
                                                    Mathematics in Chemical Engineering                   87
Each eigenvalue has a corresponding eigen-           number of terms are used, some error always
function                                             occurs. For large times a single term is ade-
                                                     quate, whereas for small times many terms are
X n ðxÞ ¼ Esin n p x=L                               needed. For small times the Laplace transform
                                                     method is also useful, because it leads to
The composite solution is then                       solutions that converge with fewer terms. For
                                                     small times, the method of combination of
                               n p x ln Dt
X n ðxÞT n ðtÞ ¼ E A sin            e                variables may be used as well. For nonlinear
                                 L
                                                     problems, the method of separation of variables
This function satisfies the boundary conditions      fails and one of the other methods must be
and differential equation but not the initial        used.
condition. To make the function satisfy the             The method of combination of variables is
initial condition, several of these solutions        useful, particularly when the problem is posed
are added up, each with a different eigenfunc-       in a semi-infinite domain. Here, only one exam-
tion, and E A is replaced by An.                     ple is provided; more detail is given in [3, 113,
                                                     114]. The method is applied here to the non-
            X
            1
                             n p x n2 p2 Dt=L2
                                                     linear problem
uðx; tÞ ¼           An sin        e
                               L
            n¼1                                                                          
                                                     @c   @        @c         @ 2 c d DðcÞ @c 2
                                                        ¼     DðcÞ      ¼ DðcÞ 2 þ
                                                     @t @x         @x         @x      dc   @x
The constants An are chosen by making u (x, t)
satisfy the initial condition.
                                                     with boundary and initial conditions
            X
            1
                             npx x
uðx; 0Þ ¼           An sin      ¼ 1                  cðx; 0Þ ¼ 0
            n¼1
                              L  L
                                                      cð0; tÞ ¼ 1; cð1; tÞ ¼ 0
Next, the Galerkin method is applied, and the        The use of the 4 and D0 makes the analysis below
residual is made orthogonal to a complete set of     simpler. The equation for c (x, t) is transformed
functions, which are the eigenfunctions.             into an equation for f (h)
                                                      @c df @h @c df @h
ZL                                                        ¼     ;   ¼
       x        mpx                                  @t dh @t @x dh @x
          1 sin     dx                                           
       L          L                                   @ 2 c d2 f @h 2 df @ 2 h
0                                                           ¼         þ
                                                      @x2 dh2 @x        dh @x2
    X1    Z     L
               mpx     npx      Am
¼       An sin     sin     dx ¼                       @h      x=2            @h     1           @2 h
                L       L        2                       ¼  pffiffiffiffiffiffiffiffiffiffiffiffi ;    ¼ pffiffiffiffiffiffiffiffiffiffi ;      ¼0
    n¼1
            0                                         @t      4D0 t3 @x            4D0 t @x2
points. Thus, the boundary conditions can be            where the matrix B is tridiagonal. The stability
combined to give                                        of the integration of these equations is governed
                                                        by the largest eigenvalue of B. If Euler’s
f ð1Þ ¼ 0                                               method is used for integration,
                                                              D    2
The other boundary condition is for x ¼ 0 or            Dt      
                                                             Dx2 jljmax
h ¼ 0,
whereas the oscillation limit is given by                    boundary value problems by letting the solution
DDt   0:25
                                                             at the nodes depend on time. For the diffusion
                                                            equation the finite element method gives
Dx2 1  u
ADxðcnþ1
     i    cni Þ ¼ DtAðJ j1=2  J iþ1=2 Þ                   Now the matrix C C is not diagonal, so that a set
                                                             of equations must be solved for each time step,
where J is the flux due to convection and                    even when the right-hand side is evaluated
diffusion, positive in the þx direction.                     explicitly. This is not as time-consuming as it
                                                             seems, however. The explicit scheme is written
             @c                               ci  ci1=2
J ¼ uc  D      ; J i1=2 ¼ ui1=2 ci1=2  D                as
             @x                                   Dx
                                                                     cnþ1  cni
   The concentration at the edge of the cell is              CC ji    i
                                                                                ¼ AAji cni
                                                                        Dt
taken as
        1                                                    and rearranged to give
ci1=2 ¼ ðci þ ci1 Þ
        2
                                                              CC ji ðcnþ1
                                                                      i    cni Þ ¼ DtAAji cni or
   Rearrangement for the case when the veloc-
                                                              CCðcnþ1  cn Þ ¼ DtAAc
ity u is the same for all nodes gives
cnþ1
 i    cni uðciþ1  ci1 Þ   D                               This is solved with an L U decomposition (see
          þ                ¼ 2 ðciþ1  2ci þ ci1 Þ
   Dt           2Dx         Dx                               Section 1.2) that retains the structure of the mass
                                                             matrix C C. Thus,
    This is the same equation as obtained using
the finite difference method. This is not always             CC ¼ LU
true, and the finite volume equations are easy to
derive. In two- and three-dimensions, the mesh               At each step, calculate
need not be rectangular, as long as it is possible
to compute the velocity normal to an edge of the
                                                             cnþ1  cn ¼ DtU1 L1 AAcn
cell. The finite volume method is useful for
applications involving filling, such as injection
molding, when only part of the cell is filled with           This is quick and easy to do because the inverse
fluid. Such applications do involve some                     of L and U are simple. Thus the problem is
approximations, since the interface is not                   reduced to solving one full matrix problem and
tracked precisely, but they are useful engineer-             then evaluating the solution for multiple right-
ing approximations.                                          hand sides. For implicit methods the same
    The finite element method is handled in a                approach can be used, and the LU decomposi-
similar fashion, as an extension of two-point                tion remains fixed until the time step is changed.
90                  Mathematics in Chemical Engineering
   The method of orthogonal collocation uses              and evaluated at each collocation point
a similar extension: the same polynomial of x
                                                                                  n
is used but now the coefficients depend on                unþ1  unj           @u 
                                                                     þ f ðunj Þ  ¼ 0
                                                           j
time.                                                        Dt                @x j
   The prototype elliptic problem is steady-                              And this is converted to the Gauss–Seidel iter-
state heat conduction or diffusion,                                       ative method.
     2                                                                           
     @ T @2T                                                                    Dx2
k         þ 2 ¼Q                                                           2 1 þ 2 T sþ1      s         sþ1
                                                                                      i;j ¼ T iþ1;j þ T i1;j
     @x 2  @y                                                                   Dy
                                                                               Dx2 s                      2 Qi;j
                                                                           þ      ðT         i;j1 Þ  Dx
                                                                                         þ T sþ1
                                                                               Dy2 i;jþ1                     k
possibly with a heat generation term per unit
volume, Q. The boundary conditions can be                                 Calculations proceed by setting a low i, com-
                                                                          puting from low to high j, then increasing i and
Dirichlet or 1st kind:              T ¼ T1 on boundary S1                 repeating the procedure. The relaxation method
                                                                          uses
Neumann or 2nd kind:                k @T
                                      @n ¼ q2 on boundary S2
                                                                                  
                                    k @T                                       Dx2                            Dx2
Robin, mixed, or 3rd kind:             @n ¼ hðT  T 3 Þ on boundary S3     2 1 þ 2 T i;j ¼ T siþ1;j þ T sþ1
                                                                                                        i1;j þ
                                                                                Dy                              Dy2
                                                                                                          Qi;j
                                                                                         i;j1 Þ  Dx
                                                                           ðT si;jþ1  T sþ1          2
   Illustrations are given for constant physical                                                           k
properties k, h, while T1, q2, T3 are known
functions on the boundary and Q is a known
function of position. For clarity, only a two-                            T sþ1     s            
                                                                                                      s
                                                                            i;j ¼ T i;j þ bðT i;j  T i;j Þ
dimensional problem is illustrated. The finite
difference formulation is given by using the                              If b ¼ 1, this is the Gauss–Seidel method. If
following nomenclature                                                    b > 1, it is overrelaxation; if b < 1, it is
                                                                          underrelaxation. The value of b may be chosen
T i;j ¼ TðiDx; jDyÞ                                                       empirically, 0 < b < 2, but it can be selected
                                                                          theoretically for simple problems like this [117,
The finite difference formulation is then                                 p. 100], [3, p. 282]. In particular, the optimal
                                                                          value of the iteration parameter is given by
T iþ1;j  2T i;j þ T i1;j T i;jþ1  2T i;j þ T i;j1
                          þ                           ¼ Qi;j =k   ð40Þ
          Dx2                        Dy2                                  lnðbopt  1Þ  R
If the boundary is parallel to a coordinate axis                          (when there are n points in both x and y direc-
the boundary slope is evaluated as in Chapter 7,                          tions). For Neumann boundary conditions, the
by using either a one-sided, centered difference                          value is
or a false boundary. If the boundary is more
irregular and not parallel to a coordinate line,
                                                                                 p2               1
more complicated expressions are needed and                               R¼
                                                                                 2n2 1 þ max½Dx2 =Dy2 ; Dy2 =Dx2 
the finite element method may be the better
method.
    Equation (40) is rewritten in the form                                   Iterative methods can also be based on lines
                                                                          (for 2D problems) or planes (for 3D problems).
        
     Dx2                           Dx2                                       Preconditioned conjugate gradient methods
2 1 þ 2 T i;j ¼ T iþ1;j þ T i1;j þ 2
     Dy                            Dy                                     have been developed (see Chap. 1). In these
                             Qi;j                                         methods a series of matrix multiplications are
ðT i;jþ1 þ T i;j1 Þ  Dx2
                              k                                           done iteration by iteration; and the steps lend
92                   Mathematics in Chemical Engineering
themselves to the efficiency available in paral-                                In these equations I and J refer to the nodes
lel computers. In the multigrid method the                                      of the triangle forming element e and the
problem is solved on several grids, each                                        summation is made over all elements. These
more refined than the previous one. In iterating                                equations represent a large set of linear equa-
between the solutions on different grids, one                                   tions, which are solved using matrix techniques
converges to the solution of the algebraic equa-                                (Chap. 1).
tions. A chemical engineering application is                                        If the problem is nonlinear, e.g., with k or Q a
given in [118]. Software for a variety of these                                 function of temperature, the equations must be
methods is available, as described below.                                       solved iteratively. The integrals are given for a
   The Galerkin finite element method                                           triangle with nodes I, J, and K in counter-
(FEM) is useful for solving elliptic problems                                   clockwise order. Within an element,
and is particularly effective when the domain
or geometry is irregular [119–125]. As an                                       T ¼ N I ðx; yÞT I þ N J ðx; yÞT J þ N K ðx; yÞT K
example, cover the domain with triangles
                                                                                     aI þ bI x þ cI y
and define a trial function on each triangle.                                   NI ¼
                                                                                            2D
The trial function takes the value 1.0 at one                                   aI ¼ xJ yK  xK yJ
corner and 0.0 at the other corners, and is
                                                                                bI ¼ yI  yK
linear in between (see Fig. 30). These trial
functions on each triangle are pieced together                                  cI ¼ xK  x J
to give a trial function on the whole domain.                                    plus permutation on I; K; J
For the heat conduction problem the method                                               2
                                                                                           1 xI yI
                                                                                                      3
gives [3]                                                                                6            7
                                                                                2D ¼ det6             7
                                                                                         4 1 xJ yJ 5 ¼ 2ðarea of triangleÞ
XX                         XX
             AeIJ T eJ ¼                F eI                             ð41Þ                 1   xK   yK
 e   J                     e        J
                                                                                aI þ aJ þ aK ¼ 1
where                                                                           bI þ bJ þ bK ¼ 0
              Z                                Z                                cI þ c J þ cK ¼ 0
AeIJ ¼           krN I  rN J dA                  h3 N I N J dC
                                                                                             k
                                               C3                               AeIJ ¼        ðbI bJ þ cI cJ Þ
                                                                                            4D
         Z                     Z                    Z
F eI ¼        N I QdA þ             N I q2 dC                                            Q                        QD
                                                         N I h3 T 3 dC          F eIJ ¼     ðaI þ bI x þ cI y Þ ¼
                                                                                          2                         3
                               C2                   C3
                                                                                     xI þ xJ þ x K       y þ y J þ yK
                                                                                x¼                 ; y ¼ I
                                                                                           3                  3
Also, a necessary condition is that
                                                                                                   2
T i ¼ T 1 on C 1                                                                aI þ bI x þ cI y ¼ D
                                                                                                   3
element and the basis functions. In two dimen-           In the finite difference method an explicit
sions, triangular elements are usually used           technique would evaluate the right-hand side at
because it is easy to cover complicated geome-        the n-th time level:
tries and refine parts of the mesh. However,
rectangular elements sometimes have advan-                      T nþ1     n
                                                                  i;j  T i;j
                                                       rC p
tages, particularly when some parts of the                            Dt
solution do not change very much and the               ¼
                                                             k
                                                               ðT n  2T ni1;j þ T ni1;j Þ
elements can be long. In three dimensions                   Dx2 iþ1;j
the same considerations apply: tetrahedral ele-             k
                                                       þ      ðT n  2T ni;j þ T ni;j1 Þ  Q
                                                           Dy2 i;jþ1
ments are frequently used, but brick elements
are also possible. While linear elements (in both
two and three dimensions) are usually used,           When Q ¼ 0 and Dx ¼ Dy, the time step is
higher accuracy can be obtained by using qua-         limited by
dratic or cubic basis functions within the ele-
                                                                Dx2 %C p    Dx2
ment. The reason is that all methods converge         Dt                or
                                                                  4k        4D
according to the mesh size to some power,
and the power is larger when higher order             These time steps are smaller than for one-
elements are used. If the solution is dis-            dimensional problems. For three dimensions,
continuous, or has discontinuous first deriva-        the limit is
tives, then the lowest order basis functions are
used because the convergence is limited by the                  Dx2
                                                      Dt 
properties of the solution, not the finite element              6D
approximation.
    One nice feature of the finite element            To avoid such small time steps, which must
method is the use of natural boundary condi-          be smaller when Dx decreases, an implicit
tions. In this problem the natural boundary           method could be used. This leads to large
conditions are the Neumann or Robin condi-            sparse matrices, rather than convenient tri-
tions. When using Equation (41), the problem          diagonal matrices.
can be solved on a domain that is shorter than
needed to reach some limiting condition (such
as at an outflow boundary). The externally
                                                      8.6. Special Methods for Fluid
applied flux is still applied at the shorter          Mechanics
domain, and the solution inside the truncated
domain is still valid. Examples are given in          The method of operator splitting is also
[107] and [131]. The effect of this is to allow       useful when different terms in the equation
solutions in domains that are smaller, thus           are best evaluated by using different methods
saving computation time and permitting the            or as a technique for reducing a larger problem
solution in semi-infinite domains.                    to a series of smaller problems. Here the
                                                      method is illustrated by using the Navier–
                                                      Stokes equations. In vector notation the equa-
8.5. Parabolic Equations in Two or                    tions are
Three Dimensions
                                                           @u
                                                      %       þ %u  ru ¼ %f  rp þ mr2 u
Computations become much more lengthy with                 @t
This can be done by using the finite difference        here follows [142]. A lattice is defined, and one
[132, p. 162] or the finite element method [133–       solves the following equation for f i ðx; tÞ, the
135].                                                  probability of finding a molecule at the point x
   In fluid flow problems solved with the              with speed ci.
finite element method, the basis functions
for pressure and velocity are often different.         @f i                f  f eq
                                                            þ ci  rf i ¼  i    i
This is required by the LBB condition (named           @t                     t
after LADYSHENSKAYA, BREZZI, and BABUSKA)
[134, 135]. Sometimes a discontinuous basis               The right-hand side represents a single time
function is used for pressure to meet this             relaxation for molecular collisions, and t is
condition. Other times a penalty term is added,        related to the kinematic viscosity. By means
or the quadrature is done using a small number         of a simple stepping algorithm, the computa-
of quadrature points. Thus, one has to be              tional algorithm is
careful how to apply the finite element method
to the Navier–Stokes equations. Fortunately,                                                    Dt
                                                       f i ðx þ ci Dt; t þ DtÞ ¼ f i ðx; tÞ       ðf  f eq
                                                                                                          i Þ
software exists that has taken this into                                                        t i
account.
                                                       Consider the lattice shown in Figure 34. The
Level Set Methods. Multiphase problems are             various velocities are
complicated because the terms in the
equations depend on which phase exists at a            c1 ¼ ð1; 0Þ; c3 ¼ ð0; 1Þ; c5 ¼ ð1; 0Þ; c7 ¼ ð0; 1Þ
particular point, and the phase boundary may           c2 ¼ ð1; 1Þ; c4 ¼ ð1; 1Þ; c6 ¼ ð1; 1Þ; c8 ¼ ð1; 1Þ
move or be unknown. It is desirable to compute
on a fixed grid, and the level set formulation         c0 ¼ ð0; 0Þ
   The physics governing the velocity of the           where the weighting functions are
interface must be defined, and this equation is            4                     1                       1
solved along with the other equations repre-           w0 ¼ ; w1 ¼ w3 ¼ w5 ¼ w7 ¼ ; w2 ¼ w4 ¼ w6 ¼ w8 ¼
                                                           9                     9                      36
senting the problem [130–137].
                                                       With these conditions, the solution for velocity
Lattice Boltzmann Methods. Another way to              is a solution of the Navier—Stokes equation.
solve fluid flow problems is based on a                These equations lead to a large computational
molecular viewpoint and is called the Lattice          problem, but it can be solved by parallel proc-
Boltzmann method [138–141]. The treatment              essing on multiple computers.
                                                      Mathematics in Chemical Engineering                95
then include approximations and correlations for     evolutionary problems, whereas Fredholm
some features that would be difficult to solve for   equations are analogous to boundary value
directly. Four widely used major packages are        problems. The terms in the integral can be
Fluent (http://www.fluent.com/), CFX (now            unbounded, but still yield bounded integrals,
part of ANSYS), Comsol Multiphysics                  and these equations are said to be weakly
(formerly FEMLAB) (http://www.comsol.com/),          singular. A Volterra equation of the second
and ANSYS (http://www.ansys.com/). Of these,         kind is
Comsol Multiphysics is particularly useful
because it has a convenient graphical user                                 Zt
interface, permits easy mesh generation and re-      yðtÞ ¼ gðtÞ þ l            Kðt; sÞyðsÞds                         ð42Þ
                                                                            a
finement (including adaptive mesh refinement),
allows the user to add in phenomena and
additional equations easily, permits solution        whereas a Volterra equation of the first kind is
by continuation methods (thus enhancing
convergence), and has extensive graphical                       Zt
output capabilities. Other packages are also         yðtÞ ¼ l        Kðt; sÞyðsÞds
available (see http://cfd-online.com/), and                     a
                                                     @T
                                                        ¼ GðT; tÞ; x ¼ 0; t > 0
                                                     @x
9.1. Classification
                                                     The solution to this problem is
Volterra integral equations have an integral with
a variable limit, whereas Fredholm integral                           Zt                1
                                                                1
equations have a fixed limit. Volterra equations                            G Tð0; sÞ; s pffiffiffiffiffiffiffiffiffiffi ex =4ðtsÞ ds
                                                                                                        2
                                                     Tðx; tÞ ¼ pffiffiffi
                                                                 p                        ts
are usually associated with initial value or                           0
                                                         Mathematics in Chemical Engineering                 97
If T (t) is used to represent T (0, t), then             problems. An eigenvalue problem is a homoge-
                                                          neous equation of the second kind.
               Zt               1
        1
T ðtÞ ¼ pffiffiffi           G TðsÞ; s pffiffiffiffiffiffiffiffiffiffi ds                         Zb
          p                       ts
               0                                          yðxÞ ¼ l           Kðx; sÞyðsÞds                   ð44Þ
                                                                        a
Thus the behavior of the solution at the bound-
ary is governed by an integral equation. NAGEL            Solutions to this problem occur only for spe-
and KLUGE [156] use a similar approach to solve           cific values of l, the eigenvalues. Usually the
for adsorption in a porous catalyst.                      Fredholm equation of the second or first kind is
   The existence and uniqueness of the solution           solved for values of l different from these,
can be proved [151, p. 30, 32].                           which are called regular values.
   Sometimes the kernel is of the form                        Nonlinear Volterra equations arise naturally
                                                          from initial value problems. For the initial value
Kðt; sÞ ¼ Kðt  sÞ                                        problem
                                                          dy           
Equations of this form are called convolution                ¼ F t; yðtÞ
                                                          dt
equations and can be solved by taking the
Laplace transform. For the integral equation
                                                          both sides can be integrated from 0 to t to obtain
                        Zt
YðtÞ ¼ GðtÞ þ l              Kðt  tÞYðtÞdt                                    Zt           
                        0                                 yðtÞ ¼ yð0Þ þ             F s; yðsÞ ds
               Zt                                                              0
Solving this for y(s) gives                               Theorem [151, p. 55]. If g (t) is continuous, the
                                                          kernel K (t, s, y) is continuous in all variables
yðsÞ ¼
           gðsÞ                                           and satisfies a Lipschitz condition
         1  kðsÞ
                                                                                    Zt
whereas a Fredholm equation of the first kind is          ynþ1 ðtÞ ¼ gðtÞ þ              K½t; s; yn ðsÞds
                                                                                    0
Zb
     Kðx; sÞyðsÞds ¼ gðxÞ
                                                            Nonlinear Fredholm equations have special
a
                                                          names. The equation
                                                                    Z1
The limits of integration are fixed, and these            f ðxÞ ¼        K½x; y; f ðyÞdy
problems are analogous to boundary value                            0
98                  Mathematics in Chemical Engineering
is called the Urysohn equation [150, p. 208].                           solve for y1, y2, . . . in succession. For a single
The special equation                                                    integral equation, at each step one must solve a
          Z1
                                                                        single nonlinear algebraic equation for yn. Typ-
f ðxÞ ¼        K½x; yF½y; f ðyÞdy                                     ically, the error in the solution to the integral
          0                                                             equation is proportional to Dtm, and the power
                                                                        m is the same as the power in the quadrature
is called the Hammerstein equation [150, p.                             error [151, p. 97].
209]. Iterative methods can be used to solve                               The stability of the method [151, p. 111] can
these equations, and these methods are closely                          be examined by considering the equation
tied to fixed point problems. A fixed point
problem is                                                                              Zt
                                                                        yðtÞ ¼ 1  l         yðsÞds
x ¼ FðxÞ                                                                                0
Local convergence theorems prove the process                            Since the integral equation can be differentiated
convergent if the solution is close enough to the                       to obtain the initial value problem
answer, whereas global convergence theorems
are valid for any initial guess [150, p. 229–231].                      dy
                                                                           ¼ ly; yð0Þ ¼ 1
The successive substitution method for non-                             dt
linear Fredholm equations is
                                                                        the stability results are identical to those for
               Z1                                                       initial value methods. In particular, using the
ynþ1 ðxÞ ¼          K½x; s; yn ðsÞds                                   trapezoid rule for integral equations is identical
               0                                                        to using this rule for initial value problems. The
                                                                        method is A-stable.
Typical conditions for convergence include that                            Higher order integration methods can also
the function satisfies a Lipschitz condition.                           be used [151, p. 114, 124]. When the kernel is
                                                                        infinite at certain points, i.e., when the problem
                                                                        has a weak singularity, see [151, p. 71, 151].
9.2. Numerical Methods for Volterra
Equations of the Second Kind
Volterra equations of the second kind are ana-                          9.3. Numerical Methods for Fredholm,
logous to initial value problems. An initial                            Urysohn, and Hammerstein Equations
value problem can be written as a Volterra                              of the Second Kind
equation of the second kind, although not all
Volterra equations can be written as initial value                      Whereas Volterra equations could be solved
problems [151, p. 7]. Here the general nonlinear                        from one position to the next, like initial value
Volterra equation of the second kind is treated                         differential equations, Fredholm equations
(Eq. 46). The simplest numerical method                                 must be solved over the entire domain, like
involves replacing the integral by a quadrature                         boundary value differential equations. Thus,
using the trapezoid rule.                                               large sets of equations will be solved and the
                                                                        notation is designed to emphasize that.
yn 	 yðtn Þ ¼ gðtn Þ
           (                                                        )      The methods are also based on quadrature
             1               X
                             n1
                                                   1
     þ Dt Kðtn ; t0 ; y0 Þ þ     Kðtn ; ti ; yi Þ þ Kðtn ; tn ; yn Þ    formulas. For the integral
             2               i¼1
                                                   2
                                                                                 Zb
This equation is a nonlinear algebraic equation                         IðwÞ ¼        fðyÞdy
for yn. Since y0 is known it can be applied to                                   a
                                                                                  Mathematics in Chemical Engineering                                    99
is a known function. The integral equation is maxax;yb jF½x; y; f ðyÞ F½x; z; f ðzÞj Kjy zj
then replaced by
                                                                                   Theorem [150, p. 214]. If the constant K is < 1
                         X
                         n
f ðxÞ ¼ gðxÞ þ                 wi Kðx; yi Þ½f ðyi Þ  f ðxÞ þ f ðxÞHðxÞ           and certain other conditions hold, the succes-
                         i¼0                                                       sive substitution method
        X
        n
                                                                                   converges to the solution of the integral
f ¼           ai fi ðxÞ
        i¼0                                                                        equations.
100                 Mathematics in Chemical Engineering
9.4. Numerical Methods for Eigenvalue                     where Green’s function is [50, p. 891]
Problems
                                                                               1
                                                          Gðr; r0 Þ ¼            in three dimensions
                                                                               r
Eigenvalue problems are treated similarly to
                                                          ¼ 2 ln r in two dimensions
Fredholm equations, except that the final equa-                      qffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi
tion is a matrix eigenvalue problem instead of a          where r ¼ ðx  x0 Þ2 þ ðy  y0 Þ2 þ ðz  z0 Þ2
set of simultaneous equations. For example,               in three dimensions
                                                                   qffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi
X
n                                                         and r ¼ ðx  x0 Þ2 þ ðy  x0 Þ2
  wi Kðyi ; yi Þf ðyi Þ ¼ lf ðyj Þ;
i¼1                                                       in two dimensions
i ¼ 0; 1; . . . ; n
                                                                For the problem
leads to the matrix eigenvalue problem                    @u
                                                             ¼ Dr2 u; u ¼ 0 on S;
                                                          @t
K D f ¼ lf                                                with a point source at x0 ; y0 ; z0
                                                            ZZZ
                                                             t
              Z1                                                             @u
                                                          þD    fðx; y; z; tÞ dSdt
TðxÞ ¼
          1
                   Gðx; yÞQðyÞdy                                             @n
          k                                                     0
              0
              8
              < xð1  yÞ      xy                        When the problem is two-dimensional,
Gðx; yÞ ¼
              :                                                       1                         2        2
                   yð1  xÞ   yx                        u ¼ pffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi e½ðxx0 Þ þðyy0 Þ =4DðttÞ
                                                                4pDðt  tÞ
                                                              ZZ
Green’s functions for typical operators are               c¼      ðuÞt¼0 f ðx; yÞdxdy
given below.
                                                            ZZt
   For the Poisson equation with solution                 þD
                                                                          @u
                                                                fðx; y; tÞ dCdt
decaying to zero at infinity                                              @n
                                                                0
  2
r c ¼ 4p%                                                  For the following differential equation and
                                                          boundary conditions
the formulation as an integral equation is                                 
                                                            1 d          dc
                                                                    xa1      ¼ f ½x; cðxÞ;
          Z                                               xa1 dx        dx
cðrÞ ¼        rðr0 ÞGðr; r0 ÞdV 0                         dc           2 dc
                                                             ð0Þ ¼ 0;       ð1Þ þ cð1Þ ¼ g
          V                                               dx          Sh dx
                                                                    Mathematics in Chemical Engineering                                                 101
where Sh is the Sherwood number, the problem                           Thus, the problem ends up as one directly
can be written as a Hammerstein integral equa-                         formulated as a fixed point problem:
tion:
                                                                       f ¼ Fðf Þ
                Z1
cðxÞ ¼ g            Gðx; y; ShÞf ½y; cðyÞya1 dy
                                                                         When the problem is the diffusion–reaction
                0
                                                                       one, the form is
Green’s function for the differential operators                                          Zx
are [163]                                                              cðxÞ ¼ g              ½uðxÞvðyÞ  uðyÞvðxÞ
                                                                                         0
a¼1
                     (       2
                                                                       f ½y; cðyÞya1 dy  avðxÞ
                         1 þ  x;       yx                                   Z1
                            Sh                                         a¼          uðyÞf ½y; cðyÞya1 dy
Gðx; y; ShÞ ¼                                    a ¼ 2Gðx; y; ShÞ
                             2
                         1 þ  y;       x<y                                   0
                            Sh
                    (2       1
                                                                       9.6. Boundary Integral Equations and
                            þ  1;     yx
                       Sh    x                                         Boundary Element Method
               ¼
                       2 1
                         þ  1;        x<y
                       Sh y                                            The boundary element method utilizes Green’s
                                                                       theorem and integral equations. Here, the
Green’s functions for the reaction diffusion                           method is described briefly for the following
problem were used to provide computable error                          boundary value problem in two or three dimen-
bounds by FERGUSON and FINLAYSON [163].                                sions
   If Green’s function has the form
                                                                                                                @f
               (                                                       r2 f ¼ 0; f ¼ f 1 on S1 ;                   ¼ f 2 on S2
                   uðxÞvðyÞ    0yx                                                                            @n
Kðx; yÞ ¼
                   uðyÞvðxÞ    xy1
                                                                       Green’s theorem (see chap. 5) says that for any
                                                                       functions sufficiently smooth
the problem
                                                                       Z                                         Z                   
                                                                                                                             @c    @f
          Z1                                                               ðfr2 c  cr2 fÞdV ¼                           f      c      dS
                                                                                                                             @n    @n
f ðxÞ ¼        Kðx; yÞF½y; f ðyÞdy                                    V                                           S
where {x0, y0} or {x0, y0, z0} is a point in the                  This is an integral equation for f on the bound-
domain. The solution f also satisfies                             ary. Note that the order is one less than the
                                                                  original differential equation. However, the
r2 f ¼ 0                                                          integral equation leads to matrices that are
                                                                  dense rather than banded or sparse, so some
so that                                                           of the advantage of lower dimension is lost.
                                                                  Once this integral equation (Eq. 48) is solved to
Z                    
             @c    @f                                             find f on the boundary, it can be substituted in
         f      c      dS ¼ 0
             @n    @n                                             Equation (47) to find f anywhere in the domain.
 S
                                                                     In the boundary finite element method, both
   Consider the two-dimensional case. Since                       the function and its normal derivative along the
the function c is singular at a point, the inte-                  boundary are approximated.
grals must be carefully evaluated. For the
                                                                       X                               N     
region shown in Figure 31, the domain is                               N
                                                                                                 @f X      @f
                                                                  f¼         fj N j ðjÞ;            ¼           N j ðjÞ
S ¼ S1þ S2; a small circle of radius r0 is placed                      j¼1
                                                                                                 @n   j¼1
                                                                                                           @n j
around the point P at x0, y0. Then the full
integral is                                                       One choice of trial functions can be the piece-
                                 
                                                                  wise constant functions shown in Figure 32.
R                @ ln r        @f                                 The integral equation then becomes
             f           ln r      dS
     S            @n           @n
                                                                          2                               3
         Z 
         u¼2p                                                                 Z               Z
                     @ ln r0         @f                               XN
                                                                          6      @ ln ri       @fj        7
þ                  f          ln r0      r0 du ¼ 0               pfi     4f j           dS       ln r i 5ds
                       @n            @n                                            @n          @n
     u¼0                                                              j¼1
                                                                                      sj                         sj
and
             Z
             u¼2p
                        @ ln r0
lim                 f           r0 du ¼ fðPÞ2p
r 0 !0                    @n
          u¼0
This set of equations is then solved for fi and                  mathematical optimization problems involving
@fi/@n along the boundary. This constitutes the                  continuous and discrete (or integer) variables.
boundary integral method applied to the Laplace                  This is followed by a review of solution meth-
equation.                                                        ods of the major types of optimization problems
   If the problem is Poisson’s equation                          for continuous and discrete variable optimiza-
                                                                 tion, particularly nonlinear and mixed-integer
r2 f ¼ gðx; yÞ                                                   nonlinear programming. In addition, we discuss
                                                                 direct search methods that do not require deriv-
Green’s theorem gives                                            ative information as well as global optimization
                                                                 methods, before reviewing extensions of these
Z                                Z
            @ ln r        @f                                     methods for the optimization of systems
        f           ln r      dS þ g ln r dA ¼ 0
             @n           @n                                     described by differential and algebraic equa-
S                                           A
                                                                 tions. These methods, which are designed for
Thus, for an internal point,                                     single-objective optimization problems, form
                                                                 the basis to tackle multicriteria optimization
                                               
                    R          @ ln r        @f                  problems and optimization under uncertainty,
2pfðPÞ ¼                     f         ln r      dS
                        S       @r           @n        ð49Þ      which are reviewed in the two final parts of this
    R                                                            chapter.
þ       A   g ln rdA
10. Optimization
We provide a survey of systematic methods for
a broad variety of optimization problems. The
survey begins with a general classification of                   Figure 33.
104          Mathematics in Chemical Engineering
x 2 n ; y 2 f0; 1gt
functions involved are nonlinear. If all functions        Strict convexity requires that the inequalities
are linear it corresponds to a mixed-integer linear       (53) and (54) be strict. Convex feasible
program (MILP). If there are no 0–1 variables,            regions require g(x) to be a convex function
then Problem (51) reduces to a nonlinear pro-             and h(x) to be linear. If Problem (52) is a
gram (52) or linear program (65) depending on             convex problem, than any local solution is
whether or not the functions are linear.                  guaranteed to be a global solution to Problem
   We first start with continuous variable opti-          (52). Moreover, if the objective function is
mization and consider in the next section the             strictly convex, then this solution x is unique.
solution of NLPs with differentiable objective            On the other hand, nonconvex problems may
and constraint functions. If only local solutions         have multiple local solutions, i.e., feasible
are required for the NLP, then very efficient             solutions that minimize the objective function
large-scale methods can be considered. This is            within some neighborhood about the solution.
followed by methods that are not based on local               We first consider methods that find only
optimality criteria; we consider direct search            local solutions to nonconvex problems, as
optimization methods that do not require deriv-           more difficult (and expensive) search proce-
atives, as well as deterministic global optimi-           dures are required to find a global solution.
zation methods. Following this, we consider the           Local methods are currently very efficient and
solution of mixed-integer problems and outline            have been developed to deal with very large
the main characteristics of algorithms for their          NLPs. Moreover, by considering the structure
solution. Finally, we conclude with a discussion          of convex NLPs (including LPs and QPs), even
of optimization modeling software and its                 more powerful methods can be applied. To
implementation in engineering models.                     study these methods, we first consider condi-
                                                          tions for local optimality.
                                                                                                   
                                                       fences [rgðx Þ] and rails [rhðx Þ] on the
                                                       ball are balanced by the force of “gravity”
                                                                
                                                       [rf ðx Þ]. This condition can be stated by
                                                       the following KKT necessary conditions for
                                                       constrained optimality:
                                                          Stationarity Condition: It is convenient to
                                                       define the Lagrangian function L(x, l, n) ¼ f(x) þ
                                                       g(x)Tl þ h(x)Tn along with “weights” or multi-
                                                       pliers l and n for the constraints. These multipli-
                                                       ers are also known as “dual variables” and
                                                       “shadow prices”. The stationarity condition
                                                       (balance of forces acting on the ball) is then given
                                                       by:
(l, n) are guaranteed to be unique at the optimal                           Min    ðx1 Þ2  4x1 þ 3=2ðx1 Þ2  7x2 þ x1 x2 þ 9  lnx1  lnx2
solution. The weaker Mangasarian–Fromovitz                                  s:t:   4  x1 x 2  0                                                 ð63Þ
constraint qualification (MFCQ) requires only                                      2x1  x2 ¼ 0
that rh(x ) have full column rank and that a
direction p exist that satisfies rh(x )Tp ¼ 0 and                          that corresponds to the plot in Figure 37. The
rgA(x )Tp > 0. With MFCQ, the KKT multi-                                   optimal solution can be found by applying the
pliers (l, n) are guaranteed to be bounded (but                             first-order KKT Conditions (56–58):
not necessarily unique) at the optimal solution.
Additional discussion can be found in [172].                                rLðx; l; nÞ ¼ rf ðxÞ þ rhðxÞl þ rgðxÞn
   Second-Order Conditions: As with un-                                                                              
                                                                                   2x1  4 þ x2  1=x1       2        x2
constrained optimization, nonnegative (posi-                                  ¼                         þ        lþ         n¼0
                                                                                   3x2  7 þ x1  1=x2      1        x1
tive) curvature is necessary (sufficient) in all                                   gðxÞ ¼ 4  x1 x2  0; hðxÞ ¼ 2x1  x2 ¼ 0                      ð64Þ
of the allowable (i.e., constrained) nonzero                                              gðxÞn ¼ ð4  x1 x2 Þn; n  0
                                                                                                       +
directions p. This condition can be stated in                                               1:4142
several ways. A typical necessary second-order                                       x ¼            ; l ¼ 1:036; n ¼ 1:068
                                                                                            2:8284
condition requires a point x that satisfies LICQ
and first-order Conditions (56–58) with multi-                              and f(x ) ¼ 1.8421. Checking the second-
pliers (l, n) to satisfy the additional conditions                          order Conditions (60) leads to:
given by:
                                                                            rxx Lðx ; l ; n Þ ¼ rxx ½f ðx Þ þ hðx Þl þ gðx Þn 
                                                                                                                                       
                                                          
pT rxx Lðx ; l; nÞp  0 for all p 6¼ 0; rhðx ÞT p ¼ 0; rgA ðx ÞT p ¼ 0                           2
                                                                                                                           2:5 0:068
                                                                ð59Þ         ¼ 2 þ 1=ðx1 Þ              1n
                                                                                                                 2 ¼
                                                                                       1n          3 þ 1=ðx2 Þ           0:068 3:125
                                                                                                                          
                                                                                                           2     2:8284
The corresponding sufficient conditions require                             ½rhðx ÞjrgA ðx Þp ¼                           p ¼ 0; p 6¼ 0
                                                                                                          1 1:4142
that the inequality in Condition (59) be strict.
Note that for the example in Figure 36, the                                 Note that LICQ is satisfied. Moreover, because
allowable directions p span the entire space for                                          
                                                                            ½rhðx ÞjrgA ðx Þ is square and nonsingular,
x while in Figure 37, there are no allowable                                there are no nonzero vectors p that satisfy the
directions p.                                                               allowable directions. Hence, the sufficient sec-
                                                                                                                 
   Example: To illustrate the KKT conditions,                                                                       
                                                                            ond-order conditions (pT rxx Lðx ; l ; n Þ > 0,
consider the following unconstrained NLP:                                   for all allowable p) are vacuously satisfied
                                                                            for this problem.
Min ðx1 Þ2  4x1 þ 3=2ðx2 Þ2  7x2 þ x1 x2 þ 9  lnx1  lnx2
                                                                ð60Þ        Convex Cases of NLP. Linear programs and
                                                                            quadratic programs are special cases of Problem
corresponding to the contour plot in Figure 36.                             (52) that allow for more efficient solution, based
The optimal solution can be found by solving                                on application of KKT Conditions (56–59).
108           Mathematics in Chemical Engineering
Because these are convex problems, any locally       developed in the late 1940s [173], although,
optimal solution is a global solution. In particu-   starting from KARMARKAR’s discovery in 1984,
lar, if the objective and constraint functions in    interior point methods have become quite
Problem (52) are linear then the following linear    advanced and competitive for large scale prob-
program (LP):                                        lems [174]. The simplex method proceeds by
                                                     moving successively from vertex to vertex with
Min    cT x                                          improved objective function values. Methods to
s:t:   Ax ¼ b                                 ð65Þ   solve Problem (66) are well implemented and
                                                     widely used, especially in planning and logisti-
       Cx  d
                                                     cal applications. They also form the basis for
                                                     MILP methods (see below). Currently, state-of-
can be solved in a finite number of steps, and the
                                                     the-art LP solvers can handle millions of vari-
optimal solution lies at a vertex of the polyhe-
                                                     ables and constraints and the application of
dron described by the linear constraints. This is
                                                     further decomposition methods leads to the
shown in Figure 39, and in so-called primal
                                                     solution of problems that are two or three orders
degenerate cases, multiple vertices can be alter-
                                                     of magnitude larger than this [175, 176]. Also,
nate optimal solutions with the same values of
                                                     the interior point method is described below
the objective function. The standard method to
                                                     from the perspective of more general NLPs.
solve Problem (65) is the simplex method,
                                                         Quadratic programs (QP) represent a slight
                                                     modification of Problem (65) and can be stated
                                                     as:
                                                                  1
                                                     Min    cT x þ xT Qx
                                                                  2
                                                     s:t:   Ax ¼ b                                 ð66Þ
xd
                                                                                  1
                                                             Min     rf ðxk ÞT p þ pT rxx Lðxk ; lk ; nk Þp
                                                                                  2
                                                                                                                         ð68Þ
                                                             s:t:    hðxk Þ þ rhðxk ÞT p ¼ 0
gðxk Þ þ rgðxk ÞT p þ s ¼ 0; s 0
Figure 40. Contour plots of convex quadratic programs        The KKT conditions of Equation (68) are given
                                                             by:
    Often these strategies are heuristics built                 one-at-a-time search and methods based on
    into particular NLP codes. While quite                      experimental designs [196]. At that time, these
    effective, most of these heuristics do not                  direct search methods were the most popular
    provide convergence guarantees for gen-                     optimization methods in chemical engineering.
    eral NLPs.                                                  Methods that fall into this class include the
                                                                pattern search [197], the conjugate direction
   Table 11 summarizes the characteristics of a                 method [198], simplex and complex searches
collection of widely used NLP codes. Much                       [199], and the adaptive random search methods
more information on widely available codes                      [200–202]. All of these methods require only
can also be found on the NEOS server                            objective function values for unconstrained
(www-neos.mcs.anl.gov) and the NEOS                             minimization. Associated with these methods
Software Guide.                                                 are numerous studies on a wide range of process
                                                                problems. Moreover, many of these methods
                                                                include heuristics that prevent premature ter-
10.3. Optimization Methods without                              mination (e.g., directional flexibility in the
Derivatives                                                     complex search as well as random restarts
                                                                and direction generation). To illustrate these
A broad class of optimization strategies does                   methods, Figure 41 illustrates the performance
not require derivative information. These                       of a pattern search method as well as a
methods have the advantage of easy imple-                       random search method on an unconstrained
mentation and little prior knowledge of the                     problem.
optimization problem. In particular, such
methods are well suited for “quick and dirty”                   Simulated Annealing. This strategy is related
optimization studies that explore the scope of                  to random search methods and derives from a
optimization for new problems, prior to                         class of heuristics with analogies to the motion
investing effort for more sophisticated model-                  of molecules in the cooling and solidification of
ing and solution strategies. Most of these                      metals [203]. Here a “temperature” parameter u
methods are derived from heuristics that nat-                   can be raised or lowered to influence the prob-
urally spawn numerous variations. As a result,                  ability of accepting points that do not improve
a very broad literature describes these meth-                   the objective function. The method starts with a
ods. Here we discuss only a few important                       base point x and objective value f(x). The next
trends in this area.                                            point x’ is chosen at random from a distribution.
                                                                If f(x’) < f(x), the move is accepted with x’
Classical Direct Search Methods. Developed                      as the new point. Otherwise, x’ is accepted
in the 1960s and 1970s, these methods include                   with probability p(u, x’,x). Options include the
114         Mathematics in Chemical Engineering
large, complex simulation models. These              variables and compares lower bound and upper
include the DAKOTA package at Sandia                 bound for fathoming each subregion. The one
National Lab [207], http://endo.sandia.gov/          that contains the optimal solution is found by
DAKOTA/software.html           and     FOCUS         eliminating subregions that are proved not to
developed at Boeing Corporation [208].               contain the optimal solution.
   Direct search methods are easy to apply to a          For nonconvex NLP problems, QUESADA and
wide variety of problem types and optimization       GROSSMANN [210] proposed a spatial branch-
models. Moreover, because their termination          and-bound algorithm for concave separable,
criteria are not based on gradient information       linear fractional, and bilinear programs using
and stationary points, they are more likely to       linear and nonlinear underestimating functions
favor the search for global rather than locally      [211]. For nonconvex MINLP, RYOO and SAHI-
optimal solutions. These methods can also be         NIDIS [212] and later TAWARMALANI and SAHINIDIS
adapted easily to include integer variables.         [213] developed BARON, which branches on
However, rigorous convergence properties to          the continuous and discrete variables with
globally optimal solutions have not yet been         bounds reduction method. ADJIMAN et al.
discovered. Also, these methods are best suited      [214, 215] proposed the SMIN-aBB and
for unconstrained problems or for problems           GMIN-aBB algorithms for twice-differentiable
with simple bounds. Otherwise, they may              nonconvex MINLPs. Using a valid convex
have difficulties with constraints, as the only      underestimation of general functions as well
options open for handling constraints are equal-     as for special functions, ADJIMAN et al. [216]
ity constraint elimination and addition of pen-      developed the aBB method, which branches on
alty functions for inequality constraints. Both      both the continuous and discrete variables
approaches can be unreliable and may lead to         according to specific options. The branch-
failure of the optimization algorithm. Finally,      and-contract method [217] has bilinear, linear
the performance of direct search methods scales      fractional, and concave separable functions in
poorly (and often exponentially) with the num-       the continuous variables and binary variables,
ber of decision variables. While performance         uses bound contraction, and applies the outer-
can be improved with the use of parallel com-        approximation (OA) algorithm at each node of
puting, these methods are rarely applied to          the tree. SMITH and PANTELIDES [218] proposed a
problems with more than a few dozen decision         reformulation method combined with a spatial
variables.                                           branch-and-bound algorithm for nonconvex
                                                     MINLP and NLP.
                                                         Because in global optimization one cannot
10.4. Global Optimization                            exploit optimality conditions like the KKT
                                                     conditions for a local optimum, these methods
Deterministic optimization methods are avail-        work by first partitioning the problem domain
able for nonconvex nonlinear programming             (i.e., containing the feasible region) into sub-
problems of the form Problem (52) that guar-         regions (see Fig. 42). Upper bounds on the
antee convergence to the global optimum.             objective function are computed over all sub-
More specifically, one can show under mild           regions of the problem. In addition, lower
conditions that they converge to an e distance       bounds can be derived from convex underesti-
to the global optimum on a finite number of          mators of the objective function and constraints
steps. These methods are generally more              for each subregion. The algorithm then pro-
expensive than local NLP methods, and they           ceeds to eliminate all subregions that have
require the exploitation of the structure of the     lower bounds that are greater than the least
nonlinear program.                                   upper bound. After this, the remaining regions
   Global optimization of nonconvex programs         are further partitioned to create new subregions
has received increased attention due to their        and the cycle continues until the upper and
practical importance. Most of the deterministic      lower bounds converge. Below we illustrate
global optimization algorithms are based on          the specific steps of the algorithm for nonlinear
spatial branch-and-bound algorithm [209],            programs that involve bilinear, linear fractional,
which divides the feasible region of continuous      and concave separable terms [210, 217].
116                 Mathematics in Chemical Engineering
                                                                                                subject to
                                                                                                                     X                            X                       X
                                                                                                ^f k ðx; y; zÞ ¼               aijk yij þ                    bijk zij þ           ^gi;k ðxi Þ
                                                                                                                   ði;jÞ2BLk                    ði;jÞ2LF k                i2C k
                                                                                                                   þ hk ðxÞ  0             k2K
                                                                                                ðx; y; zÞ 2 TðVÞ 
 R  R  Rn2  n          n1
where aij, aijk, bij, bijk are scalars with i 2 I ¼                                                   where ^gi ðxi Þ ¼ gi ðxi Þ at xi ¼ xLi , and
f1; 2;    ; ng, j 2 J ¼ f1; 2;    ; ng, and k 2                                                 xi ¼ xUi ; likewise, ^    gi;k ðxi Þ ¼ gi;k ðxi Þ at
K ¼ f1; 2;    ; mg. BL0 ; BLk ; LF 0 ; LF k are                                                    xi ¼ xLi , and xi ¼ xUi .
(i, j)-index sets, with i 6¼ j, that define the                                                    b. y ¼ fyij g is a vector of additional variables
bilinear and linear fractional terms present in                                                       for relaxing the bilinear terms in Problem
the problem. The functions hðxÞ; hk ðxÞ are                                                           (80), and is used in the following inequali-
convex, and twice continuously differentiable.                                                        ties which determine the convex and con-
C0 and Ck are index sets for the univariate twice                                                     cave envelopes of bilinear terms:
continuously differentiable concave functions
gi ðxi Þ; gi;k ðxi Þ. The set S 
 Rn is convex, and
v0Rn is an n-dimensional hyperrectangle                                                                   yij  xLj xi þ xLi xj  xLi xLj                     ði; jÞ 2 BLþ
defined in terms of the initial variable bounds                                                                                                                                                        ð82Þ
                                                                                                          yij    xU
                                                                                                                   j xi   þ    xU
                                                                                                                                i xj      xU  U
                                                                                                                                            i xj              ði; jÞ 2 BLþ
xL,in and xU,in:
                                                                                                           yij  xLj xi þ xU       U L
                                                                                                                           i xj  xi x j                 ði; jÞ 2 BL
V0 ¼ fx 2 Rn : 0  xL;in  x  xU;in ; xL;in
                                        j    >0
                                                                                                                                                                                                       ð83Þ
   ifði; jÞ 2 LF 0 [ LF k ;            i 2 I; j 2 J; k 2 Kg                                                yij  xU       L       L U
                                                                                                                  j xi þ xi xj  xi x j                  ði; jÞ 2 BL
                                                                                 Mathematics in Chemical Engineering                                                          117
                                                       s:t:     x1  x  x u
                                                                                                           ð90Þ
   For iteration k with a set of partitions Vk;j                         3ðxu  xÞ             ðx  xl Þ
and bounds in each subregion OLBj and OUBj:                     w ¼ ðxl Þ            þ ðxu Þ3
                                                                          ðxu  xl Þ          ðxu  xl Þ
1. Bound: Define best upper bound: OUB ¼               In Figure 43 we also propose subregions that
   Minj OUBj and delete (fathom) all subre-            are created by simple bisection partitioning
   gions j with lower bounds OLBj  OUB. If            rules, and we use a “loose” bounding tolerance
   OLBj  OUB  e, stop.                               of e ¼ 0:2. In each partition the lower bound, fL
2. Refine: Divide the remaining active subre-          is determined by Problem (90) and the upper
   gions into partitiions Vk;j1 and Vk;j2 . (Several   bound fU is determined by the local solution of
   branching rules are available for this step.)       the original problem in the subregion. Figure 44
3. Select: Solve the convex NLP (80) in the            shows the progress of the spatial branch-
   new partitions to obtain OLBj1 and OLBj2.           and-bound algorithm as the partitions are
   Delete partition if no feasible solution.           refined and the bounds are updated. In
                                                        Mathematics in Chemical Engineering                119
Figure 43, note the definitions of the partitions           Unlike local optimization methods, there are no
for the nodes, and the sequence numbers in each             optimality conditions, like the KKT conditions,
node that show the order in which the partitions            that can be applied directly.
are processed. The grayed partitions correspond
to the deleted subregions and at termination of             Mixed Integer Linear Programming. If the
the algorithm we see that fLj  fU  e (i.e.,               objective and constraint functions are all linear
19.85  19.7–0.2), with the gray subregions               in Problem (52), and we assume 0–1 binary
in Figure 43 still active. Further partitioning in          variables for the discrete variables, then this
these subregions will allow the lower and upper             gives rise to a mixed integer linear program-
bounds to converge to a tighter tolerance.                  ming (MILP) problem given by Equation (91).
    A number of improvements can be made to
the bounding, refinement, and selection strate-             Min Z ¼ aT x þ cT y
gies in the algorithm that accelerate the con-                s:t: Ax þ By  b                             ð91Þ
                                                             x  0; y 2 f0; 1gt
vergence of this method. A comprehensive
discussion of all of these options can be found
in [225–227]. Also, a number of efficient global            As is well known, the (MILP) problem is NP-
optimization codes have recently been devel-                hard. Nevertheless, an interesting theoretical
oped, including aBB, BARON, LGO, and                        result is that it is possible to transform it into
OQNLP. An interesting numerical comparison                  an LP with the convexification procedures pro-
of these and other codes can be found in [228].             posed by LOVACZ and SCHRIJVER [229], SHERALI
                                                            and ADAMS [230], and BALAS et al. [231]. These
                                                            procedures consist of sequentially lifting the
10.5. Mixed Integer Programming                             original relaxed xy space into higher dimen-
                                                            sion and projecting it back to the original space
Mixed integer programming deals with both                   so as to yield, after a finite number of steps, the
discrete and continuous decision variables.                 integer convex hull. Since the transformations
For simplicity in the presentation we consider              have exponential complexity, they are only of
the most common case where the discrete deci-               theoretical interest, although they can be used
sions are binary variables, i.e., yi ¼ 0 or 1, and          as a basis for deriving cutting planes (e.g. lift
we consider the mixed integer Problem (52).                 and project method by [231]).
120       Mathematics in Chemical Engineering
method. The selection of binary variables for            OSL, XPRESS. Recent trends in MILP include
branching, known as the branching rule, is               the development of branch-and-price [241]
based on a number of different possible criteria;        and branch-and-cut methods such as the lift-
for instance, choosing the fractional variable           and-project method [231] in which cutting
closest to 0.5, or the one involving the largest of      planes are generated as part of the branch-
the smallest pseudocosts for each fractional             and-bound enumeration. See also [238] for a
variable. Branching strategies to navigate the           recent review on MILP. A description of sev-
tree take a number of forms. More common                 eral MILP solvers can also be found in the
depth-first strategies expand the most recent            NEOS Software Guide.
node to a leaf node or infeasible node and
then backtrack to other branches in the tree.            Mixed Integer Nonlinear Programming The
These strategies are simple to program and               most basic form of an MINLP problem when
require little storage of past nodes. On the other       represented in algebraic form is as follows:
hand, breadth-first strategies expand all the
                                                             min z ¼ f ðx; yÞ
nodes at each level of the tree, select the              s:t: gj ðx; yÞ  0 j 2 J         ðP1Þ            ð93Þ
node with the lowest objective function, and                  x 2 X; y 2 Y
then proceed until the leaf nodes are reached.
Here, more storage is required, but generally            where f(), g() are convex, differentiable func-
fewer nodes are evaluated than in depth-first            tions, J is the index set of inequalities, and x and
search. In practice, a combination of both               y are the continuous and discrete variables,
strategies is commonly used: branch on the               respectively. The set X is commonly assumed
dichotomy 0–1 at each node (i.e., like                   to be a convex compact set, e.g.,
breadth-first), but expand as in depth-first.            X ¼ fxjx 2 Rn ; Dx  d; xL  x  xU g;           the
Additional description of these strategies can           discrete set Y corresponds to a polyhedral set
be found in [240].                                       of integer points, Y ¼ fyjx 2 Z m ; Ay  ag,
   Example: To illustrate the branch-and-bound           which in most applications is restricted to
approach, we consider the MILP:                          0–1 values, y 2 {0,1}m. In most applications
                                                         of interest the objective and constraint func-
Min Z ¼ x þ y1 þ 2y2 þ 3y3                               tions f(), g() are linear in y (e.g., fixed cost
 s:t:  x þ 3y1 þ y2 þ 2y3  0                           charges       and    mixed-logic       constraints):
                                               ð92Þ
      4y1  8y2  3y3  10
      x  0; y1 ; y2 ; y3 ¼ f0; 1g
                                                         f ðx; yÞ ¼ cT y þ rðxÞ; gðx; yÞ ¼ By þ hðxÞ.
                                                             Methods that have addressed the solution of
                                                         Problem (93) include the branch-and-bound
The solution to Problem (92) is given by x ¼ 4,          method (BB) [242–246], generalized Benders
y1 ¼ 1, y2 ¼ 1, y3 ¼ 0, and Z ¼ 7. Here we use a         decomposition (GBD) [247], outer-approxima-
depth-first strategy and branch on the variables         tion (OA) [248–250], LP/NLP based branch-
closest to zero or one. Figure 45 shows the              and-bound [251], and extended cutting plane
progress of the branch-and-bound algorithm as            method (ECP) [252].
the binary variables are selected and the bounds             There are three basic NLP subproblems that
are updated. The sequence numbers for each               can be considered for Problem (93):
node in Figure 45 show the order in which they
are processed. The grayed partitions correspond           a. NLP relaxation
to the deleted nodes and at termination of the
algorithm we see that Z ¼ 7 and an integer
                                                                 Min Z kLB ¼ f ðx; yÞ
solution is obtained at an intermediate node                    s:t: gj ðx:yÞ  0 j 2 J
where coincidentally y3 ¼ 0.                                        x 2 X; y 2 Y R               ðNLP1Þ   ð94Þ
   Mixed integer linear programming (MILP)                         yi  aki i 2 I kFL
                                                                   yi  bki i 2 I kFU
methods and codes have been available and
applied to many practical problems for more
than twenty years (e.g., [240]. The LP-based                   where YR is the continuous relaxation of the
branch-and-bound method [234] has been                         set Y, and I kFL ; I kFU are index subsets of
implemented in powerful codes such as                          the integer variables yi, i 2 I which are
122            Mathematics in Chemical Engineering
                                                                             x 2 X; y 2 Y                                                ð98Þ
                                                                Given the assumption on convexity of the func-
                                                                tions f(x, y) and g(x, y), it can be proved that the
                                                                solution of Problem (98) Z kL corresponds to a
                                                                lower bound of the solution of Problem (93).
                                                                Note that this property can be verified in
                                                                Figure 46. Also, since function linearizations
                                                                are accumulated as iterations proceed, the mas-
                                                                ter Problems (98) yield a nondecreasing
                                                                sequence of lower bounds Z 1L . . .  Z kL . . . Z kL
                                                                since linearizations are accumulated as itera-
                                                                tions k proceed.
                                                                   The OA algorithm as proposed by DURAN
                                                                and GROSSMANN consists of performing a cycle
                                                                of major iterations, k ¼ 1,..K, in which Problem
Figure 47. Major steps in the different algorithms              (95) is solved for the corresponding yK, and the
124                       Mathematics in Chemical Engineering
relaxed MILP master Problem (98) is updated                       fjjgj ðxk ; yk Þ ¼ 0g and the set x 2 X is disre-
and solved with the corresponding function                        garded. In particular, consider an outer-approx-
linearizations at the point (xk, yk) for which                    imation given at a given point (xk, yk)
the corresponding subproblem NLP2 is solved.
If feasible, the solution to that problem is used                                                     2            3
                                                                                                          x  xk
to construct the first MILP master problem;                                                       k T67
                                                                  a  f ðx ; y Þ þ rf ðx ; y Þ 4
                                                                            k     k           k
                                                                                                      5
otherwise a feasibility Problem (96) is solved                                                 y  yk
to generate the corresponding continuous point                                            2        3                                    ð100Þ
                                                                                            x  xk
[250]. The initial MILP master Problem (98)                                        k k T6
                                                                  gðx ; y Þ þ rg ðx ; y Þ 4
                                                                     k k                           7
                                                                                                   50
then generates a new vector of discrete varia-                                              y  yk
bles. The Subproblems (95) yield an upper
bound that is used to define the best current
solution, UBk ¼ mink fZ kU g. The cycle of itera-                 where for a fixed yk the point xk corresponds to
tions is continued until this upper bound and the                 the optimal solution to Problem (95). Making
lower bound of the relaxed master problem Z kL                    use of the Karush–Kuhn–Tucker conditions and
are within a specified tolerance. One way to                      eliminating the continuous variables x, the
avoid solving the feasibility Problem (96) in the                 inequalities in Problem (100) can be reduced
OA algorithm when the discrete variables in                       as follows [251]:
Problem (93) are 0–1, is to introduce the fol-
lowing integer cut whose objective is to make                     a  f ðxk ; yk Þ þ ry f ðxk ; yk ÞT ðy  yk Þ þ ðmk ÞT ½gðxk ; yk Þ
infeasible the choice of the previous 0–1 values                       þ ry gðxk ; yk ÞT ðy  yk Þ                                     ð101Þ
generated at the K previous iterations [248]:
X                X
          yi            yi  jBk j  1   k ¼ 1; . . . K   ð99Þ   which is the Lagrangian cut projected in the y-
      k
i2B              i2N k
                                                                  space. This can be interpreted as a surrogate
                                                                  constraint of the equations in Problem (100),
where Bk¼fijyki ¼ 1g; N k ¼ fijyki ¼ 0g; k ¼                      because it is obtained as a linear combination
1; . . . K. This cut becomes very weak as the                     of these.
dimensionality of the 0–1 variables increases.                       For the case when there is no feasible solu-
However, it has the useful feature of ensuring                    tion to Problem (95), then if the point xk is
that new 0–1 values are generated at each                         obtained from the feasibility subproblem
major iteration. In this way the algorithm                        (NLPF), the following feasibility cut projected
will not return to a previous integer point                       in y can be obtained using a similar procedure.
when convergence is achieved. Using the
above integer cut the termination takes place                     ðlk ÞT ½gðxk ; yk Þ þ ry gðxk ; yk ÞT ðy  yk Þ  0                  ð102Þ
as soon as Z KL  UBK .
    The OA algorithm trivially converges in
                                                                  In this way, Problem (97) reduces to a problem
one iteration if f(x, y) and g(x, y) are linear.
                                                                  projected in the y-space:
This property simply follows from the fact that
if f(x, y) and g(x, y) are linear in x and y the                     Min        Z KL ¼ a
MILP master problem is identical to the origi-
                                                                  st a         f ðxk ; yk Þ þ ry f ðxk ; yk ÞT ðy  yk Þ þ ðmk ÞT
nal Problem (93). It is also important to note
that the MILP master problem need not be                                        ½gðxk ; yk Þ þ ry gðxk ; yk ÞT ðy  yk Þ   k 2 KFS
solved to optimality.
                                                                  ðlk ÞT ½gðxk ; yk Þ þ ry gðxk ; yk ÞT ðy  yk Þ  0 k 2 KIS
can be derived from master Problem (98), in the          Convergence is achieved when the maximum
context of Problem (93), GBD can be regarded             constraint violation lies within the specified
as a particular case of the outer-approximation          tolerance. The optimal objective value of
algorithm. In fact one can prove that given the          Problem (97) yields a nondecreasing sequence
same set of K subproblems, the lower bound               of lower bounds. It is of course also possible to
predicted by the relaxed master problem is               either add to problem linearizatons of all the
greater than or equal to that predicted by the           violated constraints in the set Jk, or lineariza-
relaxed master Problem (103) [248]. This proof           tions of all the nonlinear constraints j 2 J. In
follows from the fact that the Lagrangian and            the ECP method the objective must be defined
feasibility cuts (Eqs. 101 and 102) are surro-           as a linear function, which can easily be
gates of the outer-approximations (Problems              accomplished by introducing a new variable
103). Given the fact that the lower bounds of            to transfer nonlinearities in the objective as an
GBD are generally weaker, this method com-               inequality.
monly requires a larger number of cycles or                  Note that since the discrete and continuous
major iterations. As the number of 0–1 variables         variables are converged simultaneously, the
increases this difference becomes more pro-              ECP method may require a large number of
nounced. This is to be expected since only one           iterations. However, this method shares with the
new cut is generated per iteration. Therefore,           OA method Property 2 for the limiting case
user-supplied constraints must often be added            when all the functions are linear.
to the master problem to strengthen the bounds.
Also, it is sometimes possible to generate               LP/NLP-Based Branch and Bound [251]. This
multiple cuts from the solution of an NLP                method is similar in spirit to a branch-and-cut
subproblem in order to strengthen the lower              method, and avoids the complete solution of the
bound [254]. As for the OA algorithm, the                MILP master problem (M-OA) at each major
trade-off is that while it generally predicts            iteration. The method starts by solving an initial
stronger lower bounds than GBD, the computa-             NLP subproblem, which is linearized as in (M-
tional cost for solving the master problem (M-           OA). The basic idea consists then of performing
OA) is greater, since the number of constraints          an LP-based branch-and-bound method for (M-
added per iteration is equal to the number of            OA) in which NLP Subproblems (95) are
nonlinear constraints plus the nonlinear                 solved at those nodes in which feasible integer
objective.                                               solutions are found. By updating the represen-
    If Problem (93) has zero integrality gap, the        tation of the master problem in the current open
GBD algorithm converges in one iteration once            nodes of the tree with the addition of the
the optimal (x , y ) is found [255]. This prop-        corresponding linearizations, the need to restart
erty implies that the only case in which one can         the tree search is avoided.
expect the GBD method to terminate in                        This method can also be applied to the
one iteration is that in which the initial discrete      GBD and ECP methods. The LP/NLP method
vector is the optimum, and when the objective            commonly reduces quite significantly the
value of the NLP relaxation of Problem (93) is           number of nodes to be enumerated. The
the same as the objective of the optimal mixed-          trade-off, however, is that the number of
integer solution.                                        NLP subproblems may increase. Computa-
                                                         tional experience has indicated that often
Extended Cutting Plane (ECP) [252]. The
                                                         the number of NLP subproblems remains
ECP method, which is an extension of Kelley’s
                                                         unchanged. Therefore, this method is better
cutting-plane algorithm for convex NLP [256],
                                                         suited for problems in which the bottleneck
does not rely on the use of NLP subproblems
                                                         corresponds to the solution of the MILP mas-
and algorithms. It relies only on the iterative
                                                         ter problem. LEYFFER [257] has reported sub-
solution of problem by successively adding a
                                                         stantial savings with this method.
linearization of the most violated constraint at
                                                             Example: Consider the following MINLP
the predicted point (xk, yk):
                                                         problem, whose objective function and con-
J k ¼ f^j 2 argfmax gj ðxk ; yk Þgg                      straints contain nonlinear convex terms
                 j2J
126            Mathematics in Chemical Engineering
(e.g., M-MIPF, M-GBD, M-MIQP) do not                     Since the tree searches are not finite (except for
guarantee a valid lower bound Z KL or a valid            e convergence), these methods can be compu-
bounding representation with which the global            tationally expensive. However, their major
optimum may be cut off.                                  advantage is that they can rigorously find the
   Rigorous global optimization approaches for           global optimum. Specific cases of nonconvex
addressing nonconvexities in MINLP problems              MINLP problems have been handled. An exam-
can be developed when special structures are             ple is the work of PO€ RN and WESTERLUND [264],
assumed in the continuous terms (e.g. bilinear,          who addressed the solution of MINLP prob-
linear fractional, concave separable). Specifi-          lems with pseudoconvex objective function and
cally, the idea is to use convex envelopes or            convex inequalities through an extension of the
underestimators to formulate lower-bounding              ECP method.
convex MINLP problems. These are then com-                   The other option for handling nonconvex-
bined with global optimization techniques for            ities is to apply a heuristic strategy to try to
continuous variables [209, 210, 212, 217, 219,           reduce as much as possible the effect of non-
225, 260], which usually take the form of                convexities. While not being rigorous, this
spatial branch-and-bound methods. The lower              requires much less computational effort. We
bounding MINLP problem has the general                   describe here an approach for reducing the
form,                                                    effect of nonconvexities at the level of the
                                                         MILP master problem.
Min Z ¼ f ðx; yÞ                                            VISWANATHAN and GROSSMANN [265] proposed
s:t: g j ðx; yÞ  0 j 2 J                    ð105Þ
x 2 X; y 2 Y
                                                         to introduce slacks in the MILP master problem
                                                         to reduce the likelihood of cutting off feasible
                                                         solutions. This master problem (augmented pen-
where f ; g
            , are valid convex underestimators          alty/equality relaxation) has the form:
such that f ðx; yÞ  f ðx; yÞ and the inequalities                            X
                                                                               K
g ðx; yÞ  0 are satisfied if gðx; yÞ  0. A typi-       min Z K ¼ a þ                  wkp pk þ wkq qk
cal example of convex underestimators are the                                     k¼1                           2        39
                                                                                                                    x  xk>
                                                                                                                          >
convex envelopes for bilinear terms [211].                                                          6                    7>
                                                             s:t: a  f ðxk ; yk Þ þ rf ðxk ; yk ÞT 4                    5>
                                                                                                                          >
                                                                                                                          >
                                                                                                                          >
                                                                                                                          >
    Examples of global optimization methods                                                                         yy >
                                                                                                                       k  >
                                                                                                                          >
                                                                                                                          >
                                                                                                                          >
                                                                                                                          >
for MINLP problems include the branch-and-                                         2              3                       >
                                                                                                                          >
                                                                                                                          >
                                                                                        x  xk                            >
                                                                                                                          =
reduce method [212, 213], the a-BB method                                    k T6
                                                             T k rhðxk ; y Þ 4
                                                                                                  7
                                                                                                                                 k ¼ 1; . . . K
                                                                                                  5  pk
[215], the reformulation/spatial branch-and-                                                                                 >
                                                                                                                             >
                                                                                        yy   k                              >
                                                                                                                             >
                                                                                                                             >
                                                                                                                             >
bound search method [261], the branch-and-                                                        2             3            >
                                                                                                                             >
                                                                                                                             >
                                                                                                                             >
                                                                                                      x  xk                 >
                                                                                                                             >
cut method [262], and the disjunctive branch-                                                k T6               7            >
                                                                                                                             >
                                                                                                                             >
                                                             gðx ; y Þ þ rgðx ; y Þ 4
                                                                  k   k                  k
                                                                                                                5  qk       >
                                                                                                                             >
and-bound method [263]. All these methods                                                                                    >
                                                                                                                             ;
                                                                                                            k
                                                                                                      yy
rely on a branch-and-bound procedure. The                P                P
                                                           i2Bk   yi         i2N k   yi  jBk j  1 k ¼ 1; . . . K
difference lies in how to perform the branching
                                                         x 2 X; y 2 Y; a 2 R1 ; pk ; qk  0
on the discrete and continuous variables. Some                                                                                                    ð106Þ
methods perform the spatial tree enumeration
on both the discrete and continuous variables of         where            wkp ;    wkq
                                                                          are weights that are chosen
Problem (105). Other methods perform a spa-              sufficiently large (e.g., 1000 times the magni-
tial branch and bound on the continuous vari-            tude of the Lagrange multiplier). Note that if the
ables and solve the corresponding MINLP                  functions are convex then the MILP master
Problem (105) at each node using any of the              Problem (106) predicts rigorous lower bounds
methods reviewed above. Finally, other meth-             to Problem (93) since all the slacks are set to
ods branch on the discrete variables of Problem          zero.
(105), and switch to a spatial branch and bound
on nodes where a feasible value for the discrete         Computer Codes for MINLP. Computer codes
variables is found. The methods also rely on             for solving MINLP problems include the fol-
procedures for tightening the lower and upper            lowing. The program DICOPT [265] is an
bounds of the variables, since these have a great        MINLP solver that is available in the modeling
effect on the quality of the underestimators.            system GAMS [266]. The code is based on the
128       Mathematics in Chemical Engineering
master Problem (106) and the NLP Subpro-            review of logic-based optimization can be found
blems (95). This code also uses relaxed Sub-        in [272, 273].
problem (94) to generate the first linearization       Generalized disjunctive programming in
for the above master problem, with which the        Problem (107) [269] is an extension of dis-
user need not specify an initial integer value.     junctive programming [274] that provides an
Also, since bounding properties of Problem          alternative way of modeling MILP and MINLP
(106) cannot be guaranteed, the search for          problems. The general formulation of Problem
nonconvex problems is terminated when there         (107) is as follows:
is no further improvement in the feasible NLP                 P
                                                    Min Z ¼       k2K   ck þ f ðxÞ
subproblems. This is a heuristic that works
                                                          s:t: gðxÞ  0                                     ð107Þ
reasonably well in many problems. Codes                              2             3
                                                                            Y jk
that implement the branch-and-bound method                      _
                                                               j2J k 6             7
                                                                     6 hjk ðxÞ  0 7; k 2 K
using Subproblems (94) include the code                              4             5
                                                                         ck ¼ g jk
MINLP_BB, which is based on an SQP algo-
rithm [246] and is available in AMPL, the code                 VðYÞ ¼ True
BARON [267], which also implements global
                                                                  x 2 Rn ; c 2 Rm ; Y 2 ftrue; falsegm
optimization capabilities, and the code SBB,
which is available in GAMS [266]. The code a-
ECP implements the extended cutting-plane           where Yjk are the Boolean variables that decide
method [252], including the extension by PO€ RN     whether a term j in a disjunction k 2 K is true or
and WESTERLUND [264]. Finally, the code MIN-        false, and x are continuous variables. The objec-
OPT [268] also implements the OA and GBD            tive function involves the term f(x) for the
methods, and applies them to mixed-integer          continuous variables and the charges ck that
dynamic optimization problems. It is difficult      depend on the discrete choices in each dis-
to derive general conclusions on the efficiency     junction k 2 K. The constraints g(x)  0 hold
and reliability of all these codes and their        regardless of the discrete choice, and hjk(x)  0
corresponding methods, since no systematic          are conditional constraints that hold when Yjk is
comparison has been made. However, one              true in the j-th term of the k-th disjunction.
might anticipate that branch-and-bound codes        The cost variables ck correspond to the fixed
are likely to perform better if the relaxation of   charges, and are equal to gjk if the Boolean
the MINLP is tight. Decomposition methods           variable Yjk is true. V(Y) are logical relations for
based on OA are likely to perform better if the     the Boolean variables expressed as proposi-
NLP subproblems are relatively expensive to         tional logic.
solve, while GBD can perform with some effi-           Problem (107) can be reformulated as an
ciency if the MINLP is tight and there are many     MINLP problem by replacing the Boolean var-
discrete variables. ECP methods tend to per-        iables by binary variables yjk,
form well on mostly linear problems.                          P         P
                                                    Min Z ¼       k2K     j2J k   g jk yjk þ f ðxÞ
                                                    s:t: gðxÞ  0
Logic-Based Optimization. Given difficulties
in the modeling and solution of mixed integer       hjk ðxÞ  M jk ð1  yjk Þ; j 2 J k ; k 2 K       ðBMÞ
                                                    P                                                       ð108Þ
problems, the following major approaches based         j2J k yjk ¼ 1; k 2 K
on logic-based techniques have emerged: gener-
                                                    Ay  a
alized disjunctive programming (Problem 107)
[269], mixed-logic linear programming (MLLP)        0  x  xU ; yjk 2 f0; 1g; j 2 J k ; k 2 K
[270], and constraint programming (CP) [271].
The motivations for this logic-based modeling       where the disjunctions are replaced by “Big-M”
has been to facilitate the modeling, reduce the     constraints which involve a parameter Mjk and
combinatorial search effort, and improve            binary variables yjk. The propositional logic
the handling of nonlinearities. In this section     statements V(Y) ¼ True are replaced by the
we mostly concentrate on generalized dis-           linear constraints Ay  A [275] and [276]. Here
junctive programming and provide a brief refer-     we assume that x is a nonnegative variable with
ence to constraint programming. A general           finite upper bound xU. An important issue in
                                                                                   Mathematics in Chemical Engineering                                        129
                                                                                                                 P         P
Problem (108) is how to specify a valid value                                              Min Z L ¼                 k2K       j2J k   g jk ljk þ f ðxÞ
for the Big-M parameter Mjk. If the value is too                                                             s:t: gðxÞ  0
small, then feasible points may be cut off. If Mjk                                         P                P
                                                                                      x¼            vjk ;     j2J k ljk ¼ 1; k 2 K                    ðCRPÞ
is too large, then the continuous relaxation                                                j2J k
might be too loose and yield poor lower bounds.                                        0  x; vjk  xU ; 0  ljk  1; j 2 J k ; k 2 K                         ð111Þ
Therefore, finding the smallest valid value for                                                ljk hjk ðv =ljk Þ  0; j 2 J k ; k 2 K
                                                                                                            jk
where hjk(x) are assumed to be convex and                                             Constraint Programming. Constraint pro-
bounded over x. The convex hull relaxation of                                         gramming (CP) [271, 272] is a relatively new
disjunction (Problem 109) [245] is given as                                           modeling and solution paradigm that was orig-
follows:                                                                              inally developed to solve feasibility problems,
                                                                                      but it has been extended to solve optimization
               P                               P                                      problems as well. Constraint programming is
          x¼       j2J k   vjk ;     c¼            j2J   ljk g jk
                                                                                      very expressive, as continuous, integer, and
                                        jk ;                                          Boolean variables are permitted and, moreover,
                           jk
               0v                ljk xU      j 2 Jk
P                                                                                     variables can be indexed by other variables.
  j2J k   ljk ¼ 1; 0  ljk  1; j 2 J k                             ðCHÞ   ð110Þ
                                                                                      Constraints can be expressed in algebraic form
             ljk hjk ðvjk =ljk Þ  0; j 2 J k                                         (e.g., h(x)  0), as disjunctions (e.g., [A1x  b1]
                                                                                      _ A2x  b2]), or as conditional logic statements
                 x; c; vjk  0; j 2 J k
                                                                                      (e.g., If g(x)  0 then r(x)  0). In addition, the
                                                                                      language can support special implicit functions
where vjk are disaggregated variables that are                                        such as the all-different (x1, x2, . . . xn) con-
assigned to each term of the disjunction k 2 K,                                       straint for assigning different values to the
and ljk are the weight factors that determine the                                     integer variables x1, x2, . . . xn. The language
feasibility of the disjunctive term. Note that                                        consists of Cþþ procedures, although the
when ljk is 1, then the j-th term in the k-th                                         recent trend has been to provide higher level
disjunction is enforced and the other terms are                                       languages such as OPL. Other commercial CP
ignored. The constraints ljk hjk ðvjk =ljk Þ are con-                                 software packages include ILOG Solver [282],
vex if hjk(x) is convex [278, p. 160]. A formal                                       CHIP [283], and ECLiPSe [284].
proof can be found in [245]. Note that the
convex hull (Eqs. 110) reduces to the result
by BALAS [279] if the constraints are linear.                                         10.6. Dynamic Optimization
Based on the convex hull relaxation Equations
(110), LEE and GROSSMANN [277] proposed                                               Interest in dynamic simulation and optimiza-
the following convex relaxation program of                                            tion of chemical processes has increased
Problem (107).                                                                        significantly since the 1990s. Chemical
130            Mathematics in Chemical Engineering
processes are modeled dynamically using                   those that discretize the state and control pro-
differential-algebraic equations (DAEs), con-             files (full discretization). Basically, the partially
sisting of differential equations that describe           discretized problem can be solved either by
the dynamic behavior of the system, such as               dynamic programming or by applying a non-
mass and energy balances, and algebraic equa-             linear programming (NLP) strategy (direct-
tions that ensure physical and thermodynamic              sequential). A basic characteristic of these
relations. Typical applications include control           methods is that a feasible solution of the
and scheduling of batch processes; startup,               DAE system, for given control values, is
upset, shut-down, and transient analysis;                 obtained by integration at every iteration of
safety studies; and the evaluation of control             the NLP solver. The main advantage of these
schemes. We state a general differential-                 approaches is that, for the NLP solver, they
algebraic optimization Problem (112) as                   generate smaller discrete problems than full
follows:                                                  discretization methods.
                                                              Methods that fully discretize the continuous
                                                        time problem also apply NLP strategies to solve
Min F zðtf Þ; yðtf Þ; uðtf Þ; tf ; p
                                                          the discrete system and are known as direct-
                               
s:t:   F dz=dt; zðtÞ; uðtÞ; t; p ¼ 0; zð0Þ ¼ z0           simultaneous methods. These methods can use
                                                         different NLP and discretization techniques,
       Gs ½zðts Þ; yðts Þ; uðts Þ; ts ; p  ¼ 0           but the basic characteristic is that they solve
       zL  zðtÞ  xU                             ð112Þ
                                                          the DAE system only once, at the optimum. In
                                                          addition, they have better stability properties
       yL  yðtÞ  yU                                     than partial discretization methods, especially
       uL  uðtÞ  yU                                     in the presence of unstable dynamic modes. On
       pL  p  pU                                        the other hand, the discretized optimization
                                                          problem is larger and requires large-scale
       ttf  tf  tU
                   f
                                                          NLP solvers, such as SOCS, CONOPT, or
                                                          IPOPT.
where F is a scalar objective function at final               With this classification we take into account
time tf, and F are DAE constraints, Gs addi-              the degree of discretization used by the differ-
tional point conditions at times ts, z(t) differ-         ent methods. Below we briefly present the
ential state profile vectors, y(t) algebraic state        description of the variational methods, followed
profile vectors, u(t) control state profile vec-          by methods that partially discretize the
tors, and p is a time-independent parameter               dynamic optimization problem, and finally
vector.                                                   we consider full discretization methods for
    We assume, without loss of generality, that           Problem (112).
the index of the DAE system is one, consistent
initial conditions are available, and the objec-          Variational Methods. These methods are
tive function is in the above Mayer form.                 based on the solution of the first-order neces-
Otherwise, it is easy to reformulate problems             sary conditions for optimality that are obtained
to this form. Problem (112) can be solved either          from Pontryagin’s maximum principle [285,
by the variational approach or by applying                286]. If we consider a version of Problem
some level of discretization that converts the            (112) without bounds, the optimality conditions
original continuous time problem into a dis-              are formulated as a set of DAEs:
crete problem. Early solution strategies, known
                                                          @Fðz; y; u; p; tÞ 0 @H @Fðz; y; u; p; tÞ
as indirect methods, were focused on solving                               l ¼    ¼                l     (113a)
                                                               @z0             @z      @z
the classical variational conditions for optimal-
ity. On the other hand, methods that discretize
                                                          Fðz; y; u; p; tÞ ¼ 0                           (113b)
the original continuous time formulation can be
divided into two categories, according to the
                                                          Gf ðz; y; u; p; tf Þ ¼ 0                       (113c)
level of discretization. Here we distinguish
between the methods that discretize only the
                                                          Gs ðz; y; u; p; ts Þ ¼ 0                       (113d)
control profiles (partial discretization) and
                                                       Mathematics in Chemical Engineering                           131
@H @Fðz; y; u; p; tÞ
   ¼                 l¼0                      (113e)      control parametrization n. The sequential
@y       @y
                                                          method is reliable when the system contains
                                                          only stable modes. If this is not the case, finding
@H @Fðz; y; u; p; tÞ
@u
   ¼
         @u
                     l¼0                      (113f)      a feasible solution for a given set of control
                                                          parameters can be very difficult. The time
Ztf
                                                          horizon is divided into time stages and at
      @Fðz; y; u; p; tÞ                                   each stage the control variables are represented
                        l dt ¼ 0             (113g)
            @p
0                                                         with a piecewise constant, a piecewise linear, or
                                                          a polynomial approximation [287, 288]. A
where the Hamiltonian H is a scalar function of           common practice is to represent the controls
the form H(t) ¼ F(z, y, u, p, y)Tl(t) and l(t) is a       as a set of Lagrange interpolation polynomials.
vector of adjoint variables. Boundary and jump               For the NLP solver, gradients of the objec-
conditions for the adjoint variables are given            tive and constraint functions with respect to the
by:                                                       control parameters can be calculated with the
                                                          sensitivity equations of the DAE system, given
@F             @F @Gf                                     by:
    lð t f Þ þ    þ    vf ¼ 0
@z0            @z   @z
                                              ð114Þ
@F 
  
 @Gs       @F 
 þ 
                               @F T       @F     @F      @F T               @zð0Þ
    l ts þ    vs ¼    l t                                      sk 0 þ T sk þ T wk þ      ¼ 0; sk ð0Þ ¼       k ¼ 1; . . . N q
@z0        @z      @z0 s                                  @z0        @z     @y      @qk                 @qk
                                                                                                                  ð115Þ
a solution to this boundary value problem.                Equation (115), the cost of obtaining these sensi-
Normally, the state variables are given as initial        tivities is directly proportional to Nq, the number
conditions, and the adjoint variables as final            of decision variables in the NLP. Alternately,
conditions. This formulation leads to boundary            gradients can be obtained by integration of the
value problems (BVPs) that can be solved by a             adjoint Equations (113a, 113e, 113g) [282, 285,
number of standard methods including single               286] at a cost independent of the number of input
shooting, invariant embedding, multiple shoot-            variables and proportional to the number of con-
ing, or some discretization method such as                straints in the NLP.
collocation on finite elements or finite differ-              Methods that are based on this approach
ences. Also the point conditions lead to an               cannot treat directly the bounds on state vari-
additional calculation loop to determine the              ables, because the state variables are not
multipliers vf and vs. On the other hand,                 included in the nonlinear programming prob-
when bound constraints are considered, the                lem. Instead, most of the techniques for dealing
above conditions are augmented with addi-                 with inequality path constraints rely on defining
tional multipliers and associated complemen-              a measure of the constraint violation over the
tarity conditions. Solving the resulting system           entire horizon, and then penalizing it in the
leads to a combinatorial problem that is pro-             objective function, or forcing it directly to zero
hibitively expensive except for small problems.           through an end-point constraint [291]. Other
                                                          techniques approximate the constraint satisfac-
Partial Discretization. With partial discretiza-          tion (constraint aggregation methods) by intro-
tion methods (also called sequential methods or           ducing an exact penalty function [290, 292] or a
control vector parametrization), only the con-            Kreisselmeier–Steinhauser function [292] into
trol variables are discretized. Given the initial         the problem.
conditions and a given set of control parame-                 Finally, initial value solvers that handle path
ters, the DAE system is solved with a differen-           constraints directly have been developed [288].
tial algebraic equation solver at each iteration.         The main idea is to use an algorithm for con-
This produces the value of the objective func-            strained dynamic simulation, so that any admis-
tion, which is used by a nonlinear programming            sible combination of the control parameters
solver to find the optimal parameters in the              produces an initial value problem that is
132              Mathematics in Chemical Engineering
feasible with respect to the path constraints.               techniques to enforce feasibility, like the ones
The algorithm proceeds by detecting activation               used in the sequential methods.
and deactivation of the constraints during the                   The resulting NLP is solved using SQP-type
solution, and solving the resulting high-index               methods, as described above. At each SQP iter-
DAE system and their related sensitivities.                  ation, the DAEs are integrated in each stage and
                                                             objective and constraint gradients with respect to
Full Discretization. Full discretization meth-               p, zi, and ni are obtained using sensitivity equa-
ods explicitly discretize all the variables of the           tions, as in Problem (115). Compared to sequen-
DAE system and generate a large-scale non-                   tial methods, the NLP contains many more
linear programming problem that is usually                   variables, but efficient decompositions have
solved with a successive quadratic program-                  been proposed [294] and many of these calcula-
ming (SQP) algorithm. These methods follow a                 tions can be performed in parallel.
simultaneous approach (or infeasible path                        In collocation methods, the continuous time
approach); that is, the DAE system is not solved             problem is transformed into an NLP by approx-
at each iteration; it is only solved at the opti-            imating the profiles as a family of polynomials
mum point. Because of the size of the problem,               on finite elements. Various polynomial repre-
special decomposition strategies are used to                 sentations are used in the literature, including
solve the NLP efficiently. Despite this charac-              Lagrange interpolation polynomials for the
teristic, the simultaneous approach has advan-               differential and algebraic profiles [295]. In
tages for problems with state variable (or path)             [194] a Hermite–Simpson collocation form is
constraints and for systems where instabilities              used, while CUTHRELL and BIEGLER [296] and
occur for a range of inputs. In addition, the                TANARTKIT and BIEGLER [297] use a monomial
simultaneous approach can avoid intermediate                 basis for the differential profiles. All of these
solutions that may not exist, are difficult to               representations stem from implicit Runge–
obtain, or require excessive computational                   Kutta formulae, and the monomial repre-
effort. There are mainly two different appro-                sentation is recommended because of smaller
aches to discretize the state variables explicitly,          condition numbers and smaller rounding
multiple shooting [293, 294] and collocation on              errors. Control and algebraic profiles, on the
finite elements [184, 194, 295].                             other hand, are approximated using Lagrange
    With multiple shooting, time is discretized              polynomials.
into P stages and control variables are parame-                  Discretizations of Problem (112) using col-
trized using a finite set of control parameters in           location formulations lead to the largest NLP
each stage, as with partial discretization. The              problems, but these can be solved efficiently
DAE system is solved on each stage, i ¼                      using large-scale NLP solvers such as IPOPT
1, . . . P and the values of the state variables             and by exploiting the structure of the colloca-
z(ti) are chosen as additional unknowns. In this             tion equations. BIEGLER et al. [184] provide a
way a set of relaxed, decoupled initial value                review of dynamic optimization methods using
problems (IVP) is obtained:                                  simultaneous methods. These methods offer a
                                                             number of advantages for challenging dynamic
                          
F dz=dt; zðtÞ; yðtÞ; ni ; p ¼ 0;                             optimization problems, which include:
t 2 ½ti1 ; ti ; zðti1 Þ ¼ zi                      ð116Þ       Control variables can be discretized at the
ziþ1  zðti ; zi ; ni ; pÞ ¼ 0; i ¼ 1; . . . P  1                same level of accuracy as the differential
                                                                  and algebraic state variables. The KKT
Note that continuity among stages is treated                      conditions of the discretized problem can
through equality constraints, so that the final                   be shown to be consistent with the varia-
solution satisfies the DAE system. With this                      tional conditions of Problem (116). Finite
approach, inequality constraints for states and                   elements allow for discontinuities in con-
controls can be imposed directly at the grid                      trol profiles.
points, but path constraints for the states may                  Collocation formulations allow problems
not be satisfied between grid points. This prob-                  with unstable modes to be handled in an
lem can be avoided by applying penalty                            efficient and well-conditioned manner.
                                                     Mathematics in Chemical Engineering             133
     The NLP formulation inherits stability             platforms are essential for the formulation task.
     properties of boundary value solvers.              These are classified into two broad areas: opti-
     Moreover, an elementwise decomposition             mization modeling platforms and simulation
     has been developed that pins down                  platforms with optimization.
     unstable modes in Problem (112).
    Collocation formulations have been pro-            Optimization modeling platforms provide gen-
     posed with moving finite elements. This            eral purpose interfaces for optimization algo-
     allows the placement of elements both              rithms and remove the need for the user to
     for accurate breakpoint locations of con-          interface to the solver directly. These plat-
     trol profiles as well as accurate DAE              forms allow the general formulation for all
     solutions.                                         problem classes discussed above with direct
                                                        interfaces to state of the art optimization
   Dynamic optimization by collocation meth-            codes. Three representative platforms are
ods has been used for a wide variety of process         GAMS (General Algebraic Modeling Sys-
applications including batch process optimiza-          tems), AMPL (A Mathematical Programming
tion, batch distillation, crystallization, dynamic      Language), and AIMMS (Advanced Integrated
data reconciliation and parameter estimation,           Multidimensional Modeling Software). All
nonlinear model predictive control, polymer             three require problem-model input via a declar-
grade transitions and process changeovers,              ative modeling language and provide exact
and reactor design and synthesis. A review of           gradient and Hessian information through auto-
this approach can be found in [298].                    matic differentiation strategies. Although pos-
                                                        sible, these platforms were not designed to
                                                        handle externally added procedural models.
10.7. Development of Optimization                       As a result, these platforms are best applied
Models                                                  on optimization models that can be developed
                                                        entirely within their modeling framework.
The most important aspect of a successful               Nevertheless, these platforms are widely used
optimization study is the formulation of the            for large-scale research and industrial applica-
optimization model. These models must reflect           tions. In addition, the MATLAB platform
the real-world problem so that meaningful               allows the flexible formulation of optimization
optimization results are obtained, and they             models as well, although it currently has only
also must satisfy the properties of the problem         limited capabilities for automatic differentia-
class. For instance, NLPs addressed by gradi-           tion and limited optimization solvers. More
ent-based methods require functions that are            information on these and other modeling
defined in the variable domain and have                 platforms can be found on the NEOS server
bounded and continuous first and second deriv-          www-neos.mcs.anl.gov.
atives. In mixed integer problems, proper for-
mulations are also needed to yield good lower           Simulation platforms with optimization are
bounds for efficient search. With increased             often dedicated, application-specific modeling
understanding of optimization methods and               tools to which optimization solvers have been
the development of efficient and reliable opti-         interfaced. These lead to very useful optimiza-
mization codes, optimization practitioners now          tion studies, but because they were not origi-
focus on the formulation of optimization                nally designed for optimization models, they
models that are realistic, well-posed, and              need to be used with some caution. In particu-
inexpensive to solve. Finally, convergence              lar, most of these platforms do not provide exact
properties of NLP, MILP, and MINLP solvers              derivatives to the optimization solver; often
require accurate first (and often second) deriv-        they are approximated through finite differ-
atives from the optimization model. If these            ence. In addition, the models themselves are
contain numerical errors (say, through finite           constructed and calculated through numerical
difference approximations) then performance             procedures, instead of through an open declar-
of these solvers can deteriorate considerably.          ative language. Examples of these include
As a result of these characteristics, modeling          widely used process simulators such as
134        Mathematics in Chemical Engineering
x2                                f2
                            f
              X                               f (X) Solutions dominated
          x                                         by f (x)
                                       f(x)
x1 f1
Pareto optimal if there is no other solution y                                The convexity property, which is eligible for
such that f i ðxÞ < f i ðyÞ for all components i.                          optimization problems with one objective, is
Usually, weakly Pareto optimal solutions,                                  also desirable for multicriteria problems. We
which are not Pareto optimal, are undesired,                               call a multicriteria optimization problem con-
because they are dominated, but some algo-                                 vex if the feasible set X is convex and the
rithms for obtaining Pareto optimal solutions                              objective functions fi are convex. In that case
only guarantee weak Pareto optimality. But                                 the Pareto set plus R kþ is convex. Otherwise, it
there might also be undesired Pareto optimal                               needs not to be convex and not even connected.
solutions. Assume two solutions x and y where x                               The concept of Pareto optimal solutions fits
is significantly worse than y with respect to fj                           the practical needs often better than breaking all
but at most only slightly better with respect to                           aspects down to one objective and several
any other objective, the deterioration may be                              constraints. Solving such a multicriteria opti-
not worth the improvement. So, the tradeoff                                mization problem is, however, much more
between these two components may be consid-                                involved than the solution of a single objective
ered undesired. Using a more general definition                            optimization problem: For an optimization
of dominance and Pareto optimality using so                                problem with a single objective, it is clear
called “ordering cones” one can exclude those                              that the problem is solved if an optimal solution
solutions. An equivalent formulation of “x                                 is found. Finding all Pareto optimal solutions
dominates y” is the following:                                             for a multicriteria optimization problem is in
                                                                           general not possible nor usually necessary. The
f ðyÞ  f ðxÞ 2 R kþ nf0g                                                  aim is to support the person or the group of
                                                                           persons who have eventually to select a solution
The set R kþ :¼ fz 2 R k : z  0g is the usual                             by making the decision. That person or that
“ordering cone” for defining Pareto opti-                                  group is referred to as decision-maker in the
mality. A point x is Pareto optimal if                                     following. Therefore, two issues remain: 1)
f ðxÞ  R kþ \ f ðxÞ ¼ u. By using a larger order-                         How to calculate Pareto optimal solutions;
ing cone such as:                                                          and 2) how to support the selection of a final
                                                                           solution, that is, to make a decision.
R ke :¼ fz 2 R k : distðz; R kþ Þ  e k z kg
f2 f2
        Weight vector
                    Pareto points                                                Pareto point
                                                                                 obtained by 1-norm
                                                                                                          Pareto point
                                                                                                          obtained by weighted
           Tangent planes                                                                                 Euclidean norm
                                                                              zref
f1 f1
                       X
gweighted sum ðxÞ :¼       wi f i ðxÞ                     (117)
                                                                  for given barrieres ej with ji. Solving this
with nonnegative weights wi, see Figure 51. The                   problem for any i and any ej one obtains a
solution of                                                       weakly Pareto optimal solution and if the solu-
                                                                  tion is unique it is Pareto optimal. Furthermore,
min gweighted sum ðxÞ s:t: x 2 X                                  all Pareto optimal solutions can be obtained this
                                                                  way.
is weakly Pareto optimal and Pareto optimal if
                                                                       Another reformulation is based on norms
all weights are positive or if the solution is
                                                                  (Fig. 53). Let zref 2k be a reference point, that
unique. If the original multicriteria optimiza-
                                                                  is, a point in the objective space which is
tion problem is convex, then any Pareto optimal
                                                                  desirable to be obtained but not necessarily
point is a solution of the weighted sum problem
                                                                  possible to obtain. For any norm jj  jj we can
for appropriately chosen weights. However,
                                                                  define a new objective function ggoal ðxÞ :¼
otherwise this need not be the case.
                                                                  jjf ðxÞ  zref jj and solve:
   Another approach is the e-constraint
method, see Figure 52. Here the new optimiza-                     min ggoal ðxÞ s:t: x 2 X:                                      (118)
tion problem has the following shape
enumeration variable with few values attained,        solutions. If this is combined with an intuitive
we can approximate the convex part for each           navigation mechanism it gives the decision-
value of the enumeration variable and end up          maker the feeling of having the complete Pareto
with a set of approximations.                         set at hand, even though only some point were
   If the multicriteria optimization problem is       actually calculated.
simply nonconvex, the Pareto optimal solu-
tions f ðxÞ can be approximated by hyperboxes.        10.9. Optimization under Uncertainty
This approach has the disadvantage that it
becomes very slow if the number of objectives         Karl-Heinz K€ufer, Michael Bortz
increases.
                                                      10.9.1. Introduction
10.8.5. Other Approaches
                                                      In practical applications, models are affected by
A popular approach to obtain a set of Pareto
                                                      uncertainties of their parameters. These result
optimal solutions are population-based meta-
                                                      from the incomplete knowledge of model
heuristics. Here a set of solutions is modified
                                                      parameters, stemming from the model that is
iteratively according to some rules such that
                                                      only an approximation of the reality. Thus,
it gets closer to the set of Pareto optimal
                                                      model parameters (such as parameters in
solutions. Their main advantage is that they
                                                      reaction kinetics or equilibria) are often esti-
do not rely on differentiabity or continuity
                                                      mated from experiments. In addition, there are
assumptions, hence the objective evaluation
                                                      uncertainties in process control, affecting, for
can be a black box. One of the most well-
                                                      example, flow rates or feed compositions, but
known metaheuristics is the evaluationary
                                                      also in varying commodity or energy prices and
algorithm variant NSGA-II. In each iteration
                                                      in changing environmental conditions such as
the population is modified and extended and
                                                      atmospheric temperature and pressure. Since
afterwards a subset of solutions is selected as
                                                      these factors often randomly fluctuate, one
population of the next generation, respectively,
                                                      cannot simply fix them by one value but has
iteration. Possible criteria for the selection
                                                      to consider their “whole behavior”.
are non-dominance and the diversity or spread
                                                          The impact of uncertainties on optimal solu-
of the solutions to approximate the complete
                                                      tions is of special interest. Not only is it impor-
Pareto set.
                                                      tant to know how changes in the input affect the
    It is also possible to use methods from
                                                      optimal settings, it is also desirable to compute
differential geometry to obtain Pareto optimal
                                                      settings that are robust to changes in the input
solutions: In the homotopy approach [303] the
                                                      (in this case, a local optimum can be better than
Pareto set is considered as a manifold which is
                                                      a global one, Fig. 55) or to compute settings that
iteratively sampled.
                                                      take into account the randomness of the input.
10.8.6. Decision Support                              The first aspect can be studied by using sensi-
                                                      tivity analysis, for the second, which is the
Having obtained a set of Pareto optimal solu-         actual optimization under uncertainty, tech-
tions, it is still the question what to do with it.   niques of robust and stochastic optimization
While it is sometimes enough to just calculate        can be used. These tools are discussed in
one or more Pareto optimal solutions, often the       more detail in the following.
decision-maker has to make a choice given a set           In analogy to the general NLP Problem (52)
of solutions. And it is a challenge to support the    we consider a parametric optimization problem:
decision in such a way that the decision-maker        PðjÞ : minn f ðx; jÞs:t :gðx; jÞ  0;
                                                             x2R
can make use of all information available and                                                       (121)
not just select some solution after spending                                 hðx; jÞ ¼ 0:
    Interactions: Interactions occur when the            Screening methods are sampling-based meth-
     perturbation of two or more inputs simul-        ods. Their objective is to identify input variables
     taneously causes variation in the output         which are contributing significantly to the output
     greater than that of varying each of the         variability in high-dimensional models, rather
     inputs alone.                                    than exactly quantifying sensitivity. They tend to
    Given data: While in many cases the prac-        have a relatively low computational cost com-
     titioner has access to the model, in some        pared to other approaches, and can be used in a
     instances a sensitivity analysis must be         preliminary analysis to weed out uninfluential
     performed with “given data”, i.e. where          variables before applying a more informative
     the sample points (the values of the model       analysis to the remaining set. One of the most
     inputs for each run) cannot be chosen by the     commonly used screening methods is the ele-
     analyst.                                         mentary effect method, where averaged finite
                                                      differences are calculated in order to estimate
    There are a large number of approaches to         the effect of uncertain model parameters.
perform a sensitivity analysis, many of which             Variance-based methods are a class of prob-
have been developed to address one or more of         abilistic approaches that quantify the input and
the constraints discussed above:                      output uncertainties as probability distributions
    One of the simplest and most common               and decompose the output variance into parts
approaches is that of changing one-factor-at-         attributable to input variables and combinations
a-time (OFAT or OAT), to see what effect this         of them. The sensitivity of the output to an input
produces on the output?. Thus, any change             variable is therefore measured by the amount of
observed in the output will unambiguously be          variance in the output caused by that input.
due to the single variable changed. The fixation          Further approaches are emulators (also
of all other variables to their central or baseline   known as meta- or surrogate models or response
values increases the comparability of the results     surfaces) and Fourier amplitude sensitivity test
(all “effects“ are computed with reference to the     (FAST).
same central point in space). However, this               For a detailed introduction into sensitivity
approach does not fully explore the input space,      analysis and its application to chemical engi-
since it does not take into account the simulta-      neering amongst others see [304] and [305].
neous variation of input variables. Consequently,         Many of the above mentioned methods, but
the OAT approach cannot detect the presence of        also others are available in form of software.
interactions between input variables.                 Some references are:
    Local methods involve taking the partial
derivative of the output with respect to an input         SimLab: a free software for global sensi-
factor. Similar to OAT/OFAT, they do not                   tivity analysis [306]
attempt to fully explore the input space, since           Sensitivity Analysis Excel Add-In: for sim-
they examine only small perturbations, typi-               ple sample based sensitivity analysis runs
cally one variable at a time.                              in Excel [307]
    Another simple but useful approach consists           MUCM Project: an extensive resource for
in scatter plots: There the output variables are           uncertainty and sensitivity analysis of
plotted against individual input variables, after          computationally-demanding models [308]
(randomly) sampling the model over its input              SALib: a sensitivity analysis library in
distributions. This approach can also deal with            Python (Numpy), [309]
“given data” and gives a direct visual indication
of sensitivity.
    Regression analysis, in the context of sensi-     10.9.3. Robust and Stochastic Optimization
tivity analysis, involves fitting a linear function
to the model response and using standardized          Depending on the circumstances, there are
regression coefficients as measures of sensitiv-      different approaches to treat optimization prob-
ity, which is only suitable when the model is         lems depending on uncertain parameters:
linear. The advantages of regression analysis             It is only known which values (discrete or
are its simplicity and low computational cost.              continuous) a parameter can have.
                                                           Mathematics in Chemical Engineering                                       141
in this section to inequality constraint functions          This formulation can be extended to a multi-
gðx; jÞ.                                                    stage setting by modeling the uncertainty as a
    The simplest approach to deal with random-              filtration process (successive realization of
ness in an optimization problem is to replace               parameters). Under discrete distributions, this
the random variables or the whole (objective                reduces to a scenario tree of parameter realiza-
or constraint) function by their means:                     tions. The compensation approach, however,
gðx; E½jÞ  0 or E½gðx; jÞ  0: But this leads            requires that compensating actions exist at all
to the same problems as mentioned above.                    and can be reasonably modeled, which is not
    Of the many concepts that have been dis-                always possible.
cussed over the years, two concepts have been                   Two-stage linear stochastic programs can be
prevailed.                                                  efficiently solved for a discrete set of scenarios
                                                            in many cases. The solution methods build on
Two- and Multistage Stochastic Programs                     decomposition techniques of linear program-
with Recourse. The basic idea of recourse                   ming. Multi-stage integer and nonlinear sto-
(or compensation) relies on the possibility to              chastic programs are often not or only with
adjust constraints in the system gðx; jÞ  0;               great effort solvable for practically relevant
after observation of j by later compensating                problem sizes.
actions. Accordingly in a two-stage program,
the set of variables splits into first stage deci-
                                                            Stochastic Programs with Chance/Probabilis-
sions x (to be fixed before realization of j) and
                                                            tic Constraints. If the emphasis is on the reli-
second stage or recourse decisions y (to be
                                                            ability of a system, it can be claimed that the
fixed after realization of j, see Fig. 56).
                                                            (system) constraints are not fulfilled for every
   The adjustment of the constraint violation is
                                                            j 2 X but with high probability. That means
modeled by an inequality system ~   gðx; j; yÞ  0,
                                                            gðx; jÞ  0 is replaced by
connecting all three types of variables, and it
causes additional costs vðy; jÞ for the second
                                                            P½gðx; jÞ  0  a                           (126)
stage decisions. Of course, given x and j, y
should be chosen as to minimize second stage
costs among all feasible decisions.                         where a 2 ð0; 1 is some probability level. A
   Summarizing, recourse models replace the                 constraint of this type is called chance or
original problem of minimizing the costs of first           probabilistic constraint. Of course, the higher
stage decisions under stochastic inequalities by            a the more reliable is the modeled system. On
a problem where the sum of first stage costs and            the other hand, the set of feasible values is more
expected optimal second stage costs is mini-                and more shrunken with a%, which increases
mized:                                                      the optimal value of the objective function, e.g.
                                                            the minimal costs, at the same time. The
min ff ðxÞ þ E½uðx; jÞg                         (125)      extreme case a ¼ 1 is similar to the worst-
x2R n
                                                            case approach while the case a ¼ 1=2 relates
                                                            to the expected value approach. This makes
where
                                                            use of probabilistic constraints a good compro-
uðx; jÞ ¼ minn fvðy; jÞj~
                        gðx; j; yÞ  0g
                                                            mise between these two methods.
          y2R
one, Equation (129) is nothing more than the             The root-mean-square or quadratic mean is
scalarization of the (two) objective functions by
means of a weighted sum. By this, one can                                                   vffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi
                                                                                            u N
                                                                               pffiffiffiffiffiffiffiffiffiffiffi uX
compare solutions for different l.                       Root  mean  square ¼ Eðy Þ ¼ t
                                                                                       2             y2i =N
   In the same way, one can add the variability                                                                   i¼1
The median is the middle value (or average of            The value n  1 is used in the denominator
the two middle values) when the set of numbers           because the deviations from the sample average
is arranged in increasing (or decreasing) order.         must total zero:
The geometric mean is
                                                         X
                                                         n
                                                               ðyi  y Þ ¼ 0
y G ¼ ðy1 y2 . . . yN Þ1=N                              i¼1
                                                           Mathematics in Chemical Engineering                                145
Thus, knowing n1 values of yiy and the fact                Table 12. Area under normal curve*
that there are n values automatically gives the n-                                                Rz z2 =2
                                                                                   F ðzÞ ¼ p1ffiffiffiffi
                                                                                             2p
                                                                                                    e       dz
th value. Thus, only n1 degrees of freedom n                                                      0
(y) is replaced by the sample mean y derived                  0.0                0.0000                 1.5             0.4332
from the data.                                                0.1                0.0398                 1.6             0.4452
                                                              0.2                0.0793                 1.7             0.4554
   If data are taken consecutively, running
                                                              0.3                0.1179                 1.8             0.4641
totals can be kept to permit calculation of the               0.4                0.1554                 1.9             0.4713
mean and variance without retaining all the                   0.5                0.1915                 2.0             0.4772
data:                                                         0.6                0.2257                 2.1             0.4821
                                                              0.7                0.2580                 2.2             0.4861
                                                              0.8                0.2881                 2.3             0.4893
X
n               X
                n           X
                            n
  ðyi  y Þ2 ¼   y21  2y   yi þ ðy Þ2                     0.9                0.3159                 2.4             0.4918
i¼1                        i¼1      i¼1                       1.0                0.3413                 2.5             0.4938
       X
       n                                                      1.1                0.3643                 2.7             0.4965
y ¼             yi =n                                        1.2                0.3849                 3.0             0.4987
           i¼1                                                1.3                0.4032                 4.0             0.499968
                                                              1.4                0.4192                 5.0             0.4999997
Thus,                                                         *
                                                               Table gives the probability F that a random variable will fall in the
                                                              shaded region of Figure 58. For a more complete table (in slightly
     X
     n                   X
                         n
                                                              different form), see [23, Table 26.1]. This table is obtained in
n;         y21 ; and           yi
     i¼1                 i¼1
                                                              Microsoft Excel with the function NORMDIST(z,0,1,1)-0.5.
Figure 57. Frequency of occurrence of different scores        Figure 58. Area under normal curve
146              Mathematics in Chemical Engineering
0 FðxÞ 1
Fð1Þ ¼ 0; Fðþ1Þ ¼ 1
         dFðxÞ
pðxÞ ¼
          dx
where
Thus, variables yA and yC are distributed inde-          Suppose Y, which is the sum or difference of
pendently. The joint distribution for two vari-        two variables, is of interest:
ables is
                                                       Y ¼ y A 
 yB
pðyA ; yB Þ ¼ pðyA ÞpðyB jyA Þ
               P
               n                                       s 2 ðYÞ ¼         a2i s 2 ðyi Þ
                   ðyAi  y A ÞðyBi  y B Þ                      i¼1
                                                                                                    (131)
rðyA ; yB Þ ¼ i¼1                                        Xn X
                                                            n
                       ðn  1ÞsA sB                    þ2     ai aj Covðyi ; yj Þ
                                                            i¼1 j¼iþ1
   If measurements are for independent, iden-
tically distributed observations, the errors are       If the variables are uncorrelated and have the
independent and uncorrelated. Then y varies           same variance, then
about E (y) with variance s 2/n, where n is the
                                                                               !
number of observations in y . Thus if some-                        X
                                                                    n
                                                       s 2 ðYÞ ¼           a2i s 2
thing is measured several times today and                           i¼1
every day, and the measurements have the
same distribution, the variance of the means
decreases with the number of samples in each              This fact can be used to obtain more accurate
day’s measurement n. Of course, other factors          cost estimates for the purchased cost of a
(weather, weekends) may cause the observa-             chemical plant than is true for any one piece
tions on different days to be distributed              of equipment. Suppose the plant is composed of
nonidentically.                                        a number of heat exchangers, pumps, towers,
                                                         Mathematics in Chemical Engineering                         149
                      pffiffiffi
and is then 
ð40= nÞ%. Thus the standard
deviation of the cost for the entire plant is               of significance, 0.95 ¼ 0.5 (for negative z) þ F
the standard deviation of each piece of equip-              (for positive z). Thus, F ¼ 0.45 or z ¼ 1.645. In
ment divided by the square root of the number               the two-sided test (see Fig. 61), if a single
of units. Under less restrictive conditions the             sample is chosen and z < 1.96 or z > 1.96,
actual numbers change according to the above                then this could happen with probability 0.05 if
equations, but the principle is the same.                   the hypothesis were true. This z would be
    Suppose modifications are introduced into               significantly different from the expected value
the manufacturing process. To determine if the              (based on the chosen level of significance) and
modification causes a significant change, the               the tendency would be to reject the hypothesis.
mean of some property could be measured                     If the value of z was between 1.96 and 1.96,
before and after the change; if these differ,               the hypothesis would be accepted.
does it mean the process modification caused                    The same type of decisions can be made for
it, or could the change have happened by                    other distributions. Consider Student’s t-distri-
chance? This is a statistical decision. A hypoth-           bution. At a 95 % level of confidence, with n ¼
esis H0 is defined; if it is true, action A must be         10 degrees of freedom, the t values are 
 2.228.
taken. The reverse hypothesis is H1; if this is             Thus, the sample mean would be expected to be
true, action B must be taken. A correct decision            between
is made if action A is taken when H0 is true or                     s
action B is taken when H1 is true. Taking action            y 
tc pffiffiffi
                                                                     n
B when H0 is true is called a type I error,
whereas taking action A when H1 is true is                  with 95 % confidence. If the mean were outside
called a type II error.                                     this interval, the hypothesis would be rejected.
    The following test of hypothesis or test of                The chi-square distribution is useful for
significance must be defined to determine if the            examining the variance or standard deviation.
hypothesis is true. The level of significance is            The statistic is defined as
the maximum probability that an error would be
accepted in the decision (i.e., rejecting the                      ns2
                                                            x2 ¼
hypothesis when it is actually true). Common                       s2
levels of significance are 0.05 and 0.01, and the                  ðy1  y Þ2 þ ðy2  y Þ2 þ . . . þ ðyn  y Þ2
                                                               ¼
test of significance can be either one or two                                           s2
sided. If a sampled distribution is normal, then
the probability that the z score                            and the chi-square distribution is
Table 14. Percentage points of area under chi-square distribution with n degrees of freedom*
of freedom and a 95 % confidence level, the                                 called y, with N2 samples. First, compute the
critical values of x2 are 0.025 and 0.975. Then                             standard error of the difference of the means:
 pffiffiffi       pffiffiffi                                                                 vffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi
s n        s n                                                                   uP                            P
       <s<                                                                       u N1                           N2
                                                                                 u ðxi  xÞ2 þ ðyi  y Þ2                                                 
x0:975     x0:025                                                                ti¼1                          i¼1                           1          1
                                                                            sD ¼                                                                  þ
                                                                                                N1 þ N2  2                                N1 N2
or
 pffiffiffi      pffiffiffi
                                                                            Next, compute the value of t
s n       s n
      <s<
20:5      3:25                                                                    x y
                                                                            t¼
                                                                                    sD
with 95% confidence.
   Tests are available to decide if two distribu-                           and evaluate the significance of t using
tions that have the same variance have different                            Student’s t-distribution for N1 þ N2 2 degrees
means [15, p. 465]. Let one distribution be                                 of freedom.
called x, with N1 samples, and the other be                                    If the samples have different variances, the
                                                                            relevant statistic for the t-test is
                                                                                                x  y
                                                                            t ¼ pffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi
                                                                                 varðxÞ=N 1 þ varðyÞ=N 2
   There is also an F-test to decide if two              If each measured quantity has some variance,
distributions have significantly different vari-         what is the variance in the thermal
ances. In this case, the ratio of variances is           conductivity?
calculated:                                                 Suppose a model for Y depends on various
                                                         measurable quantities, y1, y2, . . . Suppose
      varðxÞ
F¼
      varðyÞ
                                                         several measurements are made of y1, y2, . . .
                                                         under seemingly identical conditions and sev-
where the variance of x is assumed to be larger.         eral different values are obtained, with means E
Then, a table of values is used to determine             (y1), E (y2), . . . and variances s 21 , s 22 , . . .
the significance of the ratio. The table                 Next suppose the errors are small and indepen-
[23, Table 26.9] is derived from the formula             dent of one another. Then a change in Y is
[15, p. 169]                                             related to changes in yi by
                                   n     
                                     2 n1
QðFjn1 ; n2 Þ ¼ I n2 =ðn2 þn1 FÞ      ;                         @Y      @Y
                                    2 2                  dY ¼      dy þ    dy þ . . .
                                                                @y1 1 @y2 2
where the right-hand side is an incomplete beta
function. The F table is given by the Microsoft          If the changes are indeed small, the partial
Excel function FINV(fraction, nx, ny), where             derivatives are constant among all the samples.
fraction is the fractional value ( 1) represent-        Then the expected value of the change is
ing the upper percentage and nx and ny are the
                                                                   N 
                                                                   X      
degrees of freedom of the numerator and                                @Y
                                                         EðdYÞ ¼                    Eðdyi Þ
denominator, respectively.                                            i¼1
                                                                             @yi
   Example. For two sample variances with 8
degrees of freedom each, what limits will                Naturally E (dyi) ¼ 0 by definition so that E
bracket their ratio with a midarea probability           (dY) ¼ 0, too. However, since the errors are
of 90 %? FINV(0.95, 8, 8)  3.44. The 0.95 is            independent of each other and the partial deriv-
used to get both sides to total 10%. Then                atives are assumed constant because the errors
P½1=3:44  varðxÞ=varðyÞ  3:44 ¼ 0:90:
                                                         are small, the variances are given by Equation ()
                                                         [331, p. 550]
11.3. Error Analysis in Experiments                                   N 
                                                                      X     
                                                                          @Y 2
                                                         s 2 ðdYÞ ¼                   s 2i               (132)
                                                                       i¼1
                                                                              @yi
Suppose a measurement of several quantities
is made and a formula or mathematical model              Thus, the variance of the desired quantity Y can
is used to deduce some property of interest.             be found. This gives an independent estimate of
For example, to measure the thermal conduc-              the errors in measuring the quantity Y from the
tivity of a solid k, the heat flux q, the thickness      errors in measuring each variable it depends
of the sample d, and the temperature differ-             upon.
ence across the sample DT must be measured.
Each measurement has some error. The heat
flux q may be the rate of electrical heat input          11.4. Factorial Design of Experiments
Q_ divided by the area A, and both quantities            and Analysis of Variance
are measured to some tolerance. The thickness
of the sample is measured with some accuracy,            Statistically designed experiments consider, of
and the temperatures are probably measured               course, the effect of primary variables, but
with a thermocouple, to some accuracy. These             they also consider the effect of extraneous
measurements are combined, however, to                   variables, the interactions among variables,
obtain the thermal conductivity, and the error           and a measure of the random error. Primary
in the thermal conductivity must be deter-               variables are those whose effect must be deter-
mined. The formula is                                    mined. These variables can be quantitative or
      d                                                  qualitative. Quantitative variables are ones
k¼       Q
     ADT                                                 that may be fit to a model to determine the
152       Mathematics in Chemical Engineering
model parameters. Curve fitting of this type is      Table 15. Estimating the effect of four treatments
discused in Chapter 2. Qualitative variables         Treatment                                   1   2       3   4
are ones whose effect needs to be known; no
attempt is made to quantify that effect other                                                                 
                                                                                                              
than to assign possible errors or magnitudes.                                                                 
Qualitative variables can be further subdivided                                                                
into type I variables, whose effect is deter-                                                                   
mined directly, and type II variables, which                                                                     
                                                     Treatment average, y t                                  
contribute to performance variability, and           Grand average, y                                   
whose effect is averaged out. For example,
in studying the effect of several catalysts on
yield in a chemical reactor, each different type
                                                     that is, whether their means are different.
of catalyst would be a type I variable, because
                                                     The samples are assumed to have the same
its effect should be known. However, each
                                                     variance. The hypothesis is that the treatments
time the catalyst is prepared, the results are
                                                     are all the same, and the null hypothesis is that
slightly different, because of random varia-
                                                     they are different. Deducing the statistical
tions; thus, several batches may exist of what
                                                     validity of the hypothesis is done by an analysis
purports to be the same catalyst. The variabil-
                                                     of variance.
ity between batches is a type II variable.
                                                        The data for k ¼ 4 treatments are arranged
Because the ultimate use will require using
                                                     in Table 15. Each treatment has nt experi-
different batches, the overall effect including
                                                     ments, and the outcome of the i-th experiment
that variation should be known, because
                                                     with treatment t is called yti. The treatment
knowing the results from one batch of
                                                     average is
one catalyst precisely might not be represent-
ative of the results obtained from all batches of             P
                                                              nt
the same catalyst. A randomized block design,                 i¼1
                                                                    yti
incomplete block design, or Latin square             y t ¼
                                                               nt
design, for example, all keep the effect of
experimental error in the blocked variables          and the grand average is
from influencing the effect of the primary
variables. Other uncontrolled variables are                   P
                                                              k
                                                                   nt y t            X
                                                                                      k
accounted for by introducing randomization           y ¼    t¼1
                                                                             ; N¼           nt
in parts of the experimental design. To study                      N                  t¼1
Table 17. Two-level factorial design with three variables                chosen. CROPLEY [333] gives an example of
Run       Variable 1              Variable 2                Variable 3
                                                                         how to combine heuristics and statistical argu-
                                                                         ments in application to kinetics mechanisms in
1                                                                     chemical engineering.
2         þ                                                
3                                þ                         
4         þ                       þ                         
5                                                         þ            References
6         þ                                                þ
7                                þ                         þ             1 D.E. Seborg, T.F. Edgar, D.A. Mellichamp: Process Dynamics
8         þ                       þ                         þ               and Control, 2nd ed., John Wiley & Sons, New York 2004.
                                                                          2 G. Forsyth, C.B. Moler: Computer Solution of Lineart Alge-
                                                                            braic Systems, Prentice-Hall, Englewood Cliffs 1967.
                                                                          3 B.A. Finlayson: Nonlinear Analysis in Chemical Engineering,
                                                                            McGraw-Hill, New York 1980; reprinted, Ravenna Park, Seat-
and each effect is determined with the preci-                               tle 2003.
                                                                          4 S.C. Eisenstat, M.H. Schultz, A.H. Sherman: “Algorithms and
sion of a fourfold replicated difference. The                               Data Structures for Sparse Symmetric Gaussian Elimination,”
advantage of a one-at-a-time experiment is the                              SIAM J. Sci. Stat. Comput. 2 (1981) 225–237.
gain in precision if the variables are additive                           5 I.S. Duff: Direct Methods for Sparse Matrices, Charendon
and the measure of nonadditivity if it occurs                               Press, Oxford 1986.
                                                                          6 H.S. Price, K.H. Coats, “Direct Methods in Reservoir Simu-
[338, p. 313].                                                              lation,” Soc. Pet. Eng. J. 14 (1974) 295–308.
   Interaction effects between variables 1 and 2                          7 A. Bykat: “A Note on an Element Re-Ordering Scheme,” Int. J.
are obtained by comparing the difference                                    Num. Methods Egn. 11 (1977) 194 –198.
between the results obtained with the high                                8 M.R. Hesteness, E. Stiefel: “Methods of conjugate gradients
                                                                            for solving linear systems,” J. Res. Nat. Bur. Stand 29 (1952)
and low value of 1 at the low value of 2 with                               409–439.
the difference between the results obtained with                          9 Y. Saad: Iterative Methods for Sparse Linear Systems, 2nd ed.,
the high and low value 1 at the high value of 2.                            Soc. Ind. Appl. Math., Philadelphia 2003.
The 12-interaction is                                                    10 Y. Saad, M. Schultz: “GMRES: A Generalized Minimal Resid-
                                                                            ual Algorithm for Solving Nonsymmetric Linear Systems.”
12 interaction ¼ ½ðy4  y3 þ y8  y7 Þ  ½ðy2  y1 þ y6  y5 Þ=2          SIAM J. Sci. Statist. Comput. 7 (1986) 856–869.
                                                                         11 http://mathworld.wolfram.com/
The key step is to determine the errors associ-                             GeneralizedMinimalResidualMethod.html.
ated with the effect of each variable and each                           12 http://www.netlib.org/linalg/html_templates/Templates.html.
                                                                         13 http://software.sandia.gov/.
interaction so that the significance can be deter-                       14 E. Isaacson, H.B. Keller, Analysis of Numerical Methods,
mined. Thus, standard errors need to be                                     J. Wiley and Sons, New York 1966.
assigned. This can be done by repeating the                              15 W.H. Press, B.P. Flannery, S.A. Teukolsky, W.T. Vetterling:
experiments, but it can also be done by using                               Numerical Recipes, Cambridge University Press, Cambridge
                                                                            1986.
higher order interactions (such as 123 interac-                          16 R.W.H. Sargent: “A Review of Methods for Solving Non-linear
tions in a 24 factorial design). These are                                  Algebraic Equations,” in R.S.H. Mah, W.D. Seider (eds.):
assumed negligible in their effect on the                                   Foundations of Computer-Aided Chemical Process Design,
mean but can be used to estimate the standard                               American Institute of Chemical Engineers, New York 1981.
                                                                         17 J.D. Seader: “Computer Modeling of Chemical Processes,”
error [338, pp. 319–328]. Then calculated                                   AIChE Monogr. Ser. 81 (1985) no. 15.
effects that are large compared to the standard                          18 http://software.sandia.gov/trilinos/packages/nox/loca_user.
error are considered important, whereas those                               html.
that are small compared to the standard error                            19 N.R. Amundson: Mathematical Methods in Chemical Engi-
                                                                            neering, Prentice-Hall, Englewood Cliffs, N.J. 1966.
are considered due to random variations and are                          20 R.H. Perry, D. W Green: Perry’s Chemical Engineers’ Hand-
unimportant.                                                                book, 7th ed., McGraw-Hill, New York 1997.
   In a fractional factorial design, only part of                        21 D.S. Watkins: “Understanding the QR Algorithm,” SIAM Rev.
the possible experiments is performed. With k                               24 (1982) 427–440.
                                                                         22 G.F. Carey, K. Sepehrnoori: “Gershgorin Theory for Stiffness
variables, a factorial design requires 2k experi-                           and Stability of Evolution Systems and Convection-Diffusion,”
ments. When k is large, the number of experi-                               Comp. Meth. Appl. Mech. 22 (1980) 23–48.
ments can be large; for k ¼ 5, 25 ¼ 32. For k this                       23 M. Abranowitz, I.A. Stegun: Handbook of Mathematical Func-
large, BOX et al. [331, p. 235] do a fractional                             tions, National Bureau of Standards, Washington, D.C. 1972.
                                                                         24 J.C. Daubisse: “Some Results about Approximation of Func-
factorial design. In the fractional factorial                               tions of One or Two Variables by Sums of Exponentials, ” Int. J.
design with k ¼ 5, only 8 experiments are                                   Num. Meth. Eng. 23 (1986) 1959–1967.
                                                                     Mathematics in Chemical Engineering                           155
25 O.C. McGehee: An Introduction to Complex Analysis, John              53 J.D. Lambert: Computational Methods in Ordinary Differential
   Wiley & Sons, New York 2000.                                            Equations, J. Wiley and Sons, New York 1973.
26 H.A. Priestley: Introduction to complex analysis, Oxford Uni-        54 N.R. Amundson: Mathematical Methods in Chemical Engi-
   versity Press, New York 2003.                                           neering, Prentice-Hall, Englewood Cliffs, N.J. 1966.
27 Y.K. Kwok: Applied complex variables for scientists                  55 J.R. Rice: Numerical Methods, Software, and Analysis,
   and engineers, Cambridge University Press, New York                     McGraw-Hill, New York 1983.
   2002.                                                                56 M.B. Bogacki, K. Alejski, J. Szymanowski: “The Fast Method
28 N. Asmar, G.C. Jones: Applied complex analysis with partial             of the Solution of Reacting Distillation Problem,” Comput.
   differential equations, Prentice Hall, Upper Saddle River, NJ,          Chem. Eng. 13 (1989) 1081–1085.
   2002.                                                                57 C.W. Gear: Numerical Initial-Value Problems in Ordinary
29 M.J. Ablowitz, A.S. Fokas: Complex variables: Introduction              Differential Equations, Prentice-Hall, Englewood Cliffs, N.J.
   and applications, Cambridge University Press, New York                  1971.
   2003.                                                                58 N.B. Ferguson, B.A. Finlayson: “Transient Modeling of a
30 J.W. Brown, R.V. Churchill: Complex variables and applica-              Catalytic Converter to Reduce Nitric Oxide in Automobile
   tions, 6th ed., McGraw-Hill, New York 1996; 7th ed. 2003.               Exhaust,” AIChE J 20 (1974) 539–550.
31 W. Kaplan: Advanced calculus, 5th ed., Addison-Wesley,               59 W.F. Ramirez: Computational Methods for Process Simula-
   Redwood City, Calif., 2003.                                             tions, 2nd ed., Butterworth-Heinemann, Boston 1997.
32 E. Hille: Analytic Function Theory, Ginn and Co., Boston             60 U.M. Ascher, L.R. Petzold: Computer methods for ordinary
   1959.                                                                   differential equations and differential-algebraic equations,
33 R.V. Churchill: Operational Mathematics, McGraw-Hill,                   SIAM, Philadelphia 1998.
   New York 1958.                                                       61 K.E. Brenan, S.L. Campbell, L.R. Petzold: Numerical Solution
34 J.W. Brown, R.V. Churchill: Fourier Series and Boundary                 of Initial-Value Problems in Differential-Algebraic Equations,
   Value Problems, 6th ed., McGraw-Hill, New York 2000.                    Elsevier, Amsterdam 1989.
35 R.V. Churchill: Operational Mathematics, 3rd ed., McGraw-            62 C.C. Pontelides, D. Gritsis, K.R. Morison, R.W.H. Sargent:
   Hill, New York 1972.                                                    “The Mathematical Modelling of Transient Systems Using
36 B. Davies: Integral Transforms and Their Applications, 3rd ed.,         Differential-Algebraic Equations,” Comput. Chem. Eng. 12
   Springer, Heidelberg 2002.                                              (1988) 449–454.
37 D.G. Duffy: Transform Methods for Solving Partial Differen-          63 G.A. Byrne, P.R. Ponzi: “Differential-Algebraic Systems,
   tial Equations, Chapman & Hall/CRC, New York 2004.                      Their Applications and Solutions,” Comput. Chem. Eng. 12
38 A. Varma, M. Morbidelli: Mathematical Methods in Chemical               (1988) 377–382.
   Engineering, Oxford, New York 1997.                                  64 L.F. Shampine, M.W. Reichelt: The MATLAB ODE Suite,
39 H. Bateman, Tables of Integral Transforms, vol. I, McGraw-              SIAM J. Sci. Comp. 18 (1997) 122.
   Hill, New York 1954.                                                 65 M. Kubicek, M. Marek: Computational Methods in Bifurcation
40 H.F. Weinberger: A First Course in Partial Differential Equa-           Theory and Dissipative Structures, Springer Verlag, Berlin
   tions, Blaisdell, Waltham, Mass. 1965.                                  1983.
41 R.V. Churchill: Operational Mathematics, McGraw-Hill, New            66 T.F.C. Chan, H.B. Keller: “Arc-Length Continuation and
   York 1958.                                                              Multi-Grid Techniques for Nonlinear Elliptic Eigenvalue Prob-
42 J.T. Hsu, J.S. Dranoff: “Numerical Inversion of Certain                 lems,” SIAM J. Sci. Stat. Comput. 3 (1982) 173–194.
   Laplace Transforms by the Direct Application of Fast Fourier         67 M.F. Doherty, J.M. Ottino: “Chaos in Deterministic Systems:
   Transform (FFT) Algorithm,” Comput. Chem. Eng. 11 (1987)                Strange Attractors, Turbulence and Applications in Chemical
   101–110.                                                                Engineering,” Chem. Eng. Sci. 43 (1988) 139–183.
43 E. Kreyzig: Advanced Engineering Mathematics, 9th ed., John          68 M.P. Allen, D.J. Tildesley: Computer Simulation of Liquids,
   Wiley & Sons, New York 2006.                                            Clarendon Press, Oxford 1989.
44 H.S. Carslaw, J.C. Jaeger: Conduction of Heat in Solids, 2nd         69 D. Frenkel, B. Smit: Understanding Molecular Simulation,
   ed., Clarendon Press, Oxford–London 1959.                               Academic Press, San Diego 2002.
45 R.B. Bird, R.C. Armstrong, O. Hassager: Dynamics of Poly-            70 J.M. Haile: Molecular Dynamics Simulation, John Wiley &
   meric Liquids, 2nd ed., Appendix A, Wiley-Interscience, New             Sons, New York 1992.
   York 1987.                                                           71 A.R. Leach: Molecular Modelling: Principles and Applica-
46 R.S. Rivlin, J. Rat. Mech. Anal. 4 (1955) 681–702.                      tions, Prentice Hall, Englewood Cliffs, NJ, 2001.
47 N.R. Amundson: Mathematical Methods in Chemical Engi-                72 T. Schlick: Molecular Modeling and Simulations, Springer,
   neering; Matrices and Their Application, Prentice-Hall, Engle-          New York 2002.
   wood Cliffs, N.J. 1966.                                              73 R.B. Bird, W.E. Stewart, E.N. Lightfoot: Transport Phe-
48 I.S. Sokolnikoff, E.S. Sokolnikoff: Higher Mathematics for              nomena, 2nd ed., John Wiley & Sons, New York 2002.
   Engineers and Physicists, McGraw-Hill, New York 1941.                74 R.B. Bird, R.C. Armstrong, O. Hassager: Dynamics of Poly-
49 R.C. Wrede, M.R. Spiegel: Schaum’s outline of theory and                meric Liquids, 2nd ed., Wiley-Interscience, New York
   problems of advanced calculus, 2nd ed, McGraw Hill, New                 1987.
   York 2006.                                                           75 P.V. Danckwerts: “Continuous Flow Systems,” Chem. Eng. Sci.
50 P. M. Morse, H. Feshbach: Methods of Theoretical Physics,               2 (1953) 1–13.
   McGraw-Hill, New York 1953.                                          76 J.F. Wehner, R. Wilhelm: “Boundary Conditions of Flow
51 B.A. Finlayson: The Method of Weighted Residuals and Vari-              Reactor,” Chem. Eng. Sci. 6 (1956) 89–93.
   ational Principles, Academic Press, New York 1972.                   77 V. Hlavacek, H. Hofmann: “Modeling of Chemical Reactors-
52 G. Forsythe, M. Malcolm, C. Moler: Computer Methods for                 XVI-Steady-State Axial Heat and Mass Transfer in Tubular
   Mathematical Computation, Prentice-Hall, Englewood Cliffs,              Reactors. An Analysis of the Uniqueness of Solutions,” Chem.
   N.J. 1977.                                                              Eng. Sci. 25 (1970) 173–185.
156         Mathematics in Chemical Engineering
78 B.A. Finlayson: Numerical Methods for Problems with Moving        100 R. Aris, N.R. Amundson: Mathematical Methods in Chemi-
   Fronts, Ravenna Park Publishing Inc., Seattle 1990.                   cal Engineering, vol. 2, First-Order Partial Differential
79 E. Isaacson, H.B. Keller: Analysis of Numerical Methods, J.           Equations with Applications, Prentice-Hall, Englewood
   Wiley and Sons, New York 1966.                                        Cliffs, NJ, 1973.
80 C. Lanczos: “Trigonometric Interpolation of Empirical and         101 R. Courant, D. Hilbert: Methods of Mathematical Physics, vol.
   Analytical Functions,” J. Math. Phys. (Cambridge Mass.) 17            I and II, Intersicence, New York 1953, 1962.
   (1938) 123–199.                                                   102 P.M. Morse, H. Feshbach: Methods of Theoretical Physics, vol.
81 C. Lanczos: Applied Analysis, Prentice-Hall, Englewood                I and II, McGraw-Hill, New York 1953.
   Cliffs, N.J. 1956.                                                103 A.D. Polyanin: Handbook of Linear Partial Differential Equa-
82 J. Villadsen, W.E. Stewart: “Solution of Boundary-Value Prob-         tions for Engineers and Scientists, Chapman and Hall/CRC,
   lems by Orthogonal Collocation,” Chem. Eng. Sci. 22 (1967)            Boca Raton, FL 2002.
   1483–1501.                                                        104 D. Ramkrishna, N.R. Amundson: Linear Operator Methods in
83 M.L. Michelsen, J. Villadsen: “Polynomial Solution of Differ-         Chemical Engineering with Applications to Transport and
   ential Equations” in R.S.H. Mah, W.D. Seider (eds.): Founda-          Chemical Reaction Systems, Prentice Hall, Englewood Cliffs,
   tions of Computer-Aided Chemical Process Design,                      NJ, 1985.
   Engineering Foundation, New York 1981, pp. 341–368.               105 D.D. Joseph, M. Renardy, J.C. Saut: “Hyperbolicity and
84 W. E. Stewart, K. L. Levien, M. Morari: “Collocation Methods          Change of Type in the Flow of Viscoelastic Fluids,” Arch.
   in Distillation,” in A. W. Westerberg, H. H. Chien (eds.):            Rational Mech. Anal. 87 (1985) 213–251.
   Proceedings of the Second Int. Conf. on Foundations of Com-       106 H.K. Rhee, R. Aris, N.R. Amundson: First-Order Partial Differ-
   puter-Aided Process Design, Computer Aids for Chemical                ential Equations, Prentice-Hall, Englewood Cliffs, N.J. 1986.
   Engineering Education (CACHE), Austin Texas, 1984,                107 B.A. Finlayson: Numerical Methods for Problems with Moving
   pp. 535–569.                                                          Fronts, Ravenna Park Publishing Inc., Seattle 1990.
85 W.E. Stewart, K.L. Levien, M. Morari: “Simulation of Frac-        108 D.L. Book: Finite-Difference Techniques for Vectorized Fluid
   tionation by Orthogonal Collocation,” Chem. Eng. Sci. 40              Dynamics Calculations, Springer Verlag, Berlin 1981.
   (1985) 409–421.                                                   109 G.A. Sod: Numerical Methods in Fluid Dynamics, Cambridge
86 C.L.E. Swartz, W.E. Stewart: “Finite-Element Steady State             University Press, Cambridge 1985.
   Simulation of Multiphase Distillation,” AIChE J. 33 (1987)        110 G.H. Xiu, J.L. Soares, P. Li, A.E. Rodriques: “Simulation of
   1977–1985.                                                            five-step one-bed sorption-endanced reaction process,” AIChE
87 K. Alhumaizi, R. Henda, M. Soliman: “Numerical Analysis of            J. 48 (2002) 2817–2832.
   a reaction-diffusion-convection system,” Comp. Chem. Eng. 27      111 A. Malek, S. Farooq: “Study of a six-bed pressure swing
   (2003) 579–594.                                                       adsorption process,” AIChE J. 43 (1997) 2509–2523.
88 E.F. Costa, P.L.C. Lage, E.C. Biscaia, Jr.: “On the numerical     112 R.J. LeVeque: Numerical Methods for Conservation Laws,
   solution and optimization of styrene polymerization in tubular        Birkh€auser, Basel 1992.
   reactors,” Comp. Chem. Eng. 27 (2003) 1591–1604.                  113 W.F. Ames: “Recent Developments in the Nonlinear Equations
89 J. Wang, R.G. Anthony, A. Akgerman: “Mathematical simula-             of Transport Processes,” Ind. Eng. Chem. Fundam. 8 (1969)
   tions of the performance of trickle bed and slurry reactors for       522–536.
   methanol synthesis,” Comp. Chem. Eng. 29 (2005) 2474–2484.        114 W.F. Ames: Nonlinear Partial Differential Equations in Engi-
90 V.K.C. Lee, J.F. Porter, G. McKay, A.P. Mathews: “Application         neering, Academic Press, New York 1965.
   of solid-phase concentration-dependent HSDM to the acid dye       115 W.E. Schiesser: The Numerical Method of Lines, Academic
   adsorption system,” AIChE J. 51 (2005) 323–332.                       Press, San Diego 1991.
91 C. deBoor, B. Swartz: “Collocation at Gaussian Points,” SIAM      116 D. Gottlieb, S.A. Orszag: Numerical Analysis of Spectral
   J. Num. Anal. 10 (1973) 582–606.                                      Methods: Theory and Applications, SIAM, Philadelphia, PA
92 R.F. Sincovec: “On the Solution of the Equations Arising From         1977.
   Collocation With Cubic B-Splines,” Math. Comp. 26 (1972)          117 D.W. Peaceman: Fundamentals of Numerical Reservoir Simu-
   893–895.                                                              lation, Elsevier, Amsterdam 1977.
93 U. Ascher, J. Christiansen, R.D. Russell: “A Collocation Solver   118 G. Juncu, R. Mihail: “Multigrid Solution of the Diffusion-
   for Mixed-Order Systems of Boundary-Value Problems,”                  Convection-Reaction Equations which Describe the Mass and/
   Math. Comp. 33 (1979) 659–679.                                        or Heat Transfer from or to a Spherical Particle,” Comput.
94 R.D. Russell, J. Christiansen: “Adaptive Mesh Selection Strat-        Chem. Eng. 13 (1989) 259–270.
   egies for Solving Boundary-Value Problems,” SIAM J. Num.          119 G.R. Buchanan: Schaum’s Outline of Finite Element Analysis,”
   Anal. 15 (1978) 59–80.                                                McGraw-Hill, New York 1995.
95 P.G. Ciarlet, M.H. Schultz, R.S. Varga: “Nonlinear Boundary-      120 M.D. Gunzburger: Finite Element Methods for Viscous
   Value Problems I. One Dimensional Problem,” Num. Math. 9              Incompressible Flows, Academic Press, San Diego 1989.
   (1967) 394–430.                                                   121 H. Kardestuncer, D.H. Norrie: Finite Element Handbook,
96 W.F. Ames: Numerical Methods for Partial Differential Equa-           McGraw-Hill, New York 1987.
   tions, 2nd ed., Academic Press, New York 1977.                    122 J.N. Reddy, D.K. Gartling: The Finite Element Method in Heat
97 J.F. Botha, G.F. Pinder: Fundamental Concepts in The Numer-           Transfer and Fluid Dynamics, 2nd ed., CRC Press, Boca Raton,
   ical Solution of Differential Equations, Wiley-Interscience,          FL 2000.
   New York 1983.                                                    123 O.C. Zienkiewicz, R.L. Taylor, J.Z. Zhu: The Finite Element
98 W.H. Press, B.P. Flanner, S.A. Teukolsky, W.T. Vetterling:            Method: Its Basis & Fundamentals, vol. 1, 6th ed., Elsevier
   Numerical Recipes, Cambridge Univ. Press, Cambridge                   Butterworth-Heinemann, Burlington, MA 2005.
   1986.                                                             124 O.C. Zienkiewicz, R.L. Taylor: The Finite Element Method,
99 H. Schlichting, Boundary Layer Theory, 4th ed. McGraw-Hill,           Solid and Structural Mechanics, vol. 2, 5th ed., Butterworth-
   New York 1960.                                                        Heinemann, Burlington, MA 2000.
                                                                        Mathematics in Chemical Engineering                             157
125 O.C. Zienkiewicz, R.L. Taylor: The Finite Element Method,              149 C.T.H. Baker: The Numerical Treatment of Integral Equations,
    Fluid Dynamics, vol. 3, 5th ed., Butterworth-Heinemann,                    Clarendon Press, Oxford 1977.
    Burlington, MA 2000.                                                   150 L.M. Delves, J. Walsh (eds.): Numerical Solution of Integral
126 Z.C. Li: Combined Methods for Elliptic Equations with Sin-                 Equations, Clarendon Press, Oxford 1974.
    gularities, Interfaces and Infinities, Kluwer Academic Pub-            151 P. Linz: Analytical and Numerical Methods for Volterra Equa-
    lishers, Boston, MA, 1998.                                                 tions, SIAM Publications, Philadelphia 1985.
127 G.J. Fix, S. Gulati, G.I. Wakoff: “On the use of singular              152 M.A. Golberg (ed.): Numerical Solution of Integral Equations,
    functions with finite element approximations,” J. Comp.                    Plenum Press, New York 1990.
    Phys. 13 (1973) 209–238.                                               153 C. Corduneanu: Integral Equations and Applications, Cam-
128 N.M. Wigley: “On a method to subtract off a singularity at a               bridge Univ. Press, Cambridge 1991.
    corner for the Dirichlet or Neumann problem,” Math. Comp. 23           154 R. Kress: Linear Integral Equations, 2nd ed. Springer, Heidel-
    (1968) 395–401.                                                            berg 1999.
129 H.Y. Hu, Z.C. Li, A.H.D. Cheng: “Radial Basis Collocation              155 P.K. Kythe P. Purl: Computational Methods for Linear Integral
    Methods for Elliptic Boundary Value Problems,” Comp. Math.                 Equations, Birkh€auser, Basel 2002.
    Applic. 50 (2005) 289–320.                                             156 G. Nagel, G. Kluge: “Non-Isothermal Multicomponent
130 S.J. Osher, R.P. Fedkiw:, Level Set Methods and Dynamic                    Adsorption Processes and Their Numerical Treatment by
    Implicit Surfaces, Springer, New York 2002.                                Means of Integro-Differential Equations,” Comput. Chem.
131 M.W. Chang, B.A. Finlayson: “On the Proper Boundary Con-                   Eng. 13 (1989) 1025–1030.
    dition for the Thermal Entry Problem,” Int. J. Num. Methods            157 P.L. Mills, S. Lai, M.P. Dudukovic, P.A. Ramachandran: “A
    Eng. 15 (1980) 935–942.                                                    Numerical Study of Approximation Methods for Solution of
132 R. Peyret, T.D. Taylor: Computational Methods for Fluid Flow,              Linear and Nonlinear Diffusion-Reaction Equations with Dis-
    Springer Verlag, Berlin–Heidelberg–New York–Tokyo 1983.                    continuous Boundary Conditions,” Comput. Chem. Eng. 12
133 P.M. Gresho, S.T. Chan, C. Upson, R.L. Lee: “A Modified                    (1988) 37–53.
    Finite Element Method for Solving the Time-Dependent,                  158 D. Duffy: Green’s Functions with Applications, Chapman and
    Incompressible Navier–Stokes Equations,” Int. J. Num.                      Hall/CRC, New York 2001.
    Method. Fluids (1984) “Part 1. Theory”: 557–589; “Part 2:              159 I. Statgold: Green’s Functions and Boundary Value Problems,
    Applications”: 619–640.                                                    2nd ed., Interscience, New York 1997.
134 P.M. Gresho, R.L. Sani: Incompressible Flow and the Finite             160 B. Davies: Integral Transforms and Their Applications, 3rd ed.,
    Element Method, Advection-Diffusion, vol. 1, John Wiley &                  Springer, New York 2002.
    Sons, New York 1998.                                                   161 J.M. Bownds, “Theory and performance of a subroutine for
135 P.M. Gresho, R.L. Sani: Incompressible Flow and the Finite                 solving Volterra integral equations,” J. Comput. 28 317–332
    Element Method , Isothermal Laminar Flow, vol. 2, John Wiley               (1982).
    & Sons, New York 1998.                                                 162 J.G. Blom, H. Brunner, “Algorithm 689; discretized collo-
136 J.A. Sethian, Level Set Methods, Cambridge University Press,               cation and iterated collocation for nonlinear Volterra integral
    Cambridge 1996.                                                            equations of the second kind,” ACM Trans. Math. Software
137 C.C. Lin, H. Lee, T. Lee, L.J. Weber: “A level set characteristic          17 (1991) 167–177.
    Galerkin finite element method for free surface flows,” Int. J.        163 N.B. Ferguson, B.A. Finlayson: “Error Bounds for Approxi-
    Num. Methods Fluids 49 (2005) 521–547.                                     mate Solutions to Nonlinear Ordinary Differential Equations,”
138 S. Succi: The Lattice Boltzmann Equation for Fluid Dynamics                AIChE J. 18 (1972) 1053–1059.
    and Beyond, Oxford University Press, Oxford 2001.                      164 R.S. Dixit, L.L. Taularidis: “Integral Method of Analysis of
139 M.C. Sukop, D.T. Thorne, Jr.: Lattice Boltzmann Modeling: An               Fischer–Tropsch Synthesis Reactions in a Catalyst Pellet,”
    Introduction for Geoscientists and Engineers, Springer, New                Chem. Eng. Sci. 37 (1982) 539–544.
    York 2006.                                                             165 L. Lapidus, G.F. Pinder: Numerical Solution of Partial Differ-
140 S. Chen, S.G.D. Doolen: “Lattice Boltzmann method for fluid                ential Equations in Science and Engineering, Wiley-Inter-
    flows,” Annu. Rev. Fluid Mech. 30 (2001) 329–364.                          science, New York 1982.
141 H.T. Lau: Numerical Library in C for Scientists and Engineers,         166 C.K. Hsieh, H. Shang: “A Boundary Condition Dissection
    CRC Press, Boca Raton, FL 1994.                                            Method for the Solution of Boundary-Value Heat Conduction
142 Y.Y. Al-Jaymany, G. Brenner, P.O. Brum: “Comparative study                 Problems with Position-Dependent Convective Coefficients,”
    of lattice-Boltzmann and finite volume methods for the simu-               Num. Heat Trans. Part B: Fund. 16 (1989) 245–255.
    lation of laminar flow through a 4:1 contraction,” Int. J. Num.        167 C.A. Brebbia, J. Dominguez: Boundary Elements–An Intro-
    Methods Fluids 46 (2004) 903–920.                                          ductory Course, 2nd ed., Computational Mechanics Publica-
143 R.L. Burden, R. L J. D. Faires, A.C. Reynolds: Numerical                   tions, Southhamtpon 1992.
    Analysis, 8th ed., Brooks Cole, Pacific Grove, CA 2005.                168 J. Mackerle, C.A. Brebbia (eds.): Boundary Element Reference
144 S.C. Chapra, R.P. Canal: Numerical Methods for Engineers,”                 Book, Springer Verlag, Berlin 1988.
    5th ed., McGraw-Hill, New York 2006.                                   169 L.T. Biegler, I.E. Grossmann, A.W. Westerberg: Systematic
145 H.T. Lau: Numerical Library in Java for Scientists and Engi-               Methods for Chemical Process Design, Prentice-Hall Engle-
    neers, CRC Press, Boca Raton, FL 2004.                                     wood Cliffs, NJ 1997.
146 K.W. Morton, D.F. Mayers: Numerical Solution of Partial                170 I.E. Grossmann (ed.), “Global Optimization in Engineering
    Differential Equations, Cambridge University Press, New                    Design”, Kluwer, Dordrecht 1996.
    York 1994.                                                             171 J. Kallrath, “Mixed Integer Optimization in the Chemical
147 A. Quarteroni, A. Valli: Numerical Approximation of Partial                Process Industry: Experience, Potential and Future,” Trans.
    Differential Equations, Springer, Heidelberg 1997.                         I. Chem. E., 78 (2000) Part A, 809–822.
148 F. Scheid: “Schaum’s Outline of Numerical Analysis,” 2nd ed.,          172 O.L. Mangasarian: Nonlinear Programming, McGraw-Hill,
    McGraw-Hill, New York 1989.                                                New York 1969.
158          Mathematics in Chemical Engineering
173 G.B. Dantzig: Linear Programming and Extensions, Princeton       195 gPROMS User’s Guide, PSE Ltd., London 2002.
    University Press, Princeton, N.J, 1963.                          196 G.E.P. Box: “Evolutionary Operation: A Method for Increasing
174 S.J. Wright: Primal-Dual Interior Point Methods, SIAM,               Industrial Productivity,” Applied Statistics 6 (1957) 81–101.
    Philadelphia 1996.                                               197 R. Hooke, T.A. Jeeves: “Direct Search Solution of Numerical
175 F. Hillier, G.J. Lieberman: Introduction to Operations               and Statistical Problems,” J. ACM 8 (1961) 212.
    Research, Holden-Day, San Francisco, 1974.                       198 M.J.D. Powell: “An Efficient Method for Finding the Mini-
176 T.F. Edgar, D.M. Himmelblau, L.S. Lasdon: Optimization of            mum of a Function of Several Variables without Calculating
    Chemical Processes, McGraw-Hill Inc., New York 2002.                 Derivatives, Comput. J. 7 (1964) 155.
177 K. Schittkowski: More Test Examples for Nonlinear Program-       199 J.A. Nelder, R. Mead: “A Simplex Method for Function
    ming Codes, Lecture notes in economics and mathematical              Minimization,” Computer Journal 7 (1965) 308.
    systems no. 282, Springer-Verlag, Berlin 1987.                   200 R. Luus, T.H.I. Jaakola: “Direct Search for Complex Systems,”
178 J. Nocedal, S.J. Wright: Numerical Optimization, Springer,           AIChE J 19 (1973) 645–646.
    New York 1999.                                                   201 R. Goulcher, J.C. Long: “The Solution of Steady State Chemi-
179 R. Fletcher: Practical Optimization, John Wiley & Sons, Ltd.,        cal Engineering Optimization Problems using a Random
    Chichester, UK 1987                                                  Search Algorithm,” Comp. Chem. Eng. 2 (1978) 23.
180 A. W€achter, L.T. Biegler: “On the Implementation of an          202 J.R. Banga, W.D. Seider: in “Global Optimization of Chemical
    Interior Point Filter Line Search Algorithm for Large-Scale          Processes using Stochastic Algorithms,” C. Floudas, P. Parda-
    Nonlinear Programming,” 106 Mathematical Programming                 los (eds.): State of the Art in Global Optimization, Kluwer,
    (2006) no. 1, 25–57.                                                 Dordrecht 1996, p. 563.
181 C.V. Rao, J.B. Rawlings S. Wright: On the Application of         203 P.J.M. van Laarhoven, E.H.L. Aarts: Simulated Annealing:
    Interior Point Methods to Model Predictive Control. J. Optim.        Theory and Applications, Reidel Publishing, Dordrecht 1987.
    Theory Appl. 99 (1998) 723.                                      204 J.H. Holland: Adaptations in Natural and Artificial Systems,
182 J. Albuquerque, V. Gopal, G. Staus, L.T. Biegler, B.E. Ydstie:       University of Michigan Press, Ann Arbor 1975.
    “Interior point SQP Strategies for Large-scale Structured Pro-   205 J.E. Dennis, V. Torczon: “Direct Search Methods on Parallel
    cess Optimization Problems,” Comp. Chem. Eng. 23 (1997)              Machines,” SIAM J. Opt. 1 (1991) 448.
    283.                                                             206 A.R. Conn, K. Scheinberg, P. Toint: “Recent Progress in
183 T. Jockenhoevel, L.T. Biegle, A. W€achter: “Dynamic Optimi-          Unconstrained Nonlinear Optimization without Deriva-
    zation of the Tennessee Eastman Process Using the                    tives,” Mathemtatical Programming, Series B, 79 (1997)
    OptControlCentre,” Comp. Chem. Eng. 27 (2003) no. 11,                no. 3, 397.
    1513–1531.                                                       207 M. Eldred: “DAKOTA: A Multilevel Parallel Object-Oriented
184 L.T. Biegler, A.M. Cervantes, A. W€achter: “Advances in              Framework for Design Optimization, Parameter Estimation,
    Simultaneous Strategies for Dynamic Process Optimization,”           Uncertainty Quantification, and Sensitivity Analysis,” 2002,
    Chemical Engineering Science 57 (2002) no. 4, 575–593.               http://endo.sandia.gov/DAKOTA/software.html
185 C.D. Laird, L.T. Biegler, B. van Bloemen Waanders, R.A.          208 A.J. Booker et al.: “A Rigorous Framework for Optimization of
    Bartlett: “Time Dependent Contaminant Source Determination           Expensive Functions by Surrogates,” CRPC Technical Report
    for Municipal Water Networks Using Large Scale Opti-                 98739, Rice University, Huston TX, February 1998.
    mization,” ASCE Journal of Water Resource Management             209 R. Horst, P.M. Tuy: “Global Optimization: Deterministic
    and Planning 131 (2005) no. 2, 125.                                  Approaches,” 3rd ed., Springer Verlag, Berlin 1996.
186 A.R. Conn, N. Gould, P. Toint: Trust Region Methods, SIAM,       210 I. Quesada, I.E. Grossmann: “A Global Optimization Algo-
    Philadelphia 2000.                                                   rithm for Linear Fractional and Bilinear Programs,” Journal of
187 A. Drud: “CONOPT–A Large Scale GRG Code,” ORSA                       Global Optimization 6 (1995) no. 1, 39–76.
    Journal on Computing 6 (1994) 207–216.                           211 G.P. McCormick: “Computability of Global Solutions to Fac-
188 B.A. Murtagh, M.A. Saunders: MINOS 5.1 User’s Guide,                 torable Nonconvex Programs: Part I–Convex Underestimating
    Technical Report SOL 83-20R, Stanford University, Palo               Problems,” Mathematical Programming 10 (1976) 147–175.
    Alto 1987.                                                       212 H.S. Ryoo, N.V. Sahinidis: “Global Optimization of Noncon-
189 R.H. Byrd, M.E. Hribar, J. Nocedal: An Interior Point Algo-          vex NLPs and MINLPs with Applications in Process Design,”
    rithm for Large Scale Nonlinear Programming, SIAM J. Opt. 9          Comp. Chem. Eng. 19 (1995) no. 5, 551–566.
    (1999) no. 4, 877.                                               213 M. Tawarmalani, N.V. Sahinidis: “Global Optimization of
190 R. Fletcher, N.I.M. Gould, S. Leyffer, Ph. L. Toint, A.              Mixed Integer Nonlinear Programs: A Theoretical and Com-
    Waechter: “Global Convergence of a Trust-region (SQP)-filter         putational Study,” Mathematical Programming 99 (2004) no.
    Algorithms for General Nonlinear Programming,” SIAM J.               3, 563–591.
    Opt. 13 (2002) no. 3, 635–659.                                   214 C.S. Adjiman., I.P. Androulakis, C.A. Floudas: “Global Opti-
191 R.J. Vanderbei, D.F. Shanno: An Interior Point Algorithm for         mization of MINLP Problems in Process Synthesis and Design.
    Non-convex Nonlinear Programming. Technical Report SOR-              Comp. Chem. Eng., 21 (1997) Suppl., S445–S450.
    97-21, CEOR, Princeton University, Princeton, NJ, 1997.          215 C.S. Adjiman, I.P. Androulakis, C.A. Floudas: “Global Opti-
192 P.E. Gill, W. Murray, M. Wright: Practical Optimization,             mization of Mixed-Integer Nonlinear Problems,” AIChE Jour-
    Academic Press, New York 1981.                                       nal, 46 (2000) no. 9, 1769–1797.
193 P.E. Gill, W. Murray, M.A. Saunders: User’s guide for SNOPT:     216 C.S. Adjiman, I.P. Androulakis, C.D. Maranas, C.A. Floudas:
    A FORTRAN Package for Large-scale Nonlinear Program-                 “A Global Optimization Method, aBB, for Process Design,”
    ming, Technical report, Department of Mathematics, Univer-           Comp. Chem. Eng., 20 (1996) Suppl., S419–S424.
    sity of California, San Diego 1998.                              217 J.M. Zamora, I.E. Grossmann: “A Branch and Contract Algo-
194 J.T. Betts: Practical Methods for Optimal Control Using              rithm for Problems with Concave Univariate, Bilinear and
    Nonlinear Programming, Advances in Design and Control 3,             Linear Fractional Terms,” Journal of Gobal Optimization 14
    SIAM, Philadelphia 2001.                                             (1999) no. 3, 217–249.
                                                                     Mathematics in Chemical Engineering                           159
218 E.M.B. Smith, C.C. Pantelides: “Global Optimization of              239 J.F. Benders: “Partitioning Procedures for Solving Mixed-
    Nonconvex NLPs and MINLPs with Applications in Process                  variables Programming Problems,” Numeri. Math. 4 (1962)
    Design,” Comp. Chem. Eng., 21 (1997) no. 1001,                          238–252.
    S791–S796.                                                          240 G.L. Nemhauser, L.A. Wolsey: Integer and Combinatorial
219 J.E. Falk, R.M. Soland: “An Algorithm for Separable Non-                Optimization, Wiley-Interscience, New York 1988.
    convex Programming Problems,” Management Science 15                 241 C. Barnhart, E.L. Johnson, G.L. Nemhauser, M.W.P. Savels-
    (1969) 550–569.                                                         bergh, P.H. Vance: “Branch-and-price: Column Generation for
220 F.A. Al-Khayyal, J.E. Falk: “Jointly Constrained Biconvex               Solving Huge Integer Programs,” Operations Research 46
    Programming,” Mathematics of Operations Research 8                      (1998) 316–329.
    (1983) 273–286.                                                     242 O.K. Gupta, V. Ravindran: “Branch and Bound Experiments in
221 F.A. Al-Khayyal: “Jointly Constrained Bilinear Programs and             Convex Nonlinear Integer Programming,” Management Sci-
    Related Problems: An Overview,” Computers and Mathemat-                 ence 31 (1985) no. 12, 1533–1546.
    ics with Applications, 19 (1990) 53–62.                             243 S. Nabar, L. Schrage: “Modeling and Solving Nonlinear
222 I. Quesada, I.E. Grossmann: “A Global Optimization Algo-                Integer Programming Problems,” AIChE Meeting, Chicago
    rithm for Heat Exchanger Networks,” Ind. Eng. Chem. Res. 32             1991.
    (1993) 487–499.                                                     244 B. Borchers, J.E. Mitchell: “An Improved Branch and Bound
223 H.D. Sherali, A. Alameddine: “A New Reformulation-Linear-               Algorithm for Mixed Integer Nonlinear Programming,” Com-
    ization Technique for Bilinear Programming Problems,” Jour-             puters and Operations Research 21 (1994) 359–367.
    nal of Global Optimization 2 (1992) 379–410.                        245 R. Stubbs, S. Mehrotra: “A Branch-and-Cut Method for 0-1
224 H.D. Sherali, C.H. Tuncbilek: “A Reformulation-Convexi-                 Mixed Convex Programming,” Mathematical Programming 86
    fication Approach for Solving Nonconvex Quadratic Pro-                  (1999) no. 3, 515–532.
    gramming Problems,” Journal of Global Optimization 7                246 S. Leyffer: “Integrating SQP and Branch and Bound for Mixed
    (1995) 1–31.                                                            Integer Noninear Programming,” Computational Optimization
225 C.A. Floudas: Deterministic Global Optimization: Theory,                and Applications 18 (2001) 295–309.
    Methods and Applications, Kluwer Academic Publishers, Dor-          247 A.M. Geoffrion: “Generalized Benders Decomposition,” Jour-
    drecht, The Netherlands, 2000.                                          nal of Optimization Theory and Applications 10 (1972) no. 4,
226 M. Tawarmalani, N. Sahinidis: Convexification and Global                237–260.
    Optimization in Continuous and Mixed-Integer Nonlinear              248 M.A. Duran, I.E. Grossmann: “An Outer-Approximation Algo-
    Programming: Theory, Algorithms, Software, and Applica-                 rithm for a Class of Mixed-integer Nonlinear Programs,” Math
    tions, Kluwer Academic Publishers, Dordrecht 2002.                      Programming 36 (1986) 307.
227 R. Horst, H. Tuy: Global optimization: Deterministic                249 X. Yuan, S. Zhang, L. Piboleau, S. Domenech : “Une Methode
    approaches. Springer-Verlag, Berlin 1993.                               d’optimisation Nonlineare en Variables Mixtes pour la Con-
228 A. Neumaier, O. Shcherbina, W. Huyer, T. Vinko: “A Compar-              ception de Procedes,” RAIRO 22 (1988) 331.
    ison of Complete Global Optimization Solvers,” Math. Pro-           250 R. Fletcher, S. Leyffer: “Solving Mixed Integer Nonlinear
    gramming B 103 (2005) 335–356.                                          Programs by Outer Approximation,” Math Programming 66
229 L. Lovasz A. Schrijver: “Cones of Matrices and Set-functions           (1974) 327.
    and 0-1 Optimization,” SIAM J. Opt. 1 (1991) 166–190.               251 I. Quesada, I.E. Grossmann: “An LP/NLP Based Branch and
230 H.D. Sherali, W.P. Adams: “A Hierarchy of Relaxations                   Bound Algorithm for Convex MINLP Optimization Prob-
    Between the Continuous and Convex Hull Representations                  lems,” Comp. Chem. Eng. 16 (1992) 937–947.
    for Zero-One Programming Problems,” SIAM J. Discrete                252 T. Westerlund, F. Pettersson: “A Cutting Plane Method for
    Math. 3 (1990) no. 3, 411–430.                                          Solving Convex MINLP Problems,” Comp. Chem. Eng. 19
231 E. Balas, S. Ceria, G. Cornuejols: “A Lift-and-Project Cutting          (1995) S131–S136.
    Plane Algorithm for Mixed 0-1 Programs,” Mathematical               253 O.E. Flippo, A.H.G.R. Kan: “Decomposition in General Math-
    Programming 58 (1993) 295–324.                                          ematical Programming,” Mathematical Programming 60
232 A.H. Land, A.G. Doig: “An Automatic Method for Solving                  (1993) 361–382.
    Discrete Programming Problems,” Econometrica 28 (1960)              254 T.L. Magnanti, R.T. Wong: “Acclerated Benders Decomposi-
    497–520.                                                                tion: Algorithm Enhancement and Model Selection Criteria,”
233 E. Balas: “An Additive Algorithm for Solving Linear Programs            Operations Research 29 (1981) 464–484.
    with Zero-One Variables,” Operations Research 13 (1965)             255 N.V. Sahinidis, I.E. Grossmann: “Convergence Properties of
    517–546.                                                                Generalized Benders Decomposition,” Comp. Chem. Eng. 15
234 R.J. Dakin: “A Tree Search Algorithm for Mixed-Integer Pro-             (1991) 481.
    gramming Problems”, Computer Journal 8 (1965) 250–255.              256 J.E. Kelley Jr.: “The Cutting-Plane Method for Solving Convex
235 R.E. Gomory: “Outline of an Algorithm for Integer Solutions             Programs,” J. SIAM 8 (1960) 703–712.
    to Linear Programs,” Bulletin of the American Mathematics           257 S. Leyffer: “Deterministic Methods for Mixed-Integer Non-
    Society 64 (1958) 275–278.                                              linear Programming,” Ph.D. thesis, Department of Mathemat-
236 H.P. Crowder, E.L. Johnson, M.W. Padberg: “Solving Large-               ics and Computer Science, University of Dundee, Dundee
    Scale Zero-One Linear Programming Problems,” Operations                 1993.
    Research 31 (1983) 803–834.                                         258 M.S. Bazaraa, H.D. Sherali, C.M. Shetty: Nonlinear Program-
237 T.J. Van Roy, L.A. Wolsey: “Valid Inequalities for Mixed 0-1            ming, John Wiley & Sons, Inc., New York 1994.
    Programs,” Discrete Applied Mathematics 14 (1986)                   259 G.R. Kocis, I.E. Grossmann: “Relaxation Strategy for the
    199–213.                                                                Structural Optimization of Process Flowsheets,” Ind. Eng.
238 E.L. Johnson, G.L. Nemhauser, M.W.P. Savelsbergh: “Progress             Chem. Res. 26 (1987) 1869.
    in Linear Programming Based Branch-and-Bound Algorithms:            260 I.E. Grossmann: “Mixed-Integer Optimization Techniques for
    Exposition,” INFORMS Journal of Computing 12 (2000) 2–23.               Algorithmic Process Synthesis”, Advances in Chemical
160           Mathematics in Chemical Engineering
      Engineering, vol. 23, Process Synthesis, Academic Press,       282 ILOG, Gentilly Cedex, France 1999, www.ilog.com,
      London 1996, pp. 171–246.                                      283 M. Dincbas, P. Van Hentenryck, H. Simonis, A. Aggoun, T.
261   E.M.B. Smith, C.C. Pantelides: “A Symbolic Reformulation/          Graf, F. Berthier: The Constraint Logic Programming Lan-
      Spatial Branch and Bound Algorithm for the Global Optimi-          guage CHIP, FGCS-88: Proceedings of International Con-
      zation of Nonconvex MINLPs,” Comp. Chem. Eng. 23 (1999)            ference on Fifth Generation Computer Systems, Tokyo,
      457–478.                                                           693–702.
262   P. Kesavan P.P.I. Barton: “Generalized Branch-and-cut Frame-   284 M. Wallace, S. Novello, J. Schimpf: “ECLiPSe: a Platform for
      work for Mixed-integer Nonlinear Optimization Problems,”           Constraint Logic Programming,” ICL Systems Journal 12
      Comp. Chem. Eng. 24 (2000) 1361–1366.                              (1997) no.1, 159–200.
263   S. Lee I.E. Grossmann: “A Global Optimization Algorithm for    285 V. Pontryagin, V. Boltyanskii, R. Gamkrelidge, E. Mishchenko:
      Nonconvex Generalized Disjunctive Programming and Appli-           The Mathematical Theory of Optimal Processes, Interscience
      cations to Process Systems,” Comp. Chem. Eng. 25 (2001)            Publishers Inc., New York, NY, 1962.
      1675–1697.                                                     286 A.E. Bryson, Y.C. Ho: “Applied Optimal Control: Optimiza-
264   R. P€orn, T. Westerlund: “A Cutting Plane Method for Mini-         tion, Estimation, and Control,” Ginn and Company, Waltham,
      mizing Pseudo-convex Functions in the Mixed-integer Case,”         MA, 1969.
      Comp. Chem. Eng. 24 (2000) 2655–2665.                          287 V. Vassiliadis, PhD Thesis, Imperial College, University of
265   J. Viswanathan, I.E. Grossmann: “A Combined Penalty Func-          London 1993.
      tion and Outer-Approximation Method for MINLP Opti-            288 W.F. Feehery, P.I. Barton: “Dynamic Optimization with State
      mization,” Comp. Chem. Eng. 14 (1990) 769.                         Variable Path Constraints,” Comp. Chem. Eng., 22 (1998)
266   A. Brooke, D. Kendrick, A. Meeraus, R. Raman: GAMS–A               1241–1256.
      User’s Guide, www.gams.com, 1998.                              289 L. Hasdorff:. Gradient Optimization and Nonlinear Control,
267   N.V.A. Sahinidis, A. Baron: “A General Purpose Global Opti-        Wiley-Interscience, New York, NY, 1976.
      mization Software Package,” Journal of Global Optimization 8   290 R.W.H. Sargent, G.R. Sullivan: “Development of Feed
      (1996) no.2, 201–205.                                              Changeover Policies for Refinery Distillation Units,” Ind.
268   C.A. Schweiger, C.A. Floudas: “Process Synthesis, Design and       Eng. Chem. Process Des. Dev. 18 (1979) 113–124.
      Control: A Mixed Integer Optimal Control Framework,” Pro-      291 V. Vassiliadis, R.W.H. Sargent, C. Pantelides: “Solution of a
      ceedings of DYCOPS-5 on Dynamics and Control of Process            Class of Multistage Dynamic Optimization Problems,” I & EC
      Systems, Corfu, Greece 1998, pp. 189–194.                          Research 33 (1994) 2123.
269   R. Raman, I.E. Grossmann: “Modelling and Computational         292 K.F. Bloss, L.T. Biegler, W.E. Schiesser: “Dynamic Process
      Techniques for Logic Based Integer Programming,” Comp.             Optimization through Adjoint Formulations and Constraint
      Chem. Eng. 18 (1994) no.7, 563.                                    Aggregation,”. Ind. Eng. Chem. Res. 38 (1999) 421–432.
270   J.N. Hooker, M.A. Osorio: “Mixed Logical.Linear Pro-           293 H.G. Bock, K.J. Plitt: A Multiple Shooting Algorithm for Direct
      gramming,” Discrete Applied Mathematics 96-97 (1994)               Solution of Optimal Control Problems, 9th IFAC World Con-
      395–442.                                                           gress, Budapest 1984.
271   P.V. Hentenryck: Constraint Satisfaction in Logic Program-     294 D.B. Leineweber, H.G. Bock, J.P. Schl€oder, J.V. Gallitzend€or-
      ming, MIT Press, Cambridge, MA, 1989.                              fer, A. Sch€afer, P. Jansohn: “A Boundary Value Problem
272   J.N. Hooker: Logic-Based Methods for Optimization: Combin-         Approach to the Optimization of Chemical Processes
      ing Optimization and Constraint Satisfaction, John Wiley &         Described by DAE Models,” Computers & Chemical Engi-
      Sons, New York 2000.                                               neering, April 1997. (IWR-Preprint 97-14, Universit€at Heidel-
273   J.N. Hooker: Logic-Based Methods for Optimization, John            berg, March 1997.
      Wiley & Sons, New York 1999.                                   295 J.E. Cuthrell, L.T. Biegler: “On the Optimization of Differen-
274   E. Balas: “Disjunctive Programming,” Annals of Discrete            tial- algebraic Process Systems,” AIChE Journal 33 (1987)
      Mathematics 5 (1979) 3–51.                                         1257–1270.
275   H.P. Williams: Mathematical Building in Mathematical Pro-      296 A. Cervantes, L.T. Biegler: “Large-scale DAE Optimization
      gramming, John Wiley, Chichester 1985.                             Using Simultaneous Nonlinear Programming Formulations.
276   R. Raman, I.E. Grossmann: “Relation between MILP Model-            AIChE Journal 44 (1998) 1038.
      ling and Logical Inference for Chemical Process Synthesis,”    297 P. Tanartkit, L.T. Biegler: “Stable Decomposition for Dynamic
      Comp. Chem. Eng. 15 (1991) no. 2, 73.                              Optimization,” Ind. Eng. Chem. Res. 34 (1995) 1253–1266.
277   S. Lee, I.E. Grossmann: “New Algorithms for Nonlinear          298 S. Kameswaran, L.T. Biegler: “Simultaneous Dynamic Opti-
      Generalized Disjunctive Programming,” Comp. Chem. Eng.             mization Strategies: Recent Advances and Challenges,” Chem-
      24 (2000) no. 9-10, 2125–2141.                                     ical Process Control–7, to appear 2006.
278   J. Hiriart-Urruty, C. Lemarechal: Convex Analysis and Mini-   299 B. Bloemen Waanders, R. Bartlett, K. Long, P. Boggs, A.
      mization Algorithms, Springer-Verlag, Berlin 1993.                 Salinger: “Large Scale Non-Linear Programming for PDE
279   E. Balas: “Disjunctive Programming and a Hierarchy of              Constrained Optimization,” Sandia Technical Report
      Relaxations for Discrete Optimization Problems,” SIAM J.           SAND2002-3198, October 2002.
      Alg. Disc. Meth. 6 (1985) 466–486.                             300 I. Das, J.E. Dennis: “A closer look at drawbacks of minimizing
280   I.E. Grossmann, S. Lee: “Generalized Disjunctive Program-          weighted sums of objectives for pareto set generation in multi-
      ming: Nonlinear Convex Hull Relaxation and Algorithms,”            criteria optimization problems”, Structural optimization 14
      Computational Optimization and Applications 26 (2003)              (1997) no. 1, 63–69.
      83–100.                                                        301 B. Schandl, K. Klamroth, M.M. Wiecek: “Norm-based approx-
281   S. Lee, I.E. Grossmann: “Logic-based Modeling and                  imation in bicriteria programming”, Computational Optimiza-
      Solution of Nonlinear Discrete/Continuous Optimization             tion and Applications 20 (2001) no. 1, 23–42.
      Problems,” Annals of Operations Research 139 2005              302 M. Bortz, J. Burger, N. Asprion, S. Blagov, R. B€ottcher, U.
      267–288.                                                           Nowak, A. Scheithauer, R. Welke, H. Hasse, K.-H. K€ufer:
                                                                        Mathematics in Chemical Engineering                             161
      “Multi-criteria optimization in chemical process design and          321 A. Shapiro, D. Dentcheva, A. Ruszczy’nski: Lectures on
      decision support by navigation on pareto sets”, Comp. Chem.              Stochastic Programming–Modeling and Theory. 2nd ed.,
      Eng. 60 (2014) 354–363.                                                  MOS-SIAM, Philadelphia 2014.
303   C. Hillermeier: “Nonlinear Multiobjective Optimization: A            322 N.V. Sahinidis: “Optimization under uncertainty: state-of-the-
      Generalized Homotopy Approach”, International Series of                  art and opportunities”, Comp. & Chem. Eng. 28 (2004)
      Numerical Mathematics, Springer, New York 2001.                          971–983.
304   Saltelli, A., Ratto, M., Andres, T., Campolongo, F., Cariboni,       323 R. Henrion, P. Li, A. M€oller, M.C. Steinbach, M. Wendt, G.
      J., Gatelli, D. Saisana, M., Tarantola, S.: Global Sensitivity           Wozny: “Stochastic Optimization for Operating Chemical
      Analysis - The Primer. Wiley, New York 2008.                             Processes under Uncertainty”, Online Optimization of Large
305   A. Saltelli, S. Tarantola, F. Campolongo, M. Ratto: Sensitivity          Scale Systems, Springer-Verlag, Berlin Heidelberg 2001,
      Analysis in Practice: A Guide to Assessing Scientific Models,            pp. 457–478.
      Wiley, New York 2004.                                                324 AIMMS Stochastic Programming Software: http://www
306   SimLab, http://ipsc.jrc.ec.europa.eu/?id=756 (accessed 1st               .aimms.com/operations-research/mathematical-programming/
      September 2015).                                                         stochastic-programming/ (accessed 1st September 2015).
307   Sensitivity Analysis Excel Add-In: http://www.life-cycle-            325 SLP-IOR:        http://www.business.uzh.ch/professorships/qba/
      costing.de/sensitivity_analysis/ (accessed 1st September 2015).          publications/stochOpt.html (accessed 1st September 2015).
308   MUCM Project: http://mucm.ac.uk/index.html (accessed 1st             326 Stochastic Modeling Interface: http://www.coin-or.org/projects/
      September 2015).                                                         Smi.xml (accessed 1st September 2015).
309   SALib: http://jdherman.github.io/SALib/ (accessed 1st Sep-           327 NEOS solver for stochastic linear programming: http://www
      tember 2015).                                                            .neos-server.org/neos/solvers/index.html (accessed 1st Sep-
310   R. Hettich, K.O. Kortanek: “Semi-infinite programming–                   tember 2015).
      Theory, methods, and applications”, SIAM Rev. 35 (1993)              328 AMPLDev SP Edition: http://www.optiriskindia.com/products/
      380–429.                                                                 ampl-dev-sp-edition-fort-sp (accessed 1st September 2015).
311   F. Guerra Vazquez, J.-J. R€uckmann, O. Stein, G. Still:             329 S. Zenios, A. Consiglio, S.S. Nielsen: “Mean-variance portfo-
      “Generalized semi-infinite programming–A tutorial”, J.                   lio optimization”, Practical Financial Optimization–A Library
      Comp. Appl. Math. 217 (2008) no. 2, 394–419.                             of GAMS Models, Wiley, New York 2009.
312   D. Bertsimas, D.B. Brown, C. Caramanis: “Theory and appli-           330 M. Suh, T. Lee: “Robust optimization method for the economic
      cations of robust optimization”, SIAM Rev. 53 (2011) no. 3,              term in chemical process design and planning”, Ind. Eng.
      464–501.                                                                 Chem. Res. 40 (2001) no. 25, 5950–5959.
313   A. Ben-Tal, L. El Ghaoui, A. Nemirovski: Robust Optimiza-            331 G.P. Box, J.S. Hunter, W.G. Hunter: Statistics for Experiment-
      tion. Princeton University Press, 2009.                                  ers: Design, Innovation, and Discovery, 2nd ed., John Wiley &
314   A. Ben-Tal, A. Nemirovski: “Robust optimization–method-                  Sons, New York 2005.
      ology and applications”, Math. Prog. 92 (2002) no. 3,                332 D.C. Baird: Experimentation: An Introduction to Measurement
      453–480.                                                                 Theory and Experiment Design, 3rd ed., Prentice Hall, Engel-
315   AIMMS Robust Programming Software: http://www.aimms.                     wood Cliffs, NJ, 1995.
      com/operations-research/mathematical-programming/robust-             333 J.B. Cropley: “Heuristic Approach to Comples Kinetics,”
      optimization/ (accessed 1st September 2015).                             ACS Symp. Ser. 65 (1978) 292–302.
316   Robust Optimization with YALMIP: http://users.isy.liu.se/            334 S. Lipschutz, J.J. Schiller, Jr: Schaum’s Outline of Theory and
      johanl/yalmip/pmwiki.php?n=Tutorials.RobustOptimization                  Problems of Introduction to Probability and Statistics,
      (accessed 1st September 2015).                                           McGraw-Hill, New York 1988.
317   ROME - Robust Optimization Made Easy: http://www.robustopt           335 D.S. Moore, G.P. McCabe: Introduction to the Practice of
      .com/ (accessed 1st September 2015).                                     Statistics,” 4th ed., Freeman, New York 2003.
318   J.R. Birge, F. Louveaux: Introduction to Stochastic Program-         336 D.C. Montgomery, G.C. Runger: Applied Statistics and Prob-
      ming. 2nd ed., Springer, Heidelberg 2011.                                ability for Engineers, 3 rd ed., John Wiley & Sons, New York
319   P. Kall, S.W. Wallace: Stochastic Programming–Numerical                  2002.
      Techniques and Engineering Applications, Springer, Heidel-           337 D.C. Montgomery, G.C. Runger, N.F. Hubele: Engineering
      berg 1994.                                                               Statistics, 3rd ed., John Wiley & Sons, New York 2004.
320   A. Ruszczy’n ski, A. Shapiro (eds.): “Stochastic Pro-                338 G.E.P. Box, W.G. Hunter, J.S. Hunter: Statistics for Experi-
      gramming”, Handbook in Operations Research and Manage-                   menters, John Wiley & Sons, New York 1978.
      ment Science, Vol. 10, Elsevier, Amsterdam 2003.