Professional Documents
Culture Documents
A. J. Roberts1
November 26, 2009
ii
Contents
Preface
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
109
. 110
. 113
. 118
. 125
. 128
iv
Contents
3.7
3.8
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
257
. 259
. 284
. 312
. 340
.
.
.
.
343
. 344
. 364
. 368
. 370
the centre
. . . . . . .
. . . . . . .
. . . . . . .
. . . . . . .
. . . . . . .
manifold371
. . . . . . 374
. . . . . . 384
. . . . . . 402
. . . . . . 417
. . . . . . 433
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
435
437
440
457
479
Preface
Here we explore how to derive relatively simple dynamical equations that
model complex physical interactions. The book arises out of my interests in
both fluid dynamics and modern dynamical systems theory. Thus fluid flows
form many of the applications we investigate. The triple aim of the book is
to explore: algebraic techniques; some applications; and general modelling
principles.
The basis for the methodology is the theory and geometric picture of invariant manifolds in dynamical systems: in particular, we heavily use centre and
slow manifolds. The wonderful aspect of this approach are the geometric
interpretations of the modelling process. Simple geometric pictures inspire
sound methods of analysis and construction. But further, the pictures that
we draw of state space also provide a route for better assessing limitations
and strengths in a model. Geometry and algebra form a powerful partnership.
. . . duality between algebra and geometry was discovered by
Rene Descartes: every geometric object has an algebraic description, every algebraic formula determines a geometric object. Humans tend to use the algebraic version for calculation, and the
geometric one for imagination.
Fearful symmetry, Stewart & Golubitsky.
vi
Preface
Chapter 1
1.2
1.3
1.1.1
1.1.2
1.1.3
1.1.4
34
1.2.1
. . . . . . . . . . . . . . . 35
1.2.2
1.2.3
1.2.4
Exercises . . . . . . . . . . . . . . . . . . . . . . . 65
68
1.4
1.3.2
1.3.3
1.3.4
1.3.5
1.3.6
Chapter Summary . . . . . . . . . . . . . . . . . .
107
1.1
Warning:
The irony in the above quote is that this is exactly what we do: we supply
the computer with a wrong answer, albeit roughly correct; and then we
iterate to improve the answer until it becomes right!
We adopt iteration throughout this book.
1.1.1
(1.1)
Suppose we did not know the famous formula for solving such a nonlinear
equation,
p
x = 21 0.1 0.12 + 4 = 0.95125 , 1.05125 .
Forget that you know how to solve quadratics like this. Instead we could
and do now argue that the quadratic (1.1) is nearly the same as the simpler
quadratic
x2 1 = 0 .
(1.2)
The simpler quadratic (1.2) has easier solution x = 1 which for the purposes of argument we suppose we know. In a manner to become familiar,
we use the simpler problem (1.2), and its solution, to perturbatively solve
the more complicated problem (1.1).
Connect the simpler quadratic (1.2) to the original quadratic (1.1) as instances of the more general quadratic problem
x2 + x 1 = 0 .
(1.3)
We introduce the parameter so that the general quadratic (1.3) encompasses both the simpler quadratic (1.2), the case = 0 , and the original
Tony Roberts, 6 Mar 2009
quadratic (1.1), the case = 0.1 . Now we proceed to expand the solution
in powers of small parameter : this expansion empowers us to use the
simpler quadratic (1.2) as a base to solve the original quadratic (1.1).
Our aim is to use simple iteration to recover the Taylor series
x = 1 12 + 81 2
1 4
128
(1.4)
x = 1 x .
(1.5)
Why? Because this form mimics our known solution method for solving the
simple quadratic (1.2). However, an unfortunate property of this straightforward approach is that the argument is not generalisable. We have to
have some inspiration in each new problemand inspiration is often hard
to obtain.
Nonetheless, let us continue and furthermore, let us just concentrate on the
positive solution, the variable x should be near 1. Rewrite the form (1.5) as
the iteration scheme
p
(1.6)
xj+1 = 1 xj ,
where the choice of the plus alternative for the square root leads us to the
positive solution x. Every iteration needs a starting value which will be the
solution of the easy quadratic (1.2), namely x0 = +1 .
The first iteration is
x1 = 1 = 1 21 81 2
1 3
16
=
1 1
= 1 12 1 18 2 (1 )
x2 =
1 3
16 (1
)3/2 +
= 1 12 + 18 2 + 03 + ,
after invoking the binomial series repeatedly.
And so on to further approximations with higher powers in the parameter .
See the first two iterations give precisely the first terms in the Taylor series (1.4) of the exact solution to the general quadratic (1.3): we use the
base case of = 0 with iteration to obtain solutions for general parameter .
In particular, evaluate this series at = 0.1, namely x = 0.95125 , to obtain
to five digits the positive root of the original quadratic (1.1).
(1.7)
The residual is a function of variable x (and the parameter ). Our aim is
to find the variable x, as a function of parameter , for which the residual
is zero and hence to have solved the quadratic (1.3).
We start from the simple case of = 0 for which we know the positive
solution x = 1 . Correspondingly, the residual Res1.3 (1) = 12 + 0 1 1 = 0 .
Thus set the initial approximation x0 = 1 .
Tony Roberts, 6 Mar 2009
We continue with further iteration in the next subsubsection aided by computer algebra. The key is that the corrections are simply guided by the
residual of the quadratic equation (1.3) which we aim to solve.
Computer algebra makes iteration easy
Reexamine solving the quadratic (1.3) by iteration.
Suppose that at the jth iterate we have an approximation xj , expressed as
a truncated power series in the parameter . Seek a small correction x
^j so
that xj+1 = xj + x
^j is a better approximation. How can it be better? By
reducing the residual (1.7) of the quadratic equation (1.3). So substitute
x = xj + x
^j into the quadratic (1.3) to see
(xj + x
^j )2 + (xj + x
^1 ) 1 = 0
x2j + xj 1 + (2xj + )^
xj + x
^2j = 0 .
That is, we seek to use the equation
Resj +(2xj + )^
xj + x
^2j = 0
(1.8)
Algorithm 1.1 reduce code to solve the quadratic (1.3) by iteration using
successive corrections (1.9).
x:=1;
% set initial approx to root
let eps^10=>0;
% truncate power series
repeat begin
% repeat iteration
res:=x^2+eps*x-1; % current residual of quadratic
write x:=x-res/2; % correct approx using residual
end until (res=0);
% until residual effectively 0
we cannot expect to solve (1.8) exactly as that would be tantamount
to solving the original quadratic (1.3) exactly;
instead, omit all products of corrections such as x
^2j ; and
approximate all coefficients of linear terms in the correction x
^j by the
unperturbed value, here setting xj 1 and 0 so that (2xj + )^
xj
2^
xj .
Then the equation (1.8) is approximately
Resj +2^
xj = 0
correction x
^j = 12 Resj .
(1.9)
Recognise that this is effectively what we did by hand in the previous subsubsection. Now proceed to iterate with the power and reliability of computer
algebra.
The whole of the developments and operations of analysis are
now capable of being executed by machinery. . . . As soon as
an Analytical Engine exists, it will necessarily guide the future
course of science.
Charles Babbage, 1864
Look at the computer algebra code of Algorithm 1.1. To execute this code,
I recommend you type it into a text file for reduce to execute. That
way you will be able to easily repair errors in typing or conception. Type
Algorithm 1.1 into a text file, say named quad. Then most importantly
surround the code by the two lines
Tony Roberts, 6 Mar 2009
1
4
- -----*eps
128
1
4
1
6
- -----*eps + ------*eps
128
1024
1
4
1
6
- -----*eps + ------*eps
128
1024
1
4
1
6
- -----*eps + ------*eps
128
1024
11
You easily generate high orders Computing higher order terms in the
power series is, until you run out of time or memory, simply a matter of
increasing the order of truncation of the power series and rerunning the
code.
It is flexible Any problem in a wide class may be solved simply by changing the computation of the residual. For example, if you wanted to solve
the cubic x3 + x2 + x 1 = 0 , then just change the computation of the
residual accordingly. Simply changing the computation of the residual will
work provided the initial approximation is unchanged and the rule for using
the residual to give a correction remains unchanged.
Tony Roberts, 6 Mar 2009
13
Exercise 1.1:
Change the computation of the residual in the above computer algebra code to find power series solutions to the following algebraic equations for the root x 1 :
1. x3 + x2 + x 1 = 0 ;
2. (1 )x2 + 2x 1 = 0 ;
3. x3 12 x2 + x
1
2
= 0.
For each of these algebraic equations, estimate the root for = 0.1 ,
1/2 and 1. Discuss how accurate, or even valid, you consider your
estimates.
Answer:
1. 1 + 32 2 33 +
55 4
8
2. 1 12 18 2 +
3. 1 12
1 2
16
1 3
16
1 3
64
175 +
11 4
128
31
4
1024
7
5
256
91
5
4096
Exercise 1.2:
Change the computation of the residual and the rule for
updating corrections in the above computer algebra code to find power
series solutions to the following algebraic equations for the root x 1 :
1. x3 + x2 + x 2 = 0 ;
2. (1 )x2 + (1 + 2)x 2 = 0 ;
3. x4 x2 x 1 = 0 .
Use one of your series to estimate a root of x3 + x2 + x 2 = 0 ; discuss
the accuracy of your estimate. Hint: for each of the above, use a for
loop instead of a repeat loop until you know your code will converge.
Answer:
Tony Roberts, 6 Mar 2009
1
2
125
2. 1 13
1 2
27
7
3
243
3. 1 + 12
1 3
16
1 4
32
8
3
3125
1
4
15625
31
4
2187
1
5
128
154
5
1953125
41
5
19683
5
6
256
598
6
48828125
251
6
59049
Exercise 1.3:
In complicated problems we often
series ex need power
pansions for elementary functions such as 1 + , 1/ 1 + 2 and
1/(1 + )2 . Of course we could explicitly code these elementary functions using the binomial expansion. But it is just as easy to use the
same iteration to create the power series expansions of these functions
simultaneously with the iteration to solve the complicated problem.
To do this, just recast such elementary functions as the solution of an
algebraic equation and apply the techniques we have just explored.
Answer:
1. 1 + 12 18 2 +
1 3
16
2. 1 12 2 + 38 4
5 6
16
5
4
128
7
5
256
21
6
1024
1.1.2
15
(1.10)
As for all quadratics, this has two solutions. But when the parameter = 0 ,
the equation reduces to x1 = 0 and only one solution is apparent. This is an
example of a singular perturbation: the basic case = 0 differs qualitatively
from the cases 6= 0 . Interesting problems are often singular.
Problems which are not singular are termed regular .
Here, resolve the apparent contradiction between = 0 and 6= 0 by finding
the exact roots of the quadratic (1.10) and
exploring their behaviour as
0 . The two roots are, from x = (1 1 + 4)/(2) ,
x(1) =
1 + 22 53 + ,
1
x(2) = 1 + 22 + 53 .
The root x(1) is near 1 and is well behaved. The other root x(2) disappears
out of sight to as 0 ; hence x(2) is not apparent when we simply set
= 0.
Iteration finds the well behaved root
As before, use iteration to find the root x(1) .
Suppose you know an approximation xj to this root, such as the initial
approximation x0 = 1 . Seek a small correction x
^j so that xj+1 = xj + x
^j
better satisfies the singular quadratic (1.10). Substitute x = xj+1 = xj + x
^j
into the quadratic (1.10) to find
(xj + x
^j )2 + (xj + x
^j ) 1 = 0
x2j + 2xj x
^j + ^
x2j + xj + x
^j 1 = 0
Tony Roberts, 6 Mar 2009
(1.11)
where Resj = x2j + xj 1 is the residual of the quadratic (1.10) for the
currently known approximation. As before, neglect products of small corrections such as x
^2j , and replace coefficients of terms linear in x
^j by their
0 limit, here (2xj + 1)^
xj +^
xj . Thus approximate equation (1.11) for
the small corrections by
Resj +^
xj = 0
correction x
^j = Resj .
(1.12)
correction x
^j = + Resj .
(1.13)
17
Thus two iterations finds the first few terms in the power series of the singular root of the singular quadratic (1.10).
Deduce the initial approximation But how do we know to start with
x0 = 1/ ? This x0 follows because nontrivial solutions must come from a
balance of at least two dominant terms in the governing equation. Consider
the possibilities exhaustively.
Suppose the last two terms of the quadratic (1.10) are the dominant
balance. That is, x 1 = 0 is the approximate equation. Then the
initial approximation would be x 1 . The neglected term x2
is then smaller than the two terms forming the balance, namely the
x and 1, and so the initial approximation is valid. We saw this lead us
to the well behaved root.
Suppose the first two terms of the quadratic (1.10) form the dominant
balance. That is, x2 + x = 0 is the approximate equation. Then the
initial approximation would be solutions of x2 + x = 0 , namely the
equation x(x + 1) = 0 giving solutions x 0 and x 1/ .
The x 0 case is not relevant because then the neglected term 1
is larger than the retained terms in the quadratic (1.10).
However, for the x 1/ case, the neglected term 1 is smaller
than the two terms forming the balance, both of large size 1/,
and so the initial approximation is valid. We saw this case lead
us to find the singular root.
Tony Roberts, 6 Mar 2009
X2 X
+ 1=0
2
X2 + X = 0 .
(1.14)
Now solve this corresponding quadratic (1.14) for roots X using the regular
methods. This works because the rescaling of the singular quadratic (1.10)
produces a regular quadratic (1.14).
For example, use computer algebra. The reduce code in Algorithm 1.2
finds the root
X = 1 + 2 23 + 54 +
which corresponds to the singular root x(2) = X/ . See the output from
the reduce code in Dialogue 1.2.
Remember to surround the code of
Algorithm 1.2 by
Tony Roberts, 6 Mar 2009
19
Algorithm 1.2 reduce code to solve the rescaled quadratic (1.14) by iteration. Here find the root X 1 , that is, x 1/.
x:=-1;
let eps^7=>0;
repeat begin
res:=x^2+x-eps;
write x:=x+res;
end until (res=0);
- 1 - eps
2
- 1 - eps + eps
2
- 1 - eps + eps
2
- 1 - eps + eps
2
- 1 - eps + eps
2
- 1 - eps + eps
2
- 1 - eps + eps
3
- 2*eps
3
- 2*eps
3
- 2*eps
3
- 2*eps
3
- 2*eps
4
+ eps
4
+ 5*eps
4
+ 5*eps
4
+ 5*eps
4
+ 5*eps
5
6
- 6*eps + 6*eps
5
6
- 14*eps + 26*eps
5
6
- 14*eps + 42*eps
5
6
- 14*eps + 42*eps
Find the other root x(1) 1 using the same program, but start from X0 = 0
(or X0 = ) corresponding to x0 = 1 , and change the update to X from
the correction. See these two changes from Algorithm 1.2 in the reduce
code of Algorithm 1.3. Dialogue 1.3 list the output from the reduce code.
Hence the other root x(1) = X/ = 1 + 22 + as before.
Avoid singular perturbations Later analysis of dynamical systems discusses how some people view the dynamics as a combination of very fast
Tony Roberts, 6 Mar 2009
21
Algorithm 1.3 reduce code to solve the rescaled quadratic (1.14) by iteration. Here find the root X 0 , that is, x 1.
x:=0;
let eps^7=>0;
repeat begin
res:=x^2+x-eps;
write x:=x-res;
end until (res=0);
3
+ 2*eps
3
+ 2*eps
3
+ 2*eps
3
+ 2*eps
3
+ 2*eps
4
- eps
4
- 5*eps
4
- 5*eps
4
- 5*eps
4
- 5*eps
5
6
+ 6*eps - 6*eps
5
6
+ 14*eps - 26*eps
5
6
+ 14*eps - 42*eps
5
6
+ 14*eps - 42*eps
Exercise 1.7:
Rescale x to find approximations, as power series in parameter , to the two real roots of 2 x6 x4 x3 + 8 = 0 . Exhaust
all scaling possibilities.
Answer: One real root comes from assuming x is of size 1, another real
root comes from the scaling x = X/2/3 (Bender & Orszag 1981, 7.2).
1.1.3
How many of the previous computer algebra exercises did you do? If you
did more than a couple, I bet that you ran an infinite loop in at least one
of the exercises.
Why? Because it is non-trivial to deduce the rule, such as (1.9) or (1.12), for
correcting an approximation based upon a residual. The algebra derivation
contains enough detail apparently uninteresting detail that people
make little errors. At least I generally do.
Tony Roberts, 6 Mar 2009
23
Algorithm 1.4 precisely two iterations of this reduce code tells us how to
use the residuals to update the corrections. Simply choose the parameter a
so that the second residual is better than the first.
factor eps;
x:=1;
let eps^10=>0;
for it:=1:2 do begin
write res:=x^2+eps*x-1;
write x:=x+a*res;
end;
We here explore an simpler but reliable alternative: namely a variant of
the method of undetermined coefficients. The method is to recognise that
generally there is a purely linear dependence upon the residual by the desired
correction, so let the coefficient of the linear dependence be some as yet
unknown parameter and then find the unknown parameter via performing
two iterations.
Example 1.8:
Reconsider finding the root x 1 of the quadratic equation (1.3) using computer algebra. Suppose we do not know that the
correction to an approximation xj is x
^j = 12 Resj , (1.9).
However, we do know that generally there is a linear dependence between residual and correction. Thus, propose that the coefficient of
proportionality is some constant, say a, that we need to determine.
Then for this problem, execute precisely two iterations of the proposed
computer algebra code, see Algorithm 1.4. Dialogue 1.4 lists the reduce output.
See that Res1 = and that Res2 = (1 + 2a) + a(1 + a)2 . For
general coefficient a, both the first and second residuals are of size .
However, the second residual Res2 will be smaller, namely of size 2 ,
when we choose the parameter a so that the coefficient of the term
in Res2 becomes zero. That is, choose parameter a = 12 so that
1 + 2a = 0 . Then the rule for correcting an approximation from a
residual is precisely (1.9) that we deduced before with algebra.
Tony Roberts, 6 Mar 2009
The next step would then be to edit the proposed Algorithm 1.4 by
setting the parameter a to its useful value of a = 21 , then performing as many iterates as necessary to obtain the desired power series
approximation, as in Algorithm 1.1.
Recognise the enormously efficient use of your time in this method of parametrising the linear dependence of a correction upon a residual. You write the
computer algebra code you would use anyway if you knew the parameter;
execute precisely two iterations to find the useful value; then quickly edit
the code to set the value of the parameter. The only crucial part of the computer algebra code is the computation of the residual: if that computation
is correctly coded and a suitable initial approximation used, then in using
this approach you will construct a correct perturbation power series.
See this method of undetermined coefficients in action again in a second
straightforward example.
Example 1.9:
Reconsider finding the negative root X 1 of the quadratic
equation (1.14) using computer algebra. Again suppose we do not
know the correction for any given residual.
However, we do know that generally the correction is linearly proportional to the residual. Thus, propose that the coefficient of proportionality is some constant, say a, that we need to determine. Then
for this problem, execute precisely two iterations of the proposed comTony Roberts, 6 Mar 2009
25
Algorithm 1.5 precisely two iterations of this reduce code tells us how
to use the residuals to update the corrections for quadratic (1.14). Simply
choose the parameter a so that the second residual is better than the first.
factor eps;
x:=-1;
let eps^10=>0;
for it:=1:2 do begin
write res:=x^2+x-eps;
write x:=x+a*res;
end;
Dialogue 1.5 output of Algorithm 1.4.
x := -1
res :=
x :=
- eps
- 1 - eps*a
2 2
res := eps*( - 1 + a) + eps *a
2
2 3
x := - 1 + eps*( - 2*a + a ) + eps *a
puter algebra code, see Algorithm 1.5. Dialogue 1.5 lists the reduce
output.
See that Res1 = and that Res2 = (a 1) + a2 2 . The second
residual Res2 will be smaller than the first, Res1 , only when we choose
the parameter a so that the coefficient of the term in Res2 is set to
zero. That is, choose parameter a = 1 so that a 1 = 0 . Then the
rule for correcting an approximation from a residual becomes precisely
that implemented before in Algorithm 1.2.
1.1.4
In the previous subsections I assumed you could already use the computer
algebra package reduce. This subsection provides an introduction to reduce for those that need more help getting started.
You might use a free demonstration copy of reduce1 which at the time of
writing are available from:
Codemist http://west.codemist.co.uk/reduce/index.html;
Konrad-Zuse Zentrum Berlin ftp://ftp.zib.de/pub/reduce/demo;
Check that you can start and run reduce, it should open up a window
saying something like2
REDUCE 3.7, 15-Apr-1999 ...
1:
The 1: is a prompt for a command: type quit; followed by the return or
enter key for reduce to finish. If this works, you can run reduce.
1
There is always a limitation in using free, demonstration copies. Here the main restriction on the demonstration version is that garbage collection is disabled in reduce.
What that means in practise is that only small to medium amounts of computer algebra
can be done before having to restart reduce. It is probably best to solve one problem at
a time, restarting reduce in between each problem.
2
We generally use such a coloured, teletype font for computer instructions and dialogue.
Tony Roberts, 6 Mar 2009
27
http://www.zib.de/Symbolik/reduce/Overview/Overview.html is
an on-line overview to the capabilities of reduce.
http://www.uni-koeln.de/cgi-bin/redref/redr_dir.html gives extensive online help to the commands and syntax for reduce.
Explore a little reduce
Start reduce in Unix by typing reduce in a command window. To
exit from reduce type the command quit; followed by the enter
key .
Note: all reduce statements must be terminated with a semi-colon.
Do not forget. They are subsequently executed by typing a enter
key.
reduce uses exact arithmetic by default: for example to find 100! in
full gory detail type factorial(100);enter (I will not mention the
enter key again unless necessary).
Identifiers, usually we use single letters for brevity, denote either variables or expressions: in f:=2*x^2+3*x-5; the identifier x is a variable
whereas f, after the assignment with :=, contains the above expression;
similary after g:=x^2-x-6; then g contains an algebraic expression.
Expressions may be
added with f+g;
subtracted with f-g;
multiplied with f*g;
divided with f/g;
exponentiated with f^3;, etc.
Straightforward equations may be solved (by default equal to zero):
solve(x^2-x-6,x); or through using an expression previously found
such as solve(f,x); .
Tony Roberts, 6 Mar 2009
29
Linear operators distribute We need to tell reduce that the operator iint distributes over addition and multiplication by a constant so
that reduce knows, for example, that iint(2*x+5*x^2) is the same as
2*iint(x)+5*iint(x^2) . The reduce command linear does precisely
that provided we also specify that x is the variable of interest. Replace the
above definitions with
operator iint;
linear iint;
let {iint(1,x)=>x^2/2,
iint(x,x)=>x^3/6,
Tony Roberts, 6 Mar 2009
31
Build in subsidiary conditions In the above example of double integration we ignored any integration constants. Our code above always provides
an answer for which the value and its derivative at x = 0 are both zero.
That is, our above code implicitly assumed the answer has to be such that
its value and derivative at x = 0 has to be zero. What if instead we wanted
an answer that was zero at x = 0 and zero at x = 1, and there is no constraint on the derivative? We build such subsidiary homogeneous conditions
into the operator iint.
A little algebra gives the general double integral
ZZ
1
xn dx dx =
xn+2 + ax + b .
(n + 1)(n + 2)
Make this zero at x = 0 by choosing b = 0 , as we did implicitly above. Make
1
the answer zero at x = 1 by choosing a = (n+1)(n+2)
. That is, the general
Tony Roberts, 6 Mar 2009
RR
xn dx dx =
1
n+2
(n+1)(n+2) (x
x) .
operator iint;
linear iint;
let {iint(1,x)=>(x^2-x)/2,
iint(x,x)=>(x^3-x)/6,
iint(x^~n,x)=>(x^(n+2)-x)/(n+1)/(n+2) };
Then see it works on some examples such as the following:
6: y:=iint((b*x+x^2)^2,x);
2 3
2
4
5
x*(5*b *x - 5*b + 6*b*x - 6*b + 2*x - 2)
y := ---------------------------------------------60
7: sub(x=1,y);
0
In essence we have constructed the operator iint that solves the second
order differential equation d2 y/dx2 = y 00 = f such that y(0) = y(1) = 0 .
We just type y:=iint(f,x), provided the right-hand side f is a polynomial
in x.
If the right-hand side is non-polynomial, then we either code more suitably
general rules or we resort back to the powerful general reduce integration
operator int. But remember that a specifically coded operator,like our iint,
is very much quicker for reduce to execute.
Exercise 1.11:
Code an operator that constructs particular solutions
of the differential equation x2 y 00 + 2y = f for polynomial right-hand
sides f.
Exercise 1.12:
Code an operator that constructs solutions of the differential equation x2 y 00 + xy 0 4y = f for polynomial right-hand sides f.
The subsidiary conditions are that solutions must be well behaved near
x = 0 , whereas y = 0 at x = 1 . Assume that the polynomial f never
Tony Roberts, 6 Mar 2009
33
1.2
Warning:
The techniques introduced in this section are prototypes for all the asymptotic and perturbative analysis of dynamical systems. The techniques build
on from those introduced in Section 1.1 for solving algebraic equations. Discover much more depth in the approximate solution of differential equations
by reading Chapter 3 in the book by Bender & Orszag (1981). Here we
limit attention to introducing those techniques later pertinent in modelling
dynamical systems.
1.2.1
35
Computers are extremely good are repeating the same thing many times
over. We use this aspect to find power series solutions of some simple differential equations, and then some horrible nonlinear differential equations.
The ideas are developed by example.
Example 1.13:
The solution to y 00 + y = 0, y(0) = 1 and y 0 (0) = 0 is
y = cos x. Find the Taylor series solution by iteration first by hand
and secondly using computer algebra.
00
Solution: Rearrange
RR this ode to y = y and then formally integrate twice to y = y dx dx. These integrals on the right-hand side
are indefinite integrals so implicit constants of integration, say a + bx,
should appear on the right-hand side. But we know that the cosine
solution to y 00 + y = 0 has y(0) = 1 and y 0 (0) = 0 so surely we should
set a = 1 and b = 0 to account for these initial conditions. Thus
ZZ
y = 1 y dx dx
(1.15)
where here the integrals are implicitly the definite integral from 0 to x.
This rearrangement incorporates the information of the ode and its
initial conditions.
In this form we readily find its power series solution by iteration: given
an approximation yn (x) we find a new approximation by evaluating3
ZZ
yn+1 = 1 yn dx dx .
First try by hand starting from y0 = 1:
RR
y1 = 1 1 dx dx = 1 21 x2 ;
RR
y2 = 1 1 21 x2 dx dx = 1 12 x2 +
1 4
24 x .
Interestingly, this is Picard iteration that is also used to prove existence of solutions
to odes.
Tony Roberts, 1 Mar 2008
Example 1.14:
Find the general Taylor series solution to y 00 + y = 0
using computer algebra (reduce).
37
Algorithm 1.6 use iteration to find the general solution of the ode in
Example 1.14.
factor a,b;
y:=a+b*x;
for n:=1:4 do write
y:=a+b*x-int(int(y,x),x);
Now let us try something rather hardin fact almost impossible to quantitatively solve except via power series methods. We now use precisely the
same iteration to solve a nonlinear ode!
Example 1.15:
Find the Taylor series solution to the nonlinear ode
00
2
y = 6y , y(0) = 1 and y 0 (0) = 2.
Before solving this as a power series (by my design its exact solution
just happens to be y = 1/(1 + x)2 ), investigate it qualitatively using
techniques developed in courses on systems of first-order differential
equations. Introduce z(x) = y 0 then the equivalent system is
y0 = z ,
z 0 = 6y2 .
Hence the evolution in the phase plane is dictated by the arrows shown
below with the particular trajectory starting from the initial condition
(1, 2) shown in green:
Tony Roberts, 1 Mar 2008
39
0
-0.2
-0.4
z=y'
-0.6
-0.8
-1
-1.2
-1.4
-1.6
-1.8
-2
-0.5
0.5
y2
41
Algorithm 1.7 Use reduce to find the power series solution to the nonlinear odeof Example 1.15 via its integral representation (1.16).
let x^8=>0;
y:=1-2*x;
for n:=1:5 do write
y:=1-2*x+6*int(int(y^2,x),x);
1.2.2
What we have done is rather remarkable. Recall learning about fixed point iteration
as a method of solving linear and nonlinear equations. We have done precisely fixed point
iteration here. The remarkable difference is that previously you have simply found the
one number that satisfies a given equation; here you have found the function, via its
power series, that satisfies the given differential equationa much more difficult task.
Nonetheless the strategy of appropriately rearranging the equation and iterating works.
Tony Roberts, 1 Mar 2008
43
discuss why this strategy works at all, and what extension we need to solve
a very wide range of differential equations.
The iteration works because integration is basically a smoothing operation.
This smoothing by integration tends to reduce errors in a power series. For
example, suppose an error was O x3 , so that it is roughly about 103
when
x = 0.1 say: then integrating it twice will lead to an error O x5 in the
integral which is much smaller in magnitude, roughly 105 when x = 0.1.
Conversely differentiation
derivatives of an error O x3
magnifies errors: two
1
becomes an error O x which, at roughly 10 when x = 0.1, is much larger.
To make errors smaller, equivalently to push them to higher powers in x, we
generally need to integrate. Thus an integral reformulation of an ode is the
basis for a successful iterative solution.
How do we know how many iterations should be performed? One answer is
simple: keep iterating until there is no more change to the solution. One
consequence of the answer though is that we have to keep track of the change
in the approximations. A good way to find the change in an approximation
is to solve for it explicitly. That is, in analogy with the solution of algebraic
equations in Section 1.1.1, we solve for corrections to each approximation.
But first we have to find an equation for the small corrections to an approximate solution at each iteration. This leads us to a powerful iterative
framework, based upon the residual of the ode, which we develop further
and explore by example.
Example 1.16: Legendre functions.
Use iteration to find the general
Taylor series solutions to Legendres equation, written here as
(1 x2 )y 00 2xy 0 + ky = 0
for k = n(n + 1) ,
to an error O x10 for initial conditions y(0) = 1 and y 0 (0) = 0.
Solution:
45
Example 1.17:
Find the Taylor series solution to errors O x8 to the
nonlinear ode y 00 + (1 + x)y 0 6y2 = 0 such that y(0) = 1 and
y 0 (0) = 1.
47
first correction
1 4
x ,
2
49
Algorithm 1.9 reduce code to solve the nonlinear ode in Example 1.17.
Note: it is safest to specify the integrals are from 0 so that we know the
corrections satisfy y
^ (0) = y
^ 0 (0) = 0 .
y:=1-x;
let x^8=>0;
repeat begin
res:=df(y,x,x)+(1+x)*df(y,x)-6*y^2;
res:=-int(int(res,x),x);
write y:=y+res;
end until res=0;
11
3
1
4
- ----*x + ---*x
6
2
3
31
4
- 3*x + ----*x 8
3
25
4
- 3*x + ----*x 6
3
25
4
- 3*x + ----*x 6
3
25
4
- 3*x + ----*x 6
3
25
4
- 3*x + ----*x 6
3
25
4
- 3*x + ----*x 6
121
5
-----*x
40
169
5
-----*x
40
257
5
-----*x
60
257
5
-----*x
60
257
5
-----*x
60
257
5
-----*x
60
199
6
83
7
+ -----*x - ----*x
60
42
83
6
8543
7
+ ----*x - ------*x
16
1680
787
6
789
7
+ -----*x - -----*x
144
140
219
6
9551
7
+ -----*x - ------*x
40
1680
219
6
1433
7
+ -----*x - ------*x
40
252
219
6
1433
7
+ -----*x - ------*x
40
252
51
(1.17)
dx
x
16
x4
.
=
64
Hence y2 = 1 x2 /4 + x4 /64 .
Tony Roberts, 1 Mar 2008
53
1
4
+ ----*x
64
1
4
1
6
+ ----*x - ------*x
64
2304
1
4
1
6
1
8
+ ----*x - ------*x + --------*x
64
2304
147456
1
4
1
6
1
8
+ ----*x - ------*x + --------*x
64
2304
147456
Using residuals to improve approximate solutions, getting computers to do the tedious algebra, may be adapted to a wide variety of problems.
Tony Roberts, 1 Mar 2008
55
1.2.3
All the series we have generated so far are convergent. However, mostly we
generate divergent series. Somehow we must make sense of such divergent
series. In this subsection we explore the concept of asymptotic series and
how they may form useful approximations.
Lets start the exploration with an innocuous looking example that almost
immediately presents new challenges.
Example 1.20: Stieltjes series Find the Taylor series, about x = 0 , of
the solution to the linear Stieltjes ode
x2 y 00 + (1 + 3x)y 0 + y = 0
such that
y(0) = 1 .
(1.18)
57
Algorithm 1.12 use reduce to find the power series solution to the Stieltjes ode (1.18).
let x^9=>0;
y:=1;
repeat begin
res:=x^2*df(y,x,x)+(1+3*x)*df(y,x)+y;
res:=-int(res,x);
y:=y+res;
end until res=0;
write y;
Dialogue 1.13 reduce output of Algorithm 1.12.
4: in stieltjes$
y := 1
2
3
4
5
6
7
8
1 - x + 2*x - 6*x + 24*x - 120*x + 720*x - 5040*x + 40320*x
where Res1.18 is the residual of the ode (1.18) at the current approximation. Now approximate the operator acting on the correction y
^n
to give the equation
y
^ n0 = Res1.18 :
3x 1 for x near zero and so is omitted;
x2 y
^ n00 is of size y
^ n , both are smaller than y
^ n0 , and so both are
omitted.
Use Algorithm 1.12 to find the power series solution to the Stieltjes
ode (1.18). Dialogue 1.13 lists the reduce output, now without trace
printing in the loop. See these are the first nine terms in the Stieltjes
series
X
y=
(1)n n!xn .
(1.19)
n=0
Exercise 1.36 asks you to confirm this infinite sum satisfies the Stieltjes
ode (1.18).
Tony Roberts, 1 Mar 2008
Exercise 1.36
et
dt ,
1 + xt
(1.20)
x(1 + xt)2 et dt
t=0
0
Z
= 1x
(1 + xt)2 et dt .
y(x) =
59
This integration by parts gives the first term in the Stieltjes series (1.19).
Further integration by parts gives more terms in the Stieltjes series:
Z
y(x) = 1 x
(1 + xt)2 et dt
0
Z
2
(1 + xt)3 et dt
= 1 x + 2x
0
Z
2
3
= 1 x + 2x 6x
(1 + xt)4 et dt
0
..
.
= 1 x + 2!x2 3!x3 + + (1)N N!xN + N (x) ,
where
N (x) = (1)
N+1
N+1
(N + 1)!x
(1 + xt)N2 et dt .
See the Stieltjes series appear; but importantly, the series appears with
a definite expression for the error N (x) in the Nth partial sum.
The error is bounded as 1 + xt 1 for x, t 0 , consequently (1 +
xt)N2 1 and hence
Z
N+1
|N (x)| (N + 1)!x
et dt = (N + 1)!xN+1 .
0
Although the series is divergent, this bound on the error of some partial
sum may be small enough for practical purposes. For example, the four
term partial sum 1 x + 2x2 6x3 approximates the Stieltjes integral
to two decimal places when 24|x|4 < 0.01 , that is, for |x| < 0.14 .
Observe that the error N (x) in the partial sums of the Stieltjes series is
bounded by the magnitude of the next term in the series. Experience suggests that errors of partial sums are often about the size of the first, lowest
order, neglected term in the series. Occasionally we invoke the following
conjecture.
Tony Roberts, 1 Mar 2008
61
means
lim
f(x)
= 0;
g(x)
lim
f(x)
= 1;
g(x)
xx0
means
xx0
means
f(x)
is bounded as x x0 .
g(x)
In these limits, the limit point x0 may be 0, or a one sided limit such
as 0+.
Example 1.22: Asymptotic relations:
x log x as x 0 ;
e1/x x2 as x 0+ (but not as x 0);
(log x)3 x1/5 as x + ;
ex ex + x as x + ;
sin x x as x 0 ;
1 cos2 x = O x2 as x 0 .
m
X
an n (x) m (x)
as x x0
for all m.
(1.21)
n=0
as x 0 .
63
1
2
2 5
15 x
3
sin x +
3
8
sin5 x + .
Nonuniqueness There are many functions with the same asymptotic series. For example, consider
X
1
xn
1x
as x 0 ,
n=0
but also
X
1
2
+ e1/x
xn
1x
as x 0 .
n=0
Two
with a zero asymptotic series,
P functions that differ by something
1/x2 above, wil have the same
0
(x)
,
such
as
the
function
e
n
n=0
asymptotic series. That is, any given asymptotic series represents a
whole class of functions. However, the difference between members
in the class of functions is much smaller than all elements in the sequence n (x).
Equating coefficients We may occasionally write
X
n=0
an n (x)
bn n (x)
as x x0 ,
n=0
1.2.4
65
Exercises
Exercise 1.24: Modify the iteration of Example 1.13 to find the Taylor
series solution to the ode y 00 2y =0 such that y(0) = 1 and y 0 (0) = 0
using reduce and to errors O x10 .
Answer: y = 1 + x2 + 61 x4 +
1 6
90 x
1
8
2520 x
+ O x10
1
9
12960 x
1
12
1710720 x )
+ O x15
O x10
b2 ( 12 x4 + x6 +
45 8
28 x )
+ b3 ( 17 x7 +
5 9
14 x )
+ O x10
+ O x10
159 9
4480 x )
+O x
17 6
80 x
663 8
4480 x )
+ b(x 16 x3
3 5
40 x
10
Exercise 1.30: Modify the arguments and the reduce computer algebra
of Example 1.16 to find the Taylor series, to an error O x10 , of the
general solution to the following three odes:
1. (x 2)y 0 = xy ;
2. (1 x2 )y 0 = 2xy ;
3. y 00 4xy 0 + (4x2 2)y = 0 .
Exercise 1.31: Modify the computer algebra
code for Example 1.18 to find
the Taylor series, to errors O x10 , of the well-behaved solution of the
nonlinear ode x2 y 00 + x2 y 0 + xy3 = 0 such that y(0) = 2.
Answer: y = 2 4x + 4x2
1199 8
1350 x
4142 9
6561 x
+O x
10
32 3
9 x
26 4
9 x
56 5
25 x
3404 6
2025 x
832 7
675 x
Exercise 1.32: Usereduce to help you find the power series about x = 0,
to errors O x10 , of the well-behaved solutions of the ode x2 y 00 +x3 y 0 +
(x2 2)y = 0. Hint: x2 y 00 2y = (x4 (y/x2 ) 0 ) 0 . Then modify your
reduce code to find the power series of the one parameter family of
well-behaved solutions to the nonlinear ode x2 y 00 + x3 y 0 + (x2 2)y +
y2 = 0.
Tony Roberts, 1 Mar 2008
67
3 4
10 x
1
8
144 x
10
1 6
Answer: Well behaved solutions are proportional to y = 1 51 x3 + 80
x
1
9
2640 x
1
12
147840 x
1
15
12566400 x
1
18
1507968000 x
+ O x20
Exercise 1.34: Find the power series expansions about x = 0, to errors O x10 ,
for the two parameter general solution to the linear ode x2 y 00 sin(x)y 0 +
y = 0, with the aid of computer algebra. Hint: expand sin x in a Taylor series and write x2 y 00 xy 0 in the form x2p (xp y 0 ) 0 .
1 3
7
89
6721
Answer: y = (a+b log x)(x 24
x + 3840
x5 1161216
x7 + 2229534720
x9 )+
1
b( 23040
x5 +
11
7
11612160 x
5951
9
44590694400 x )
+ O x10 .
Exercise 1.35: Algorithm 1.13 lists some reduce code to iteratively find
a power series solution to an ode: what is the differential equation it
purports to solve? and its initial conditions? what is the value of y
after the first iteration of the loop? what is the order of error in the
computed power series after the loop terminates?
Tony Roberts, 1 Mar 2008
Exercise 1.36: Use algebra to show that the divergent Stieltjes series (1.19)
formally solves the Stieltjes ode (1.18). Show that the Stieltjes integral (1.20) satisfies the the Stieltjes ode (1.18). Argue that the eight
term partial sum of the Stieltjes series (1.19) approximates the Stieltjes
integral to four decimal places for |x| < 0.08 .
Exercise 1.37: For any specified positive x, estimate the number of terms
in the Stieltjes series that will give the smallest error boundthe
optimal truncation. Hence, and using Stirlings formula for a factoriral, deduce
the Stieltjes series can be summed to an error
p that
e1/x .
|optimal | 2x
Exercise 1.38: Argue that (Bender & Orszag 1981, Ex. 3.64)
Z
X
et
(1)n (2n)! xn x 0 + .
dt
2
1
+
xt
0
n=0
(x)
and
g(x)
n
n
n=0
n=0 bn n (x) ,
P
then f + g n=0 (an + bn )n (x) as x x0 .
Rx
2. Let n (x) = x0 n (t) dt and suppose all the functions n (x) are
P
positiveRnear x0 . Deduce
that
f(x)
n=0 an n (x) as x x0
P
x
implies x0 f(t) dt
a
(x)
as
x
x0 .
n
n
n=0
1.3
Warning:
Rapid oscillations are a simple form of dynamics that we understand. Typically we want to know the long term evolution of the oscillations: do they
Tony Roberts, 16 Apr 2009
1.3. The normal form of oscillations give their amplitude and frequency
69
decay in amplitude? do they persist for all time? do they blow-up? does
their frequency evolve over long times? For example, a wave approaching a
beach is recognisably a wave up to the moment it breaks into tumultuous
surf. Yet as waves propagates towards the beach they slow down and simultaneously increase in amplitude (like depth1/4 ) to breaking, yet retain
the same frequency. Normal form transformations help us to uncover such
behaviour.
Classic nonlinear oscillators yield their secrets with our normal form methods. We investigate the frequency shift in the Duffing oscillator (Hinch 1991,
6.2); the subtleties of Morrisons problem; the appearance of a limit cycle
in the van der Pol oscillator (Hinch 1991, 7.1); and the stability of forced
oscillations in the Mathieu equation (Hinch 1991, 7.2). We start with a
relatively simple system.
1.3.1
(1.22)
(1.23)
where is some small positive parameter. See the plot of an example solution x(t) in Figure 1.1.
First, see the amplitude and frequency may be extracted from the linear
dynamics. Neglect the products and cross-products of x, y and the small
to describe the linear part of the dynamics:
x = y and y = x .
These describe simple harmonic oscillations. The general solution is x =
A cos(t + B) and y = A sin(t + B) for any constants A and B. The key step
is to alternatively describe these oscillations in terms of their amplitude r
and phase as
x = r cos
and y = r sin ,
(1.24)
0.25
0.20
0.15
0.10
0.05
0.00
0.05
0.10
0.15
0.20
0.25
0
10
20
30
40
50
60
70
80
90
100
t
Figure 1.1: solution x(t) of the simple nonlinear oscillator (1.221.23) for
parameter = 0.06 . See the oscillation amplitude grows until it reaches
about 0.25 . This growth matches well with the amplitude r(t) of the oscillations predicted by the model (1.26).
1.3. The normal form of oscillations give their amplitude and frequency
where
r = 0
and = 1 .
71
(1.25)
and = 1 + r2 .
(1.26)
From the amplitude equation r = r r3 readily see there are two fixed
0.3
0.2
0.1
0.0
y 0.1
0.2
0.3
0.3
0.2
0.1
0.0
0.1
0.2
0.3
x
Figure 1.2: a trajectory in the xy-plane of the simple nonlinear oscillator (1.221.23) for parameter = 0.06 . See the oscillation spirals to a
perfect circle.
1.3. The normal form of oscillations give their amplitude and frequency
73
1.0
0.8
0.6
0.4
0.2
0.0
x 0.2
0.4
0.6
0.8
1.0
0
10
t 15
20
25
30
Figure 1.3: three solutions x(t) of the Duffings ode (1.27) for parameter
= 1 . See the perpetual oscillations for each amplitude.
and y = r sin ,
1.0
0.5
0.0
y
0.5
1.0
1.00.80.60.40.20.0
x 0.2 0.4 0.6 0.8 1.0
Figure 1.4: three trajectories in the xy-plane of the Duffings ode (1.27).
See in the largest oscillation the non-circular shape that is crucial to the
nonlinear dynamics.
1.3.2
(1.27)
When the amplitude of the oscillations are small enough, the ode reduces
+ x = 0 . This has general solution x(t) =
to the classic linear oscillator x
A cos(t + B) for arbitrary constants A and B. What happens when we take
into account the effects of the nonlinear terms x3 ?
The answer here is that the frequency of the oscillations changes a little and
the shape of the oscillations change a bit as you see in Figure 1.4. To see
this algebraically we transform the Duffing equation to radius and phase
Tony Roberts, 16 Apr 2009
1.3. The normal form of oscillations give their amplitude and frequency
75
pair of odes r = 0 and = 1 . From these last two odes, read off that the
linear oscillator has constant amplitude r and a fixed frequency = = 1 .
In the amplitude-phase variables it is easy to interpret the solution of the
ode without actually solving the ode; we only transformed the ode.
Now explore the complicated algebra of the similar transform of the nonlinear Duffings ode (1.27). Assume the amplitude r of the oscillations are
small enough so that in this initial analysis we neglect terms quartic in r,
or higher powers.
Transform to amplitude-phase variables The fundamental transformation to amplitude-phase variables seeks solutions of Duffings ode (1.27)
in the form
such that
x = r cos + x 0 (r, )
r = g 0 (r) and = 1 + h 0 (r) ,
(1.28)
(1.29)
41 r3 (cos 3 + 3 cos ) = 0 ,
upon also omitting products of small corrections x 0 and the relatively small
amplitue r (as in the omission of the term 3r2 cos2 x 0 ). The r cos terms
Tony Roberts, 16 Apr 2009
1.3. The normal form of oscillations give their amplitude and frequency
77
cancel as these satisfy the linear oscillator part of Duffings ode (1.27):
0
2g 0 sin 2r cos h 0 + x
+ x 0 14 r3 (cos 3 + 3 cos ) = 0 .
(1.30)
x r cos
such that
r 0
and
1 3
32 r cos 3
1 38 r2 .
(1.31)
(1.32)
2x + 2x 4x = E ,
where E is an integration constant. Since 12 x 2 looks like a kinetic energy
term, and V(x) = 12 x2 41 x4 could be a potential energy, we identify E as
the conserved total energy. That we can integrate Duffings ode (1.27) in
this way shows that energy is conserved, and hence expect the oscillations
to have constant amplitude.
The following theorem (Takens 1974), based upon the construction we employ for Duffings ode (1.27), assures us that the amplitude and frequency
model for nonlinear oscillators is generally applicable.
Theorem 1.4 Autonomous, second order, nonlinear oscillators satisfying
odes of the form
+ 2 x + f(x, x)
= 0,
x
(1.33)
for smooth, Cp , strictly nonlinear functions f,7 can always be put into the
normal form of amplitude and phase angle variables, r and respectively,
x = r cos + X(r, ) + o(rp ) such that
r = g(r) + o(rp ) and = + h(r) + o(rp1 ) ,
(1.34)
1.3. The normal form of oscillations give their amplitude and frequency
79
such that r =
g(r) and = + h(r) satisfies the ode (1.33) to a residual
k
Res = O r , k p . Seek a correction to the coordinate transform to
x = r cos + X(r, ) + x 0 (r, ) such that r = g(r) + g 0 (r) and = + h(r) +
h 0 (r) that reduces the residual, that is, increases the order of the residual.
Substitute into the ode (1.33), linearise in the corrections to derive
0
2g 0 sin 2r cos h 0 + 2 (x
+ x 0 ) + Res = 0 .
Exercise 1.41:
Modify the normal form transformation of Duffings
ode (1.27) to interpret the long term evolution of the nonlinear oscillations (Bender & Orszag 1981, pp.5523, e.g.) governed by the
ode
+ x + x 3 = 0 .
x
(1.35)
is
Hint: the algebra of the expansion of the second time derivative x
identical.
1.3.3
1.3. The normal form of oscillations give their amplitude and frequency
81
1 3
32 r cos 3
1 2 5
1024 r (21 cos 3
+ cos 5) + O r7 .
(1.36)
More importantly, from the computer algebra output the normal form for
the dynamics of the amplitude and phase of the oscillations of Duffings
ode (1.27) is
15 4
123 6
r = 0 and = 1 83 r2 256
r 8192
r + O r8 .
(1.37)
That is, the amplitude of the Duffing oscillations is conserved, r = 0 ; the
nonlinear term x3 in Duffings ode (1.27) only modifies the frequency by
an amplitude dependent amount.
You might regard the program of Algorithm 1.14 as rather long for such
a simple problem. However, a wide variety of oscillation equations may be
modelled with this one program simply by modifying the one line computation
of the residual. Is that not marvellous? Try it in the exercises.8
Algorithm 1.14 implements the iterative procedure to the reduce output
in Dialogue 1.14. The main four parts of Algorithm 1.14 are the following.
Define that variable names r and q depend implicitly upon time. Also
use a let statement to tell reduce to replace time derivatives of these
variables by whatever asymptotic approximation we will have found
to date.
See in the discussion after equation (1.30) that we need to solve the
0 +x 0 +Res = 0 for the correction x 0 forced by the current residode x
ual Res. Here the residual will be of the form of a sum of cos n terms.
For each such term in the residual, we introduce into the correction x 0
the component cos n /(n2 1) to satisfy this ode.
The linear operator linv defines this process for us: being linear, the
operator distributes over addition and multiplication by any factor
independent of its second argument q which denotes the phase angle .
8
Furthermore, other algorithms for normal forms, such as that proposed by Rand &
Armbruster (1987) [Chapter 3], are considerably more involved to apply.
Tony Roberts, 16 Apr 2009
Then the pattern matching of the let statement will map residuals to
the corresponding corrections.
Set the linear approximation to the oscillations, that x = r cos such
that g = r = 0 and h = = 1 .
Iterate until the computed residual is zero to the specified O r8 error.
Compute the residual of Duffings ode (1.27), ensuring all products of sines and cosines are combined.
Extract any component in the residual in cos and sin , as these
would otherwise generate unallowable secular terms in the correc
tion, and use these components to update the evolution of r and .
Lastly, use the pattern matching of linv to correct the shape of
the oscillations.
The algorithm only terminates when the residual is zero to the specified erTony Roberts, 16 Apr 2009
1.3. The normal form of oscillations give their amplitude and frequency
83
ror. Thus the computation of the residual is the critical part of the algorithm
to ensure a correct normal form transformation.
1.3.4
Exercise 1.42:
Modify the computer algebra of Algorithm 1.14 to interpret the long term evolution of the nonlinear oscillations governed
by (1.35). Hint: include a pattern matching rule for the terms that
arise in the residual involving sin n.
9 4
and 1 + 256
r +
decay slowly to zero amplitude, roughly r t1/2 .
Answer: r 38 r3
351 7
8192 r
1755 8
262144 r
. Oscillations
Exercise 1.43:
Modify the computer algebra of Algorithm 1.14 to find
the frequency change in the nonlinear pendulum governed by the ode
+ sin x = 0 ,
x
Tony Roberts, 16 Apr 2009
(1.38)
1.3. The normal form of oscillations give their amplitude and frequency
85
where x is the angle of the pendulum from the vertical. Hint: when
computing the residual for any given approximation to the pendulums
angle x(r, ), use an appropriate Taylor series of sin x rather than the
sine function explicitly.
+ (g/`) sin x = 0
Extension: recall that the pendulum equation is x
where g is gravity and ` is the length of the pendulum; use your computer algebra to find how one should, in principle, vary the length `
with amplitude r of the oscillation so that the frequency remains precisely 1 (to some order of error in amplitude r).
Further extension: how can you vary the length ` with pendulum
angle x so that the frequency remains precisely 1 (to some order of
error in amplitude r).
Answer: x r
1 2
16 r
1
4
1024 r
. `1
1 3
192 r cos 3 such
3 4
18 r2 + 512
r or
Exercise 1.44:
Use a normal form transformation to explore the behaviour of the van der Pol equation (Kreyszig 1999, 3.5, e.g.)
+ x + (x2 1)x = 0 ,
x
(1.39)
for small values of the nonlinearity parameter . Describe the qualitative behaviour of its solutions. Hint: instead of discarding high
powers of amplitude r, discard high powers of parameter .
7 4
r ) . Oscillations
Answer: r ( 12 r 81 r3 ) and 1 + 18 2 (1 + r2 32
tend to a limit cycle of radius 2 with a slightly lower frequency than that of
the linear dynamics.
Exercise 1.45:
Describe the qualitative behaviour of solutions to Morrisons equation
+ x + x 3 32 x = 0 ,
x
(1.40)
Tony Roberts, 16 Apr 2009
9
2 4
256 r
. Solutions tend to a
limit cycle of radius 2 with a very slightly higher frequency than that of the
linear oscillations. The 2 term in r destabilises the origin to form the limit
cycles.
1.3.5
such that
r = r r3
and = 1 + r2 .
1.3. The normal form of oscillations give their amplitude and frequency
87
1 i it
1 i it
e
2 re e + 2 re
it
it
e
= ae + a
1 i
+ 12 rei i
2 re
3 i
1
+ 12 rei ir2
2 (r r )e
1 i
+ (1 + i)r2
2 re
2
= a[ + (1 + i)4|a| ] ,
using r2 = 4a
a = 4|a|2 . In summary, we may choose to write the
solution of odes (1.22)(1.23) as
eit
x = aeit + a
such that
a = a[ + (1 + i)4|a|2 ] .
Again, the reason the analysis is straightforward for this Example 1.46 is
that the shape of the oscillations are essentially circularsee the trajectories
in Figure 1.2. The analysis is significantly more complicated for typical
nonlinear oscillations because their shape may be far from circular. Lets
revisit the nonlinear oscillations of equation (1.35) in Example 1.41.
Furthermore, lets immediately turn to the computer to do the tedious algebra involved.
Example 1.47:
We use the residual of the ode (1.35) to drive corrections to the complex amplitude normal form of the ode. Here we
deduce that the nonlinear oscillations take the form
eit + i 18 a3 ei3t + a
3 ei3t
x = aeit + a
3
ei3t + 9a
5 ei5t
+ 64
a5 ei5t + 9a4 a
a4 ei3t a
+ O |a|6
(1.41)
where the complex amplitude evolves according to
9
a = 23 |a|2 a + i 16
|a|4 a + O |a|6 .
(1.42)
Before proceeding to derive this model, see that the real part, 32 |a|2 a,
controls the slow decay of the oscillation to zero amplitude, whereas the
9
|a|4 a governs the nonlinear frequency correction.
imaginary part, i 16
To see these two roles, simply substitute a = 21 r(t) exp i(t) into the
complex amplitude model (1.42) which then becomes
1 i
i = 3 r3 ei + i 9 r5 ei + O r6 .
+ i 12 re
2 re
16
512
Divide by exp i and equate real and imaginary parts to deduce
= 9 r4 + O r5 .
and
r = 38 r3 + O r6
256
Tony Roberts, 16 Apr 2009
1.3. The normal form of oscillations give their amplitude and frequency
89
After division by exp i the real part of the left-hand side is 21 r and
consequently the real
the imaginary part of the left-hand side is 21 r;
and imaginary parts of the right-hand side control the amplitude and
These are the typical roles for
phase evolution respectively (r and ).
In principle, we could try to account for the evolution of the complex amplitude a
in seeking a correction for any given residual, but this would be a very complicated task.
Instead, prefer to correct the transformation in many approximate steps, rather than few
accurate steps. This is a little appreciated but important shift in philosophy from that
insisted on by other methods such as multiple scales and averaging.
Tony Roberts, 16 Apr 2009
Algorithm 1.15 reduce code to find the complex modulation of the nonlinear oscillations of the ode (1.35).
factor i,zz;
% complex exponential cis(u)=exp(i*u)
operator cis;
let { df(cis(~v),~u)=>i*cis(v)*df(v,u)
, cis(~u)*cis(~v)=>cis(u+v)
, cis(~u)^2=>cis(2*u)
, cis(0)=>1 };
% to solve x_tt+x+res=0
operator linv; linear linv;
let { linv(1,cis)=>-1
, linv(cis(~n*t),cis)=>cis(n*t)/(n^2-1) };
% complex amplitude variables
depend a,t;
depend b,t;
let { df(a,t)=>ga, df(b,t)=>gb };
% initial linear approximation
x:=zz*(a*cis(t)+b*cis(-t));
ga:=gb:=0;
% iterate to nonlinear transformation
let zz^6=>0;
repeat begin
res:=df(x,t,t)+x+df(x,t)^3;
ga:=ga+i/2/zz*(ca:=coeffn(res,cis(+t),1));
gb:=gb-i/2/zz*(cb:=coeffn(res,cis(-t),1));
x:=x+linv(res-ca*cis(t)-cb*cis(-t),cis);
end until res=0;
1.3. The normal form of oscillations give their amplitude and frequency
91
1.3. The normal form of oscillations give their amplitude and frequency
93
1.3.6
So far we have discussed how nonlinearities affect the amplitude and frequency of oscillations. Periodic forcing also modifies oscillatory dynamics,
and in the case of the Mathieu equation generates growing instabilities in
certain parameter regions. Similarly, oscillations involving slow variations
in properties, such as slowly changing the length of a pendulum, also modifies the oscillations. The complex amplitude normal form empowers us to
analyse both of these situations.
Similar but generalised analysis applies to travelling waves and other more
complicated oscillatory dynamics. But here we restrict attention to simple
oscillations as an introduction to the methodology.
The Mathieu equation displays resonant instabilities
The linear Mathieu equation
+ (1 + 2 cos t)x = 0 ,
x
(1.43)
is an example of an ode whose coefficients are periodic; here the parameter the size of the periodic variations and is their frequency relative
Tony Roberts, 16 Apr 2009
1.3. The normal form of oscillations give their amplitude and frequency
95
and b = i 21 a + O 2 .
(1.45)
Having constant coefficients, this pair of odes for the complex amplitude
are much simpler than Mathieus ode (1.43). To find the instability, seek
solutions to this pair of coupled odes proportional to et and deduce the
characteristic equation 2 14 2 = 0 , and hence the complex amplitudes
are a linear combination of the two components exp( 21 t). Consequently
we will see the instability on a time scale of 1/ from the growing component exp( 12 t).
Synchronous fluctuations also excite instability What happens when
the forcing is cos t rather than cos 2t? Find out simply by setting w:=1 in
Tony Roberts, 16 Apr 2009
1.3. The normal form of oscillations give their amplitude and frequency
97
1.3. The normal form of oscillations give their amplitude and frequency
99
Algorithm 1.16, then execute the reduce code to see the output in Dialogue 1.17. Thus deduce that the oscillations of Mathieus ode (1.43) with
= 1 have shape
h
i
x = aeit + beit + (a + b) + 31 aei2t + 31 bi2t
2 i3t
+
ae + bei3t + O 3 .
(1.46)
24
More importantly, the complex amplitude of the oscillations evolve according
to the normal form
a = i2 ( 31 a + 12 b) + O 3
and b = +i2 ( 21 a + 13 b) + O 3 . (1.47)
2
See that here the coefficients
in the odes are of O , whereas in the
= 2 case they are of O : thus any instability is weaker in this case of
= 1 . To find the instability, seek solutions to this pair of coupled odes
5 4
proportional to et and deduce the characteristic equation 2 36
= 0 , and
hence the complex amplitudes are a linear combination of exp( 52 t/6).
Consequently we will see
instability on a long time scale of 1/2 from
the
2
the growing mode exp( 5 t/6).
Where are the stability boundaries? We have seen that exactly at
these critical frequencies, the response of Mathieus oscillator is that of growing oscillations. But such growth will only occur close enough to the critical
cases. We discover how close by detuning the natural oscillations and seeing
how the growth rate depends upon the detuning parameter.
Consider the = 2 case and modify Mathieus ode (1.43) to
+ (1 + + 2 cos 2t)x = 0 ,
x
for a small detuning parameter . The natural oscillations of the
ode, without any imposed fluctuations = 0 , then have frequency 1 + . This
natural frequency is a little different from the subharmonic of the fluctuations cos 2t: the parameter controls the amount of the difference. Correspondingly modify the reduce code of Algorithm 1.16:
Tony Roberts, 16 Apr 2009
Hence the growth-rates are = 12 2 2 . Only for || > || will there
be instability. The boundaries of the instability are = .
These boundaries are only approximate. Retaining higher order terms in parameters and gives a characteristic equation that predicts more accurate
stability boundaries.
Tony Roberts, 16 Apr 2009
1.3. The normal form of oscillations give their amplitude and frequency
101
Equation (1.48) is our first example of a multivariable asymptotic expansion. These will occur more often as we consider physical problems
with more competing physical processes that need parameterisation. We
have not established the properties for the manipulation of multivariable
asymptotic series; there are some subtleties which we will blithely ignore
for now.
Hinch (1991) [7.5] identifies twelve people who each played an important part in the
development of wkb theory: Liouville, Green, Horn, Rayleigh, Gans, Jeffrey, Wentzel,
Kramers, Brillouin, Langer, Olver and Meyer.
Tony Roberts, 16 Apr 2009
(1.49)
For example, in the potential well of a quantum particle, with t representing the spatial coordinate, 2 is the difference between the energy of the
particle and the potential energy at any position t. Now, if is roughly
constant near any time, then surely the solution x should locally be a linear
combination of exp(it). Actually this is not quite correct. A better
apR
proximation is that the Rsolution x is a linear combination of exp(i dt) :
the phase angle (t) = dt does indeed change by 2 over a local period
t 2/ , and so exp(i(t)) are global smooth oscillations that locally
have the correct frequency.
Thus we approach wkb solutions to the varying oscillator (1.49) by posing
the solutions are of the form
x = aei + bei
such that
a = ga (, a)
and b = gb (, b) , (1.50)
where the complex amplitudes a and b evolve relatively slowly and similar
to the rate at which the natural frequency varies. For real solutions x, the
complex amplitudes a and b will be complex conjugates; if derived correctly,
the evolution (1.50) will maintain this complex conjugate symmetry.
Computer algebra derives the wkb solution by adapting ideas in Algorithms
1.14 and 1.15. Algorithm 1.17 is the result.
The first innovation is to write the oscillations in terms of the phase
(represented by q in the reduce code) such that d/dt = as implemented by the let rule df(q,t)=>w.
But we need to account for derivatives of the varying frequency . Use
the dummy parameter eps to count the number of time derivatives
in each term. Insert the count by splitting the time derivative of
the frequency w into two via the two commands: depend w,tt; and
depend tt,t;. Then count with the pattern match and replacement
Tony Roberts, 16 Apr 2009
1.3. The normal form of oscillations give their amplitude and frequency
103
df(w,t)=>eps*df(w,tt).11
To form the asymptotic expansion, assume that the variations of the
frequency are slow enough so that terms with n derivatives of
are asymptotically smaller than those with n 1 derivatives and hence
we may discard terms with more than a set number
of derivatives. As
3
coded in Algorithm 1.17 we work to errors O , that is, we retain
only terms with none, one or two derivatives of frequency . This
limit may of course be increased as you wish.
Finally, in this linear problem there is no need to correct the form of
the oscillations, hence we do not need to change x and neither do we
need the pattern matching of linv.
Execute the above reduce code to get the output of Dialogue 1.19. That
is, the wkb normal form for the linear varying oscillator (1.49) is that x =
11
1.3. The normal form of oscillations give their amplitude and frequency
aei + bei for phase =
according to
105
2
a+i
a + O 3t ,
(1.51)
3
2
2
8
4
2
b =
bi
b + O 3t ,
(1.52)
3
2
2
8
4
where the error O 3t encompasses all terms with three or more derivatives of the slowly varying frequency .
a =
For example, we use this model to predict the conservation of action. Con
(1.53)
= () d = E ( + )2 d = 31 E[(t + )3 3 ] .
0
For example, this ode might describe the oscillations of a quantum particle in a
quantum potential well; in which case the eigenvalues E represent the discrete energy
levels of the quantum dynamics.
Tony Roberts, 16 Apr 2009
.
n
n
3
494
Bender & Orszag (1981) [p.492] show that this approximation for the
eigenvalues is amazingly good. We expect it to be good for large n
as then the variations in natural frequency is relatively small per
oscillation: indeed E10 is accurate to a relative error of 0.13%. But
even the fundamental eigenvalue E1 is estimated by the above formula
to a relative error of no more than 8.1%.
Bender & Orszag (1981) [Chapter 10] discuss many more aspects of the
wkb approximation, including conditions of its validity. Note that the normal form derived above is equivalent to the wkb approximation they discuss.
However, our approach empowers us to analyse much more difficult problems
such as nonlinear oscillators with large variations in coefficients provided the
variations take place slowly. For example, Exercise 1.52 asks you to analyse a nonlinear Duffing oscillator with a varying frequency. The necessary
modifications to the computer algebra are straightforward. This one normal
form approach fantastically allows us to solve easily a wide range of difficult
problems.
Furthermore, our normal form approach also closely parallels aspects of homogenisation methods and theory; for example, Bensoussan et al. (1978) or
Rosencrans (1997) may introduce you to aspects of homogenisation. Many
practical situations involve the dispersion or other macroscopic transport
through a material with complicated microscopic structureground water
flow through soil for example. Homogenisation analyses the microstructure and the transport through the microstructure to derive an equation for
the effective macroscopic transport. Here we similarly analyse the microscopic shape of oscillations to derive equations governing the evolution of
Tony Roberts, 16 Apr 2009
107
the macroscopic amplitude and phase of the oscillations over long macroscopic times. Our analysis is simply based upon a normal form change in
the coordinates. The normal form approach subsumes or illuminates many
other modelling methods.
Exercise 1.51:
Use all the terms found in the normal form (1.51) for
the linear varying oscillator (1.49) to deduce more accurate estimates
for the eigenvalues En of the ode (1.53).
Exercise 1.52:
Construct a wkb model for the slowly varying version
of Duffings oscillator
+ 2 x + x3 = 0 .
x
For simplicity let the parameter measure the magnitude of the slow
variations in time of the frequency and the magnitude of the nonlinearity x3 . Use Algorithm 1.17 as a base, adapt and incorporate
the linv operator of Algorithm 1.15 to update the shape x of the
oscillations. Interpret your results.
1
a3 ei3 +
82
1
3 i3
2
b e
+ O where the complex amplitude a evolves according to
82
2
2
2
i3 2
3 2
1
3 2
3
a = 2 a + 2 a b + 9
a b + i
15
16 a b + 8 a 4 a + O
43
3
1.4
Chapter Summary
Chapter 2
Flow description . . . . . . . . . . . . . . . . .
Conservation of mass . . . . . . . . . . . . . .
Conservation of momentum . . . . . . . . . .
2.3.1 Stress tensor . . . . . . . . . . . . . . . . . .
2.3.2 Momentum equation . . . . . . . . . . . . . .
2.3.3 The Euler equations for an ideal fluid . . . .
2.3.4 The NavierStokes equations . . . . . . . . .
2.4 The state space . . . . . . . . . . . . . . . . . .
2.5 Chapter summary and exercises . . . . . . .
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
110
113
118
. 119
. 121
. 123
. 124
125
128
Here we review how to describe the the flow of a fluid such as water or
air. Physical quantities, such as fluid velocity, are generally unknown in
any application; so we derive differential equations which they must satisfy.
The equations are based on physical principles. In any particular problem
it is these differential equations which are solved to predict a flow pattern.
Lastly, we look at some useful properties deduced directly from the equations
of motion.
Most of the principal concepts of hydrodynamics are more easily understood
in one-dimensional dynamics rather than in the three-dimensional dynamics
109
110
2.1
Flow description
Sorry:
We describe fluid flow primarily in terms of the velocity, density and pressure
fields. Of these, the velocity field, the actual motion of the fluid, is the one
of prime importance. Other physical properties arise in certain applications.
We adopt a so-called Eulerian description of field variables. That is, we
describe how physical quantities vary in space and time by considering them
at specific points in space-time.
For example, the pressure p is considered as a function of space and
time, which in Cartesian coordinates is written p(x, y, z, t). You may
see isobars on weather maps; these are contours of the pressure field
on the Earths surface. They show how the pressure field varies on
the surface. Over successive days the isobars move indicating that the
pressure field varies in time. But note that our observations of the
pressure field is largely derived from fixed observation stations, that
is, we record the pressure at a specific point as time variesthis is the
essence of the Eulerian description.
By contrast, a Lagrangian description of the flow describes conditions
experienced by particles of the fluid as they move through space. But
we do not adopt the Lagrangian approach.
Similarly the density is described by its space-time dependence, (x, y, z, t).
The velocity of a fluid flow is a vector field q. This is because at
every point in space the fluid moves with some speed in some direction, and consequently is described as a vector. As above we consider
Tony Roberts, 1 Mar 2008
111
112
(a) uniform
(c) vortex
Figure 2.1: four example flow fields: (a) velocity field of uniform flow in the
x-direction; (b) velocity field of solid body rotation; (c) velocity field of the
bathtub vortex; (d) velocity field of a corner flow.
113
large such a block does not rotate as it revolves around the centre of
the vortex, it typically has the same orientation; this is an example of
an irrotational flowan important class of flows in fluid dynamics. In
contrast, a building block placed on a turntable rotates as it revolves
showing that its motion is rotational.
As a last example the velocity field q = xi + yj may be seen as flow
turns a corner due to some obstruction. A quadrant of the velocity
field shown in Figure 2.1(d) may be seen in a corner of the flow over
a rectangular obstruction shown in Plate 5 of Van Dyke Dyke (1982).
These ideas of velocity and density fields overlook the fact that real fluids
are made of molecules. In principle the physical fields we use are defined
by averages over volumes, each volume centred on some particular point
in space-time. Thus
(x, y, z, t) =
1 X
(mass of particles in V) ,
kVk
2.2
Conservation of mass
Warning:
114
We need to find equations which reflect how the fluid fields such as q and
evolve in time. The evolution come from physical laws of conservation.
First, the material making up the fluid can neither appear or disappear.
This conservation of mass leads to the so-called continuity equation.
An extremely quick and dirty method is to adapt the one-dimensional
continuity equation (Roberts 1994, Eqn (2.1)),
(u)
+
= 0,
t
x
to three-dimensions. Simply we recognise that time derivatives and density remains unchanged, whereas the one-dimensional scalar velocity u
becomes a vector quantity q in three-dimensions, and one-dimensional
spatial gradients /x become one of the vector differential operators
grad, , or div, . Thus the above one-dimensional continuity equation
translates to
+ (q) = 0 .
t
The choice of the divergence is indicated because the time derivative shows
that the equation is a scalar equation, and the div operator is the appropriate way to produce a scalar quantity from the vector flux q.
We now turn to a more acceptable argument for deriving the above continuity equation.1 However, as in one-dimension (Roberts 1994, 2.1), I prefer
an integral argument.
Consider any fixed
R volume V within the fluid. The mass of fluid inside the
volume is m = V dV. Over time this mass varies as fluid is carried into
or out off the volume by the flow of the fluid. The rate of change of mass
is, by definition,
Z
dm
=
dV .
dt
V t
But this must be equal to net rate of fluid entering the volume across the
surface, S, of the volume. Across any small area dS of the surface, fluid
^ is the unit vector normal
mass crosses into V at a rate q (^
n) (where n
to dS but pointing out of V), namely
1
See Dean & Dalrymple (1991)[2.2.1] for a similar development but in differential
terms.
Tony Roberts, 1 Mar 2008
115
Equating these two expressions for dm/dt and combining them into the one
integral leads to the penultimate result that
Z
+ (q) dV = 0 ,
V t
for all volumes V in the fluid.
Now, the only way that the integral of a particular continuous integrand can
be zero for all volumes V is if the integrand is identically zero. Consider, if
the integrand is ever non-zero, then by continuity it must be the same sign in
some small volume V 0 around that point, and so the integral over V 0 would
be non-zero. But this contradicts the fact that the integral is always zero.
Thus the integrand must be zero everywhere. Consequently the previous
result assures us that everywhere in the fluid
+ (q) = 0 .
t
(2.1)
This is the general continuity equation. However, in most familiar circumstances it simplifies somewhat. For example, if the density of the fluid is
everywhere the same for all time, is constant, then all the derivatives of
are zero and the continuity equation reduces to the requirement that the
velocity field must be divergence free:
q = 0.
(2.2)
In fact this equation is much more general than I have suggested so far;
as well as applying to constant density fluids, it also applies to so-called
Tony Roberts, 1 Mar 2008
116
+ q + q = 0 .
t
Now the rate of change of density of the parcel of fluid is, by the chain rule,
d
dt
=
=
=
dx dy dz
+
+
+
t x dt y dt
z dt
+
u+
v+
w
t x
y
z
+ q ,
t
as appears in the first two terms of the expanded continuity equation above.
Thus if each and every parcel of fluid is itself incompressible, its density does
Tony Roberts, 1 Mar 2008
117
not change, d/dt = 0 throughout the fluid, and so the general continuity
equation (2.1) reduces to the simpler (2.2). I emphasize that this occurs
even though the density may vary in space and time, as it does in the ocean,
so long as each parcel of fluid is itself incompressible.
Each of the four example velocity fields given in the previous section are
divergence free. For example, the vortex flow field satisfies
x
y2x
x2y
q=
+
= 2
+ 2
= 0.
2
2
2
2
2
2
x x + y
y x + y
(x + y )
(x + y2 )2
Hence they could each occur in the flow of an incompressible fluid.
Appearing in the above derivation is the very important concept of the
material derivative. Namely the derivative with respect to time of some
quantity as seen a parcel of the material; in the above we discussed the
density as seen by a fluid parcel. Where needed the material derivative of
some quantity, c say, will be denoted by
dc
c
=
+ q c .
dt
t
(2.3)
118
2.3
Conservation of momentum
Sorry:
+u
,
=F+
t
x
x
to three-dimensions. On the left-hand side we observe the material derivative of the velocity. On the right-hand side, F denotes body forces, such as
gravity g, and so becomes a vector, whereas the end forces could remain as the scalar pressure field, p, which exerts a normal forces across
any surface in the fluid. Thus the above one-dimensional momentum
equation translates to
dq
q
=
+ q q = F p .
dt
t
However, this equation is only valid for fluids which are not sticky (not
viscous). In a viscous fluid, the fluid particles exert tangential forces across
surfaces; for example, they experience a tangential drag when slipping
over a solid substrate. For this common case the stress becomes a rank 2
tensor, denoted . Note: a rank n tensor is some physical quantity with n
directional properties; for example, a scalar is a rank 0 tensor, and a vector
is a rank 1 tensor. Then for a viscous fluid the above momentum equation
is generalised to
q
dq
=
+ q q = F + .
dt
t
Interestingly, this momentum equation is directly analogous to Newtons
2nd law which asserts that (mass) (acceleration) = (net force). In the
above:
mass is represented by the density ;
Tony Roberts, 24 Apr 2009
119
2.3.1
Stress tensor
Consider a small tetrahedral parcel of fluid with three faces normal to the
^ . Across each face, the
coordinate axes and the fourth face with normal n
fluid outside the tetrahedron exerts a force, or traction, on the fluid inside;
call the forces Ax T x , Ay T y , Az T z , and An T n respectively where Ai is the
area of the ith face and T i are the stresses.
The stress is linear in the normal. Firstly, if there is any imbalance
in the net force applied by these stresses, then the fluid tetrahedron must
accelerate. Newtons 2nd law then indicates that
dq
V
= Ax T x + Ay T y + Az T z + An T n ,
dt
where V is the volume of the tetrahedron. Hence, crudely, the acceleration
must be proportional to (surface area)/(volume) which becomes infinite for
arbitrarily small volumes. Such infinite acceleration cannot be permitted,
and hence there cannot be any net imbalance of the forces on the tetrahedron. Thus
Ay
Ay
Ax
Tx
Ty
Tz
Tn =
An
An
An
by projection of areas
= (^
n i)T x + (^
n j)T y + (^
n k)T z .
^ is linear
From this we see that the stress across a plane with any normal n
^ : for some stress tensor
in n
^ .
Tn = n
(2.4)
Tony Roberts, 24 Apr 2009
120
xx xy
= yx yy
zx zy
xz
yz
zz
where ij is the component of the stress in the jth coordinate direction upon
a plane with normal in the direction of the ith coordinate.
x
3 Ax T x
j = Vxy ;
y
3 Ay T y
(i) = Vyx .
xx xy xz
p + xx
xy
xz
p + yy
yz .
= yx yy yz = yx
zx zy zz
zx
zy
p + zz
Tony Roberts, 24 Apr 2009
121
The deviatoric stress tensor ij measures the departure from isotropy of the
stress tensor. If ij are all zero, then the stress tensor = pI and hence
^ , is = n
^ (pI) = p^
the stress across any surface, normal n
n. That is,
the force across any surface is directly inwards and of magnitude p. This
is precisely the accepted understanding of pressure. If not all ij are zero,
then a local change of coordinates2 will put the stress tensor in diagonal
form (always possible for a symmetric matrix),
0
p + xx
0
0
0
,
0
0
p + yy
0
0
0
p + zz
and so we see that in some directions the stresses will be stronger than in
others, causing a deformation of parcels of fluid.
2.3.2
Momentum equation
=
(q) dV .
dt
t
V
But this must be equal to net rate of momentum
entering the volume across the surface, S, of the volume;
generated internally by external body forces such as gravity; and
induced by stresses imposed on the surface of the volume by the fluid
immediately outside.
Consider each of these in turn. Across any small area dS of the surface, fluid
^ is the unit vector
momentum crosses into V at a rate u(^
n) q (where n
normal to dS but pointing out off V), namely
2
The pressure, being proportional to the trace of the stress tensor, is invariant to such
changes in coordinates.
Tony Roberts, 24 Apr 2009
122
Equating these two expressions for dp/dt and combining them into the one
integral leads to the penultimate result that
Z
(q) + (qq) F dV = 0 ,
V t
for all volumes V in the fluid. As before, such an integral can only be zero
for all volumes V if the integrand is everywhere zero. Thus we deduce the
basic form of the momentum equation
(q) + (qq) = F + .
(2.5)
t
However, in virtually all cases we only use this equation in conjunction with
the continuity equation (2.1). In which case, after expanding the derivatives
of the products on the left-hand side as
q
(q) + (qq) =
+q
+ q (q) + (q )q ,
t
t
t
recognise that the middle two terms are simply q times the continuity equation and so vanish identically. This gives us the more usual form analogous
to Newtons second law:
dq
q
=
+ (q )q = F + .
(2.6)
dt
t
3
2.3.3
123
(2.7)
1 2
1
1
0 =
+
|q| + + p ,
t
2
124
1 2 p
=
+ |q| + +
.
t
2
The only way that the gradient of a quantity is zero in some spatial domain
is if it is constant in space; that is, we obtain Bernoullis equation
p 1 2
+
+ |q| +
= B(t) ,
t
2
(2.8)
for some function B(t). This equation gives the pressure field for any given
flow field of an ideal fluid.
Three of the example flows are potential flows. Consider two.
(a) q = Ui has velocity potential = Ux and the pressure is reduced from
hydrostatic: p = 21 U2 .
(c) The vortex flow q = (yi + xj)/(x2 + y2 ) has velocity potential =
arctan(y/x) = in cylindrical coordinates, and a pressure p =
1
2
2 /r .
This flow is seen in the swirling water of the bathtub vortex, for
example see (Dyke 1982, Plate ??). The funnel shape of the central
core of air is determined once you know that in effect air exerts a
constant pressure, p0 say, everywhere upon the free-surface of the
water. If the body force potential is due to gravity, = gz, and
hence the free-surface of the water must be such that p0 = gz
1
2
2 /r . Rearranging shows that the height of the free-surface in the
funnel must be of the form z = A B/r2 .
2.3.4
125
qk
.
x`
(2.10)
where = / is called the kinematic viscosity. For water and air under
common conditions the kinematic viscosity is 0.0114 cm2 / sec and 0.145 cm2 / sec
respectively. Observe that the term due to viscosity, 2 q, describes an essentially Fickian diffusion of momentum. This is reflected in many of the
flow properties that we later explore.
2.4
Warning:
7
126
In this book we investigate fluid dynamics using tools of analysis from the
modern theory of dynamical systems. Consequently we here put the equations of fluid mechanics into the context of a dynamical system.
The key concept, introduced by Poincare, is that of a state space. The
state space of a systems is the set of all possible states of the system. For
example, as explained in 2.1 of Abraham & Shaw (1983), the state of an
ideal pendulum oscillating in two physical dimensions is its position and
velocity. When we measure these by the angle from the vertical and the
angular velocity , then the state space is the 2D set of all possible values
of (, ).
However, if the pendulum, instead of being swung on the end of a thin rod,
is hung by an elastic spring, then two more variables are needed to describe
its state: namely, the extension of the spring, say `, and the rate of change
of the extension, say w. Thus the state space is enlarged to the set of all
possible values of (, , `, w), a 4D space.
Now, at any given instant of time, a dynamical system will be in precisely
one state. In other words, the state of the dynamical system at any time
is represented by one point in state space, call such a state u in general.
Evolution of the system in time, its state changing, is then represented by
movement of a point through state space (Abraham & Shaw 1983, 1.2),
then u = u(t). This is the basis of the abstract view of an evolving system:
no matter how complex the original physics, we represent its evolution by
the movement of a point through the state space.
If the mathematical model of the system is consistent, it can only evolve in
one way from any given state. After all, if the system could evolve in more
than one way, then we could not use the model to make predictions because
any one of the possible futures may occur. Indeed this is a modelling issue
we only address those models which are predictive; that is, only those with
a unique future from any given state.
Consequently, at any time and at any state the dynamical system must be
moving through state space in some definite direction and at some definite
speed. Thus the rate-of-change of state can be described by a vector function
Tony Roberts, 1 Mar 2008
127
of position:
du
= f(u) ,
dt
where f is some definite function characteristic of the particular dynamical
system. We will analyse such evolution equations throughout this book.
u =
128
the pressure field may be viewed as being used to keep the velocity field
divergence free.
2.5
Generally we will deal with the NavierStokes and continuity equations, together describing the motion of a viscous incompressible fluid.
The primary field of interest is the velocity field q(x, y, z, t), and to
some extent the pressure field p(x, y, z, t). Typically we treat incompressible fluids except in special circumstances.
Another flow property of great interest in fluid dynamics is the vorticity. But for purposes of simplicity I have chosen not to expound on it
and its effects in this introduction.
By placing fluid dynamics within the context of a dynamical system,
albeit in an infinite dimensional state space, we use tools of modern
dynamical systems theory to analyse fluid dynamics. In particular, we
create low dimensional models of the dynamics of fluids.
To follow as needed: boundary conditions, heat and mass transfer,
sample solutions and dynamics.
Exercise 2.1: Lookup and write a one page summary of vorticity, irrotational flow and the Kelvin circulation theorem.
Exercise 2.2: Find out and write down the continuity, Euler and Navier
Stokes equations in component form in both Cartesian and spherical
coordinates.
Exercise 2.3: Do Problems 2.1, 2.3(a), 2.4, 2.5, 2.8 and 2.9 in Dean &
Dalrymple (1991), Water wave mechanics for engineers and scientists.
Exercise 2.4: In a viscous fluid of constant density with no body forces,
consider a swirling flow which in cylindrical coordinates has velocity
components u = w = 0, v = v(r, t), p = p(r, t). What do the Navier
Stokes and continuity equations reduce to? Seeking a solution of the
Tony Roberts, 1 Mar 2008
129
form v(r, t) = A(t) exp r2 /D(t) determine A(t) and D(t). Explain
what this solution may physically describe.
130
Chapter 3
3.2
3.3
3.4
Couette flow . . . . . . . . . . . . . . . . . . . . .
134
3.1.1
3.1.2
3.1.3
3.1.4
3.1.5
Exercises . . . . . . . . . . . . . . . . . . . . . . . 148
. . . . . . . . . . 136
150
154
3.3.1
3.3.2
3.3.3
3.3.4
3.3.5
3.3.6
193
3.4.1
3.4.2
131
132
3.5
3.6
3.7
3.8
3.4.3
3.4.4
212
3.5.1
3.5.2
3.5.3
Exercises . . . . . . . . . . . . . . . . . . . . . . . 238
3.6.2
3.6.3
Summary . . . . . . . . . . . . . . . . . . . . . . . 251
252
3.7.1
3.7.2
3.7.3
3.7.4
Summary . . . . . . . . . . . . . . . . . . . . . . . 255
Chapter summary . . . . . . . . . . . . . . . . . .
255
This chapter aims to develop the basic concepts of centre manifolds and their
use in modelling dynamical systems. Subsequent chapters explore important
classes of applications as well as further illuminating theory.
The closeness of the agreement between his [Taylors] theoretical
and experimental results was without precedent in the history of
fluid mechanics.
(Drazin & Reid 1981, p105)
Here we look at an important experiment in fluid mechanics, the Taylor
Couette flow, and determine the basic dynamics of the development of a
nontrivial flow. Ideas in dynamical systems theory which are associated
with this flow are those of: bifurcation, symmetry breaking, and pattern
formation.
Tony Roberts, 1 Mar 2008
133
We look at a simple fluid flow between two rotating cylinders, called Couette flow. Then we find that it loses stability, as the speed of rotation of the
inner cylinder is increased, and wonder what happens to the flow dynamics
after that. Pictures of experiments in the various flow regimes are shown
in Plates 1278 of Dyke (1982). Indeed, this was one of the early experiments to confirm chaos in an actual fluid flow (Gleick 1988, pp128131). By
introducing the ideas and tools of centre manifold theory via some simple
examples, we analyse the TaylorCouette dynamics in a little detail.
quasi ??
The flexible power of the centre manifold approach to creating dynamical
models comes from three theorems introduced in Section 3.3. The three
theorems assure us of three key aspects:
a model exists depending upon the linear picture of the dynamics;
the model is relevant because all neighbouring dynamics are attracted
to the model;
we may construct the model to a controlled order of error.
The key is to find a domain where there is a separation of time scales between
long lasting modes of interest, and rapidly decaying modes that provide
relatively uninteresting details. In principle, the models may be refined to
arbitrarily high order of accuracy because we can control the error.
Construct chapter begin??
colan??
Centre manifold theory also underpins global models, Section 3.6.1. Although introduced as being local to an equilibrium, the theory also applies
when there is a parametrised family of equilibria, a subspace of equilibria for
example: the theory then supports a dynamic model that applies uniformly
across the family to generate a model which is local in some variables but
is global in other variables. One application is to chemical reactions. Other
applications include many that others call singularly perturbed systems.
Also the theory applies in a finite domain, and so rigorously supports modTony Roberts, 1 Mar 2008
134
elling for finite sized parameters and variables. Thus to estimate the model
accurately at finite values of parameters and variables, we need to compute
asymptotic approximations to different orders in the parameters and variables. The Newton diagrams introduced in Section 3.6.2 empower us to very
flexibly create and use models.
??irregular
3.1
Couette flow
Warning:
3.1.1
As a precursor to Couette flow we briefly look at the shear flow between two
parallel plates.
Consider two plates, infinite in horizontal extent, separated by a distance h
where the bottom plate is stationary and the top plate is moving with
speed U in the horizontal x-direction. Let the bottom plate be at y = 0
and consequently the top plate is at y = h. The top plate will drag the
fluid with it, whereas the bottom plate will tend to retard the flow. Thus a
shearing flow is established between the plates.
By translational symmetry in the horizontal the flow will be independent
of both horizontal directions. So we assume that q = q(y) and p = p(y)1
only. Also by symmetry, there should be no flow in the z-directionthere
1
135
is no driving force in that direction and so viscosity will typically damp any
such flowhence we take w = 0.
The continuity equation then reduces to
q=
v
= 0,
y
and hence v = const. But no fluid can cross the solid boundaries2 at
y = 0 and y = h, v = 0 there, so that v = 0 everywhere. Hence the
velocity field is just q = u(y)i.
The vertical component of the NavierStokes equation,
v
1 p
+ q v =
+ 2 v ,
t
y
reduces to simply
0=
1 p
,
y
2 u
,
y2
136
Such a linear variation in velocity across the stream is a uniform shear flow.
The viscous fluid separating the two plates transmits a frictional drag from
one plate to the other. This drag is the yx component of the deviatoric
stress tensor, namely the traction in the x-direction acting across a surface
with normal in the y-direction. From the Newtonian stress/rate-of-strain
relation (2.9)
u v
U
yx =
+
=
.
y x
h
Observe this drag behaves reasonably: it is proportional to the viscosity;
proportional to the velocity difference; and inversely proportional to the
plate separation.
3.1.2
In essence we now turn the previous shear flow on its side and wrap it around
a cylinder to form Couette flow. Consider two co-axial vertical cylinders with
fluid in between:
the inner cylinder of radius R1 and rotating with angular velocity 1 ;
and
the outer cylinder of radius R2 and rotating with angular velocity 2 .
Because of the cylindrical geometry it is convenient to resort to cylindrical
coordinate system (r, , z) where:
r is the radial distance from the vertical axis;
is the angular distance around the cylinder; and
z is the vertical position.
Tony Roberts, 5 Apr 2009
137
For convenience we again call the velocity components (u, v, w), that is, the
velocity q = uer + ve + wk.
By using symmetry, we again look for a simple solution of the fluid dynamics
equations. Because of the circular symmetry around the cylinder, and the
translational symmetry along the cylinder, we expect to find solutions that
are functions purely of the radial coordinate r. Expect that w = 0 because
there is no mechanism to drive an axial flow.
The continuity equation
1 (ru) 1 v w
+
+
= 0,
r r
r
z
then determines that the radial velocity u = C/r. But, as for the
parallel plate flow, there can be no flow across the cylinder walls at
R1 and R2 , so C = 0, and hence u = 0 throughout.
That is, the flow must be purely around the cylinder, termed azimuthal, but possibly of different speeds at locations ranging from
the inner cylinder to the outer cylinder. In order to turn the fluid
particles to keep them in their circular paths we will need to impose
a radial pressure gradient.
The angular component of the NavierStokes equation, when put in
cylindrical coordinates (Batchelor 1979, Appendix), reduces to
1
0=
r r
v
v
r
2 .
r
r
This is an Euler ODE. Substituting v = rn leads to the indicial equation n2 1 = 0 with solution n = 1. Thus
v = Ar +
B
,
r
138
2 R22 1 R21
R22 R21
r+
1 2
1/R21 1/R22
1
.
r
v2
1 p
=
.
r
r
v2
dr ,
r
3.1.3
Non-dimensional equations
139
Re
+q q
= p + 2 q ,
t
q = 0 ,
where Re = 1 R21 / is a Reynolds number. Reynolds numbers are typically
defined as Re = (velocity)(length)/ which is precisely the same here as
1 R1 is the velocity of the surface of the inner cylinder, and R1 its radius.
Consider the range of Reynolds numbers:3
Re < 100 are typically viscously dominated flows;
Re > 10000 typically are very turbulent flows full of fine structure and
detail;
100 < Re < 10000 is roughly the parameter regime where interesting
transitions take place in the destruction of order in favour of chaos.
As shown in Iooss & Adelmeyer (Iooss & Adelmeyer 1992, p89), many interesting transitions occur from the above Couette flow as the Reynolds
number (Re1 in their figure) is increased.
Henceforth we do two simplifying things. Firstly, we drop the stars from
the non-dimensional quantities. It will be just as if we solve the physical
problem with R1 = 1 = = 1. Secondly, and more fundamentally we make
the outer cylinder stationary (as is usually done in experiments). Thus the
boundary condition on the outer cylinder is simply
q = 0 on r = R.
Whereas that on the inner cylinder is
q = e
on r = 1.
Then in cylindrical coordinates, the non-dimensional NavierStokes and continuity equations become, in component form,
u
u v u
u v2
p
2 v
u
Re
+u
+
+w
+ 2 u 2
,
=
t
r
r
z
r
r
r r2
3
These ranges give the generic correspondance between Re and flow regime.
Tony Roberts, 5 Apr 2009
140
v
v v v
v uv
1 p
2 u
v
Re
+u +
+w +
=
+ 2 v + 2
2,
t
r r
z
r
r
r r
w v w
w
p
w
+u
+
+w
=
+ 2 w ,
Re
t
r
r
z
z
u 1 v w u
+
+
+
= 0,
r
r
z
r
1 2
where 2 = 1r r
r r
+ r2 2 + z
z .
In this non-dimensionalisation and with a fixed outer cylinder, Couette flow
is, see Figure 3.1,4
2
1
R
q = v(r)e = 2
r e .
R 1 r
As the inner cylinder spins faster, increasing Reynolds number Re, the fluid
near the inner cylinder increasingly wants to spin out towards the outer
cylinder. At some rotation rate, this overcomes viscosity and the inner fluid
moves towards the outer cylinder, and the fluid near the outer cylinder must
contrapositively move inwards. These two opposing motions organise into
a pattern of cell-like vortices wrapped around the cylinder. It is this flow
which we examine.
3.1.4
Linear instability
In a dynamical system with a fixed point, Couette flow for example, the
dynamics in the immediate neighbourhood of the fixed point is determined
0
by linearisation.
If u = f(u) is the dynamical system with fixed0 point u0
0
(so that f u = 0), then small time dependent perturbations, u (t), to u
satisfy an equation of the form u 0 = Lu for some linear operator L.
The eigenvalues of L determine the stability of the fixed point. If all the
eigenvalues of L have real part negative, then the fixed point is stable. If
at least one eigenvalue of L has real part positive, then the fixed point is
unstable.
4
141
Figure 3.1: velocity field of Couette flow with fixed outer cylinder and R = 2.
142
In the analysis of Couette flow, we aim directly for the linearised eigenvalue
problem by seeking solutions of the form
q v0 e + q 0 (r, , z)et ,
p p0 + p 0 (r, , z)et ,
where v0 and p0 are the fields determined for Couette flow. Substitute
these into the NavierStokes and continuity equations and neglect nonlinear
terms, products of primed quantities, to obtain
v0 u 0 2v0 0
p 0
2 v 0 u 0
0
Re u +
v
=
+ 2 u 0 2
2,
r
r
r
r
r
0
0
0
0
0
0
v 0 v v
v
1 p
2 u
v0
Re v 0 +
u +
+ u0
=
+ 2 v 0 + 2
2,
r
r
r
r
r
r
0
0
0
v w
p
Re w 0 +
=
+ 2 w 0 ,
r
z
u 0 1 v 0 w 0 u 0
+
+
+
= 0.
r
r
z
r
Note that the only nonlinear terms in the NavierStokes equations are the
advection terms (q q). In the linearisation, the terms retained as important represent the advection of the perturbations by the azimuthal Couette
flow, terms such as (v0 /r)u 0 /, and the movement of the Couette flow by
the perturbations, terms such as u 0 v0 /r.
Now, these equations for the perturbations are constant coefficient with
respect to angle . Moreover, the annular geometry assures us that we are
only interested in solutions 2-periodic in . Thus we may seek solutions
proportional to exp(im) for integral m.
The equations are also constant coefficient with respect to the axial variable z, and so we may seek exponential solutions in z. However, in most
experiments the cylinders are tall (when compared to the thickness of the
fluid or to the radius of the cylinders), see Plate 127 in Dyke (1982) for example. Thus we are not interested in exponentially growing solutions in z,
in either direction; we only want solutions proportional to exp(ikz). In principle k is determined by the length of the cylinder L, perhaps something like
k = n/L for integral n. For long cylinders these wavenumbers are so closely
Tony Roberts, 5 Apr 2009
143
p 0 = p 00 (r)ei(m+kz) ,
1d
D=
r dr
d
r
dr
w 00
,
r2
= 0,
k2
1
,
r2
144
0
Recall
0 that
Couette velocity profile is v = Ar + B/r; with this the term
0
v
dv
= 2A a constant. Thus we use the last equation above to elimidr + r
nate u 00 to give just one eigen-problem for the azimuthal velocity perturbation:
2
1
4Av0 00
D Dv 00 k2
v = 0.
Re
r
In general, this eigen-problem has to be solved numerically for the eigenvalues and corresponding eigen-functions v 00 . However, to show the expected
patterns I now make a crude approximation.5 Approximate by:
assuming a thin gap so that 2A 1/b where R = 1 + b;
2
where ` = n/b.
145
2
1 2
k2
2
+
(` + k ) (`2 + k2 ) +
= 0,
Re
b
(3.1)
with solution
1 2
k
` + k2 p
,
2
Re
b(` + k2 )
as shown in Figure 3.2. This expression for the eigenvalues is made of two
parts:
=
3 32
Rec =
, at kc = .
3/2
2b
2b
Re2 = b
146
-1
-2
-3
-4
-5
-6
-7
3
k
Figure 3.2: crude approximation to the eigenvalues of perturbations to Couette flow for R = 2 (b = 1) and a Reynolds number of 35. The eigenvalues
are plotted as a function of axial wavenumber k for various radial wavenumbers ` = n/b. Observe that the top eigenvalue branch has just become
unstable for a band of axial wavenumbers k 3.
147
450
400
350
Reynolds number
300
250
200
150
100
50
6
k
10
12
148
As soon as any one mode in a dynamical system becomes unstable, the linearisation becomes inconsistent. The linear picture of the dynamics, that of
exponential behaviour, would predict an exponential growth in the unstable
mode. So very soon the mode would become large enough to invalidate the
smallness assumptions made in the linearisation. To model the dynamics of
the system for these interesting parameter ranges we must investigate the
action of the nonlinearity in the system. This is centre manifold theory.
3.1.5
Exercises
Exercise 3.1:
Consider the axisymmetric flow between a rotating cone
and a fixed disc where the tip of the cone just touches the plate (at
the origin of the coordinate system).6 Solve the NavierStokes and
continuity equations to determine the flow and the fluid stresses. What
is the torque needed to drive this flow at an angular velocity when
the disk and cone have maximum radius R?
Exercise 3.2:
RayleighBenard convection. Temperature T (x, y, z, t) evolves
according to
t
+ (qT ) = 2 T ,
T
6
149
fluid = [1 (T T )]
6
-x
hot, T = T + T
ignore density variations everywhere except where multiplied by
gravity;
suppose only 2D motion in x and y alone (no z dependence and
w = 0);
the top and bottom boundaries are to be stress-free, j = 0.
1. Find the pure conduction state (one of no motion).
2. Non-dimensionalise the equations with respect to reference length
d, reference time d2 /, reference density , and reference temperature T . Write the equations in terms of the Rayleigh number Ra = gTd3 /(), and the Prandtl number Pr = /.
3. Derive and solve the resultant eigen-problem for linearised dynamics near the pure conduction state. (Seek eigen-modes of the
form v sin(ny) cos(kx)et , and similarly for other quantities.)
4. What is the critical Rayleigh number, Rac , such that the conduction state is stable for Ra < Rac , and is unstable for Ra > Rac ?
What is the corresponding horizontal wavenumber k and vertical
mode number n?
150
Figure 3.4: a three-state continuous time Markov chain (the numbers appearing here have been chosen so that the expressions we deal with are very
simple).
3.2
Sorry:
Let pi (t) denote the probability that the system is in state Si at time t. The
dynamical evolution of the system is then governed by the following set of
linear ordinary differential equations:
dp1
dt
Tony Roberts, 24 Apr 2009
= 2p2 ,
=
1
=
2
151
1
1
+ p2 + p3 ,
2
2
1
p2 p3 ;
2
or in matrix form
dp
= Qp
dt
p1
where p = p2
p3
0
2
0
and Q = 0 12 12 .
0 12 12
(3.2)
The column sums of the rate matrix Q are zero; this has two related
implications which must always occur in a Markov chain. Firstly, the evolution of the probability vector p is such as to conserve the total probability
p1 + p2 + p3 = 1 . This is clearly desirable. Secondly, there must exist a
zero eigenvalue of Q. This zero eigenvalue must correspond to a stationary
probability distribution which is a fixed point of the evolution; here it is
p = 1 = (1, 0, 0)8 corresponding to being certainly in S1 .
Ultimately the system ends up being absorbed in S1 , that is, p() = 1 .
However, if is very small then the time to absorption is expected to be very
long and the system has time to reach a (near) balance, here the balance
occurs between states S2 and S3 . This balance is then a significant feature of
the long-term behaviour; indeed sometimes it is the only relevant feature
for example, in models of chemical reactions ?. The question we address is:
how can we find such a long-lasting, though transient, quasi-stationary
probability distribution?
Here proceed via the exact solution of the governing equation (3.2). It is a
linear system and the rate matrix Q has three eigenvalues, namely 1 = 0 ,
2 = , and 3 = 1 . Corresponding eigenvectors to these eigenvalues
7
152
1
e2 = 21 ,
1
2
and e3 = 21 .
12 +
3
X
cn en exp(n t)
n=1
= c1 e1 + c2 e2 exp(t) +
c3 e3 exp(t)
|
{z
}
(3.3)
where the cn are some constants which depend only upon the initial state
of the system.
The key to rational modelling is to throw away any rapidly decaying transients, leaving just those few modes which contribute significantly to the
picture of the long-term evolution.
Here we consider parameter to be small and so the above term in exp(t)
is a term of importance at large time, along with the constant term. However, the term involving exp(t) decays very rapidly and we view it to be of
little importance, that is, negligible.9 Thus we may assert that, independent
of the initial state, the system evolves so that
p(t) c1 e1 + c2 e2 exp(t)
as
t .
Now write the eigenvector e2 = 1 + 2 where 2 is a probability vector (which can always be done by suitably scaling e2 , as e2 must be orthogonal to the left-eigenvector of the the zero eigenvalue (1, 1, 1)). Here
2 = (0, 21 , 12 ) . Thus, upon rearranging the large-time form of p(t), we find
p(t) [c1 c2 exp(t)] 1 + c2 exp(t) 2
|
{z
}
|
{z
}
=s1 (t)
9
as
t .
=s2 (t)
153
Figure 3.5: the trajectories of the Markov process (3.2) showing the approach to a state which is a combination of the absorbing state and a quasistationary state.
Figure 3.6: the simple two-state Markov process which models the long-term
behaviour of (3.2).
where s1 and s2 are functions of timebehaving so that ds1 /dt = +c2 exp(t) =
s2 and ds2 /dt = c2 exp(t) = s2 . To summarise, we can say that
the probability vector, p, exponentially quickly approaches a state of quasibalance
p(t) s1 1 + s2 2 as t ,
(3.4)
where s1 and s2 evolve much slower, according to
d s1
s1
0
.
=
s2
0
s
dt 2
(3.5)
Figure 3.5 shows this geometrically: the plotted trajectories of the dynamical system (3.2) demonstrate the exponentially quick approach to a state
of the form (3.4), and then the slow evolution towards the absorbing state
according to (3.5).
The pleasing interpretation of (3.4) is that s1 is the probability of being
in the absorbing state 1 , and s2 is the probability of being in the quasistationary state 2 . Furthermore, the evolution equation (3.5) is appropriate for the simple two-state Markov process illustrated in Figure 3.6, and
describes the slow ultimate absorption.
The import of all this is as follows. Based on the rationale of being able to ignore rapidly decaying transients, we have replaced the full system (3.2), with
3 degrees of freedoms, by the system (3.5), with two degrees of freedoms.
We view (3.5), in conjunction with (3.4), as a simple model or approximation of (3.2) in that we have thrown away a degrees of freedom. In
larger problems we may throw away many degrees of freedoms and derive
a considerably simplified model.
Tony Roberts, 24 Apr 2009
154
Figure 3.7: a schematic diagram of the steps taken to derive the low degrees
of freedom approximation to the Markov chain; also shown across the top is
the short-cut to be developed via centre manifold theory.
The big question is: how can we do this reduction in the number of degrees
of freedoms, this simplification, without all the intermediate steps which
relied heavily upon explicit exact solutions? The answer comes from centre
manifold theory and, for more general situations, invariant manifold theory.
Our aim is to develop a short route to these simplifying approximations as
illustrated in Figure 3.7; furthermore, a route that can be taken for problems
without a known exact solution, nonlinear problems for example.
3.3
Sorry:
3.3.1
(3.6)
155
0.5
0.4
0.3
0.2
0.1
M
0
-0.1
-0.2
-0.3
-0.4
-0.5
-1
-0.8
-0.6
-0.4
-0.2
0
x
0.2
0.4
0.6
0.8
and y(t) =
1+
1
y0
.
1 exp(t)
For the above simple dynamical system observe,10 in both the algebraic
solutions and in the trajectories, that provided y0 < 1 the system tends
exponentially quickly to the line y = 0 , labelled M in Figure 3.8. Ignoring
this rapidly decaying transient behaviour, the behaviour that emerges from
the whole system is that of one degree-of-freedom, evolving according to x =
ax3 and all the while staying on M. This long-term evolution is generally
10
156
(3.7)
we cannot decide so easily what is the long-term behaviour near the origin
of this dynamical system. Linearizing about the origin we find x = 0 and
y = y and so see that approximately y is exponentially decaying, and that
x does not evolve. But this picture is not satisfactory; because, as in the
system (3.6), x is not constant but evolves algebraically.12 The nonlinear
terms in the equations must modify the too simple a picture obtained from
linearisation.
A coordinate transform simplifies By waving a magic wand (to be
explained eslewhere), I introduce a coordinate transform from x and y vari11
12
157
0.6
0.6
0.4
0.3
0.2
0.6
0.3
0.0
0.3
0.2
0.6
0.3
0.4
0.6
0.6
0.4
0.2
0.0
0.2
0.4
0.6
x
Figure 3.9: coordinate curves of the new (X, Y)-coordinate system that transforms the dynamical system (3.7) into the decoupled system (3.8). Labels
are corresponding coordinate values of some of the coordinate curves.
158
and y = Y Y 2 X3 + ,
(3.8)
These new variables highlight important aspects. First, the new variable Y
must decay exponentially quickly to zero from all initial conditions in some
finite domain around X = 0 as, by continuity, the rate 1+X2X3 +2X4 +
must be negative in some finite domain around X = 0 . Thus Y does not
appear in the long term dynamics, only the dynamics of X emerges in the
long term. Secondly, after Y 0 we see that x = X , so from the X dynamics
the emergent long term dynamics are x ax3 x5 . Lastly, these dynamics
occur for original y X3 x3 which is the curved attractor seen in
numerical simulations.
Problem Such beautiful coordinate transforms are impractical in applications. the reason is that there are usually a large, often infinite, number
of exponentially decaying transients in applications. Correspondingly, there
are usually only relatively few modes that describe the long term dynamics.
It is impractical to transform a large/infinite number of transient modes
which we ultimately just set to zero, when all we want is the relatively
few emergent modes. Centre manifold theory provides an equivalent, but
economical method.
Focus only on the emergent dynamics The centre manifold theory
which we now turn to asserts that (3.7) has an exponentially attractive
curve M on which all the long-term evolution takes place (Y = 0 in Fig
ure 3.9). The curve M may be described by y = h(x), with h(x) = O x2 as
x 0 . In essence, the theory says that the linear picture is largely correct
except that it is bent by nonlinear terms as seen in Figure 3.10.
Tony Roberts, 27 Aug 2009
159
0.5
0.4
0.3
0.2
0.1
0
-0.1
M
-0.2
-0.3
-0.4
-0.5
-1
-0.8
-0.6
-0.4
-0.2
0
x
0.2
0.4
0.6
0.8
Figure 3.10: trajectories of the dynamical system (3.7) showing the rapid
approach to the curved centre manifold M (3.9).
160
as x 0 ,
x = ax + x h(x) ax x5 .
3
(3.9)
Furthermore, the theory asserts that for all solutions of the full system (3.7),
with initial point (x0 , y0 ) sufficiently near to the origin, there exists a solution, x
^(t) say, of (3.9) such that13
x(t) = x
^(t) + O et
y(t) = h(^
x(t)) + O et
as
t .
3.3.2
(3.10)
We often say that the system settles onto M exponentially quickly, that is, the centre
manifold M emerges quickly.
14
These ensure that there is a linear coordinate transform, u (x, y) , such that the
linear system, u = Lu , separates into the centre modes x = Ax and the exponentially
decaying (stable) modes y = By .
Tony Roberts, 27 Aug 2009
161
6=()
v
v
<()
v
v
v
Figure 3.11: the spectrum of the linearized version of the dynamical system (3.10).
m eigenvalues with zero real part, 1 , . . . , m , and associated eigenvectors, e1 , . . . , em (eigenvalues repeated in the list according to their
multiplicity, and any generalised eigenvectors included);
the remaining eigenvalues, m+1 , . . . , n , have negative real part that
are bounded away from 0 by .
Thus the spectrum of the linearized problem is as shown in Figure 3.11, We
normally assume that the functions f is at least twice differentiable at the
fi
= 0 at the origin.
origin; whence f is a nonlinear function if fi = u
j
Definition 3.1 A set M Rn is said to be an invariant manifold of
the dynamical system (3.10) if for any u0 M the solution of (3.10) with
u(0) = u0 stays in M for some finite time, that is, u(t) M for 0 t <
T .15
15
Some people require the trajectory to stay in M for all time. But this seems overly
restrictive.
Tony Roberts, 27 Aug 2009
162
Note: in its most abstract, this definition says the invariant manifold M is
just the union of a set of trajectories of the system (3.10). But it is usually
used with some implicit properties such as smoothness, completeness or
analyticity near the origin.
For example, if f = 0 , so that the above dynamical system (3.10) is linear,
then there exists two interesting invariant manifolds:
the centre subspace Ec = span{e1 , . . . , em } (= {x}) , consisting of all
the non-decaying modes; and
the stable subspace Es = span{em+1 , . . . , en } (= {y}) , consisting of all
the exponentially decaying modes.
The centre subspace is interesting because for all solutions to the linear
problem, starting from any initial condition, tend exponentially quickly, at
least as fast as exp(t), to the centre subspace Ec . Once on Ec they
evolve according to some equation x = Gx where G is the restriction of L
to Ec . That is, the behaviour on the low dimensional (m dimensional) centre
subspace determines the long-term dynamics of the original linear system.
This linear picture is qualitatively the same as that discovered for Couette
flow exactly at the critical Reynolds number Re. There, all modes decay
except for one mode (m = 1) of axial wavenumber kc .16
As discussed informally in the previous subsection, analogous results hold
for the fully nonlinear system (3.10).
Theorem 3.2 (existence) There exists a neighbourhood U of the origin
an m dimensional invariant manifold for (3.10), M, with tangent space Ec
at the origin, in the form u = v(s) , that is locations on M are parameterized
by s. The flow on M is governed by the m dimensional dynamical system
s = Gs + g (s) ,
(3.11)
163
We aim to model large scale systems with many interacting modes. The
best understood of these are partial differential equations.
Example 3.4: reaction-diffusion
Consider the reaction-diffusion pde
u
2 u
t = x2 + f(u) for some strictly nonlinear reaction f(u), that is,
f(0) = f 0 (0) = 0. Suppose this reaction-diffusion takes place on a finite interval, say [0, ] for definiteness, with insulating boundary conditions; that is, u
x = 0 at x = 0, .
Since f(0) = 0 the origin u = 0 is an equilibrium. Since the reaction is
strictly nonlinear, the pde linearises about the origin to the diffusion
Tony Roberts, 27 Aug 2009
164
u
u
pde u
t = x2 with insulating boundary conditions x = 0 at x = 0, .
Standard separation of variables shows that this system has linearly
independent solutions uk = ek t cos kx for integer k = 0, 1, 2, . . . and
eigenvalues k = k2 . Since there is one zero eigenvalue, there exists a
slow manifold tangent to the slow subspace of spatially constant solutions cos 0x. Here the dynamics are sufficiently simple that we can
assert the following: if the nonlinearity f has no explicit x dependence,
then the slow manifold that emerges is u = s(t) (constant in x) such
that s = f(s) where this evolution has just one degree of freedom, s(t),
instead of the infinite number of modes inherent in the original pde.
165
2
u
Example 3.6: infinite domain diffusion Consider diffusion u
t = x2
on an infinite domain with boundary conditions that the field u be
bounded. Separation of variables gives a Fourier integral decomposition that independent modes are et+ikx for eigenvalue = k2 for,
now, all real wavenumbers k. Although there is a zero eigenvalue,
with wavenumber k = 0, there is no centre manifold as there are other
modes with eigenvalues, albeit all negative, arbitrarily close to zero.
Thus there is no clear separation of a slow centre manifold away from
the other decaying dynamics in the system.
u
nonlinear diffusion u
=
(u
)
t
x
x with insulating boundary condiu
tions x = 0 at x = 0, . Here there is a whole subspace of equilibria:
u = U = constant. Consider perturbations u 0 to each equilibria:
0
2 u 0
u(x, t) = U + u 0 (x, t). Then the linearised pde becomes u
t = U x2
0
such that u
x = 0 at x = 0, . Separation of variables gives modes
uk0 = ek t cos kx for wavenumber k = 0, 1, 2, . . . and eigenvalues k =
Uk2 . That is, provided U > 0, there is a centre (slow) manifold
about each an every equilibria. Necessarily, the equilibria with U > 0
forms at least part of the slow manifold.
166
167
y = y .
(3.13)
3.3.3
y = y ,
(3.14)
Tony Roberts, 27 Aug 2009
168
To show that this approach to a model solution is not obviously the case,
consider the following example.
Example 3.10:
y = y ,
The discrepancy in the evolution in the previous example occurs despite the
exponential attraction to y = 0 . However, the following theorem (Iooss &
Adelmeyer 1992, e.g.) guarantees the emergence and long term accuracy of
models based upon centre manifolds.
Theorem 3.4 (relevance) The neighbourhood U may be chosen so that
all solutions of (3.10) staying in U tend exponentially to some solution
17
169
of (3.11). That is, for all solutions u(t) U for all t 0 there exists
a solution s(t) of the model (3.11) such that
0
u(t) = v (s(t)) + O e t
as t ,
(3.16)
where 0 may be estimated as , the upper bound on the negative eigenvalues of the linear operator L.
This theorem is crucial; it asserts that for a wide variety of initial conditions
the full system decays exponentially quickly to a solution predicted by the
low degrees of freedom model. That is, the centre manifold emerges.
The trajectories u(t) have to be sufficiently small for this theorem to apply,
but in practice sufficiently small can be quite generous. For example, consider the system (3.6). If a 0 , then the origin is stable and we observe
that all trajectories with initial y(0) < 1 asymptote exponentially quickly
to the slow manifold y = 0 . Thus the entire half-plane y < 1 are initial conditions to which the conclusions of this theorem apply; the conclusions need
not just apply to some small neighbourhood of the origin. Every solution in
some finite domain is exponentially quickly modelled by the model.
In constrast, the theory of singular perturbations provides significantly less
assurance. Let the (small) parameter measure the ratio in timescales of
the decay of the fast variables (y) and the decay of the slow variables (x).
Tikhonovs theorem (Verhulst 2005, p.99, e.g.) asserts that in the limit
as 0 the solution of the full dynamics tends to the solution of the low
dimensional model in a finite time interval. Whereas the centre manifold
Relevance Theorem 3.4 asserts the exponential attractiveness on a specific
timescale, of the low dimensional model for finite (characteristic of the
size of the neighbourhood U), and for all time that the dynamics stays
within U. I contend that centre manifold methods are much more powerful
than singular perturbation methods.
170
manifold pushes the solution a long way away in a finite time, here to infinity.
I contend that even if the origin is unstable, in many cases the low degrees
of freedom model is still relevant, albeit perhaps for a limited time and for
limited initial conditions.
I record an older version (Carr 1981, e.g.) of the relevance theorem as two
lemmas.
Lemma 3.5 If the origin of (3.10) is stable, then for all u(0) sufficiently
small there exists a solution s(t) of the model (3.11) such that the exponential
attraction (3.16) holds.
This follows immediately from Theorem 3.4 because if the origin is stable,
then all sufficiently small u(0) lead to trajectories that stay in U, hence
satisfies the requirements of the theorem, and so must exponentially quickly
approach a solution of the model.
Lemma 3.6 The origin of (3.10) is stable if and only if the origin of (3.11)
is stable.
This lemma requires more work to prove.
Some researchers, such as Robinson (1996), use the term asymptotic completeness to address this issue of relevance of a model.
Many want to create and use models of slow dynamics by neglecting fast
oscillations, rather than fast decay. It is straightforward to create such
models using the methods described next. However, there is no emergence
theorem to support the precise relevance of such a slow model.
Example 3.11: ignoring fast oscillations Consider the simple example
x = x3 + xy2 + xz2 , y = z , z = y .
Linearising, x is a slow variable, and y and z undergo fast oscillations.
The slow manifold, because the fast variables are here independent of
the slow x, is simply to set the fast modes to zero, y = z = 0 . Hence
the slow model is x = x3 and predicts algebraic decay to zero.
Tony Roberts, 27 Aug 2009
171
However, such decay is not typical for any nearby solution. Convert
y and z to radius and angle coordinates r(t) and (t) then the system
becomes
x = x3 + xr2 , r = 0 , = 1 .
Observe that solutions in the neighbourhood of the slow manifold
y = z = 0, that is r = 0 , are not quantitatively modelled, nor even
qualitatively modelled, by the slow manifold solutions: instead nearby
solutions are attracted, like ert , to one of the two finite amplitude
equilibria at x = r . The presence of fast oscillations generally causes
full solutions to drift away from predictions of the slow model.
3.3.4
y = y + y2 + xy x3 ,
such that
x ax3 x5 .
Tony Roberts, 27 Aug 2009
172
v
[Gs + g (s)] .
s
(3.17)
173
where T v(s) = span{v/sj } is the tangent space of v at the point parameterized by s. More
crudely, this requires that v is quadratically near Ec :
2
v Ec + O |s| as s 0 . This condition ensures that the constructed
manifold truly contains the whole of the centre modes, and nothing but the
centre modes. Without it, the solution of (3.17) could be based on an almost arbitrary mixture of linear modes. Indeed, other invariant manifolds
of note satisfy (3.17) but are tangent to different vector subspaces as s 0 .
For example: the stable manifold has tangent space Es ; the centre-unstable
manifold has tangent space Ec Eu ; and lastly the slow manifold has tangent space E0 = span{e0j | j = 0} , that is, the slow manifold is composed
of all modes corresponding to precisely zero eigenvalues, not the real-part
zero. Most centre manifolds we consider actually will be slow manifolds.
It is typically impossible to find exact solutions to (3.17) as finding exact
solutions is nearly equivalent to solving the original system (3.10). However, in applications we may approximate M to any desired accuracy by
approximately solving (3.17).
The errors in the solution are determined by the residual of the governing
equations. For functions : Rm Rn (imagine that approximates v)
and : Rm Rm (imagine that approximates Gs + g(s)) define the
residual
174
Example 3.13:
Consider (3.7) and seek the slow manifold parametrised
by x, that is, y = h(x) (equivalently, x = s and y = h(s)) where
h is at least quadratic in x so that the constructed centre manifold is
tangent to Ec , the x-axis, at the origin. Then the x equation asserts
the evolution will be x = ax3 + x2 h , whence the y equation becomes
i
h h 3
ax + x2 h = h + h2 + xh x3 .
x
Rearrange so that the dominant linear term on the right-hand side is
by itself on the left-hand side:
h=
i
h h 3
ax + x2 h + h2 + xh x3 .
x
This simple sort of iteration will work for any dynamical system in the
separated form (3.12).
The approximation theorem guarantees us that the order of the difference
between two successive iterates is the order of the error in the first of the
two iterates. For example, here we are assured that h(1) describes
the centre
4
(2)
5
manifold M with errors O x , whereas h has errors O x . Often, by
iterating in this manner we obtain corrections accurate to one higher order
in x each iteration; with symmetry we may obtain two orders of accuracy
each iteration. However, sometimes, particularly associated with generalised
eigenvectors in the centre subspace, we need two or more iterations for each
order improvement in the accuracy.
Tony Roberts, 27 Aug 2009
175
3.3.5
If it were not for one implicit generalisation, all the above theory would be
extremely academic. After all, how many systems of interest are exactly
at critical? that is, with the specifically required spectrum? The answer is
almost none. For example, in the TaylorCouette problem it is only precisely
at the critical Reynolds number that in the linear dynamics about Couette
flow the spectrum has all eigenvalues negative except one which is zero:
for any smaller Reynolds number all eigenvalues are negative; for any larger
Reynolds number there are positive eigenvalues as well as negative and zero.
At first sight, centre manifold theory cannot be applied usefully. The issue
is one of relevance.
The requirement that the real parts of the eigenvalues be precisely zero in
order for the corresponding mode to be important in the long term evolution
is too restrictive in practice. Instead, surely modes which decay exponentially at a small or very small decay rate will be important over long times.
Tony Roberts, 27 Aug 2009
176
A simple bifurcation
Suppose we wanted to analyse the definite system
x = 0.1 x xy
and y = y + x2 .
and y = y + x2 ,
(3.19)
177
Let us seek a slow centre manifold of the form (a, x, y) = (a, x, h(a, x)). The
slow manifold is a two dimensional surface in the axy-space, as shown in FigTony Roberts, 27 Aug 2009
178
3
2
1
0
1.0
0.5
0.6
0.0
0.2
x
0.2
0.6
a
1.0
0.5
Figure 3.12: approximate centre manifold of (3.19) in axy-space. Also superimposed is the pitchfork of fixed points.
179
ure 3.12, because there are two zero eigenvalues in the linearized equations.
Substituting this form we deduce that h must satisfy
h = x2
h
x(a h) .
x
The finite neighbourhood U of attraction to the slow manifold extends out into the
y direction, but for simplicity Figure 3.14 does not attempt to plot the y structure.
21
A Landau equation for the amplitude x = s .
Tony Roberts, 27 Aug 2009
180
0.2
0.0
0.2
0.4
0.6
1.0 0.8 0.6 0.4 0.2
0.0
0.2
0.4
0.6
0.8
1.0
x
Figure 3.13: X and Y coordinate curves in the xy-plane that simplify the dynamics of the example system (3.19) for parameter a = 0.1 . The particular
coordinate curves plotted are 0.5 : 0.1 : 0.5 .
then the origin is unstable, and the solutions are attracted to one of the
181
A coordinate transform establishes domain of attraction For simple systems like (3.19), normal form coordinate transforms illustrate more
details. By arguments developed elsewhere, consider the system (3.19) in
new variables X and Y where, in an asymptotic expansion,
x = X + XY + 12 XY 2 + 16 XY 3 + 2X3 Y + ,
y = Y + (1 2a + 4a2 )X2 2X2 Y 2 + 2X4 + .
(3.20)
Figure 3.13 plots the coordinate curves of the XY-system. Some tedious algebra developed elsewhere shows that in these new variables the system (3.19)
becomes
X = aX (1 2a + 4a2 )X3 2X5 + ,
Y = (1 + 2X2 + 4X4 + )Y .
(3.21)
Initial conditions a long way outside this region are also attracted to the slow manifold.
However, their transient dynamics are not simple decay, they are complicated. Thus their
attraction to the slow manifold lies outside this theory.
Tony Roberts, 27 Aug 2009
182
turn the mathematics, the predictions and domain are essentially the same
as the following two examples indicate.
Example 3.14: equivalent global manifold As in many other asymptotic methods, one might choose to scale the dependent variables, here
x and y, to obtain a global model. Lets try it, focussing on just
a 0 for simplicity. Here the natural scaling is to change variables to
X(t) and Y(t) where x = aX(t) and y = aY(t). Substitute into (3.19)
and Y = Y + X2 .
0
0
0
0 .
L = s s3 0
0
2s 1
Being a triangular matrix the eigenvalues are along the diagonal, namely
0 (twice) and 1: that is, at each equilibria there is a slow manifold.
The finite neighbourhoods of validity along the set of equilibria overlap and merge to form a globally valid slow manifold in X; the finite
neighbourhoods in the direction are still just local in a. Being global
in X sounds wonderful, surely we have gained something. But alas no.
The slow manifold will be Y = X2 + O a so that the evolution is
X = a(X X3 ) + O a2 which as before predicts the bifurcation to
There is also the equilibria (a, 0, 0) but these have eigenvalues {0, a, 1} which only
have a nontrivial slow manifold in the case a = 0 that is encompassed by the body of this
example.
23
183
1.5
1.5
1.0
1.0
0.5
0.5
0.0
0.0
0.5
0.5
1.0
1.0
1.5
2.01.51.00.50.0 0.5 1.0 1.5 2.0
1.5
2.01.51.00.50.0
p0.5 1.0 1.5 2.0
X = x/
|a|
Figure 3.14: schematic diagram of the finite domain U of validity of the slow
manifold model of the bifurcation in the system (3.19): left, shows the local
domain of the pitchfork appropriate to straightforward analysis; right, shows
the same global domain of the pitchfork when scaled as in Example 3.14.
Example 3.15: singular perturbation In many applications one recognises that some physical processes are much more rapid than others;
for example, some chemical reaction may take place on microseconds
whereas other interesting reactions take place on milliseconds. The
singular perturbation approach is to replace the large rates of the fast
dynamics by a parameter such as 1/. Then one deduces a model in
the limit of small parameter . Lets do the equivalent here.
Here the fast rates are the order 1 rates of the y equation in the
Tony Roberts, 27 Aug 2009
184
1.5
1.0
0.5
0.0
1.50
0.50
0.50
a
1.50
Figure 3.15: schematic diagram of the finite domain U, below the curved
surface, of validity of the singularly perturbed version of the bifurcation
in (3.19). The physically relevant = 1 (shown) results in a finite domain
of physical validity in the ax-plane. For reference, the pitchfork bifurcation
is shown in the ax-plane.
185
and y =
1
(y + x2 ),
dx
= (ax xy)
d
and
dy
= y + x2 .
d
0
0
0
0 .
L = as s3 0
0
2s 1
The same conclusions follow: there exists a relevant slow manifold,
approximately y x2 , globally valid in x and local in . A difference
here is that this existence and relevance is true for all parameters a.
Figure 3.15 schematically shows this global domain of validity of this
singular perturbation model. But the physically relevant parameter
is = 1 , shown by the flat plane in Figure 3.15: the intersection of
this physical plane with the global domain of singular perturbations
results in the same finite domain of physical validity in the ax-plane
as determined earlier.
The finite domain of validity of the centre manifold theorems arise from the
intrinsic dynamics of the system. Thus expect that the mystic incantations
of various methodologies are not going magically change the validity of the
modelling when evaluated at corresponding physical parameters.
Tony Roberts, 27 Aug 2009
186
The master and slaved variables of Hakens synergetics Haken (1983, 1996).
187
2. The usual initial linear approximation to the shape of the slow manifold is simply the tangent approximation at the origin, namely that
y:=0;
3. Now iterate in a loop. Control the truncation of the asymptotic approximation by telling reduce to discard any factor in a2 or higher
and any factor in x4 or higher; thus all expressions are computed to
an error of O a2 , x4 .25
let {a^2=>0, x^4=>0};
repeat begin
. . .
end until res=0;
Within the above loop:
(a) The y equation must be satisfied by driving its residual to zero.
For whatever approximation y contains, since reduce knows how
x depends upon time, the residual is computed as
res:=df(y,t)+y-x^2;
(b) Then, because of the simplicity of this example, the correction to
the shape of the slow manifold is simply
y:=y-res;
Executing this program recomputes the results determined earlier by hand.
Of course with computer algebra one may partake in an orgy of computation
just for the fun of it. Here we compute the asymptotic expressions to excruciatingly high-order by changing the order at which terms are discarded.
For example, using let {a=>0,x^19=>0}; we get the computer to tell us
y = x2 + 2x4 + 12x6 + 112x8 + 1, 360x10 + 19, 872x12 + 335, 104x14
+ 6, 359, 040x16 + 133, 560, 576x18 + O x20 , a .
25
reduce is precisely literal in how it matches its patterns except for this one case of
the pattern of a simple variable raised to some power being replaced by zero. In this case
it also replaces all higher powers by zero.
Tony Roberts, 27 Aug 2009
188
(3.22)
where
is a vector of ` parameters;
L is independent of the parameters and u and has the same eigenstructure as before;
the nonlinear terms f are quadratic at the origin of the parametervariable space, (, u).
Clearly, this extended system satisfies the requirements for the existence of
a centre manifold. Seeking the centre manifold in the form
u = v(, s)
where s = G + g(, s) ,
we use the following generalised theorem to test the accuracy of approximations (, s) and (, s), (Li 1999, Chapt. 2).
Theorem 3.8 (approximation) If T (0, 0) = R` Ec and Res(, ) =
O q + sp as (, s) 0 for some p, q > 1 (where Res is as before and
and s denote || and |s| respectively), then
v(, s) = (, s) + O q + sp
and Gs + g(, s) = (, s) + O q + sp as (, s) 0 .
The equivalent statement for errors O q , sp is also true.
Tony Roberts, 27 Aug 2009
189
As you will see, in using computer algebra we usually find it most convenient
to work to residuals and errors O p , sq .
3.3.6
Current centre manifold theorems support the construction of low dimensional models of dynamical systems. The existence of a model is based
upon the structure of the eigenvalues (the spectrum) of an equilibrium. The
Tony Roberts, 27 Aug 2009
190
relevance theorem ensures that the model is valid apart from initial transients. Approximations can be found so straightforwardly that we proceed
to program computers to handle most of the messy details.
One apparent restriction is that the theorems are local to the anchoring fixed
point. However, in applications sufficiently small may be quite generous
in size. Importantly, the theory guarantees the existence and relevance of
the model for all time in a finite sized domain; in contrast, other methodologies only assert properties for a finite time and often only with some
uncontrollable error.
Exercise 3.17: A slow manifold
Consider
x = xy ,
y = 2y + z + x2 ,
z = y z + x2 .
Deduce that there exists a one dimensional slow manifold: y
2x2 and z 3x2 .
What are the next order corrections to the shape of the slow
manifold?
What is the corresponding evolution on the slow manifold? Does
theory absolutely guarantee that this 1D model is valid exponentially quickly?
191
x = ax3 + x2 y ,
y = y + y2 + xy x3 ,
and
a = 0 ,
x = ax3 + x2 y ,
y = y + y2 + xy x3 .
Write and run a reduce program to construct approximations to the
slow manifolds of these two systems. Compare and contrast the assurances the three centre manifold theorems give to the two slow manifold
models.
192
193
3.4
Warning:
194
3.4.1
Burgers pitchfork
u(0) = u() = 0 ,
(3.23)
for some field u(x, t). View this as an infinite dimensional dynamical system,
the state space being the set of all functions u(x) on [0, ].
For all values of the parameter there is a fixed point at the origin, that is,
a trivial equilibrium state is u = 0. The linearisation of the equation about
this equilibrium, namely ut = (1 + )u + uxx , has constant coefficients
in x and t. Hence it has the trigonometric modes sin kx with associated
eigenvalues k = 1 k2 + for wavenumbers k = 1, 2, . . .. Thus the origin
u = 0 becomes unstable as crosses zero, because the k = 1 mode, sin x,
begins to grow exponentially, and the system undergoes a bifurcation.
To find the details of this pitchfork bifurcation is a straightforward task for
centre manifold theory. Linearly, exactly at critical, = 0, all modes decay
exponentially quickly except for the critical mode sin x; it has a zero decay
rate and therefore is long lasting; by the Existence Theorem 3.2 we are assured that there exists a slow manifold. Nonlinearly, and for and u(x)
near 0, all modes decay exponentially except for the critical mode which
slowly evolves. Thus, neglecting the exponentially decaying transients, we
accurately model the dynamics solely in terms of the evolution of the amplitude of the sin x mode; define a to be its amplitude. By the Relevance
Theorem 3.4, after appending
= 0,
t
Tony Roberts, 24 Apr 2009
195
(3.24)
(a, ) ;
such that a g
u
u
2 u
+u
(1 + )u 2
t
x
x
v 2 v
v
(1 + )v + v
g
.
a
x x2
(3.25)
For example, in the first iteration, starting from the initial approximation (3.24), the Res = a sin xa2 sin x cosx . In any iteration the residual
is of some magnitude, say Res = O p + aq , for some exponents p and q.
The aim of each iteration is to improve the residual (increase p and q) so
that, by the Approximation Theorem, we improve the accuracy of the model.
We seek to find small corrections, indicated by primes, so that
u v(a, ) + v 0 (a, )
(a, ) + g 0 (a, ) ,
such that a g
196
v 2 v
v 0
v
v 0 2 v 0
+ 2 + (1 + )v 0 v 0
v
v0
+
,
x x
x
x
x
x2
that is,
, )+
Res (v, g
v 0 0
v
v 0
v 0 2 v 0
v 0 v 0
+
g +
g
g = (1+)v 0 v 0 v
v 0
+
.
a
a
a
x
x
x x2
It is impossible to solve this for the perfect corrections in one step. We seek
an approximate equation for the corrections of O p + aq by the following.
1. Ignore products of corrections (primed quantities) because they will
be much smaller, O 2p + a2q , than the sought corrections. Then,
for example, in the first iteration we would like to solve
Res + sin x g 0 = (1 + )v 0 a cos x v 0 a sin x
v 0 2 v 0
+
.
x
x2
But such an equation is still too difficult for analytic solution in general
as it involves the variables a and in its coefficients of the unknown
corrections v 0 .
2. Instead, also recognise that near the anchoring fixed point at the
origin, both a and are small. Thus terms such as v 0 and a cos x v 0
are small in comparison to v 0 terms and may be neglected to lead to
v0 +
2 v 0
= sin x g 0 + Res .
x2
2 v 0
1
= g 0 sin x a sin x + a2 sin 2x .
2
x
2
0 , with boundary
We wish to solve such equations for v 0 . However, v 0 + vxx
conditions v 0 (0) = v 0 () = 0 , is singular as the critical mode, sin x, is
197
1 3 2
1
a sin 2x + a3 sin 3x + O 2 + a4 ,
6
32
(3.26)
1 3
a + O 2 + a4 .
12
(3.27)
3.4.2
See burger.red
Tony Roberts, 24 Apr 2009
198
Algorithm 3.2 outline of the general iteration to construct a centre manifold model.
1: preliminaries;
2: initial linear approximation;
3: repeat
4:
compute residual;
5:
find solvability condition;
6:
compute correction to the slow manifold;
7:
update approximations;
8: until residual is small enough.
tern matching and replacement capabilities through its operator and let
statements.
1. The preliminaries are the following.
Improve the appearance of the reduce printed output.
on div; off allfac; on revpri; factor a;
Define the operator linv to act as the inverse of L:
operator linv; linear linv;
let linv(sin(~k*x),x) => sin(k*x)/(1-k^2);
declaring it linear tells reduce to expand sums and products in the first argument and to only leave functions of the
second argument inside the operator, for example, linv(a sin x+
2a2 sin 2x,x) is expanded to
alinv(sin x,x)+2a2 linv(sin 2x,x);
the let statement defines the action of the operator as the
1
0 = sin kx, namely v 0 =
solution to v 0 + vxx
sin kx, the
1k2
tilde before the k on the left-hand side matches any pattern
(no action is defined for the singular case k = 1 because the
pattern sin(~k*x) does not match sin(x)any appearance
of linv(sin(x),x) usefully signals an error).
Tony Roberts, 24 Apr 2009
199
200
a3
+ O a5 .
12(1 + /3)
Completely different dynamical systems with the same linear structure may
be analysed simply by changing the computation of the residual.
By way of comparison, Rand & Armbruster (1987) macsyma code
[pp.2734] for constructing the slow manifold of a finite-dimensional dynamical system, based upon power series manipulation, uses 53 lines of
active code (although some are for input of the dynamical equations)
whereas the above algorithm has only 14 lines of active code.
3.4.3
201
(3.28)
where the columns of E = [ej ] span the slow eigenspace E0 , and where G
may be chosen in Jordan form in the case of generalised eigenvectors; G is
zero if there are no generalised eigenvectors. In order to model the nonlinear
Tony Roberts, 24 Apr 2009
202
6Res
``
b
`
a
ba
a
bb
bb
b
b
c
b
c
cb
@bb
pb b
@
p b bL
@
p eb b
p
b
p
e bb b
p
p
e sM bbbbp
p
p
b
e
b
p
b
p
e
b e
p
b @
p
b
p
b@
@ pp
b
c
b
c
c
, g
203
(s, ) ;
such that s g
v
Lv
f(v
, ) = O p + sq , (3.30)
g
s
for some order of error, p and q, and where s and denote |s| and || respectively. Seek small corrections, indicated by primes, to the approximation
so that
(s, ) + v 0 (s, )
uv
such that
(s, ) + g 0 (s, ) ,
s g
29
204
v
f 0
v
0
+
= Lv
+ Lv + f(v
, ) +
g +
g
g
v ;
s
s
s
u v
and replacing tilde quantities by their zeroth order approximation, in
s and , wherever they are
multiplied by a correction factor, introducing errors O p+1 + sq+1 .30
Thus we wish to solve
v
v 0
+ Eg 0 +
+ Lv 0 + f(v
, ) .
g
Gs = Lv
s
s
In general, the term v 0 /s Gs must be retained because the operator
/s Gs is of zeroth order in smultiplication by s raises the order by one,
but the gradient /s drops the order by one. Thus, rearranging and
v 0
v
, g
, ) =
f(v
, ) ,
Gs Eg 0 = Res (v
Lv
s
t
(3.31)
for the primed correction quantities. The great advantage of this iterative
approach is that the right-hand side, which drives the corrections, is simply
30
The approximation of replacing tilde quantities by their zeroth order slows the iteration convergence to linear. The iteration could converge quadratically if we retained the
tilde quantities. However, I know of no practical problem where this quadratic convergence
can be realised for slow manifolds.
205
the residual of the governing equation (3.28) evaluated at the current approximation. Thus at any iteration we just deal with physically meaningful
expressions; all the detailed algebraic machinations of asymptotic expansions
in other procedures is absent.
It is not obvious, but in the case of the m critical eigenvalues having no
imaginary component, provided it is arranged so that G is in Jordan form,
as is often physically appealing, we may significantly simplify the algorithm
0
by also neglecting the term vs Gs at a cost of increasing the number of
iterations needed by a factor no more than m, the multiplicity of the zero
eigenvalue of L. In the remaining analysis, assume this neglect for simplicity.
The more general case is deferred until Chapter 6 on Hopf bifurcations.
The main detail is then to solve equations of the form
Lv 0 Eg 0 = Res ,
(3.32)
for some given residual Res. Recognise that there are more unknowns than
components in this equation; its solution is not unique. The freedom comes
from the fact that we can parameterize the slow manifold via the amplitudes s in an almost arbitrary manner. The freedom is only resolved by giving a precise meaning to the m amplitudes s. For example, often one does
define s to be precisely the modal amplitudes, that is s = hzj , ui for eigenvectors zj of the adjoint of L; in which case we seek corrections v 0 which are
orthogonal to the eigenvectors zj , that is, 0 = hzj , v 0 i. More general definitions, such as an energy related amplitude, give rise to similar considerations
to those that follow. There are two approaches to solving (3.32).
1. Numerically, it is easiest to adjoin the amplitude condition to the
equation and solve
L E v 0
Res
=
,
ZT 0
g0
0
where Z = [zj ].
2. However, algebraically it is usually more convenient to adopt the following procedure (which may be familiar to you as part of other asymptotic methods). Rewrite (3.32) as Lv 0 = Eg 0 + Res and recognise that
Tony Roberts, 24 Apr 2009
206
The last step of each iteration is to update the approximations for the slow
manifold shape and the evolution thereon.
Example 3.22: quasi-stationary distribution Consider a continuoustime Markov chain with three states, labelled S1 , S2 and S3 . Let pi (t)
denote the probability that the system is in state Si at time t. The
dynamical evolution of the system is then governed by the following
set of linear ordinary differential equations:
p 1 = 2p2 ,
p 2 = 12 + p2 + 21 p3 ,
p 3 = 21 p2 12 p3 .
Adjoin = 0 and write in matrix-vector form:
p = Lp + f
(3.33)
where
p1
p=
p2 ,
p3
0
0
L=
0
0
0 0
0 0
0 21
0 12
0
0
1
2
12
0
2
and f =
1 p2 .
1
207
and s 2 = s2 + O 2 .
Tony Roberts, 24 Apr 2009
208
3.4.4
argue for the existence of a slow manifold based upon the linear
dynamics near the origin;
find a linear approximation for the shape of the slow manifold;
Tony Roberts, 24 Apr 2009
209
Exercise 3.25:
For what values of parameter does the nonlinear advection and diffusion stabilise the nonlinear reaction in
u
u
2 u
+u
= u + u3 + 2 ?
t
x
x
Modify the computer algebra from Section 3.4.2 to answer.
Tony Roberts, 24 Apr 2009
210
Exercise 3.26:
1
32 (1
)a3 sin 3x + O a4 , 2 such
i
1h
1 + cos ( )2 /2 + (2 + 2 cos ( ) sin ) cos x
0,
0<x<
+
,
x sin(x ) , < x <
R
whence the solution is u 0 = 0 G(x, )R() d. If you have done everything else right, the boundary conditions and amplitude conditions
will hold.
3
Answer: u = a sin x+ 13+7
a2 sin 2x+ 139
sin 3x+O a4 , (1)2
144
3608 a
such that a = (1 )a
137 3
288 a
+ O a4 , (1 )2 .
211
For some values of the parameters these equations exhibit chaos. But
before the onset of chaos, the zero solution loses stability as the parameter crosses 1.
1. Argue for the existence of a slow manifold for near 1.
for the shape of the slow
2. Find a linear approximation u = u
manifold given that a parameter of the slow manifold is to be
s = x.
3. Apply the iteration scheme to derive a low dimensional model
for the Lorenz equations for near 1, for the particular values
= = 1, and where a parameter of the description is s = x.
Hence describe qualitatively the structure of the bifurcation that
occurs as crosses one.
= sin + 2 ,
t
x
(0) = () = 0 .
Find the critical value of the dissipation for which the spectrum indicates centre manifold theory may be applied. Construct the Landau
equation model that displays the pitchfork bifurcation as the dissipation is varied through the critical value.
212
3.5
Sorry:
We now turn attention back to TaylorCouette flow, and seek to analyse the
nonlinear dynamics near the onset of instability. We construct a model for
the dynamics of Taylor vortices and show that the amplitude of the vortices
saturates in the manner of a pitchfork bifurcation.
We would like to use computer algebra to do the tedious parts of the analysis.
By assuming a thin gap between the cylinders and by neglecting some lesser
Tony Roberts, 24 Apr 2009
213
terms in the fluid equations and boundary conditions, this is possible. This
rough model is developed in 3.5.1 and indeed models crudely the dynamics
of Taylor vortices.
However, computer algebra packages are still not sufficiently flexible to meet
all our needs. Thus, as here, to solve precisely the problem required we often need to use more human analysis. In 3.5.2 we explicitly construct the
asymptotic solution and numerically solve the problems for the spatial structure of the physical fields on the slow manifold. Thus we derive a numerically
exact model predicting the pitchfork bifurcation of Taylor vortices.
The form of the fluid equations
Here we compare the fluid equations with the general form assumed by
centre manifold theory. Despite the differences in detail, we may still apply
the techniques with almost no modification.
Consider the non-dimensional fluid equations given in 3.1.3:
q
Re
+ q q = p + 2 q ,
t
q = 0.
To analyse the nonlinear dynamics near the onset of the Taylor vortex flow,
we shift to equations describing the perturbations (primed quantities) to
Couette flow (superscript 0):
2
R
1
0
r e .
q = 2
R 1 r
31
Seeking
q = q0 + q 0 (r, , z, t) ,
p = p0 + p 0 (r, , z, t) ,
214
We now put these equations into the form (well almost) for the use of centre
manifold theory.
First, recognise that it is only exactly at the critical Reynolds number Rec that the spectrum can fit that required for the theory. Thus
we introduce a parameter = Re Rec measuring the departure from
critical of the actual system. As varies we will discern a pitchfork
bifurcation exactly as in the simple system of the previous section.
Second, rearrange the terms so that we separate the linear and nonlinear parts of the equations, viz
u
z }| {
q 0
Rec
=
t
f(, u)
Lu
z
}|
{
}|
{
z
0
2 0
(Rec +)q 0 q 0
p + q
h 0
q + q0 q 0 + q 0 q0
Rec q0 q 0 + q 0 q0 t
0 =
q0 .
(3.34)
Observe that the equations certainly are not in the precise form,
u = Lu + f(u), for direct application of the theory. However, the
main properties that we want for a slow manifold is that of exponential attraction to the neutral modes, which we ensure by basing the
analysis at the critical Reynolds number, and that the slow manifold
is made up of trajectories of the original system, which we ensure by
solving the fluid dynamical equations to some order in small parameters. Variations on the theme as exhibited here of no real moment,
the centre manifold techniques still apply.
3.5.1
215
p 0 = p 0 (r, z, a, ) ,
a = g(a, ) .
From our earlier crude analysis of the linear problem we seek solutions, the
Taylor vortices, which have no azimuthal variation, no dependence. If the
outer cylinder is counter-rotating this is no longer true; the critical mode
has variations around the cylinder and a more sophisticated analysis is then
needed. Thus seek solutions depending upon only r and z, and upon time
implicitly through the amplitude a.
Recall the linear analysis of the stability of Couette flow from 3.1.4. We
found that at the critical Reynolds number
3 32
3 3
Rec =
3` ,
=
2
2b3/2
all modes of the fluid dynamics decayed except for the mode
v 0 sin[`(r 1)] cos(kz) ,
for the critical axial wavenumber k = kc = / 2b = `/ 2 . In these expressions the critical radial; wavenumber ` = /b in terms of the separation b
between the inner and outer cylinder. Thus we now construct a slow manifold model for the nonlinear evolution of the dynamics of the Taylor vortices
based upon this critical mode.
Tony Roberts, 24 Apr 2009
216
217
rc:=3/2*sqrt(3*pi*el^3);
re:=rc+eps;
the amplitude a evolves slowly and so to a linear approximation
depend a,t;
let df(a,t)=>g;
g:=0;
To my taste we improve the appearance of the printed output by
on div;
on revpri;
off allfac;
on list;
factor a,eps,sin,cos;
ueq:=re*( df(u,t)+u*df(u,r)+w*df(u,z)-(2*v0*v+neglect*v^2)*rr
+df(p,r) -(rr*df(rad*df(u,r),r)+df(u,z,z)-neglect*rr^2*u
veq:=re*( df(v,t)+u*(v0r+df(v,r))+w*df(v,z)+neglect*u*(v0+v)*
-(rr*df(rad*df(v,r),r)+df(v,z,z)-neglect*rr^2*v);
weq:=re*( df(w,t)+u*df(w,r)+w*df(w,z) )
+df(p,z) -(rr*df(rad*df(w,r),r)+df(w,z,z));
ceq:= rr*df(rad*u,r) +df(w,z);
In the above equations several terms are neglected as indicated by
multiplication by neglect because I set
neglect:=0$
Tony Roberts, 24 Apr 2009
218
27`4 2 u 0
9
= res + g 0 Rec `4 sin[`(r 1)] cos(kz) .
2
4 z
2
(3.35)
Computer algebra packages always have difficulty deciding what is simple. reduce
219
sin(~a)*sin(~b)=>(-cos(a+b)+cos(a-b))/2
cos(~a)*cos(~b)=>( cos(a+b)+cos(a-b))/2
sin(~a)*cos(~b)=>( sin(a+b)+sin(a-b))/2
sin(~a)^2
=>(-cos(2*a)+1)/2
cos(~a)^2
=>( cos(2*a)+1)/2 };
220
vd:=fish(veq+rc*(sqrt(3*el/pi)*cos(kz)*sin(lr)*gd+v0r*ud),rz)
v:=v+vd;
Also given the correction ud we determine wd from the continuity equation using the inbuilt integrator (though execution is faster if we write
a specific operator):
221
wd:=-int(ceq+df(ud,r),z);
w:=w+wd;
Lastly the pressure correction may be partially determined by either
the radial or the axial components of the NavierStokes equations.
But generally not all of it is determined by either one. This is directly
analogous to the problem of finding a scalar potential, here p, for a
vector field, here the known parts of the equationsnot all the scalar
potential can be determined from any one component because spatial
differentiation in one variable loses information about anything constant with respect to that variable. Thus in general we have to use
both equations to find the p correction as in
pd:=-int(ueq+rc*(gd*cos(kz)*sin(lr)-2*v0*vd)
-(rr*df(rad*df(ud,r),r)+df(ud,z,z)),r);
p:=p+pd-int( weq-rc*sqrt(2)*gd*sin(kz)*cos(lr)+df(pd,z)
-(rr*df(rad*df(wd,r),r)+df(wd,z,z)) ,z);
Observe how the known part of the correction has to be incorporated
in the forcing of the second equation.
To unfold the pitchfork bifurcation we repeat the iterative loop until
the residuals, and hence the errors by the approximation theorem, are
at most O a4 + 2 . In reduce
it is more convenient to iterate to the
slightly smaller error O a4 , 2 and so we wrap the above computation
of corrections with
let { eps^2=>0, a^4=>0 };
repeat begin
. . .
showtime;
end until (ueq=0)and(veq=0)and(weq=0)and(ceq=0);
to iterate until all equations are effectively satisfied.
This completes all the pieces of the reduce computer algebra program35
apart from sundry comments needed to clarify the program.
35
222
The model
Running the reduce program gives the model
a
2
3 3
=
a
2 2` a3 + O a4 , 2 ,
t
9`
64
where = Re Rec . This is the classic Landau equation for a pitchfork
bifurcation. This model predicts that for Reynolds numbers above critical,
stable Taylor vortices form of amplitude
8 Re Rec
a
.
3(273 `5 )1/4
Should discuss the physical fields and draw some graphs??
Comment on timings and the combinatorial explosion in algebra.??
h
i(iterations)
(work) (deg of nonlin)(no. of params)
until we start discarding terms.
3.5.2
The stages of a slow manifold analysis are to: substitute the ansatz and
group terms of the same order; solve the hierarchy of equations to some
order; assemble the results to give a dynamical model.
The slow manifold model
To construct a slow manifold model we should first analyse the linear problem to find the spectrum and the critical value of parameters such as the
Reynolds number. However, here we have already done so, albeit crudely,
and so we know to seek a slow manifold parameterised by an amplitude a
Tony Roberts, 24 Apr 2009
223
and the bifurcation parameter . Thus here we set about the slow manifold analysis immediately. The first stage then turns out to be precisely the
linear analysis we would normally have done first anyway. Lets see how it
works out.
To start, instead of considering the fluid fields to depend directly upon
space and time, we simply consider them to be a function of space and
the amplitudes, (, a), and then let the amplitude vary in time. That is,
substitute
q 0 = q 0 (r, z, , a) ,
p 0 = p 0 (r, z, , a) ,
a = Ga + g(, a) ,
into the fluid equations. From our earlier crude analysis of the linear problem
we seek solutions, the Taylor vortices, which have no azimuthal variation,
no dependence. If the outer cylinder is counter-rotating this is no longer
truethe critical mode has variations around the cylinderand a more
sophisticated analysis is then needed. Thus seek solutions depending upon
only r and z, and upon time implicitly through the amplitude a.
I have also left Ga in the analysis for a while. This is not necessary as
we expect G = 0 as the earlier linear analysis shows that the critical mode
corresponds to one simple zero eigenvalue. But note that G plays the role
that an eigenvalue in a linear analysis (as a = Ga has solutions a =
exp(Gt)). Thus including G will make the linear problem explicitly appear
in what we do here.
By the theorems on approximation and on smoothness a power series solution in a and is reasonable. Thus we seek approximations explicitly in the
multinomial power series form.
q 0 = q1,0 a + q2,0 a2 + q3,0 a3 + q1,1 a + ,
p 0 = p1,0 a + p2,0 a2 + p3,0 a3 + p1,1 a + ,
a = Ga + g2,0 a2 + g3,0 a3 + g1,1 a + .
where superscripts m, n denotes the coefficient of a term in am n (and
G = g1,0 ), and where qm,n and pm,n depend only upon r and z. As we
expect a pitchfork bifurcation we only include the terms needed to find the
Tony Roberts, 24 Apr 2009
224
O a + a3 structure in the model. Before substituting these expansions,
note that the time derivative becomes
a
a
t
[Ga + g(, a)]
=
a =
0
0
q
q
q 0
=
q1,0 + 2q2,0 a + 3q3,0 a2 + q1,1 +
Ga + g2,0 a2 + g3,0 a3 + g1,1 a +
= q1,0 Ga + 2q2,0 G + q1,0 g2,0 a2 + 3q3,0 G + 2q2,0 g2,0 +
3
1,1
1,0 1,1
1,0 3,0
a + .
a + q G+q g
+q g
Substitute these expansions into the governing dynamical equations (3.34),
and equate terms of the same order together to find the following hierarchy
of equations to be solved.
Order a
i
h
p1,0 + 2 q1,0 = Rec q1,0 G + q0 q1,0 + q1,0 q0 ,
q1,0 = 0 .
Order a2
h
p2,0 + 2 q2,0 = Rec 2q2,0 G + q1,0 g2,0 + q0 q2,0 + q2,0 q0
i
+q1,0 q1,0 ,
q2,0 = 0 .
Order a3
h
p3,0 + 2 q3,0 = Rec 3q3,0 G + q1,0 g3,0 + q0 q3,0 + q3,0 q0
i
+2q2,0 g2,0 + q1,0 q2,0 + q2,0 q1,0 ,
q3,0 = 0 .
Tony Roberts, 24 Apr 2009
225
Order a
i
h
p1,1 + 2 q1,1 = Rec q1,1 G + q1,0 g1,1 + q0 q1,1 + q1,1 q0
h
i
+ q1,0 G + q0 q1,0 + q1,0 q0 ,
q1,1 = 0 .
The linear problem appears
Observe that the leading order equation, that of order a, is precisely the
linear eigenvalue problem that we considered in ?? for the linear stability
of the fixed point of Couette flow. This is always the case in a slow manifold
analysis.
The only difference is that here the eigenvalue is replaced by G. This
is completely natural because G is the restriction of the linear operator to
the centre subspace, and hence must be the eigenvalue corresponding to
that subspace. Here the relevant eigenvalue exactly at the critical Reynolds
number is 0, and hence G = 0. As commented earlier, only in other types of
problems, with eigenvalues of non-zero imaginary part or with degenerate
eigenvalues, does G become non-zero.
To solve this leading order eigen-problem we could persevere with the crude
approximation of ??. However, I feel bound to do a better job. Here we
investigate a numerical solution (via Matlab) which is later generalised to
deal with the higher order equations.
For simplicity assume that there is only one critical mode, so that
00
u (r) cos(kz)
q1,0 = v 00 (r) cos(kz)
w 00 (r) sin(kz)
p1,0 = p 00 (r) cos(kz) ,
(3.36)
226
1d
D=
r dr
d
r
dr
k2
1
.
r2
227
0.5
-0.5
-1
-1.5
-2
-2.5
228
-0.6
-0.4
-0.8
-0.2
0.9
0.8
0.7
0.6
0
0.5
0.4
0.3
0.2
0.2
0.8
0.4
0.1
0
0.6
1.1
1.2
1.3
1.4
1.5
r
1.6
1.7
1.8
1.9
229
800
700
600
Re
500
400
300
200
100
cal resolution, namely N = 16, gives Rec = 68.85 for a critical wavenumber
kc 3.
The upshot of of these numerics is that we have identified the critical
Reynolds number Rec on which to base slow manifold analysis. Further
we now have numerical approximations to the critical mode, given by (3.36)
with k = kc . The slow manifold we proceed to find is then parameterised
by the amplitude of this critical mode, all other modes being exponentially
decaying, and by , the departure from critical conditions. The leading order term of the evolution on the manifold Ga = 0 as the critical eigenvalue
has no imaginary part.
230
= Rec u g + u
r
z
r
00
1
du
k
= Rec u 00 cos(kz)g2,0 + (1 + cos 2kz) u 00
(1 cos 2kz) w 00 u 00
2
dr
2
00
2
1
(v )
(1 + cos 2kz)
,
2
r
0
v2,0
u2,0 v0
2 2,0
2,0 v
v 2 Rec u
+
r
r
r
1,0
1,0
u1,0 v1,0
1,0 2,0
1,0 v
1,0 v
= Rec v g + u
+w
+
r
z
r
00
1
dv
k
= Rec v 00 cos(kz)g2,0 + (1 + cos 2kz) u 00
(1 sin 2kz) w 00 v 00
2
dr
2
00
00
1
u v
+ (1 + cos 2kz)
,
2
r
p2,0
+ 2 w2,0
z
1,0
1,0
1,0 w
1,0 2,0
1,0 w
+w
= Rec w g + u
r
z
00
k
k
00
2,0
00 dw
00 2
= Rec w sin(kz)g + sin(2kz)u
+ sin(2kz)w
,
2
dr
2
231
r r
+ z2 .
where 2 = 1r r
It would appear that we may find a solution in the form (and similarly for
the pressure)
q2,0 = q2,0
+ q2,0 + q2,0
0 2,0 1 22,0
2,0
u0 (r)
u1 (r) cos kz
u2 (r) cos 2kz
2,0
2,0
.
= v2,0
0 (r) + v1 (r) cos kz + v2 (r) cos 2kz
2,0
2,0
2,0
w0 (r)
w1 (r) sin kz
w2 (r) sin 2kz
2,0
Indeed, there is no problem finding the q2,0
0 and v2 components. How2,0
ever, the equation for the q1 component is singular. Fortunately we may
eliminate all of this component in the right-hand side by setting
g2,0 = 0 .
In principle, we now have solutions (albeit numerical) for order 2, 0 quantities. Numerical solution fields for q2,0 are shown in Figure 3.20.37
Order a3 In cylindrical coordinates, with no azimuthal dependence, the
order a3 equations become
37
u3,0
p3,0
2v0 v3,0
+ 2 u3,0 2 + Rec
r "
r
r
1,0
u
u 2,0
u 1,0
u 2,0
= Rec u1,0 g3,0 + u2,0
+ u1,0
+ w2,0
+ w1,0
r
r
z
z
2v2,0 v1,0
r
= Rec u 00 cos(kz)g3,0
232
2
1
-2
0.9
0.8
0.7
0
0
0.6
-8
0.5
-6
0.4
0.3
-4
0.2
-2
0.1
0
2
1
1.1
1.2
1.3
1.4
1.5
r
1.6
1.7
1.8
1.9
233
1 2,0 du 00
1 2,0 du 00
2,0
+ u0 + u2
cos(kz) + u2
cos(3kz)
2
dr
2
dr
1 2,0
1 du2,0
d
2,0
00
00
2
u0 + u2
u cos(kz) +
u cos(3kz)
+
dr
2
2 dr
k 2,0 00
k 2,0 00
+ w2 u cos(kz) + w2 u cos(3kz)
2
2
h
i
00 2,0
+ kw 00 u2,0
cos(kz)
+
kw
u
cos(3kz)
2
2
h
i
00
v
2,0
2,0
2,0
2v0 + v2 cos(kz) + v2 cos(3kz)
,
r
0
v3,0
u3,0 v0
2 3,0
3,0 v
v 2 Rec u
+
r
r
r
"
v 2,0
v 1,0
v 2,0
v 1,0
= Rec v1,0 g3,0 + u2,0
+ u1,0
+ w2,0
+ w1,0
r
r
z
z
1,0
2,0
2,0
1,0
u v
u v
+
+
r
r
= Rec v 00 cos(kz)g3,0
1 2,0 dv 00
1 2,0 dv 00
+ u2,0
u
cos(kz)
+
u
cos(3kz)
+
0
2 2
dr
2 2 dr
d
1 2,0
1 dv2,0
2,0
00
00
2
v0 + v2
u cos(3kz)
+
u cos(kz) +
dr
2
2 dr
k 2,0 00
k 2,0 00
+ w2 v cos(kz) + w2 v cos(3kz)
2
2
h
i
00 2,0
+ kw v2 cos(kz) + kw 00 v2,0
cos(3kz)
2
00
1 2,0
v
1 2,0
2,0
u0 + u2
cos(kz) + u2 cos(3kz)
+
r
2
2
00
u
1 2,0
1 2,0
2,0
+
v0 + v2
cos(kz) + v2 cos(3kz)
,
r
2
2
234
p3,0
+ 2 w3,0
z "
w 2,0
w 1,0
w 2,0
w 1,0
+ u1,0
+ w2,0
+ w1,0
= Rec w1,0 g3,0 + u2,0
r
r
z
z
= Rec w 00 sin(kz)g3,0
1 2,0 dw 00
1 2,0 dw 00
2,0
+ u0 u2
sin(kz) + u2
sin(3kz)
2
dr
2
dr
d
1 2,0
1 dw2,0
2,0
00
00
2
+
w0 + w2
u sin(kz) +
u sin(3kz)
dr
2
2 dr
k
k 2,0 00
00
+ w2,0
w
sin(kz)
w
w
sin(3kz)
2 2
2 2
h
i
00 2,0
+ kw 00 u2,0
sin(kz)
kw
u
sin(3kz)
,
2
2
u1 (r) cos kz
u3 (r) cos 3kz
3,0
.
= v3,0
1 (r) cos kz + v3 (r) cos 3kz
3,0
3,0
w1 (r) sin kz
w3 (r) sin 3kz
But it is not so straightforward. Again there appears terms in cos kz and sin kz
in the right-hand side, indicating that the singular nature of the linear operator will cause trouble, but now simply setting g3,0 = 038 will not make
them disappear. Instead we must be more careful.
However, in principle we can choose g3,0 so that the right-hand side is in
the range of the singular linear operator L; then there will exist a solution
for q3,0 . Indeed this requirement is called the solvability condition. There
are two ways to choose g3,0 .
38
235
More generally one takes the inner product with eigen-vectors of the adjoint operator.
See Matlab program tcvort.m
41
This non-uniqueness was glossed over earlier but now must be addressed.
40
236
This second approach is ideal for numerical solutions, the first is generally
better for analytic solutions by hand.
Order a In cylindrical coordinates, with no azimuthal dependence, the
order a equations become
p1,1
u1,1
2v0 v1,1
+ 2 u1,1 2 + Rec
r
r
r
0
2v 1,0
= Rec u1,0 g1,1
v
r
2v0 00
v cos kz ,
= Rec u 00 cos(kz)g1,1
r
0
v1,1
u1,1 v0
2 1,1
1,1 v
v 2 Rec u
+
r
r
r
0
0
v
v
= Rec v1,0 g1,1 +
+
u1,0
r
r
0
v
v0
00
1,1
= Rec v cos(kz)g +
+
u 00 cos(kz) ,
r
r
p1,1
+ 2 w1,1
z
42
237
u1 (r) cos kz
1,1
.
q1,1 = q1,1
1 = v1 (r) cos kz
1,1
w1 (r) sin kz
Exactly the same arguments apply here as for the order a3 terms. Indeed,
exactly the same arguments hold at all orders in the asymptotic analysis,
other than the leading order linear problem.
Observe that the same linear operator, which we called L in general, occurs
on the left-hand side in each equation of the hierarchy of equations (see
their first lines). Furthermore, at order ap q the linear operator acts upon
the unknown q 0 and p 0 of that order; all other factors are of lower order in
and a and appear elsewhere in the equations. All? Not quite.
On the right-hand side of the equations at order ap q there appears an as
yet unknown quantity gp,q (recall that G = 0 so terms involving G vanish).
However, and this is critical, L has a zero eigenvalueL is singularand
so gp,q is used to ensure that the right-hand side of the equations is in the
range of L so that a solution for qp,q may be found. This is generally called
the solvability condition.
Numerically we here find43
g1,1 = 0.0054 .
A straightforward way to satisfy the solvability condition is to find the
left-eigenvectors (adjoint eigenvectors) corresponding to the critical modes.
43
238
3.5.3
Exercises
Exercise 3.30:
RayleighBenard again. Obtain the Landau equation for
the evolution of 2D RayleighBenard convection for Rayleigh numbers
near critical.
3.6
239
Sorry:
240
88
1920
16
192
12
- n
Theorem 3.8 says that one may approximate the centre manifold M, and
the long-term dynamics on M, to an order of accuracy which is different for
the parameters and the dynamical variables of the model. Such flexibility
is useful because although sometimes we are only interested in the leading
order effect of the parameter, occasionally we are interested in high order
expansions in an artificial parameter and low order with respect to other
variables.
3.6.1
Global models often occur in chemical reactions. There some chemical rate
constants are large corresponding to rapid reactions, and some are small
corresponding to slow reactions. Adopting the timescale of the fast reactions,
write the chemical system as44
x = f(x, y) and y = g(x, y) ,
44
Many call such systems singularly perturbed system and apply simpler methods of
analysis and approximation to obtain leading order approximate models.
Tony Roberts, 16 Apr 2009
241
m
5 6 Ss
4
3
2
1
0
S 3.
s S s
s
s
s
S
s
s
s
Qs 2. Ss
Q
S
Q
PP
s
s
s
s
Qs S
PP Q S
1. PPQPS
QPsf
s
s
s
Q
SPPP
Q
P
S
s SQQs PPP
sf
P
Q
S
s
s
s
s
s
s
- n
Figure 3.22: identifies the non-zero coefficients in h(2) h(1) , circles, which
indicate the error in the approximation h(1) to M; also plotted are three
lines each indicating a different allowable statement of the error in h(1) :
1, m/2 + n/4 = 1 ; 2, m/3 + n/3 = 1 ; and 3, m/5 + 2n/5 = 1 .
242
where is the small ratio of the timescales, and where, for fixed x, the
y equation would have solutions rapidly decaying to some manifold of quasiequilibria. The x equation then governs the long-term evolution along this
manifold.
Carr (1981) describes a realistic chemical example.
Example 3.32:
and y = y + x2 .
(3.37)
At = 0 there is a whole manifold of equilibria M0 , y = x2 . Centre manifold theory supports the model around each and every point
on M0 , just adjoin = 0 . Approximately, then, the evolution x =
x(1 x2 ) + O 2 . From which a further iteration gives the slow
manifold
y = x2 2x2 (1 x2 ) + O 2 ,
and thus corresponding evolution
x = x(1 x2 )(1 + 2x2 ) + O 3 .
There is no component of O xp in the errors as the model is global
in x: the order of error is uniform in M0 .45
However, see that in the evolution the O 2 error is approximately
22 x3 (1 x2 ) and so the multiplicative constant of the error varies
with x, and in particular, it becomes very large for large x. The
neighbourhood of support of theory will correspondingly vary in size
across M0 , generally being small when the error constant is large.
Example 3.33:
Consider x = xy , y = y xy + , for small parameter . When = 0 the variable y decays exponentially quickly
to y = 0 (provided x > 1); once on y = 0 then the variable x no
45
243
longer evolves. Thus y = 0 for every x > 1 is a subspace of attractive equilibria. One may construct the slow manifold based around
the = x = 0 . However, more useful is the global slow manifold
y=
2 x
3
+
O
.
1 + x (1 + x)4
x
+ O 2 ,
1+x
3.6.2
This initial example introduces how to view information flow across a Newton diagram. Such information flow determines valid truncations of the
multivariate asymptotic approximations to complex models.
Example 3.34:
Reconsider Example 3.32. Now make the parameter
vary slowly in time, at rate say, so that the system (3.37) is here
= ,
= 0 ,
x = (x xy)
and y = y + x2 .
(3.38)
244
2x2 (1 x2 )
x2
2x2 (1 x2 )
0
1
Figure 3.23: a type of partial Newton diagram (Murdock 2003) with the arrows denoting how two terms in the multivariate asymptotic approximation
are derived from the earlier term at order 1 0 .
From Example 3.32, an approximation to the slow manifold is
y = h(1) (x, , ) = x2 2x2 (1 x2 ) .
Let us explore the next step in the iterative construction of the slow
manifold. The residual of the y equation is
h
i
Res3.38 = 2x2 (1 x2 ) + 2 (2 6x2 ) + O 3 .
In this problem, simply add this residual to h(1) to improve the approximation of the slow manifold: h(2) = h(1) + Res3.38 . Figure 3.23
graphically shows how the term 2x2 (1 x2 ) in h(1) generates two
new terms in the approximation: the nonlinearity in the equations
generates terms to the right as shown by the right pointing arrow;46
whereas the slowly varying replaces factors with factors and
hence generates terms to the above left as shown by another arrow.
That is, these two arrows outline how information and influences flow
across the Newton diagram.
The arrows of Figure 3.24 determine valid truncations of the multivariate asymptotic expansions.47 A finite truncation of the multivariate
46
`
The nonlinearity also generates terms O 3 which would appear further to the right
on the diagram. However, for our purposes only the leading order effect in the various
directions are significant.
47
At the very least, the arrows determine valid truncations when computing the multivariate expansions.
Tony Roberts, 16 Apr 2009
3
2
1
0
245
..
..
..
..
.
.
.
.
0
1
2
3
Figure 3.24: information flow across the Newton diagram with arrows showing how terms of various multivariate orders depend upon other terms: the
bullets, , denote some algebraic expressions in the global variable x.
246
H
H@
H
@ -H
@
@
1 H
H
@HH@
-@ H@
HH
0 @
@
H
@
@
0
1
2
3
A
-A 1 AA
AA
0
0
1
2
3
Figure 3.25: valid and invalid truncations of a multivariate asymptotic expansion drawn on the Newton diagram: the green, cyan and red lines give
the boundaries of computed termsterms below and to the left are those
retained. In the valid truncations of (a), the ignored terms to the right
and abovethe green and cyan
lines may be algebraically expressed as errors
O 3 + 3 and O 4 + 2 , respectively.
247
(3.39)
v 0
Gs Eg 0 = Res ,
s
(3.41)
248
s32
s22
s12
s02
..
..
..
..
.
.
.
.
% % % %
% % % %
% % % %
% % % %
s01
s11
s21
s31
Figure 3.26: information flow across the Newton diagram with arrows showing how terms of various multivariate orders influence other terms through
the nonlinearity f(u) in the dynamical system (3.39).
Firstly, through the linear operator L in the dynamical system (3.39), any
change at order sp1 1 spmm has a direct affect at that order. Indeed, this is
how the iterative construction works: when a term of a given order occurs
in the residual, the homological equation (3.41) guides us to choose modifications that cancel out the terms in the residual. The linear operators
L and E on the left-hand side of the homological equation (3.41) reflect this
direct dependence.
Second, the strictly nonlinear function f(u) in the dynamical system (3.39)
influences only higher order terms. Any modification in the approximation
u = v(s) at a term sp1 1 spmm will, because of the necessarily multiplicative
nonlinearities in f, cause the function f to generate terms of higher order
in the residual. Here, higher order means a term sq1 1 sqmm where the
exponents qj pj and at least one of these inequalities is a strict inequality.
Thus on a Newton diagram, two dimensional for simplicity, the nonlinearity
represented by the nonlinear function f only influences upwards and/or to
the rightFigure 3.26 illustrates these typical influences.
Figure 3.26 shows that direct nonlinearity places relatively mild restrictions
on the truncation of the asymptotic approximations: we just need to consisTony Roberts, 16 Apr 2009
249
tently neglect terms that are of higher order in at least one variable of the
centre manifold model.
Third, the time derivative u has subtle effects that deserve the careful consideration and study of the next two sub-subsections.
Linear time dependence dominate possible truncations
The previous sub-subsection considered some of the effects of modifying
some term in the asymptotic approximation to the centre manifold model (3.40).
In particular it explored the dependencies generated through the linear
term Lu and nonlinear terms f(u) in the dynamical system (3.39). Now
we explore the intriguing dependencies generated through the time derivative u in the dynamical system (3.39).
Suppose the residual has such a term: given such a term in the residual
we modify, through choosing v 0 or g 0 , the asymptotic approximation to
the centre manifold in order to cancel the term in the residual: However,
depending upon operator G, time derivatives, represented by the term vs0 Gs
of the homological equation (3.41), does couple modifications.
Consider some low dimensional examples.
The operator G being diagonal is the simplest case. In particular, for
the common case of approximating slow manifolds the operator G = 0
which is even simpler. In this case the influence of any change is always
to strictly higher order terms, up and to the right in a 2D Newton
diagram, and so we may be extremely flexible about truncation.
The operator G is in Jordan form.
Example 3.35:
For example, = 0 , x = xy and y = y + x2 ,
for which, with slow variables s = (, x) ,
0 0
G=
.
1 0
Hence a modification in a term xp q generates, via the time
a term in xp1 q+1 . An influence flows upwardsderivative x,
Tony Roberts, 16 Apr 2009
250
0 0
G = 1 0
0 0
s = (, , x) , the operator
0
0 .
0
G is block diagonal. Then the variables in each block are fully coupled
together, determined simultaneously, and the error expressed in their
overall evenly weighted amplitude.
?? This version of the theorem overturns much current practise. With
currently extant methods, such as that of multiple scales, people have to try
very hard to get the relative scaling of one variable correct compared to
another. For example, to analyse the pitchfork bifurcation one normally is
forced to say things like x2 at the very start of the analysis. With
centre manifold theory we need assert no such thing, it just turns out to
be a natural consequence of the dynamics in the model. This theorem, in
essence, allows one to scale different effects in a very general way. That is,
the centre manifold model is valid over a wide range of scaling regimes.
Tony Roberts, 16 Apr 2009
251
3.6.3
Summary
??
Exercise 3.38:
Revisit Example 3.33: x = xy , y = y xy + ,
for small parameter . Construct slow manifolds and models for
small (, x) and also just for small . Compare and contrast the
two different approximations for the dynamics and their theoretical
support.
252
3.7
Sorry:
Centre manifold theory is even more flexible. Although it is restricted to dynamical systems with a linear part that has a good spectral gap, innovative
transformations of the time variable empower us to apply it to other problems. For example, later ?? we see how transforming to log-time separates
the continuous spectrum of diffusion on an infinite domain into a discrete
spectrum for modelling via centre manifold theory: the algebraic decay of
diffusion in an infinite domain becomes exponential decay in log-time.
For now we restrict attention to two variable dynamical systems. Without
loss of generality, suppose a region of interest surrounds an equilibrium at
the origin. The simple normal form for a slow manifold (Elphick et al.
1987, Cox & Roberts 1995) near the equilibrium, if regular, is
x = g(x)
and y = (x)y ,
(3.42)
where to date g(0) = g 0 (0) = 0 and the decay-rate (0) > 0 . These conditions and the form of (3.42) ensure there exists the slow manifold M : y = 0 .
Further, as the decay-rate (0) > 0 , the slow manifold M is exponentially
quickly attractive to all neighbouring solutions. We begin to investigate
the possibilities when at the equilibrium the decay-rate (0) = 0 or even
(0) = . Theory is useless in a direct application. Nonetheless, the analysis herein suggests that the usefulness of centre manifold theory extends to
at least some irregular situations.
3.7.1
Compress time
253
See that as x evolves slowly to zero, the decay-rate to the invariant manifold y = 0 becomes very small. Thus we cannot be sure that solutions
off y = 0 are well modelled by solutions on y = 0 .
However, similar to the work by Wayne (1997) for diffusion problems in
unbounded domains, introduce a compressed (asymptotically slower) time
Z
= x2 dt t1/2 .
(3.44)
I use dashes to denote derivatives in this new time, d/d, so that now
y = y 0 d/dt = x2 y 0 . Similarly, x = x2 x 0 . Hence the ode (3.43) becomes
x 0 = x3
and y 0 = y .
(3.45)
3.7.2
Now try the same transformation for the more general normal form (3.42).
Introduce the new time
Z
= (x) dt ,
(3.46)
whence y = (x)y 0 and similarly for x so that in the new time the dynamical
system (3.42) is
x 0 = g(x)/(x) and y 0 = y .
(3.47)
Evidently, provided g(x)/(x) is asymptotically small enough, usually O x2 ,
near the equilibrium, then there exists a relevant slow manifold which is exponentially quickly attractive in . This proves that for equations of the
Tony Roberts, 5 Apr 2009
254
normal form (3.42), no matter how slow (or fast if (0) = ) the decayrate near the fixed point, a useful slow manifold exists provided the rate
of evolution of x is significantly slower.
For example, if xp and g xq then y = 0 is a useful slow manifold
provided q > p + 1 . The exponentially quick attraction in of y = 0
corresponds to an attraction in time roughly as
h
i
exp t1p/(q1) .
This is sub-exponential decay, but still faster than algebraic, if p > 0; and
conversely is super-exponential decay in the singular case of p < 0 .
3.7.3
and y = 2y2 + 21 x2
(3.48)
which is Dmitrys turbulent shear model system without the bouyancy term.
Physically relevant values of variables x and y are positive; we will not
consider negative values. Transform to new variables
x = 2 + and y = + 2 ;
(3.49)
the system becomes
= 2 + 32 2
and = 3
5 2 .
2
(3.50)
and 0 = 3
5 2 / .
2
(3.51)
Provided 2 / is small, see the system does not have a slow manifold model
as the two linear eigenvalues are 1 and 3. However, the variable does
decay three times faster than the variable, hence the trajectories will curve
noticeably to the invarient manifold = 0 . Consequently, = 0 is not a
slow manifold but might be termed instead a semi-slow manifold.
Tony Roberts, 5 Apr 2009
3.7.4
255
Summary
3.8
Chapter summary
items:for:the:chapter:summary
quasi??
Current centre manifold theorems support the construction of low dimensional models of dynamical systems.
The existence of a model is based upon the eigenstructure near an
equilibrium or a set of equilibria.
The model is relevant over a finite domain in state space.
Approximations can be found so straightforwardly that we easily program computers to handle most of the messy details.
In applying the techniques to the TaylorCouette problem, one now
appreciates that there are many details to decide. However, the overall
plan is the same no matter what problem is approached. The final
result in this application is a Landau equation for the growth and
saturation of the Taylor vortices. The model describes these onedimensional dynamics for this infinite dimensional fluid dynamical
system.
In a remarkable application of these techniques, Arneodo et al. (1985a)
proved the existence of chaos in fluid dynamics. They examined triple
convection where heat, salt and momentum are advected by and diffuse through the fluid. By varying the parameters of the problem, one
can find a set where the linearised problem has three 0 eigenvalues.
Tony Roberts, 5 Apr 2009
256
Chapter 4
257
259
. 259
. 259
. 265
284
. 285
. 292
. 298
. 304
312
. 313
. 325
. 333
. 337
340
258
Numerical solutions of partial differential equations (pde) sample the solution field at discrete points or on discrete elements. There are many fewer
modes in such discrete approximations. Thus such numerical approximations deal with low dimensional approximations of high dimensional systems.
Surely we must be able to adapt our slow manifold techniques to construct
and improve numerical discretisations.
For example, consider the diffusion equation for a field u(x, t) on a finite
domain:
u
2 u
=
such that u(0, t) = u(, t) = 0 .
(4.1)
t
x2
Classic separation of variables tells us there are an infinite number of modes
in the general solution of the full diffusion equation (4.1):
u(x, t) =
ck ek t sin kx .
k=1
259
spectral approximations, or Galerkin projection, (4.1.2) which approximates solutions of a pde as a sum of a finite number of global
modes in space, and then integrates in time; and
finite element methods (4.1.3), where the discretisation in space uses
interpolation and projection over local modes, and then time integration.
Centre manifold theory provides a marvellous alternative method for deriving accurate discretisations with wonderful support. To introduce the
fundamental ideas, Section 4.2 divides the solution domain into just two
finite elements. The key is to couple the solution fields within each element with a coupling parameter controlling the flow of information between
the two elements. The result is a model expressed in the evolution of two
grid values, one for each element, but which resolves the subgrid scale field
in each element by systematically approximating solutions of the governing
pde. This contrasts with traditional discretisation methods which impose
assumed subgrid fields.
Stuff:for:the:chapter:introduction
4.1
Sorry:
Here:give:the:section:introduction:followed:by:subsections
4.1.1
4.1.2
As named, spectral methods are built from the natural modes of the linear dynamics of a problem. Obtaining the eigenmodes (eigenvectors) of
Tony Roberts, 4 Jun 2009
260
the linear dynamics, one builds approximations that are global in the domain. In spatial problems there are an infinite number of such eigenmodes
so the numerical approximation is to choose a useful finite number of modes.
Another approximation is that often the true eigenmodes are hard to find
so that often one uses approximate eigenmodes to form a pseudo-spectral
approximation.
Example 4.1: toy example Our first example focusses on the first approximation, that of choosing a subset of useful modes. This toy example illustrates some of the issues. Consider the three dimensional
system
x = ax y2 + x z ,
y = y + x2 + xz ,
z = 4z + xy ,
(4.3)
where a is some smallish parameter. The aim of a spectral approximation is to approximate the full three dimensional dynamics by that of
just one or two modes. Here the linearised system is diagonal (x ax ,
y = y and z = 4z) so the modes are represented by the primitive
variables x, y and z.
A first question: how can we measure the quality of an approximation?
One possibility we employ here is to compare the slow manifold of the
approximations with the slow manifold of the original system that
occurs for small parameter a:
y = (1 2a)x2 + ,
3
11
16 a)x +
53
( 34 16
a)x4
z = ( 14
such that
x = ax
(4.4)
261
where we keep just the x mode; and one where we keep both the
x and y modes.
1. First explore the spectral approximation obtained by just keeping
the x mode only. Then the spectral approximation is to set y =
z = 0 in the x equation of the system (4.3) to construct the
spectral model of simply x = ax.
This spectral model is almost useless. True, it does predict the
fixed point x = 0 and its loss of stability as parameter a crosses
zero. But the model contains no useful finite amplitude dynamics.
2. Second we turn to the spectral approximation of the x and y
modes together. Then the spectral approximation is to set z = 0
in the x and y equations of the system (4.3) to construct the
spectral model
x = ax y2 ,
y = y + x2 .
53
be at 3 a/( 34 16
a). The error is largely due to the ratio between the largest retained eigenvalue, here 1 from the y mode,
and the smallest neglected, here 4 from the z mode. Hence we
see the relative error of about 1/4.
The general principle seen in this example is that the error in a spectral
approximation is roughly the ratio between the rate of evolution in the
spectral model and the smallest of the rates of decay of the neglected modes.
Tony Roberts, 4 Jun 2009
262
u(0, t) = u(, t) = 0 .
The parameter a controls stability through the linear growth term +au.
First find a spectral basis for approximations. Here the linear terms in
the right-hand side are uxx +au. Thus solve the eigenvalue/eigenmode
differential equation uxx + au = a with boundary conditions of u = 0
at x = 0, . Being constant coefficient the solutions are trigonometric
and/or exponential functions. A little first year algebra leads one
to discover the eigenmodes are sin kx for integer k corresponding to
eigenvalues k = a k2 . The modes sin kx are the eigenmodes of the
spectral basis.
Second, represent solutions to the partial differential equation in the
form of a truncated sum of time dependent components of these infinite
number of eigenmodes: seek
u(x, t) =
K
X
uk (t) sin kx
k=1
for some functions uk (t) where we keep the first K eigenmodes in the
spectral approximation. Let us just look at the linear dynamics first.
Substitute the spectral sum into the differential equation:
K
X
k=1
u k (t) sin kx =
K
X
(a k2 )uk (t) sin kx .
k=1
263
for k = 1, . . . , K .
m=1
K
X
m,n=1
K
X
n=1
m,n=1
K
X
1
2 sin kx
k=1
k1
X
nun ukn +
n=1
Kk
X
n=1
nun uk+n
K
X
!
nun unk
n=k+1
n=1
n=k+1
264
The previous example showed that using basis functions, here sinusoids, that
are linear eigenmodes greatly simplifies the algebraic representation of the
linear dynamics of a system. Moreover, and very importantly, we are also
then empowered to order the basis functions in order of their decay rates.
Consequently, a simple truncation to the first K modes consistently retains
all modes with slowest decay and neglects consistently all quickly decaying
modes. We need to be consistent.
Dealing with nonlinear terms in the equations is then more involved. The
rearrangements seen here for the modified Burgers equation are about as
simple as they get. Generally expect more complicated rearrangements.
Sometimes rearrangement is not possible and one may have to resort to
explicit integration in order to find the component of a particular mode.
The next example is one such case.
Example 4.3: psuedo-spectral approximation
u
2 u
u
=
+ axu u
,
2
t
x
x
u(0, t) = u(, t) = 0 .
A spectral basis is sin kx for integer k, but only basis for linear operator when one ignores the inhomogeneous term axu. Hence, explicitly
project, rather than simply equating coefficients as in the previous
example.
One typically has to explicitly project, via appropriate integrals, for inhomogeneous pdes.
4.1.3
265
The finite element method is a numerical approach which can be used for the
accurate solution of complex engineering problems. Just some engineering
problems which use the finite element method are listed in Table 4.1. It is
the method of choice for much of the engineering profession.
In the finite element method, the solution region is considered as being built
up of many small, interconnected subregions called finite elements. In
many uses the small elements are triangles or tetrahedra which when pasted
together form the whole of the region under consideration.
Initially, the finite element method was developed to handle problems where
the desired configuration is obtained via minimising the potential energy.
However, it was quickly generalised to other problems where such a minimising principle does not holdthe method then using what is known as
weighted residuals. We shall follow a similar path: first via minimisation
and later developing the methods of weighted residuals.
Specific objectives
To know the six basic steps of the finite element method for both the
potential energy minimisation and weighted residual approach to a
statics problem.
To use the finite element method in one dimension; and solve the
resultant equations by hand or on a computer as appropriate.
To be able to extend the basic linear interpolation of the finite element
method to higher-order interpolations according to the principles of
the method in an application.
266
Table 4.1: some engineering applications of the finite element method (?,
p78).
Area of study
Civil engineering
structures
Aircraft structures
Geomechanics
Hydraulic and
water resources
engineering;
hydrodynamics.
Equilibrium
problems
Static analysis of
trusses, frames,
folded plates, shell
roofs, shear walls,
bridges and
prestressed
concrete.
Static analysis of
aircraft wings,
fuselages, fins,
rockets, spacecraft
and missile
structures
Eigenvalue
problems
Natural frequencies
and modes of
structures; stability
of structures.
Propagation
problems
Propagation of
stress waves;
response of
structures to
aperiodic loads.
Natural
frequencies, flutter,
and stability of
aircraft, rocket,
spacecraft and
missile structures.
Analysis of
excavations,
retaining walls,
underground
openings, rock
joints and soil
structure
interactions; stress
analysis in soils,
dams, layered piles
and machine
foundations.
Analysis of
potential flows, free
surface flows,
boundary layer
flows, viscous
flows, transonic
aerodynamic
problems.
Natural frequencies
and modes of
dam-reservoir
systems and
soil-structure
interaction
problems.
Response of
aircraft structures
to random loads;
dynamic response
of aircraft and
spacecraft to
aperiodic loads.
Time-dependent
soil-structure
interaction
problems; transient
seepage in soils and
rocks; stress wave
propagation in soils
and rocks
Natural periods
and modes of
shallow basins,
lakes and harbours;
sloshing of liquids
in rigid and flexible
containers.
Analysis of
unsteady fluid flow
and wave
propagation
problems; transient
seepage in aquifers
and porous media;
rarefied gas
dynamics; magnetohydrodynamics.
l1
-
A1 , E1
- u0
267
l2
A2 , E2
- u1
- F2
- u2
Figure 4.1: a stepped bar under axial load F2 applied on the right end, and
fixed at the left end.
detract less from the development of ideas.
We investigate the stresses in the stepped bar shown in Figure 4.1.
The bar is made of two lengths of rod, each having cross-sectional area A1
and A2 , and made of materials which have Youngs modulus E1 and E2
respectively. Unloaded, the two parts of the bar have lengths l1 and l2 .
However, the bar is subject to an extensional load of F2 on the right end
and so stretches. Our aim is to find the deformation of the bar. Wherever
we want to be definite, we use the following data: A1 = 2 cm2 , A2 = 1 cm2 ,
l1 = l2 = 10 cm, E1 = E2 = E = 2 106 N / cm2 , and F2 = 1 N.
1. Discretisation We consider that the bar is made of just two elements
as shown by Figure 4.1. By assuming that the bar is one dimensional,
we only consider the axial displacements along the rod. The discretisation is that we only consider the three displacements, uj , of the ends
of the elements as our unknowns.
2. Interpolation The rod is a continuum, thus we describe the deformation of the entire rod in terms of the three displacements uj . This
is an interpolation task. Here it is appropriate to assume that there is
a linear variation in the axial displacement. Let x measure distances
from the fixed wall when the rod is unloaded, then the displacement
field in each of the two elements is
u = u0 + (u1 u0 )x/l1
268
where Ti denotes the strain energy of the ith element, and where F0 is
the unknown reaction on the bar of the immovable wall at the left end
of the bar.1 The strain energy in the ith element is
Z
1
Ti =
Ai i ei dx
i 2
Z
1
=
Ai Ei e2 dx
as i = Ei ei
2
i
Z 2
u
1
=
Ai Ei
dx
as ei = u/x
2
x
i
Ai Ei
(ui ui1 )2
=
2li
where the integration is over the entire element. This expression for Ti ,
a function of the unknowns ui1 and ui only, may be written in matrix
form as
ui1 ui
Ai Ei
ui1 ui
Ti =
ui1 + ui
2li
1 T
ui1
u Ki ui
where ui =
=
ui
2 i
and where
Ai Ei 1 1
Ki =
1 1
li
is called the stiffness matrix of the ith element.
i = 0, 1, 2 .
There is no work done by the unknown reaction F0 at the fixed left end as the displacement there is u0 = 0.
Tony Roberts, 4 Jun 2009
269
Since T = T1 (u0 , u1 ) + T2 (u1 , u2 ) u0 F0 u2 F2 , in terms of the contributions arising from each element this gives
T
T1
=
F0 =
u0
u0
T1
T2
T
=
+
=
u1
u1 u1
T
T2
=
F2 =
u2
u2
A1 E1
(u0 u1 ) F0 = 0
l1
A1 E1
A2 E2
(u0 + u1 ) +
(u1 u2 ) = 0
l1
l2
A2 E2
(u1 + u2 ) F2 = 0
l2
These are put together to form a set of linear equations which may be
written in matrix form as
A E
Al11E1
0
+ l11 1
F0
u0
A1 E1
A1 E1
A2 E2
A2 E2
+
+
0
u
=
l1
1
l1
l2
l2
A2 E2
A2 E2
F2
u2
0
l2
+ l2
or more concisely2
Ku = F
where the overall stiffness matrix K is formed from the stiffness matrices of the elements by adding appropriately
K1 0
0 0T
K= T
+
;
0 0
0 K2
this is generally the easiest approach. Using the example data gives
2 2 0
F0
5
K = 2 10 2 3 1
and F = 0 .
0 1 1
1
5. Solution The naive approach is then simply to assert that the unknown displacements are u = K1 F. However, the stiffness matrix K
is singular, it does not have an inverse. The error in trying to invert
the matrix K is that we have not explicitly included the boundary
2
270
u2 = 7.5 106 cm
F0 = 1 N .
You may have been told other ways to solve this sort of equationchoose a method
that works.
Tony Roberts, 4 Jun 2009
271
d2 u du
+ u = f(x) ,
dx2
dx
0<x<1
1
(x + 1)
2
u
x
2
1
+ u2 fu dx
2
using two equi-sized elements over the domain. Clearly label the five
main steps of the finite element method in your analysis. To simplify
the analysis, use the boundary conditions as early as possible, for
example, in the interpolation.
Answer: Follow the standard plan laid out in the notes.
Exercise 4.6: For a higher order of accuracy with few elements we use a
higher order of interpolation within each element. However, each element needs enough unknowns associated with it to match the number
of coefficients of the interpolation. Quadratic interpolation in one dimension may be obtained by having three nodes on each element: one
at each end, and one usually at the midpoint. If the ends of an element
are at x1 and x2 , then the natural coordinates (see later for a formal
definition) are L1 (x) = (x2 x)/(x2 x1 ) and L2 (x) = (xx1 )/(x2 x1 ).
Show that L1 (2L1 1), L2 (2L2 1) and 4L1 L2 are all quadratic in x and
are each take the value zero on two nodes and one on the other node.
Hence argue that u(x) = u1 L1 (2L1 1) + u3/2 4L1 L2 + u2 L2 (2L2 1)
is the quadratic interpolation through (x1 , u1 ), ((x1 + x2 )/2, u3/2 )
and (x2 , u2 ).
Tony Roberts, 4 Jun 2009
272
273
274
275
276
where D(x, t) is the longitudinal turbulent diffusivity, and q(x, t) is the rate
at which salt seeps into the river at location x. We analyse the steady
seepage of salt, or discharge of contaminant, into the river.
Boundary conditions also need to be specified. Assume that the inflowing
water at the upstream end of the river is free of salt, then the boundary
condition there is that the concentration c = 0. At the downstream end of
the river, although arguable, we use the condition that there is no diffusive
flux upstream, that is c/x = 0 at the exit.
For definiteness, take the reach of the river of interest to be 400 m long and
the seepage to occur into the middle half of the river, 100 m < x < 300 m, at
a constant rate of q = 0.1 m3 / s / m. Say the area is a constant A = 10 m2 ,
the mean advection velocity is U = 1 m / s, and the diffusion is D = 10 m2 / s.
That is, solve
c
2 c
1
+ 10 2 =
H(x)
(4.5)
x
x
100
where H(x) = 1 for 100 < x < 300 and H(x) = 0 otherwise.
1. Discretisation. We divide the river into four elements with end
points x0 = 0, x1 = 100, . . ., x4 = 400. Nodes are placed at these
endpoints and we seek the solution in terms of the concentration at
these nodes, namely c0 , . . . , c4 .
2. Interpolation. We employ a linear interpolation between these nodal
values to estimate the concentration field within any element. Within
the ith element, c = ci1 + (ci ci1 )(x xi1 )/100. For the weighted
residual method it is convenient to use the boundary conditions early
in the analysis.
(a) The upstream condition that c = 0 at x = 0 is easily implemented
by fixing c0 = 0.
(b) The downstream condition that c/x = 0 at x = 4 requires that
the linear interpolant in the last element be constant. This is
ensured by setting c4 = c3 .
Then, in terms of the three unknowns, c1 , c2 and c3 , the solution may
Tony Roberts, 4 Jun 2009
1 6
@
@
@
@
100
@
@
@
277
200
@
@
@
300
-x
400
Figure 4.2: the three basis functions i (x) for the four element analysis of
the dispersion of salt along a reach of river.
be written as
c1 x/100,
2
c3 ,
0 < x < x1
x 1 < x < x2
x 2 < x < x3
x3 < x < x 4
(4.6)
278
Z x4
2 c
c x4
i c
i 2 dx = i
dx .
x
x x0
x0 x x
279
(b) In the second element, only 1 (x) and 2 (x) are non-zero. Thus
there are contributions to the first two equations:
R200
R200
1
i. 100 [1 1010 ] [c1 10 + c2 20 ] dx 100 1 100
dx = =
3
1
3
5 c1 + 5 c2 2
R200
R200
1
ii. 100 [2 1020 ] [c1 10 + c2 20 ] dx 100 2 100
dx = =
2
2
1
5 c1 + 5 c2 2
iii. nothing
(c) In the third element, only 2 (x) and 3 (x) are non-zero. Thus
there are contributions to the last two equations:
i. nothing
R300
R300
1
dx = =
ii. 200 [2 1020 ] [c2 20 + c3 30 ] dx 200 2 100
3
3
1
5 c2 + 5 c3 2
R300
R300
1
iii. 200 [3 1030 ] [c2 20 + c3 30 ] dx 200 3 100
dx = =
2
2
1
5 c2 + 5 c3 2
(d) In the fourth element, only 3 (x) is non-zero. Thus there is only
a contribution to the last equation:
i. nothing
ii. nothing
R400
iii. 300 [3 1030 ] c3 30 dx = 0 as 3 is constant there.
4. Assembly of equations Including all the above contributions to the
three equations, that is, adding all terms itemised by i. into one equation, all those itemised by ii. into a second, and so on, we find
2
5 c1
35 c1 + 53 c2
25 c1 + 52 c2
1
2
1
2
53 c2 + 35 c3
25 c2 + 25 c3
1
2
1
2
=0
=0
+0 = 0
which becomes
3
1
c1 + c2
5
5
1
2
Tony Roberts, 4 Jun 2009
280
1
q
q
0
300
400
q
100
200
-x
Figure 4.3: the approximate concentration field c(x) for a steady discharge
into the region 100 < x < 300.
2
1
c1 c2 +
5
5
2
c2 +
5
3
c3 = 1
5
2
1
c3 =
5
2
5
0.31 ,
16
c2 =
15
0.94 ,
16
c3 =
35
2.19 .
16
in a domain D
281
for some given right-hand side f. First, approximate the solution by a sum
u(x) =
N
X
ui i (x)
i=1
where ui are unknown coefficients and i are so-called basis functions which
each satisfy the boundary conditions of the problem. In a finite element
context ui are typically the nodal values of the unknown field u(x). The
basis function i would then be assembled from the natural coordinates
of those elements sharing the node i. For example, in one dimension with
linear interpolation between end-element nodes:
0
if x < xi1 or x > xi+1
xxi1
if
xi1 < x < xi
i =
xi xi1
xi x
xi xi1 if xi < x < xi+1
as shown by 1 and 2 in Figure 4.2.
For any given approximation, corresponding to a set of values for ui , the
residuals, r(x) = Lu f(x), measures how well the governing differential
equation is satisfied. Since the basis functions i all satisfy the boundary condition, then so will u(x) and we only need address the differential
equation. A good approximate solution will make the residual as small as
possible. Galerkins criterion is to require that the residual is orthogonal to
the basis functions themselves. That is, for each i
Z
i (Lu f) dx = 0 .
D
Z
N
X
=
i L
uj j f dx
D
j=1
282
N
X
i
uj Lj f dx
j=1
Z X
N
uj i Lj fi dx
D j=1
N
X
N
X
i Lj dx
uj
D
j=1
Z
fi dx
D
kij uj Fi
j=1
Z
and
Fi =
fi dx
D
Since the above equation has to be true for all i, these form a set of linear
equations, Ku = F, to be solved for the unknowns u.
Typically, most of the interaction coefficients kij are zero. This is because the
basis functions are non-zero only in the small number of elements adjoining a
given node. For example, in one dimension Ri is identically zero outside the
interval xi1 < x < xi+1 ; hence the integral D i Lj dx = 0 for |i j| > 1
and consequently K is tridiagonal (as before).
Another issue is that in order to compute the interaction coefficients by
the above integral, we need basis functions with a high enough degree of
continuity for Lj to be bounded; however, this is tedious to achieve. To
circumvent this difficulty it is usual to perform integration by parts in order
to reduce the degree of differentiation of j . This is done at the expense
of differentiating i , so that the break even point is when the orders of
differentiation of i and j differ by no more than one. For example, in one
dimension
Z
Z
3 j i 2 j
4 j
2 i 2 j
i 4 dx = i 3
+
dx
2 x2
x
x
x x2
D x
D
Tony Roberts, 4 Jun 2009
283
To see the application of this method to other examples, see for example the
book by ? [4.2].
Exercise 4.8: Modify the analysis of dispersion in a river by formulating
the equations for four elements in the case where there is a point
discharge of contaminant at location x = 150 m of 10 m3 / s. Hint: this
only involves modifying the right-hand sides of the equations.
Answer: The point discharge only influences the first two linear equations
as only 1 and 2 are non-zero at the discharge site. The right-hand side of
the other equation is zero. A point discharge into the river is zero everywhere
except at the discharge site, and so the integrals involving the discharge
reduce to simply a contribution of the integrand at the discharge site x =
150 m.
0<x<1
Answer: Use the basic weighted residual method as discussed in the text.
Discretise the domain, choose linear interpolation, compute the element integrals remembering to integrate by parts the second derivative term, assemble
the equations, and solve.
d2 u
+ u = x2 ,
dx2
0<x<2
284
Exercise 4.11: The deflection of a beam on an elastic foundation is governed by the equation
d4 u
+ u = 1,
dx4
0<x<1
4.2
Warning:
Aim: explore the process of analysing a pde to produce a numerical discretisation. For example, we might want to discretise the nonlinear diffusion
Tony Roberts, 9 Jun 2009
285
by
The plan:
1. introduce some key concepts in an example with just one numerical
grid point, one finite element, in the domain, see 4.2.1;
2. explore an example of relatively naive coupling between adjacent finite
elements when the domain is split into many, see 4.2.2;
3. introduce a nonlocal inter-element coupling that ensures consistency
as well as rigorous support from centre manifold theory, see 4.2.4.
4.2.1
such that
u(1, t) = 0 .
(4.7)
The diffusion on the spatial interval [1, 1] will smooth out any spatial
variations in the field u(x, t): the small scale spatial variations will smooth
the fastest. However, the smoothing is nonlinear: at locations where the
field u is largest the diffusion smooths the quickest.
Tony Roberts, 9 Jun 2009
286
But the field u also decays over long times. For example, the parabolic
initial condition
u(x, 0) = a0 (1 x2 )
u(x, t) =
a0 (1 x2 )
1 + 2a0 t
is the solution for all time. That is, the algebraically decaying parabola is
an exact solution of the nonlinear diffusion equation (4.7). Now rewrite this
algebraically decaying parabola in the equivalent form
u = a(t)(1 x2 )
where
a = 2a2 .
(4.8)
Interpret this form as a low dimensional model of the full system (4.7) as
the model (4.8) replaces the spatio-temporal dynamics of the field u(x, t)
by simply the dynamics of the scalar a(t). The long term evolution of the
field u(x, t) should tend to this parabolic form because any wiggles in the
initial shape of the field u should smooth by diffusion to this parabolic
shape. This subsection construct the model (4.8) as a slow manifold. But
very importantly, centre manifold theory guarantees that starting from a
wide range of initial conditions u(x, 0) the evolution of the field u(x, t) will
be described, exponentially quickly in time, by the model (4.8).
The key to applying centre manifold theory is to identify simple equilibria
at a critical parameter value. Physically, equilibria usually correspond to
some conserved mode such as total amount of material or momentum. Here
for the nonlinear diffusion problem (4.7) there is no exactly conserved mode.
Thus adopt the artifice of inventing one. Such invention is a glorious fudge,
but I aim to convince you it is effective and valuable.
We base the analysis upon a nearby system that does form a tractable basis
for centre manifold theory, and perturb it to obtain the original system (4.7).
Schematically
invented base systems original physical system (4.7)
=0
= 1.
We use the parameter to move from the invented artificial system to the
original physical system: = 0 corresponds to the artificial system; and =
Tony Roberts, 9 Jun 2009
287
u
= 2u
x
on x = 1 .
(4.9)
See that this boundary condition reproduces the physical case u = 0 when
the artificial parameter = 1 , and reduces to a conservative boundary
condition ux = 0 when the artificial parameter = 0 . The choice of signs,
, in the right-hand side is determined by requiring the same physical
meaning of the boundary condition at each end of the domain: at x = +1
the flux of u out of the domain is proportional to ux ; whereas at x = 1
the flux of u out of the domain is proportional to +ux . There is an enormous
scope in inventing other bases for the analysis; we just explore the dynamics
of the pde (4.7) with these boundary conditions (4.9).
Proceed with slow manifold analysis: find equilibria; check for decay in the
linear dynamics; construct the model to some order in the parameters; and
finally interpret the resulting model.
288
2 u 0
x2
such that
u 0
= 2 0 a on x = 1 .
x
(4.10)
Thus the diffusion causes all wiggles in the field u to decay exponentially
quickly. Due to the particular nonlinear diffusion, the decay rate to the equilibria varies, the decay is more rapid for larger amplitudes a, and vanishes
as a 0 . See the interestingly unusual feature that a = 0 is a singular
equilibria in the analysis in that all the decay rates become zero. It is only
for finite a, non-zero u, that the slow manifold picture is valid.
But, the neutral modes corresponding to zero eigenvalue, are not simple.
We must remember to look for generalised eigenmodes. Solving (4.10) with
eigenvalue = 0 we find u 0 is an arbitrary constant only when 0 = 0 ; we
apparently cannot find a two dimensional slow subspace. But we require a
two dimensional slow subspace: one dimension for the freedom to specify
the parameter ; and the other dimension to form the space in which the
amplitude a will slowly evolve. Consequently, the slow subspace must contain generalised eigenmodes. Now (u 0 , 0 ) (1, 0) satisfies (4.10) with zero
eigenvalue. To seek a generalised eigenmode, try to solve (4.10) with u 0 replaced by a constant b corresponding to the constant first component of the
ordinary eigenmode; then find that u 0 = bx2 /(2a) and 0 = b/(2a2 ) . Setting the arbitrary b = 2a2 B, the slow subspace is the linear combination
of the eigenmode and generalised eigenmode
u 0 = A Bax2
Tony Roberts, 9 Jun 2009
and 0 = B .
289
A
0 2a2 A
.
=
B
0
0
B
This Jordan form identifies in the B term that a small non-zero parameter
causes the field u 0 to leak out of the domain across the boundaries. The
critical aspect is that there is nonetheless a two dimensional slow subspace
parametrised by a overall measure of the field u and by the trivial artificial
parameter . This slow subspace forms the basis of the two dimensional
slow manifold.
by seeking
such that a = g 0 (a, ) ,
for small corrections u 0 and g 0 . Substitute into the pde (4.7) and the boundary conditions (4.9), and then omit products of primed quantities with themselves or to require the corrections satisfy
g0 = a
2 u 0
x2
and
u 0
= 2a
x
on x = 1 .
and g 0 = 2a2 .
Determine the arbitrary integration constant A by precise definition of amplitude a. To match amplitude a of the earlier quoted exact solution (4.8),
define
a(t) = u(0, t) ,
(4.11)
so u 0 (0, t) = 0 , and thus A = 0 . The first approximate model is consequently
u a ax2 such that a 2a2 .
(4.12)
Tony Roberts, 9 Jun 2009
290
Convinced? This example shows one case where we can manipulate boundary conditions to our advantage by empowering centre manifold theory to
support modelling at some finite boundary condition parameter. If you are
convinced that this sort of manipulation is valid, then skip to Section 4.3
where a similar trick couples finite elements to empower centre manifold theory to support discretisations in space of pdes. If you are not yet convinced
by this boundary condition manipulation, then keep reading this section.
Exercise 4.12:
The magical model (4.12) is no accident: why implement
the factor of two in the right-hand side of the artificial boundary condition (4.9)? Let us see that such invention can range from the wonderful
to the disastrous. Change the artificial boundary condition (4.9) to
(1 )
u
= Eu on x = 1 .
x
(4.13)
Then show that u = a(t)(1 Bx2 ) such that a = 2Ba2 , for some B
depending upon the parameters E and , forms an exact solution of the
nonlinear diffusion pde (4.7) with boundary condition (4.13). Deduce
B as an exact analytic function of E and . Confirm that setting = 1
gives the correct B = 1 independent of the artificial parameter E.
Tony Roberts, 9 Jun 2009
291
Exercise 4.14:
Create a one variable model for the long term dynamics
of linear diffusion of the field u(x, t) in a small element. Aim to model
u
2 u
= 2
t
x
such that
ux (0, t) = 0
and u(h, t) = 0 ,
292
4.2.2
The previous Section 4.2.1 introduced modelling the dynamics on a spatial domain by creatively fudging the boundary conditions. But numerical
discretisations of spatio-temporal dynamics divide the spatial domain into
many finite elements or employ a finite sized grid. Thus our next step to
use centre manifold theory to support numerical discretisations is to move
from one element on the domain to two elements on the domain, before we
deal with many elements.
In this subsection, let us explore two elements that are coupled together.
Focus on the coupling by exploring the simple diffusion pde: seek to create
a model for the dynamics of
u
2 u
= 2
t
x
on h < x < h
such that
u
= 0 on x = h .
x
(4.15)
2 = bU1 + aU2 ,
U
(4.16)
293
slow manifold modelling. The exact solution over the whole domain involves
an expansion in orthogonal Fourier modes:
u(x, t) =
Ak cos
k=0
2 k2
k
(x + h) exp
t
,
2h
4h2
for some constants Ak depending upon initial conditions. A long term model
must resolve the slowest dynamics in this solution. The more rapidly decaying modes, k 2 , are the ignored exponential transients in the rapid
evolution to the slow manifold. That is, we want to resolve
h
i
2
u = A0 + A1 cos
(x + h) exp 2 t + ignored transients , (4.17)
2h
4h
and ignore the other rapid transients. Consequently,
since cos((x1 +
Observe that U1 U2 = 2A1 exp 2 t/(4h2 ) and so a little algebra
gives the time derivatives
2
1 = (U1 + U2 )
U
2
8h
2
2 = (+U1 U2 ) .
and U
2
8h
(4.18)
This is the exact model we wish to obtain. Centre manifold theory together with innovative use of boundary conditions provides the systematic
methodology to derive this model.
Create an internal boundary Place an internal boundary at x = 0 and
then couple across this internal boundary the field u in the two elements on
either side.
Tony Roberts, 9 Jun 2009
294
+
and u
x = ux ,
and u+
x = 0.
+
+
and (1 )(u
x + ux ) = 2(u u )/h ,
(4.19)
295
296
2 = (U1 U2 ) .
and U
h2
(4.21)
For = 1 this approximates the exact model (4.18) as the exact crossdiffusion coefficient 2 /8 = 1.2337 1 which is the coefficient appearing in
this first approximation.
Follow the usual routine. Suppose that at some iteration the approximate
= g(U, ) , then
field in the left element is u = u1 (x, U, ) such that U
seek small improvements to u1 and g. Substitute u = u1 + u10 such that
= g + g 0 into the pde (4.15):
U
2
(u1 + u10 )(g1 + g10 ) +
(u1 + u10 )(g2 + g20 ) = 2 (u1 + u10 ) .
U1
U2
x
Drop products of small quantities and rearrange to
u10
2 u 0
u1
u1
2 u1
u1 0 u1 0 u10
g1
g2
g1
g2 + 21 =
g1 +
g2 2 .
U1
U2
U1
U2
x
U1
U2
x
Observe the right-hand side is just the residual of the pde (4.15) evaluated
at the current approximation; in this first iterate Res4.15 = 0 . Simplify this
equation for the small improvements by approximating the coefficients of
6
It may be interesting to consider the slow subspace of each of these equilibria. It
eventuates that we do not need to span the slow subspace, only to be able to parametrise
the slow subspace. A parametrisation of the slow subspace is then improved by iteration
to become a parametrisation of the slow manifold.
u1
U1
297
1,
2 u10
= Res4.15 .
x2
u1
U2
0 and
(4.22)
We need boundary conditions for the improvement u10 : the coupling condition (4.19) is nontrivial; similar analysis leads to requiring that at the
internal boundary x = 0
u10
u20
=
x
x
and
u10
u20
= Res4.19 ,
x
x
(4.23)
Res4.15 dx 12 Res4.19 .
g10 =
h
h
In the first iterate, this correction to the evolution gives the first approximation (4.21) to the exact model (4.18) for the left-hand element. Similarly
for the right-hand element.
Now determine the solution field within each element. Integrate (4.22) twice
to determine the improvement to the field: firstly,
Z
u10
1 x
=
Res4.15 +g10 dx
x
h
u 0
h/2
h/2 h
Tony Roberts, 9 Jun 2009
298
ensures that u10 = 0 at the mid-element grid point x = h/2 so that the
iteration maintains u1 (h/2, t) = U1 (t) . In the first iterate, this formula
updates the left-hand field to
u1 U1 + 18 (1 + 2x/h)(3 + 2x/h)(U2 U1 ) .
Similarly for the field in the right-hand element. Evaluate at = 1 to obtain
the model of the original diffusion pde (4.15).
We can make the first step. Derive more accurate approximations with
further iterations. But such algebra is very tedious and so best assigned to
a computer by the algorithm of the next subsection. We find high order
accurate approximations.
4.2.3
1 3
45
13 4
189
5
809
14175 ) h2 (U1
+ U2 ) + O 6 ; (4.26)
299
Algorithm 4.1 derive the coupled model for diffusion between two finite
elements, the pde (4.15).
factor h,nu,u;
% parametrise the model in evolving u(1) and u(2)
operator u; depend u,t;
let { df(u(1),t)=>g1, df(u(2),t)=>g2 };
% initial solution
u1:=u(1); u2:=u(2);
g1:=g2:=0;
% iterate to this order of error
let gam^6=>0;
repeat begin
% compute residuals
res1:=df(u1,t)-nu*df(u1,x,x);
res2:=df(u2,t)-nu*df(u2,x,x);
cc:=sub(x=0,(1-gam)*(df(u2,x)+df(u1,x))-2*gam*(u2-u1)/h);
% update left element from residuals
g1:=g1+(gd:=-int(res1,x,-h,0)/h-nu*cc/2/h);
u1:=u1+int(int(res1+gd,x,-h,x),x,-h/2,x)/nu;
% update right element from residuals
g2:=g2+(gd:=-int(res2,x,0,+h)/h+nu*cc/2/h);
u2:=u2+int(int(res2+gd,x,+h,x),x,+h/2,x)/nu;
end until {res1,res2,cc}={0,0,0};
% check other requirements
ccc:=sub(x=0,df(u1,x)-df(u2,x));
bc1:=sub(x=-h,df(u1,x));
bc2:=sub(x=+h,df(u2,x));
amp1:=sub(x=-h/2,u1)-u(1);
amp2:=sub(x=+h/2,u2)-u(2);
300
Parametrise the model Use the variable u1 to store the field u1 (x, t) of
the left-hand element, and the variable u2 to store the field of the right-hand
element. Then u1 and u2 depend upon position x and upon the evolving grid
values Uj denoted by u(j) for j = 1, 2 . Tell reduce that u(j) evolves in
time with the depend statement. Then use a let statement to tell reduce
to replace time derivatives of Uj by an expression gj (U) stored in g1 and g2.
Initial solution Initialise the fields in each element to be simply a constant in space x, namely the grid value Uj . The initial approximation to the
evolution in time is that there is none; hence gj = 0 .
Compute residuals Inside the iterative loop, the first task is to compute
the residuals of the nontrivial equations. Here there are three nontrivial
equations: the pde (4.15) in the left-hand element; the pde (4.15) in the
right-hand element; and the coupling condition (4.19).
Update left and right elements from the residuals After computing the residuals, improve the current approximation using the formulae
(4.24) and (4.25). See these coded in the reduce of Algorithm 4.1.
Check other requirements The last five lines in Algorithm 4.1 confirm
that the straightforward coupling, boundary and amplitude conditions are
indeed met by the solution derived in the iteration loop.
Outcome: executing the computer algebra Algorithm 4.1 gives not only
the model (4.26), but also the fields internal to each element. These are the
subgrid fields we need as identified in the exact model (4.17) for example.
Figure 4.4 plots these subgrid fields. See how beautifully smoothly the
subgrid fields match in the neighbourhood of the internal boundary at x = 0 .
Despite the crude chopping up of the domain into finite elements, high order
analysis pieces the solution field smoothly back together.
Tony Roberts, 9 Jun 2009
301
1.4
1.2
1
u_j(x)
0.8
0.6
0.4
0.2
0
-0.2
-0.4
-1
-0.5
0
x/h
0.5
Figure 4.4: the left-hand field u1 (x) (solid) and the right-hand field u2 (x)
(dashed) for the physical = 1 when grid values U1 = 0 and U2 = 1 .
See that the extrapolation from the left-hand field does a reasonable job of
extrapolating into the right-hand element, and vice versa.
302
Observe also in Figure 4.4 that the extrapolation of the left-hand field into
the right-hand element very nearly goes through the grid value U2 at x =
h/2 . Similarly, the extrapolation of the right-hand field into the left-hand
element very nearly goes through
the grid value U1 at x = h/2 . Indeed
6
truncating to errors O and evaluating at = 1 and the grid point
x = h/2, the left-hand field at the right-hand grid point u1 (h/2, U) =
0.028 U1 + 0.922 U2 . This is within 8% of the correct grid value. Higher
order analysis gets closer and closer. Perhaps we can use the requirement
that the left-hand field should extrapolate through the right-hand grid value
as an alternative to the coupling conditions (4.19). The next section shows
we can, and subsequent proofs show why it is good to do so.
Exercise 4.15:
Derive some two element models of various accuracy for
the nonlinear Burgers equation
u
u
2 u
+u
=
t
x
x2
such that
u
= 0 on x = h ,
x
(4.27)
on the domain [h, h]. Modify the reduce program of Algorithm 4.1.
The easiest way to control the combinatorial explosion of terms generated by the nonlinearity is to simply multiply the nonlinearity by
some power of the coupling parameter : that is, solve
u
u
2 u
+ p u
=
,
t
x
x2
to some order in the coupling/nonlinearity parameter . Then evaluating at = 1 is not only a fully coupled model, but is also a model for
the original Burgers equation (4.27). Investigate the nonlinear terms
in the model for various orders of truncation.
Answer: For an example, set the exponent p = 1 . Then as well as similar
1 and u1 (x, t), the evolution and the field in the right-hand
expressions for U
element is
2
U
u2 (x, t)
1
1
(U1 U2 ) + 2 U2 (U1 U2 ) + ,
2
h
2h
U2 + (U1 U2 ) 38 ( hx ) + 12 ( hx )2
( + 13 2 )
5 x 2
1 x 4
14 ( hx ) 24
( h ) + 13 ( hx )3 12
(h)
2
1
1 x 2
1 x 3
+ hU2 (U1 U2 ) 24 4 ( h ) + 6 ( h ) + .
9
64
303
Exercise 4.16:
Repeat the previous Exercise 4.15 but now control the
nonlinearity independently of the coupling by instead solving
u
u
2 u
+ u
=
.
t
x
x2
Truncate to various powers of and and compare the terms in the
models.
Answer: As well as similar expressions for U 1 and u1 (x, t), the evolution
and the field in the right-hand element is
2
U
u2 (x, t)
1
1
(U1 U2 ) + U2 (U1 U2 ) + ,
2
h
2h
U2 + (U1 U2 ) 38 ( hx ) + 21 ( hx )2
1
14 ( hx )2 + 16 ( hx )3 + .
+ hU2 (U1 U2 ) 24
=
=
Exercise 4.17:
Make other modifications of Algorithm 4.1 to model
the nonlinear diffusion pde (4.7) by the dynamics in two elements
on the domain [1, 1]. Firstly, model the dynamics with insulating
boundary conditions at x = 1 , then secondly make further significant
modifications of the computer algebra to model the dynamics with
the original fixed value boundary conditions of (4.7). Investigate its
predictions.
304
4.2.4
Inspired by Figure 4.4, let us replace the coupling conditions (4.19) between
the two elements with
u1 (h/2, U) u1 (h/2, U) = (U2 U1 )
and
(4.28)
Why? First, look at the case = 1 . Then, because from definition the
grid value u1 (h/2, U) = U1 and similarly for U2 , these nonlocal coupling
conditions reduce to
u1 (h/2, U) = U2
and u2 (h/2, U) = U1 .
That is, the left-hand field should extrapolate through the right-hand grid
value, and the right-hand field should extrapolate through the left-hand grid
value. If the two subgrid fields are to merge very smoothly, then surely this
condition is a suitable requirement. Second, look at what happens when
= 0 . Then these nonlocal coupling conditions reduce to
u1 (h/2, U) = U1
and u2 (h/2, U) = U2 .
(4.29)
That is, the extrapolation of the left-hand field is the left-hand grid value,
and vice versa. Thus at = 0 , not only are the two elements decoupled from
each other, as in the previous section, but the piecewise constant equilibrium
solution (4.20) still holds. Thus this family of equilibria still act as the basis
for a slow manifold model of the dynamics.
The outcome: before looking at the theoretical support and the minor
changes to the algorithm, let us see some of the results. Solving the diffusion
pde (4.15) with the nonlocal coupling conditions (4.28) leads to the model
1 = ( + 1 2 + 2 3 + 1 4 + 8 5 ) (U1 + U2 ) + O 6 . (4.30)
U
6
45
70
1575
2
h
This model converges slightly quicker in coupling parameter : summing
these terms at = 1 gives the inter-element interaction constant as 1.2305
which agrees with the exact model (4.18) to an error of 0.003 .
Tony Roberts, 9 Jun 2009
305
1.4
1.2
1
u_j(x)
0.8
0.6
0.4
0.2
0
-0.2
-0.4
-1
-0.5
0
x/h
0.5
Figure 4.5: the left-hand field u1 (x) (solid) and the right-hand field u2 (x)
(dashed) for the physical = 1 and the nonlocal coupling conditions (4.28)
when grid values U1 = 0 and U2 = 1 . See that the extrapolation from the
left-hand field does a reasonable job of extrapolating into the right-hand
element, and vice versa.
306
Algorithm 4.2 derive the coupled model for diffusion between two finite
elements, the pde (4.15), utilising the nonlocal coupling conditions (4.28).
I only present the iterative loop as the only differences to Algorithm 4.1 lie
within the loop.
repeat begin
% compute residuals
res1:=df(u1,t)-nu*df(u1,x,x);
res2:=df(u2,t)-nu*df(u2,x,x);
cc1:=sub(x=+h/2,u1-u(1)-gam*(u(2)-u(1)));
cc2:=sub(x=-h/2,u2-u(2)-gam*(u(1)-u(2)));
% update left element from residuals
g1:=g1+(gd:=-int(res1,x,-h,0)/h-nu*cc1/h^2);
u1:=u1+int(int(res1+gd,x,-h,x),x,-h/2,x)/nu;
% update right element from residuals
g2:=g2+(gd:=-int(res2,x,0,+h)/h-nu*cc2/h^2);
u2:=u2+int(int(res2+gd,x,+h,x),x,+h/2,x)/nu;
end until {res1,res2,cc1,cc2}={0,0,0,0};
The subgrid scale fields u1 (x, U) and u2 (x, U) are more impressive. Figure 4.5 plots the field of both
the left-hand and the right-hand element for
the model with errors O 6 evaluated at = 1 : see that the extrapolation
of each field agrees with each other much better than that of the previous coupling condition, shown in Figure 4.4. Evidently the nonlocal coupling conditions (4.28) better causes neighbouring subgrid fields to smoothly
merge.
Derive the model with computer algebra Only small changes are
required in the reduce code to use the new coupling conditions (4.28). Algorithm 4.2 lists the iterative loop with the required changes: change the
computation of the coupling condition; modify the update of the correction gj0 ; and only terminate the loop when both residuals of the coupling
condition are zero. Easy!
Tony Roberts, 9 Jun 2009
307
such that
and
u1
2 u1
= 2 ,
t
x
u1
= 0 at x = h ,
x
u1 (h/2, t) = u1 (h/2, t) .
(4.31)
The pde and the first boundary condition require all solutions to be of the
form u1 = exp(t) cos[k(x + h)] for some wavenumbers k and consequent
growth-rate = k2 . The nonlocal condition then implies cos(3kh/2) =
cos(kh/2) . Expanding and rearranging leads to the requirement
2 sin(kh/2) sin(kh) = 0 .
(4.32)
This requirement is satisfied whenever kh = n , hence the family of wavenumbers are k = n/h , with corresponding growth-rates = 2 n2 /h2 , as
in (4.2.2) of the previous coupling condition. As before, for both elements
combined, but uncoupled, the spectrum has two zero eigenvalues, and all
other eigenvalues are negative and < 2 /h2 . Consequently, centre manifold theory assures us that a slow manifold model exists, is relevant and can
be constructed asymptotically.
Exercise 4.18:
Repeat Exercises 4.15 and 4.17 but with the nonlocal
coupling (4.28).
308
u2 (x, t)
4
(U1 U2 ) + O 2 ,
2 e1 + e3
h
U2 +
(1 e1 2x + e2(x1) )U1
2 e1 + e3
i
(1 2x e1 + e2(x1) )U2 + O 2 .
2 u
x2
u
= 0 at x = h
x
Rh/2
h
309
2 v
x2
v
= 0 at x = h and v(h/2, t) = 0 ,
x
h/2+
v
v
h/2+
=
and [v]h/2 = 0 .
x h/2
x h/2
Exercise 4.21:
Difficult: ponder the theoretical implications of the double roots at wavenumbers k = 2m/h of the requirement (4.32), and
the adjoints in Exercise 4.20.
Exercise 4.22: Ensure continuity The approach described in this section creates models based upon piecewise constant functions. That is,
in each element, the starting approximation is that the field u is constant. Surely we can create models from a piecewise linear basis.
Let us model the linear diffusion pde, ut = uxx , on the domain [1, 1]
by dividing the domain into two artificial elements and creating a
piecewise linear basis. I set the diffusion coefficient and each element
size to one for simplicity. Let the physical boundary conditions be
that u(1) = 0 .
Tony Roberts, 9 Jun 2009
310
Exercise 4.23:
Write, test and execute a computer algebra program to
construct the slow manifold model of Exercise 4.22 to arbitrary order
Tony Roberts, 9 Jun 2009
311
312
coupling
Figure 4.6: this conceptual diagram shows traditional finite difference modelling approaches (right arrows) physical problems (discs) via consistency as
the grid size h 0 (left circles); whereas the holistic method approaches
(forward arrows) physical problems via asymptotics in nonlinearity and the
inter-element coupling (from right circle). The holistic approach supports
the model from both directions.
4.3
Sorry:
Section 4.2 shows we can divide a domain into two finite elements, couple
them together again, and rigorously support a model of the dynamics using centre manifold theory. This section divides a spatial domain into any
Tony Roberts, 14 Jul 2009
313
number of elements and uses special coupling conditions to empower centre manifold theory to support the resulting discretisation of the dynamics.
Interspersed throughout this section we prove that nonlocal coupling conditions inspired by (4.28) ensures the discrete model is also consistent with
the original pde as the element size h 0 . That is, the discretisation and
the pde are the same to some order of error in h, with higher orders in
the coupling parameter resulting in better order of error in h. This dual
support for the discrete models we derive, from both consistency and centre
manifold theory as shown schematically in Figure 4.6, is remarkable.
The approach developed here is based purely upon the local dynamics on
relatively small elements while maintaining fidelity with the solutions of the
original pde. Being local, the approach will flexibly cater for complicated
geometries, varying resolutions, and physical boundaries.
4.3.1
t
h2
where the grid values Uj (t) = u(jh, t). Most importantly, we aim to underpin these models with centre manifold theory in a manner so that the
theory will also support the modelling of complex systems pdes.
Introduce a regular grid As shown in Figure 4.7, place grid points a
distance h apart, Xj = jh for example. Express the field in the jth element
point by u = uj (x, t) so that the fields uj in all elements also satisfy the heat
equation (4.34). We do not restrict the field uj to just the jth element, but
extend it analytically out to the adjacent grid points, as shown in Figure 4.7.
Tony Roberts, 14 Jul 2009
314
Uj+1
Uj1
Xj2
s ` uj (x, t)
Uj a b
cs
Xj1
h
b
Xj
= 21
Xj+1
Xj+2
61
Figure 4.7: schematic picture of the equi-spaced grid, Xj spacing h, the unknowns Uj , the artificial internal boundaries between each element (vertical
lines), and in the neighbourhood of Xj the field uj (x, t) which extends outside the element and, if = 1, is aimed to pass through the neighbouring
grid values Uj1 .
Suppose that we form m elements in the physical domain. For definite
simplicity, assume the field u(x, t) is spatially periodic with period thus
being L = mh .
(4.35)
That is, the field of the jth element when extrapolated to the surrounding
gridpoints, uj (Xj1 , t), is a weighted combination of two vital extremes:
when = 1, uj (Xj1 , t) is to be the field at those grid points, uj1 (Xj1 , t), to
in effect recover the physical continuity, as shown schematically in Figure 4.7;
but when = 0, the extrapolated value uj (Xj1 , t) is to be just identical to
the mid-element value uj (Xj , t) so that each element becomes isolated from
all neighbours. The parameter controls the weighting and hence controls
the coupling between neighbouring elements.
315
316
use a wide variety of variables to parametrise the slow manifold. Make the
convenient choice to use to be the evolving grid values Uj (t) defined by
Uj (t) = uj (Xj , t).
(4.36)
317
slow manifold corresponds to evolution of the grid values, and hence gives
the discrete model. From the computer algebra construction of the slow
manifold we find the corresponding evolution of the grid values, in terms of
centred differences 2 Uj , 4 Uj and so on, to be
i
h
j = 1 2 1 2 4 + 1 3 6 Uj + O 4 , j = 1, . . . , m .
(4.38)
U
12
90
2
h
When truncated at any order in coupling and then evaluated at full coupling, = 1 , the slow manifold evolution (4.38) forms a classic finite difference approximation to the heat equation (4.34). For two examples:
O 2 ,
O 3 ,
j = 1 2 Uj ;
U
h2
h
j = 1 2 Uj
U
h2
1 4
12 Uj
These right-hand sides are the classic second and fourth order accurate finite difference approximations to the second derivative in the heat equation (4.34). The higher the order in coupling of the truncation, the more
accurate the model.
Also see this increase of accuracy in the equivalent differential equation
for (4.38). Replace differences according to = 2 sinh 12 hx , the discrete
slow manifold model (4.38) transforms into the equivalent differential equation
U
2 U
= 2 +
t
x
4 U
1 2
h
(1
)
12
x4
6 U
4
1
h
(1
)(1
4)
360
x6
+ O h6 , 4 .
(4.39)
For full coupling = 1 this equivalent differential equation reduces to the
heat equation (4.34). Truncating to lower orders in coupling parameter ,
results in an error of lower order in grid spacing h because the (1 )
factors in the equivalent differential equation will not have been fully formed.
Computing to higher orders in coupling parameter results in an error of
higher order in h. This consistency between the slow manifold discrete
model (4.38) and the original differential heat equation (4.34) follows from
the specific form of the coupling conditions (4.35). As Figure 4.6 illustrates,
our modelling has dual support: consistency and centre manifold theory.
Tony Roberts, 14 Jul 2009
318
where the centred difference x on the left applies to the spatial dependence
in x whereas the on the right applies to the element index j. That is,
the coupling condition (4.35) empowers us to replace differences of the subgrid field by differences of the element amplitudes, albeit moderated by the
coupling parameter .
Now consider the heat equation. A classic operator relationship is that the
spatial derivative x = h2 sinh1 12 x (National Physical Laboratory 1961,
p.65, e.g.). Hence the heat equation (4.34) on the jth element is equivalent
to
2
2
1 1
2
sinh 2 x uj .
t uj = x uj =
h
Inverting the operator function on the right-hand side derives that the heat
equation is equivalent to
h
p i2
2 sinh h2 t uj = 2x uj .
Tony Roberts, 14 Jul 2009
319
= Uj =
sinh 2 Uj .
dt
h
Evaluating at coupling parameter = 1, this right-hand side is precisely
the correct centred difference operator formula for the heat equation (4.34).
However, generally we truncate at
some power in coupling , as in (4.38):
when truncating to errors O p the right-hand
side contains the correct
2p
terms up to centred differences
of O . Applied to a smooth solution, the
difference 2p = O h2p for grid spacing h. Then noting the division
by h2
2p2
in the above displayed equation, the resulting error is O h
. That is,
the discrete, slow manifold, model is consistent with the heat equation (4.34)
to this order of error.
This is a really neat proof. Its power is that it is not constructive; nowhere
in the proof did we need to construct any of the subgrid fields. Thus we
readily generalise the proof to a more extensive theorem. The following
theorem is independent of whether one is seeking a slow manifold model or
not: it addresses useful general consistency that follows from the specific
coupling conditions (4.35).
P
2k be
Theorem 4.2 (isotropic consistency) Let L = `(2 ) =
k=0 `2k
any linear, constant coefficient, isotropic operator with coefficient `2 6= 0 .
Then the discretisation of ut = Lu upon elements with interelement coupling
condition (4.35), when truncated to errors O p and evaluated at = 1 ,
is consistent with ut = Lu to errors O `2p 2p .
2
For example, Lemma 4.1 is the special case when L = 2x = h2 sinh1 21 .
Tony Roberts, 14 Jul 2009
320
Proof:
Recall that the coupling condition (4.35) empowers us to replace
differences of the subgrid field by differences of the element amplitudes, albeit moderated by the coupling parameter , as in equation (4.40). Consider
the dynamic equation ut = Lu . On the jth element it is equivalent to
t uj = Luj = `0 uj + ` 0 (x )2 uj
where
` 0 (2x ) =
`2k 2k .
k=1
Put `0 uj on the left and invert the operator ` 0 on the right-hand side (the
inverse formally exists as `2 6= 0) to derive that the equation is equivalent to
`0
(t `0 )uj = 2x uj .
(t `0 )Uj = 2 Uj .
Now revert the operator relationship to deduce the evolution of the grid
values is
dUj
j = `0 Uj + ` 0 (2 )Uj = `(2 )Uj .
=U
dt
Evaluating at coupling parameter = 1, this right-hand side is precisely the
correct centred difference operator
formula for the equation ut = Lu . So
p
when truncating to errors O the right-hand
side contains the correct
terms up to centred differences of O `2p 2p . That is, the discrete model is
consistent with the equation ut = Lu to this order of error.
(4.41)
321
Uj = 2
1 2 4
h
12
n
3
1
1
+
1 2
4 2 6 Uj + O 4
(4.42)
360
n
n
Evaluate at coupling = 1 to recover the physical model. This example demonstrates novel and powerful support for mapping fine scale
discrete dynamics to coarser scale discrete dynamics. Potentially one
could map to even coarser scale dynamics, and so on, to generate a
hierarchy of models across a wide range of scales!
Note that we should check the spectrum on decoupled elements to
confirm that such a slow manifold model should indeed emerge from
9
To apply Algorithm 4.3 here, imagine the subgrid variable i/n so changing fine
grid index i by one is the same as changing subgrid variable by 1/n. Then just change the
computation of the residual of the governing subgrid scale equation from the continuum
heat equation (4.34) to the fine scale discrete equation (4.41):
nu:=n**2/h**2;
pde:=-df(uj,t)+nu*(sub(xi=xi+1/n,uj)-2*uj+sub(xi=xi-1/n,uj));
The modified algorithm successfully terminates. Because the original updates are not
crafted for this particular subgrid scale operator, the number of iterations is larger. But
because the two subgrid scale operators are approximately the same, the algorithm does
terminate.
Tony Roberts, 14 Jul 2009
322
323
324
The output of Algorithm 4.3 is not in the form recorded in the slow manifold
model (4.37)(4.38): these equations express results using centred discrete
operators. The post-processing Algorithm 4.4 transforms the output into
the form reported. It replaces all the shift operations by the equivalent
discrete operator form
i|p|
h
Uj ,
Uj+p = 1 + sign(p) + 12 2
and invokes the identity 2 = 1 + 14 2 .
The post-processing Algorithm 4.3 also generates the equivalent differential
equation (4.39), for sufficiently smooth grid value fields, using the Taylors
series
X
n U
1
(ph)n n .
Uj+p = U(Xj + ph, t) =
n!
x
n=0
Exercise 4.26:
Modify Algorithm 4.3 to generate the slow manifold discrete model of the advection-diffusion equation (4.43)
for small ad
vection speeds c: that is, construct to errors O c3 say. Explore the
modelling in view of the results of the next Section 4.3.2.
325
4.3.2
Here we explore perhaps the simplest significant example pde: the linear
advection-diffusion equation
u
u 2 u
= c
+ 2,
t
x
x
(4.43)
326
k ukj (x, U) ,
such that
j = gj =
U
k=0
k gkj (U) ,
k=1
(4.44)
where the superscripts on uj and gj are an index and do not denote exponentiation. Usually we construct models using iteration in computer algebra.
However, here we resort to explicit hand calculations and so invoke the explicit expansion (4.44) in the coupling parameter .
Now look at each equation. Substitute (4.44) into the advection-diffusion
equation (4.43) and equate like powers of to determine
k
XX
i
l=1
gli
ukl
j
Ui
= c
ukj
x
2 ukj
x2
k = 0, 1, 2, . . . .
(4.45)
Similarly substituting (4.44) into the amplitude condition (4.36) and equating powers of requires
u0j (Xj , U) = Uj ,
for k 1 .
(4.46)
u0j
x
2 u0j
x2
= 0,
327
u0j (Xj , U) = Uj ,
u0j (Xj1 , U) u0j (Xj , U) = 0 .
These have solutions which are piecewise constant: in the jth element
u0j (x, U) = Uj .
(4.48)
As before, these piecewise constant solutions form the base of the subgrid
fields in the slow manifold model.
Move on to find the O corrections by solving the k = 1 version of equations (4.45)(4.47):
c
u1j
2 u1j
= g1j ,
x
x2
u1j (Xj , U) = 0 ,
+
2 Uj
4 sinh2 (ch/2) 2 sinh (ch/2)
+ Uj
(4.49)
2
Uj + 1 2 Uj ,
h
h
ch cosh (ch/2)
.
2 sinh (ch/2)
and g1j =
where
1 =
(4.50)
(4.51)
(4.52)
Tony Roberts, 14 Jul 2009
328
E +
+
E
,
h
12
h
2 12
2 12
ut = cux + uxx +
which increases the weight of the upwind grid point (E1/2 is upwind if
c is positive). Either interpretation, as enhanced dissipation or upwind
correction, is well known to be stabilising for finite advection speeds c.
It is interesting to explore the large ch limit when there is strong advection
across each element. From (4.51), 1 ch/2 as ch and is indeed
within a few percent of this value for ch > 4 , see Figure 4.8. Thus for large
advection speed c on a finite width grid, the slow manifold analysis promotes
the model10 (written in terms of the backward difference operator =
1 E1 )
j c Uj = c (Uj Uj1 ) .
U
(4.54)
h
h
This evolution of the grid values is not, and need not be, consistent with
the advection-diffusion pde (4.43) as h 0 because it applies for finite ch.
That the upwind model should be relevant to (4.43) comes from the slow
manifold expansion in the coupling parameter , albeit evaluated at = 1
(via the holistic arrows in Figure 4.6).
10
If the advection velocity is negative c < 0 , then various signs change and the large ch
model remains an upwind model, but is consequently written in terms` offorward differences. This also occurs for the later model (4.59) accurate to errors O 3 .
Tony Roberts, 14 Jul 2009
329
1
62
42
4.5
4
3.5
3
2.5
2
1.5
1
0.5
0
10
Figure 4.8: coefficients of the slow manifold models (4.52) and (4.57) as a
function of advection speed and grid spacing, ch. The dotted lines are the
1
large ch asymptotes: 1 12 ch ; 2 41 ch 12 ; 2 21 ch
.
330
j
Then see that the moment generating function G(z, t) =
j=0 z Uj (t) is
that for a Poisson probability distribution with parameter ct/h, namely
G(z, t) = exp[(z 1)ct/h] . Hence the mean location and variance of Uj are
j = 2j =
ct
h
x = ct
and 2x = cht .
(4.55)
Thus for ch not small : the extreme upwind model (4.54) has precisely the
correct advection speed c; and although the variance is quantitatively wrong,
cht instead of 2t, at least it is qualitatively correct for finite ch. The slow
manifold model (4.52) is consistent for small h and has the virtue of being
always stable and will always maintain non-negativity of solutions no matter
how large the advection speed c.
Second order coupling Second, explore the holistic finite difference model
with second order
interactions between adjacent elements, that is, the er
rors are O 3 . The details of u2j are of little direct interest here. The finite
difference model depends directly upon
c
1
g2j = + 2 3 Uj + 2 2 1 2 2 4 Uj ,
(4.56)
h
h
ch cosh (ch/2) 1 ch cosh (ch/2)
1
where 2 =
+
,
3
2
4 sinh (ch/2)
2
8 sinh (ch/2)
4 sinh (ch/2)
1
cosh (ch/2)
1
and 2 =
+
.
2
2 2 sinh (ch/2) ch sinh (ch/2)
j = g1 +2 g2 evaluated
Observe the marvellous feature that when we form U
j
j
at = 1 the terms in 1 (ch)2 neatly cancel. The model for the linear
advection-diffusion thus reduces to
j = c 2 3 Uj + 1 2 2 4 Uj .
(4.57)
U
h
h2
As it involves third and fourth differences, this model is a five point wide
discretisation.
Tony Roberts, 14 Jul 2009
331
h4
(c x )3 uxxx + O h6 ,
90
(4.58)
and so is consistent to O h4 , independent of advection speed c, with the
advection-diffusion pde (4.43). With the specific coupling conditions (4.35),
as we compute to higher order in coupling parameter it appears that we
may get higher order consistency with the original pde (4.43).
For large advection speed or grid size, large ch, the discretisation (4.57)
behaves astonishingly well. Using the large ch approximations plotted in
Figure 4.8 for 2 and 2 , the model (4.57) reduces to simply
c
1 2
1
Uj =
+ Uj + 2 2 Uj
h
2
h
c
1
(Uj2 4Uj1 + 3Uj ) + 2 (Uj2 2Uj1 + Uj ) . (4.59)
=
2h
h
This large ch model uses only backward differences to incorporate secondorder upwind estimates of the derivative, hx + 21 2 , and the second
derivative, h2 2x 2 . To show its good properties,11 consider again a
point release
j = 0 at time t = 0 . The moment generating function
Pfrom
j
G(z, t) = j=0 z Uj (t) for the evolution governed by (4.59) is readily shown
to be
t
ct
2
G(z, t) = exp (z 1)(z 3) + 2 (z 1) .
(4.60)
2h
h
Then since
G
ct
=
z z=1
h
and
2
ct
ct 2t
2 G
=
+ 2,
2
z z=1
h
h
h
11
The upwind difference model (4.59) is only stable for ch > 2/3 . However, from
Figure 4.8 the approximation (4.59) is only applicable to (4.57) for ch greater than about 4;
thus its instability for very much smaller ch is irrelevant.
Tony Roberts, 14 Jul 2009
332
ct
2t
and 2j = 2
h
h
x = ct and 2x = 2t .
(4.61)
This predicted mean and variance following a point release are exactly correct for all time for the advection-diffusion pde (4.43). This specific result
applies to all finite advection speeds c and all finite grid spacings h whenever
ch is large enough.
Summary Creating finite differences which, as shown in Figure 4.6, are
both consistent for small element size h and also holistically derived via
centre manifold theory lead to models which are remarkably accurate and
stable over a wide range of parameters.
The interelement coupling models linear dynamics consistently
The slow manifold discretisation of advection-diffusion is apparently consistent with the pde as the grid spacing h 0: equation (4.53) demonstrates empirically that the O 2 model is O h2 consistent;
whereas
equa3
4
tion (4.58) demonstrates empirically that the O model is O h consis
tent; further computation suggests that the O p model is O h2p2 consistent. The following theorem, complementing Theorem 4.2 on isotropic
dynamics, establishes that we expect such consistency for quite general linear microscale dynamics.
Theorem 4.3 (consistency) ??
Proof:
Consider the microscale evolution equation ut = Lu (differential
or fine scale discrete) where the linear operator has formal expansion
0
L = `0 + ` (D) =
`k Dk
for operator Dx = c1 x x + c2 2x .
k=0
333
D = c1 + c2 2 .
That is, the specific coupling conditions (4.35) transform first and second
differences of the field into corresponding differences of the grid values, but
moderated by the coupling parameter .
Then for each subgrid field uj the evolution equation t uj = Luj is equivalent to t uj = `0 uj +` 0 (Dx )uj . Hence, by operator algebra, ` 0 1 (t `0 )uj =
Dx uj -the inverse of operator ` 0 formally exists as `1 6= 0 . Evaluate at
the central grid point of the jth element, x = Xj : the left-hand side only
involves time derivatives and so commutes with the evaluation; the righthand side simplifies by the specific coupling conditions (4.35); the result is
` 0 1 (t `0 )Uj = DUj . Revert the operator algebra to deduce the equation for the grid values t Uj = `0 Uj + ` 0 (D)Uj . Evaluated at full coupling,
= 1 , this is precisely consistent with the original microscale evolution
equation ut = Lu .
However, in practise we approximate
the grid model t Uj = `0 Uj +` 0 (D)Uj
p
by truncating to errors O . Evaluating
such a truncation at full coupling,
4.3.3
(4.62)
334
and test many numerical and analytical methods. The nonlinearity parameter controls the importance of the nonlinear advection term (Exercises
4.15 and 4.16 show how this parameter helps control approximations). This
section shows how to derive holistic discretisations of Burgers equation. The
methods are easily adaptable to discretise other reaction-advection-diffusion
pdes in one space dimension.
See also (Roberts 2001).??
Algorithm 4.5 derive the slow manifold holistic discretisation for Burgers
equation (4.62) on finite elements. Simply make the following modifications
to Algorithm 4.3.
Improve printing by also including factor alpha.
Change the order of error to also truncate in small nonlinearity
parameter by, say, let {gamma^3=>0, alpha^3=>0};
Change the computation of the residual to Burgers pde (4.62):
pde:=-df(uj,t)-alpha*uj*df(uj,x)+df(uj,x,2);
= gj0 + Resj
such that
(4.63)
In terms of the subgrid variable = (x xj )/h for the jth element these
equations are
2 0
1 uj
= gj0 + Resj
h2 2
Integrate this pde for uj0 over an extended jth element with weight function
w() = 1 || :
Z1
2 0
1 uj
(1 ||) 2
d =
h 2
1
Z1
1
(1 ||) gj0 + Resj d .
335
Integrate the left-hand side by parts twice, and simplify the right-hand side
slightly, gives
Z1
1
0
1
0
[(1 + )u u]1 + [(1 )u + u]0 = gj +
(1 ||) Resj d .
h2
1
Since u , equivalently hux , must be continuous at the centre of the element
= 0 , the left-hand side of this equation becomes
lhs =
=
1 0
0
0
u
|
2u
|
+
u
|
=1
=0
=+1
j
j
j
h2
1
{Res+ + Res } ,
h2
1
= 2 (Res+ + Res )
h
Z1
(1 ||) Resj d .
(4.65)
336
Corollary 4.4 There exists an exponentially attractive slow manifold discrete model of the pde
u
2 u
=
+ f(u, ux , . . .)
t
x2
(4.66)
with the nonlocal coupling conditions (4.35), for smooth enough f. With
the domain divided into m overlapping elements by (4.35), and periodic
boundary conditions on u, u(x + mh, t) = u(x, t) , the m dimensional, slow
manifold, discrete model may be written as
u = vj (x, U, , )
such that
j = gj (U, , ) ,
U
(4.67)
where Uj measures the amplitude of the field u in the jth element. The
discrete model (4.67) of the pde (4.66) may be constructed as an asymptotic expansion in two small parameters: the coupling parameter , and the
nonlinearity parameter .
Proof:
When parameters = = 0 there exists an m dimensional subspace of equilibria of piecewise constant subgrid fields: uj = constant. Linearised about each of these equilibria, and assuming constant grid spacing h for simplicity, the spectrum of the linear dynamics is that of the diffusion equation, namely, {0, 2 /h2 , 42 /h2 , 92 /h2 , . . .} with the zero
eigenvalue having multiplicity m. Centre manifold theory then assures us
that for smooth enough f: there exists a m dimensional slow manifold,
parametrised by grid values Uj say, that is exponentially attractive, roughly
like exp(2 t/h2 ), and may be constructed globally in grid values Uj but
locally in the small parameters and .
Cater for finite nonlinearity For some problems, such as Burgers equation, the nonlinearity vanishes for the entire subspace of piecewise constant
fields. Then the nonlinearity parameter need not be forced small. As the
nonlinearity vanishes in the m dimensional subspace, then the m eigenvalues of zero remain zero for finite parameter . By continuity, the negative
eigenvalues in the spectrum will remain negative for some finite domain in
parameters. Hence the slow manifold will exist and be attractive globally
Tony Roberts, 14 Jul 2009
337
4.3.4
Exercises
Exercise 4.28: Burgers equation Write a page on the theoretical support for discrete models of the fairly general class of nonlinearly modified diffusion dynamics
u
2 u
=
+ f(u, ux )
t
x2
with periodic boundary conditions that u(x, t) = u(x + mh, t) . Break
the domain into m elements of size h with nonlocal coupling conditions
u(xj1 , t) u(xj , t) = (Uj1 Uj ) .
Write computer algebra to construct the consequent discrete models of
Burgers equation obtained when the nonlinear term f = u u
x . Write
to truncate to specified powers in small parameters and . Compare
with the models produced by http://www.maths.adelaide.edu.au/
anthony.roberts/holistic1.html (your models should be the same).
338
2 uj0
x2
= Resj
where
Recalling Exercise 4.20, find the null space of the adjoint of this problem. Hence determine the solvability condition (4.65) on the above
residuals for this problem.
(2 Uj1 2 Uj ) .
h2
Write computer algebra to construct the consequent discrete models
of this KuramotoSivashinsky equation. I expect the solvability condition for updates to the evolution will be of the form
Z1
gj0
(1 ||)Residual d + .
1
339
4 uj0
= Resj
x4
uj0 (xj1 , t) uj0 (xj , t) = Res
0 ,
=
0
0
ujxx
(xj1 , t) ujxx
(xj , t) = Res
2 .
with
340
4.4
Summary
basicmeth??
Artificially changing boundary conditions empowers us to rigorously
model pdes on finite domains, 4.2.1.
Chopping up a domain into finite elements using coupling conditions
between the elements similarly empowers us to form discrete models
of pdes, 4.2.2. Wonderfully, the governing pde itself determines the
physical field within each finite element. In contrast, traditional finite differences and finite elements impose an assumed shape for the
subgrid scale field.
Tony Roberts, 14 Jul 2009
4.4. Summary
341
342
Chapter 5
344
0.5
0.0
y
0.5
1.0
1.00.80.60.40.20.0 0.2 0.4 0.6 0.8 1.0
x
Figure 5.1: some trajectories of the pair of coupled odes (5.1) evolve towards
the stable equilibrium at the origin.
introduction is given, but, for example, we see how linearisation is indeed
valid in two dimensional systems.
??separate
?? resonance
5.1
Warning:
345
1.0
0.5
0.0
0.5
1.0
1.00.80.60.40.20.0 0.2 0.4 0.6 0.8 1.0
Figure 5.2: some trajectories of the pair of coupled odes (5.1) in the transformed -plane.
odes
x = 2x ,
and y = y + x2 ;
(5.1)
Figure 5.1 shows some trajectories. Your exploration of the odes may have
followed the following path: the only equilibrium (fixed point) of the dynamical system is x = y = 0 ; the linearisation about this equilibrium is
x 2x and y y ; both of these describe decaying dynamics and so the
equilibrium (x, y) = (0, 0) is a stable node. But how do we really know that
we can neglect the nonlinear term x2 in the y-equation? True, the argument that the term x2 is negligible near (0, 0) is plausible; but is it correct?
Normal form transformations provide a sound route to answer this question.
A transformation to a normal form tries to simplify the description of the
dynamics so that we easily deduce relevant features such as stability or
classification. For the example system (5.1) try changing to new coordinate
Tony Roberts, 27 Jun 2008
346
and y = 13 2 .
(5.2)
This coordinate transform slightly relabels the points in the xy-plane: the
origin is still the origin; but for example, the point with xy-label ( 21 , 1) has
-label (1, 1). The coordinate transform is a slight relabelling because
near the equilibrium of interest, the origin, the coordinate transform is a
near identity, x and y . But the nonlinear part of the transform
bends the xy-plane upwards to straighten out the dynamics in the plane as you may see Figure 5.2. Now find the dynamics of (5.1) in the new
coordinates.
First, = x so = x = 2x = 2 . The variable evolves according
to the linearisation derived earlier.
Second, from the equation (5.1) y = y + x2 = + 31 2 + 2 =
+ 43 2 , whereas from the transformation y = 23 =
2
4 2
347
and y = y ,
(5.3)
348
1.0
0.5
0.5
0.0
0.0
0.5
0.5
1.0
1.00.80.60.40.20.0 0.2 0.4 0.6 0.8 1.0
1.0
1.00.80.60.40.20.0 0.2 0.4 0.6 0.8 1.0
Figure 5.3: trajectories of the pair of coupled odes (5.3) approach the line
y = 0 (left), but seen in the transformed -plane (right) they appear to do
so in a straight fashion.
5.1.1
Example 5.2:
Find a near identity change of variables, near the origin,
to place the system
x = 2x + y2
and y = y ,
(5.4)
into its normal form = 2 and = . That is, find the change of
variable that straightens the dynamics into the classic linear saddle
as shown in Figure 5.4.
Solution: The y equation is already in its normal form, so let us just
try to transform the x variable: pose x = + X(, ) and y = for
some as yet unknown modification to the change of variables X(, ).
Since this change of variables is to be a near identity we want X(, ) to
be small (compared with the dominant linear term x ). Substitute
into the right-hand side of the x equation:
2x + y2 = 2 + 2X + 2 .
Tony Roberts, 27 Jun 2008
349
1.0
1.0
0.5
0.5
0.0
0.0
0.5
0.5
1.0
1.00.80.60.40.20.0 0.2 0.4 0.6 0.8 1.0
1.0
1.00.80.60.40.20.0 0.2 0.4 0.6 0.8 1.0
Figure 5.4: trajectories of the pair of coupled odes (5.4) (left) are a bent
version of the classic saddle shown in the -plane (right) after the coordinate transform (5.5).
But also x = + X(, ) so, using the chain rule, the left-hand side of
the x equation is
X X
X
X
x = +
+
= 2 + 2
X
X
2X = 2 .
This is your first example of a homological equation. Such homological equations may be quite tricky to solve. Here use the following
argument:
If the modification X has any dependence, then X
would be
X
non-zero, hence the term on the left 2 would generate some
dependent terms on the left-hand side. But there are no dependent terms on the right-hand side so we discard any thought
of dependence in the modification X.
Tony Roberts, 27 Jun 2008
350
and y =
(5.5)
transforms the ode (5.4) into the simple normal form = 2 and
= . Consequently, the dynamics of the ode (5.4) are simply a
curved or bent version of those of the simple linear system = 2
and = .
This linear system is a saddle, since grows exponentially as decays
exponentially, and thus the origin in the ode (5.4) is a similar saddle.
In coupled pairs of nonlinear odes, there are special curves in state space
that guide the shape of the overall evolution. In Example 5.2:
the line y = 0 attracts all solutions of the odes (5.4)we call the line
y = 0 the unstable manifold (of the equilibrium at the origin); whereas
the curve x = y2 /4 separates the solutions that grow x and
those that grow to x we call the curve x = y2 /4 the stable
manifold (of the equilibrium at the origin).
Interestingly, converse statements hold when we run time backwards: the
stable manifold is attractive in reversed time; and the unstable manifold
separates solutions between y in reversed time. Also, the stable and
unstable manifolds of a equilibrium have other roles in other situations: for
example, in chaotic maps, they tangle together in a way that characterises
crucial aspects of the chaos. But their complementary nature always holds
and underlies their definition (Kuznetsov 1995, 2.2).
Definition 5.1 Consider a dynamical system of odes
u = f(u) ,
Tony Roberts, 27 Jun 2008
u Rn ,
(5.6)
351
with an equilibrium at u :
the set of initial conditions whose subsequent evolution approaches the
equilibrium is called the stable manifold,
W s = {u(0) : u(t) u , t +} ;
(5.7)
(5.8)
Note: at this stage we have no right to call these sets manifolds because we
do not know that they are smooth nor that they have a definite dimension;
Theorem 5.3 provides the assurance.
Example 5.3: Stable and unstable manifolds
In the system (5.4), the unstable manifold W u of the origin is
the line y = 0 , whereas the stable manifold W s is the curve
x = y2 /4 .
In the system (5.3), there is no non-trivial unstable manifold of
the origin (W u = {0}), whereas the stable manifold W s is the
curve x = y2 /2 .
In the system (5.1), there is again no non-trivial unstable manifold of the origin (W u = {0}), whereas the stable manifold W s is
the entire xy-plane.
Observe in these examples that the stable and unstable manifolds are
indeed smooth curves in the plane and hence can be indeed called
manifolds. Also observe that the non-trivial manifolds are tangent
to the eigenvectors of the linearised evolution near the equilibrium.
These are general properties formalised in the next theorem.
352
5.1.2
Now lets return to coordinate transforms. Reconsider the stable and unstable manifolds of Examples 5.1 and 5.2 but now see them clearly in the
normal form of the dynamics.
The normal form of the system (5.4) is = 2 and = . The general solution of this pair of linear equations is (, ) = (0 e2t , 0 et ).
Thus: the stable manifold of solutions 0 as t + is simply
0 = 0 , namely the -axis; and similarly the unstable manifold of
solutions 0 as t is simply 0 = 0 , namely the -axis.
Then, since we know the coordinate transform (5.5) that x = 14 2
and y = : we know the stable manifold 0 = 0 becomes x0 = 14 y20
as commented upon earlier; similarly, the unstable manifold 0 = 0 is
simply y0 = 0 .
The normal form of the system (5.3) is = 0 and = . The general
solution of this pair of linear equations is (, ) = (0 , 0 et ). Thus:
Tony Roberts, 27 Jun 2008
353
5.1.3
We have explored coupled pairs of odes that are sufficiently simple that
we could straightforwardly construct the normal form transformation. In
general, a normal form is harder to construct. Moreover, in general, a normal form has to be constructed as asymptotic series. Here we explore how
iteration empowers us to asymptotically approximate normal forms.
For simplicity we restrict attention to pairs of odes. Let us start with a
modified version of Example 5.2.
Example 5.4: first approximation
Use iteration to find a near identity change of variables, in a neighbourhood of the origin, to place the
system
x = 2x + y2 and y = y + x3 ,
(5.9)
into its normal form = 2 and = . That is, find the change of
variable that straightens the dynamics into the classic linear saddle
as shown in Figure 5.5. First we appear to succeeed, but second show
we actually fail, and lastly we show how to always succeed.
354
1.0
0.5
0.5
0.0
0.0
0.5
0.5
1.0
1.00.80.60.40.20.0 0.2 0.4 0.6 0.8 1.0
1.0
1.00.80.60.40.20.0 0.2 0.4 0.6 0.8 1.0
Figure 5.5: trajectories of the pair of coupled odes (5.9) (left) are a bent
version of the classic saddle shown in the -plane (right) after the coordinate transform (5.10).
Substitute into the x ode of (5.9), using the chain rule for time derivatives:
x = 2x + y2
X 0 X 0
2
+
+
= 2 + 2X 0 + 2 + 2Y 0 + Y 0
X 0
X 0
2
= 2 + 2X 0 + 2 + 2Y 0 + Y 0
2 + 2
X 0
X 0
2
+2
2X 0 = 2 + 2Y 0 + Y 0
X 0
X 0
+2
2X 0 2
355
Now do the analogous analysis for the y ode of (5.9). Using the chain
rule for time derivatives deduce
y = y + x3
Y 0 Y 0
2
3
+
+
= Y 0 + 3 + 32 X 0 + 3X 0 + X 0
Y 0
Y 0
3
2
+ 2
= Y 0 + 3 + 32 X 0 + 3X 0 + X 0
Y 0
Y 0
3
2
+2
+ Y 0 = 3 + 32 X 0 + 3X 0 + X 0
Y 0
Y 0
+2
+ Y 0 3
and y + 17 3
(5.10)
transforms the ode (5.9) into the simple normal form = 2 and
= . Consequently, the dynamics of the ode (5.9) are simply a
curved or bent version of those of the simple saddle system = 2
and = as shown in the transformation of Figure 5.5.
But the transformation is not quite so simple as seen in this example. See
that at one stage for each ode in the pair (5.9) we neglected small terms
in the corrections. This neglect implies the transformation (5.10) is only
approximate, as indicated. Seek higher order approximations by iteration
with the aid of computer algebra.
Example 5.5: computer algebra iterates effortlessly.
The iteration, whether by computer algebra or not, depends upon solving homological equations of the form
2
X 0
X 0
2X 0 = Res5.9,x
356
357
Y 0
Y 0
+ Y 0 = Res5.9,y .
2 3
21
1 2 2
4
+ 12 2 3 + O 6 + 6 ,
3
16
4 + O 6 + 6 .
(5.11)
(5.12)
See that Algorithm 5.1 computes the residual for each equation, then
updates the transformation by:
placing a term p q /(2p q 2) into the x transform for each
term p q in the x residual;
Tony Roberts, 27 Jun 2008
358
This little program is beautiful. It provides a simple and flexible base for
dealing with almost any pairs of odes. Unfortunately, the program is also
wrong!
Try executing it to construct an asymptotic approximation to sixteenth order
by using let zz^17=>0: fifteenth order is fine, but when you attempt to find
the sixteenth order you encounter a fatal error through an attempt to divide
by zero!
Example 5.6: resonance affects the normal form.
Explore this critical issue of zero division in the normal form of the odeS (5.9). Recall
that the iteration depends upon solving homological equations of the
form
X 0
X 0
2X 0 = Res5.9,x
Y 0
Y 0
2
+ Y 0 = Res5.9,y .
and
359
Such problems zero divisors occur frequently in asymptotic approximation of dynamics. Generically they are called resonances because
historically these zero divisors were first encountered in analysing oscillations and the zero divisors physically signalled a resonant interaction
between component oscillators.
The resolution of the zero divisor problem is that we must leave some
of the terms in the odes for x and y in the odes for the evolution
of the new variables and . That is, we can no longer insist that
the normal form is precisely that of a linear saddle. Instead there are
inescapable nonlinear modifications to the evolution.
Solution:
pose
such that
for some as yet unknown near identity transform X(, ) and Y(, ) ,
and for some minimal set of terms remaining in the evolution =
F and = G . Now adopt the approach of supposing we know an
approximation to the transformation and seeking corrections.
Suppose that at some iteration we have a current approximation to
the transform and also to the normal form odes. Seek corrections
X 0 (, ) and Y 0 (, ) to the transform and corrections F 0 (, ) and G 0 (, )
to the evolution such that after transformation x = X + X 0 and y =
Y +Y 0 the evolution = F+F 0 and = G+G 0 better describes the original dynamics of the odes (5.9). The major complication with substituting the coordinate transformation into (5.9) is the time derivatives:
consider the x derivative
X X 0
X X 0
new =
[x]
+
+
X X 0
X X 0
0
=
+
(F + F ) +
+
(G + G 0 )
360
|
{z
}
=x
X 0
X 0 0
F0 +
G
|
{z
}
negligible
current + F 0 + 2
[x]
X 0
X 0
upon recognising
the time derivative x for the current approximation,
that the product of small corrections are negligible, and
approximating the coefficients of corrections by their first order
X
approximations ( X
1 , F 2 , 0 and G ).
0
X
Note that in the retained terms on the left-hand side, 2 X
2X 0 , are all of size X 0 as the multiplictions by small and are
countered by the divisions by and in the derivatives. The righthand side of the x equation is straightforward:
[2x + y2 ]current + 2X 0
upon neglecting products of corrections, and approximating coefficients of corrections by their order 0 approximation (2Y 0). Equate
the two sides and rearrange to
F 0 + 2
X 0
X 0
2X 0 = Res5.9,x
Y 0
Y 0
361
Y 0
Y 0
+ Y 0 = Res5.9,y
which is exactly as before except we now avoid dividing by zero by placing any troublesome terms in the evolution via G 0 rather than trying
to absorb them into the coordinate transform via Y 0 .
Execute Algorithm 5.2 to find the normal form of the odes (5.9) is
6 10
17
17
= 2 + 235829
677376 + O +
21439 5 11
+ O 17 + 17 .
(5.13)
and
= 112896
Procedures homx and homy compute the homological operators which
are used in several places to flag whether a particular term will give
rise to a zero divisor or not. This algorithm works to arbitrarily high
order; it is limited only by computer memory and speed.
The conclusion is: we cannot smoothly transform the odes (5.9) into
the linear saddle = 2 and = ; terms such as those in (5.13)
may be inescapable.1
However, recall that the terms which cannot be removed from the
evolution are those of the form p q for 2p = q + 2 and since
exponent q 0 thus p 1 . Consequently, = 0 is invariant in
the normal form evolution (5.13) as the variable will always be a
Hence the stable manifold of the normal form is precisely
factor in .
the linear space = 0 . The stable manifoldin the xy-plane then
comes from (5.11),
namely x = 14 y2 + O y6 as (5.12) reduces to
6
y = + O when = 0 .
1
Allowing non-smooth logarithms into the coordinate transform typically enables us
to remove even these terms (just as logarithms enable us to form power series solutions
of odes when the indicial equation has repeated roots). However, we do not pursue this
route here.
362
Algorithm 5.2 correct reduce code to find the normal form transformation to put odes (5.9) into the normal form (5.13). Resonance occurs
at O 6 10 + 5 11 .
% xi & eta are new dynamical variables
depend xi,t; depend eta,t;
let { df(xi,t)=>g, df(eta,t)=>h };
% dummy inverse operator
operator linv; linear linv;
procedure homx(r); 2*xi*df(r,xi)-eta*df(r,eta)-2*r;
procedure homy(r); 2*xi*df(r,xi)-eta*df(r,eta)+r;
% linear approximation
factor zz;
x:=xi*zz; y:=eta*zz;
g:=2*xi; h:=-eta;
% iterate to solve
let zz^17=>0;
repeat begin
resx:=df(x,t)-2*x-y^2;
g:=g+(gd:=-(linv(resx,t) where {
linv(~r,t)=>r when homx(r)=0,
linv(~r,t)=>0 when homx(r) neq 0 }))/zz;
x:=x-(linv(resx+gd,t) where linv(~r,t)=>r^2/homx(r));
resy:=df(y,t)+y-x^3;
h:=h+(hd:=-(linv(resy,t) where {
linv(~r,t)=>r when homy(r)=0,
linv(~r,t)=>0 when homy(r) neq 0 }))/zz;
y:=y-(linv(resy+hd,t) where linv(~r,t)=>r^2/homy(r));
end until {resx,resy}={0,0};
363
Conversely, recall that the terms which cannot be removed from the
evolution are those of the form p q for q = 2p + 1 and since
exponent p 0 thus q 1 . Consequently, = 0 is invariant in the
normal form evolution (5.13) as the variable will always be a factor
Hence the unstable manifold of the normal form is precisely the
in .
linear space = 0 . The unstable manifold
in the xy-plane then comes
from (5.12), namely y = 17 x2 +O x6 as (5.11) reduces to x = +O 6
when = 0 .
Conclude: although we have not demonstrated it in general, similar characteristics hold for general systems of pairs of coupled nonlinear odes: a
smooth nonlinear coordinate transform can straighten the stable and unstable manifolds.
Exercise 5.7:
of odes
and + B3 .
Exercise 5.8:
Modify Algorithm 5.2 to find the normal form transformation of the pair of odes
x = 2x + x2 + y2 ,
y = 3y + x2 y2 .
Hence deduce asymptotic approximations to the shape of its stable
and unstable manifolds.
364
5.2
Sorry:
y = y + x2 2y2 .
and y = + Y(, ) ,
where X and Y are nonlinear functions of their arguments. The evolution in these new variables will be of the form
= g()
and = + h(, ) ,
Y
h+Y
X
X
h
,
Y
Y
= ( + X)2 2( + Y)2 g
h
.
= ( + X)( + Y) g
= ,
365
= 2 22 .
Simplify the evolution by putting as much as we can in the coordinate transform X and Y, and as little as possible into the evolution
g and h. See that X = and g = 0 satisfies the first x-equation.
For the second y-equation, the right-hand side is quadratic in
and so let us see what a quadratic in Y can do for us: try
Y = a2 + b + c2 then the equation becomes
h + a2 c2 = 2 22 ,
so we choose a = 1 , c = 2 and the evolution h = 0. The first
approximation is then
x = + ,
y = + 2 + 22 ,
= 0 ,
= .
Y
h+Y
= 22 3 + h.o.t. ,
= 2 22 22 43 + h.o.t. .
366
y = +2 +22 +23 ,
= 3 ,
= 22 .
Now explore crucial features of this system and the normal form transformation that relates it to the original.
1. See that = 0 is an invariant manifold: specifically, it is the slow
manifold of the system. Indeed = 0 will always be the slow
manifold as we will only ever put terms in the -equation which
^
are linear in : it has the form = (1 + h())
no matter what
order the truncation. Find the shape of the slow manifold in the
original variables by setting = 0 in the transformation to see
x = and y = 2 to this order of error.
Also set = 0 in the -equation to see the evolution on the slow
manifold is = 3 , noting that x = .
2. The relevance of the model follows because the evolution is identical for all . Thus the evolution of is independent of whether
a state is on or off the slow manifold = 0 . Consequently, all
states starting with the same 0 have exactly the same long term
evolution provided they are indeed attracted to the centre manifold.
In the orginal variables, this says that an initial condition (x0 , y0 )
should be projected along curves of constant to the slow manifold in order for the slow manifold model to make correct long
term predictions.
3. We can obtain indictaions of the likely extent of attraction to the
slow manifold model by looking for where the separation of time
scales and/or the coordinate transform breaks down.
Tony Roberts, 1 Mar 2008
367
J=
=
.
Y
2
1 + 4 + 63
1 + Y
(1 + + 22 )(1 + 4 + 62 )
.
2(1 + 2)
Summary
368
5.3
Sorry:
y = 2x + y + 2xz ,
Consider
z = y2 .
1 1 0
2
1 0
0
0 0
which has eigenvalues i and 0. Thus there is an oscillating mode and
a slow mode. We seek the normal form of the dynamics that separates
the oscillations from the slow dynamics as far as possible.
Linearly, although the slow mode z is decoupled from the oscillation
variables x and y which simplifies the analysis, the oscillation variables
x and y are not in a cannonical form for oscillations and this complicates aspects of the algebra. In particular the fast oscillations are elliptical in x and y. To see this construct an energy-like invariant of the
oscillations: seek E = ax2 + bxy + cy2 which is constant in time in the
linear dynamics; then E = (2a + 2b)x2 + (2a + 4c)xy + (b + 2c)y2
Tony Roberts, 1 Mar 2008
369
y = Y(, , ) ,
z = Z(, , )
= g(, , ) ,
= h(, , ) ,
where being near identity transform means that the linear components
in these transforms and evolution are
X ,
Y ,
Z ,
f ,
g 2 + ,
h 0.
370
Y 0 = C + D .
X 0 = ,
g0 = Y 0 = 0 .
5.4
Chapter summary
Chapter 6
6.3
6.4
374
384
6.2.1
6.2.2
6.2.3
6.2.4
Modulation of oscillations . . . . . . . . . . . . .
402
6.3.1
6.3.2
6.3.3
6.3.4
417
6.4.1
6.4.2
371
372
6.5
6.4.3
6.4.4
433
The:aim:of:this:chapter
Besides water, the worlds oceans have two major constituents: heat and
salt. Both these constituents alter the density of the water and so may drive
buoyant flows. In different parts of the oceans water masses have markedly
different combinations of these two characteristics. Such variations may
lead to some surprising dynamics such as that of the salt fountain (briefly
discussed by Turner (1973) [p.252]).
Another interesting effect due to the two constituents is in the melting of
icebergs. It has been proposed to tow icebergs near to the shore of arid
population area such as south Australia. There the iceberg would melt, the
melt water being fresh would float to the top of the sea water and be collected
for drinking or other purposes. However, experiments show that when the
side of an iceberg melts circulation cells of large horizontal extent and fairly
thin in the vertical are formed which carry the fresh water out horizontally
away from the iceberg and then mixed into the sea water (Turner 1973,
Fig 8.10). The presence of two differently diffusing components can have
surprising effects.
Here we instead investigate the oscillatory instability that may occur when
cool fresh water overlies warm salty water. Even when the net density
gradient is stable, with light fluid lying above denser fluid, it may be that
an instability occurs. Consider a small parcel of warm saline water displaced
upwards (Turner 1973, pp.252253). It loses heat much more rapidly than
it does salt and so quickly becomes much denser than its surroundings and
falls. But it is now cooler than its original surroundings, and the subsequent
heat gain when it falls back into the warm saline water, lags behind the
displacement of the fluid parcel. Thus the parcel overshoots, and oscillations
are initiated.
In all the examples of Chapter 3, the dynamics on the centre manifold
have been based on the eigenvalue of precisely zero, not the more general
Tony Roberts, 1 Mar 2008
373
case where just the real part of the eigenvalue is zero but the imaginary
part is non-zero. As we see in 6.2, in this latter case constructing the
centre manifold is more difficult technically. However, because an eigenvalue
with a non-zero imaginary part is always associated with oscillations, such
a case is important in practise as cases arises in the common transition from
steady to oscillatory dynamics. Many of the same issues also arise in wave
propagation.
When the critical eigenvalues are pure imaginary, the basic dynamics on
a centre manifold is that of oscillation. These oscillations are usually relatively fast at some finite frequency determined by the imaginary eigenvalues.
Rather than resolving the details of these oscillations, we are usually much
more interested in the relatively slower evolution of the amplitude of the
oscillations. The same is true for spatially propagating waves. Recall from
Section 1.3 that a coordinate transformation to complex amplitudes puts
oscillations into a readily interpretted normal form. Correspondingly Section 6.3 explores how to construct the centre manifold model directly in
complex amplitude variables. The result is a model that is directly in a
readily interpretted normal form.
Besides water, the worlds oceans have two major constituents: heat and
salt. Both these constituents alter the density of the water and so may drive
buoyant flows. In different parts of the oceans water masses have markedly
different combinations of these two characteristics. Such variations may
lead to some surprising dynamics such as that of the salt fountain (briefly
discussed by Turner (1973) [p.252]).
Another interesting effect due to the two constituents is in the melting of
icebergs. It has been proposed to tow icebergs near to the shore of arid
population area such as south Australia. There the iceberg would melt, the
melt water being fresh would float to the top of the sea water and be collected
for drinking or other purposes. However, experiments show that when the
side of an iceberg melts circulation cells of large horizontal extent and fairly
thin in the vertical are formed which carry the fresh water out horizontally
away from the iceberg and then mixed into the sea water (Turner 1973,
Fig 8.10). The presence of two differently diffusing components can have
Tony Roberts, 1 Mar 2008
374
fluid
d
= 0 [1 (T T0 ) + (S S0 )]
z
6
-x
6.1
Warning:
375
of the fluid is
= 0 [1 (T T0 ) + (S S0 )] ,
where T0 , S0 and 0 are convenient reference values. and are the coefficients of expansion, generally positive for temperature and salinity changes
at constant pressure:
1
=
,
S,p
1
=+
.
T,p
The equations we wish to solve are the NavierStokes and continuity equations together with equations describing the transport and diffusion of heat
and salt. From conservation principles (in conjunction with the continuity
equation)1
T
+ q T = 2 T ,
t
S
+ q S = S 2 S ,
t
since the flux of salt and thermal energy is qS S S and qT T
respectively. Strictly speaking we are considering the flow of a compressible fluid because variations in heat and salinity cause density variations.
However, in practise the primary influence of such density variations are in
its buoyancy effects through the gravitational field, the compressibility is
otherwise of no dynamical importance.2 Thus we consider the fluid to be
incompressible except where density fluctuations alter gravitational effects.
This is called the Boussinesq approximation.
1.4 103 cm2 / s for heat, S 1.1 105 cm2 / s for common salt NaCl.
Compare this with 1.1 102 cm2 / s for momentum.
2
D. Gough has justified the Boussinesq approximation rigorously as an anelastic approximation based on the separation of time-scales between compressible sound waves
and the dynamics of convection.
Tony Roberts, 1 Mar 2008
376
1
Pr
q
q
+ q q
t
+ q
t
+ q
t
= 0
= p + (Ra Rs )k + 2 q ,
= w + 2 ,
(6.1)
= w + 2 .
Ra = gTd
is the Rayleigh number measuring the temperature dif4
ference between top and bottom and how much effect it has on the
buoyancy as compared to the dissipative effects of viscosity and diffusion.
Tony Roberts, 1 Mar 2008
377
Rs = gSd
is an equivalent salinity Rayleigh number measuring
4
the salinity difference between top and bottom and its effect on the
buoyancy. Note that for overall convenience it is , the coefficient of
thermal diffusivity, which appears in the denominator and not S , the
coefficient of salt diffusion.
Pr = / is the Prandtl number 3 of the fluid characterising the different rates of diffusion of momentum and of temperature. In water this
is approximately Pr = 10 and we shall use this value later to simplify
the details on the analysis.
= S / is the ratio of the diffusivities of salt and temperature. This
is always less than one, and is generally small. In water 0.01,
salt is much less free to diffuse than heat. We simplify our subsequent
analysis by the approximation that = 0in4 essence the only saline
dynamics is the advection of the background salinity field given above.
The boundary conditions mentioned earlier are also chosen for analytic simplicity. In non-dimensional quantities they are
z
z
w=
=
= = = 0 on z = 0, .
u
v
Use the stream function in two-dimensional flows
For simplicity restrict attention to two-dimensional solutions to the equations (with effectively no salt diffusion, = 0):
x
z
+
= 0
u w
u
u
x
1 u
+u
+w
=
+ 2 u ,
Pr t
x
z
p
1 w
w
w
z
+u
+w
=
+ Ra Rs + 2 w ,
Pr t
x
z
p
3
In air Pr 0.7, whereas in liquid metals and the sun Pr is tiny because electrons and
photons respectively transport heat very quickly.
4
It turns out that the difference between small and = 0 has interesting consequences
later.
378
+u
+w
t
x
z
+u
+w
t
x
z
= w + 2 ,
= w.
Now reduce the set of equations using two tricks which are quite common
in fluid mechanics: we eliminate the continuity equation and eliminate the
pressure by combining the two components of the NavierStokes equation.
In two-dimensional fluid mechanics it is often convenient, as it is here,
to introduce a stream-function (x, z, t) such that
u=
z
,
w=
x
.
(6.2)
2
and hence we satisfy the continuity equation using such a streamfunction. In steady flow , /t = 0, the stream-function has a very
important physical meaning: curves = const are fluid particle paths.5
This is seen from the material derivative
d
dt
+ q
t
= 0+u
+w
x
z
=
+
z x
x z
= 0.
=
Having introduced the streem function, consider that the velocity components, wherever written, are determined from the stream function
according to (6.2). This replaces u and v as independent variables
by , and automatically satisfies the continuity equation.
5
This is also true in some sense for exponential or sinusoidal time dependence.
379
ering x
(w-eqn) z
(u-eqn) to lead to6
1
+ q
= Ra
Rs
+ 2 ,
Pr t
x
x
+ q =
+ 2 ,
(6.3)
t
x
+ q =
,
t
x
where the vorticity is denoted by = 2 . Corresponding boundary
conditions are that
===
2 z
z = 0,
2
on z = 0, .
The first equation here says that horizontal gradients of density, x and x , generate
vorticity = 2 .
Tony Roberts, 1 Mar 2008
380
conditions on the top and bottom boundaries, we may also suppose a sinusoidal dependence in the vertical z.7 Thus seek
= Aet sin kx sin nz ,
= Bet cos kx sin nz ,
(6.4)
= Ce cos kx sin nz ,
for some constants A, B and C. The horizontal wavenumber k may vary over
all real values because we have not imposed any boundary in the horizontal.
The vertical wavenumber n must be integral in order for the top and bottom
boundary conditions to be satisfied. Substituting into the previous three
equations leads to the eigen-problem8
a2
A = Ra kB + Rs kC + a4 A ,
Pr
B = kA a2 B ,
(6.5)
C = kA ,
where a2 = k2 + n2 . Nontrivial solutions for these equations only exist when
the corresponding matrix has a zero determinant. The resultant characteristic equation is cubic in ; upon setting Pr = 10 as for water,
h
i
3 + 11a2 2 + 10 a4 (Ra Rs)k2 /a2 + 10 Rs k2 = 0 .
(6.6)
For a given set of values for Ra and Rs, corresponding to any one physical
situation, we may solve this cubic for all values of k and n to determine the
growth-rate (decay-rate if negative) of the corresponding mode.
However, our interest is in the first transition from the quiescent conduction
state to a state of nontrivial motion. Thus we seek the region of stability in
the Ra Rs-plane, and the curves where this stability is lost to some mode.
Note that the bottom-right quadrant of the Ra Rs-plane, as shown in Figure 6.2, is expected to be stable because both components, salt and heat, are
7
As in TaylorCouette flow, there is no qualitative difference between the mathematically idealised case and the physically realisable case.
8
This is a family of eigen-problems, one for each k and n.
Tony Roberts, 1 Mar 2008
salty cool
fresh warm
381
fresh cool
salty warm
Ra
unstable
unstable
Hopf bifurcation
stable
Rs
unstable
stable
salty warm
fresh cool
fresh warm
salty cool
Figure 6.2: regions of the Ra Rs-plane showing the curves across which the
quiescent solution loses stability. The line Ra = Rs is drawn to separate the
two regions where less dense fluid lies above more dense fluid, and vice-versa.
382
10
a6
Rs + 2 ,
11
k
as shown schematically in Figure 6.2. We concentrate on the oscillatory instability, the so-called Hopf bifurcation, that sets in as Ra increases across such a value for some fixed Rs.
9
If we allow salt to diffuse, 6= 0 then this line moves off Rs = 0 to the left-halve plane.
383
Note that the above line in the Ra Rs-plane is actually a family of lines:
one for each value of the wavenumbers k and n. As Ra increases we are
only interested in the first such line as it indicates the first loss of stability.
Fortunately the family of lines has a very simple form: they all have the same
slope, 10/11, they just have different intercepts. Thus minimising Ra10 for
fixed Rs is the same as minimising the intercept
a6
(k2 + n2 )3
=
.
k2
k2
10
27
Rs + .
11
4
384
As in TaylorCouette flow, as soon as any one mode, here a complex conjugate pair, in a dynamical system becomes unstable, the linearisation becomes inconsistent. The linear picture of the dynamics, here of exponential
growing oscillations, predicts an exponential growth in the unstable modes.
To model the dynamics of the system for these interesting parameter ranges
we must investigate the action of the nonlinearity in the system, and we do
it via centre manifold theory.
6.2
Sorry:
In all the examples of Chapter 3, the dynamics on the centre manifold have
been based on the eigenvalue of zero, and hence is called a slow manifold.
The more general case is where just the real part of the eigenvalue is zero but
the imaginary part is non-zero. This chapter explores this latter case. When
the critical eigenvalues are non-zero, it is more difficult to construct the
centre manifold. However, because an eigenvalue with a non-zero imaginary
part is always associated with oscillations, such cases are important as they
arise in the common transition from steady to oscillatory dynamics. Many
of the same issues also arise in wave propagation.
Example 6.1:
Before moving on to discuss generic issues, let us construct a centre manifold and the evolution thereon for an example.
We twist this special example to analyse as before. Consider the three
variable system
u = wu v ,
v = +u wv ,
= w + u2 + v2 .
w
(6.7)
Very near the equilibrium at the origin, the w variable decays exponentially quickly to zero. Then the u and v variables would oscillate
according to u = v and v = +u . As shown by the trajectories of
the system in Figure 6.3, we derive the more precise statement that
near the origin, the system is exponentially quickly attracted to the
parabolic-like centre manifold
w u2 + v2 2(u2 + v2 )2 ,
Tony Roberts, 18 Sep 2009
(6.8)
385
0.3
0.2
0.1
0.0
0.3
0.3
0.1
0.1
0.1
0.1
u
0.3
0.3
Figure 6.3: Trajectories of the simple oscillatory system (6.7) showing the
quick attraction of the bowl-like centre manifold on which the solutions
undergo long lasting oscillations.
386
v (u2 + v2 )v + u .
(6.9)
The previous example fell to our analysis because its strong angular symmetry transforms it in polar coordinates. Such simplification is not normal.
Generically we must undertake a more complicated analysis as suggested by
the following example.
Example 6.2:
Tony Roberts, 18 Sep 2009
387
0.30
z
0.05
0.20
0.8
0.80
0.4
0.0
y
0.4
x
0.8
0.60
Figure 6.4: Three trajectories of the system (6.10) showing the rapid attraction to a deformed bowl shape centre manifold on which the oscillations
evolve in deformed near-circles.
388
y = x xz ,
z = z + 5x2 .
(6.10)
Trajectories plotted in Figure 6.4 show that the dynamics settle onto
a curved surface. The surface is evidently a deformed bowl. Because
of the lack of angular symmetry, transforming to polar coordinates is
no help here. Instead we analyse directly.
such that x = y ,
y = x xh .
(6.11)
h
h
h
x
= 5x2 + xh
.
x
y
y
(6.12)
(6.13)
(6.14)
389
In the second iteration, substitute the above shape h into the righthand side of (6.12) to seek to solve governing z equation in (6.10) and
rearrange to
h+y
h
h
x
= 5x2 + 2 (6x4 + 16x3 y 12x2 y2 + 8xy3 ) . (6.15)
x
y
142 4
85 x
368 3
85 x y
156 2 2
17 x y
3
448
85 xy
448 4
85 y
(6.16)
The technical difficulty Observe that in constructing the centre manifold approximations (6.13) and (6.16) we had to find all the quadratic terms
simultaneously and all the quartic terms simultaneously. This simultaneous
deduction is not hard in such a simple example, but can be confoundingly
difficult in problems of real complexity and interest.
Having to solve all terms of a given order simultaneously arises from the
coupling in the form of the homological operator h + yhx xhy : see that
the terms yhx couples with terms of one order higher in x and one order lower
in y whereas the term xhy couples terms of one order lower in x and one order
Tony Roberts, 18 Sep 2009
390
Algorithm 6.1 iteratively construct the centre manifold of the system (6.10). The method of undetermined coefficients works well in such
simple problems.
factor alf;
depend x,t; depend y,t;
let { df(x,t)=>y, df(y,t)=>-x-x*z };
z:=0;
let alf^3=>0;
n:=2*deg((1+alf)^9,alf); % truncation controls order in (x,y)
operator c;
cs:=for i:=2:n join for j:=0:i collect c(i-j,j)$
h:=for i:=2:n sum for j:=0:i sum c(i-j,j)*x^(i-j)*y^j$
hom:=h+y*df(h,x)-x*df(h,y)$
repeat begin
write res:=-df(z,t)-z+5*alf*x^2;
eqns:=for i:=0:n join coeff(coeffn(hom-res,x,i),y);
z:=z+sub(solve(eqns,cs),h);
showtime;
end until res=0;
391
higher in y. The homological operator always has this coupling for non-zero
eigenvalues in the centre manifold. In simple examples the coupling is plain
to see; in real problems it is usually hidden; but the coupling is always there
causing significant technical difficulties in constructing a centre manifold
model.
In the next Section 6.3, we explore how using complex amplitude variables in
the centre manifold avoids the difficult coupling in the homological equation.
Consequently, most people use complex amplitude variables for modelling
oscillatory dynamics. However, dynamics may be modelled over a wider
range of parameters using the Cartesian parameters such as the x and y
above. We explore this modelling a little further.
Develop a general approach Most physical problems do not have the
oscillating modes linearly decoupled from the other modes as used in Example 6.2. Thus we develop an approach for the general case when the modes
are implicit in the mathematical description, not explicit. Let us develop
the method intertwined with a specific example. As this second example, we
explore the dynamics of the ordinary differential equations (Wiggins 1990,
p.239, Ex. 2.1j)12
1 1 0
u23
1
0 u + u2 u23 ,
u = 2
(6.17)
1
2 1
0
|
{z
}
|
{z
}
L
f(u,)
where is a control parameter. The parameter is analogous to the Rayleigh
numbers in double diffusive convection. Our first task is then to determine
eigenvalues of the coefficient matrix L of the linear terms.13 It is straightforward to discover that it has eigenvalues i and 1 (when = 0, the
critical value). Thus there exists a centre manifold corresponding to the two
eigenvalues i. The mode with eigenvalue 1 is representative of the many
exponentially decaying modes we find in real applications.
12
13
2
4
and = 1.
Tony Roberts, 18 Sep 2009
392
To unfold the dynamics near the critical value of = 0 , formally adjoin the
differential equation = 0 as discussed earlier in unfolding the pitchfork
bifurcations. This allows us to treat as a small factor in the analysis.
In particular we now group any terms into the nonlinear parts of the
dynamics as already shown in (6.17).
6.2.1
1
0
e1 = 1 , e2 = 1 .
3/2
1/2
A linear approximation to the centre manifold is then
Ec = {xe1 + ye2 | for all x, y} .
Check that the linear subspace Ec is correct by applying L to vectors in Ec
and seeing that the results is still in Ec :
L(xe1 + ye2 ) = = xe2 ye1 .
This linear subspace is then invariant to the linear evolution because if
u = xe1 + ye2 , then
1 + ye
2 = u = Lu = xe2 ye1 .
xe
14
15
Later we find it more convenient to actually use this complex valued description.
Actually we use (w+ + w )/2 and i(w+ w )/2.
393
and y = x .
These two coupled equations describe the dynamics of the classic linear
oscillator with frequency 1. This must be expected because these particular
dynamics corresponds precisley to the two complex conjugate eigenvalues,
= i, of the original system exactly at the critical conditions.
Now we need to seek the nonlinear centre manifold and the long-term evolution thereon. To parameterise the centre manifold Mc we clearly need to
use s = (x, y). But what do x and y mean precisely? One possibility is
to choose, as we do here, that x = u1 /2 and y = (u1 + u2 )/2. This agrees
with what we have used for the linear dynamics. We now insist that this
definition of x and y also hold for the nonlinear analysis. Then the nonlinear
shape of Mc will follow from the third component of u:
3
1
u3 = x + y + h(s, ) ,
2
2
(6.18)
6.2.2
The next task is to find the nonlinear corrections to the shape of the centre
manifold and the evolution thereon. This will then form a more useful and
accurate model of the long-term dynamics.
Tony Roberts, 18 Sep 2009
394
Now substitute the assumption (6.18) and (6.19) into the original governing
equations (6.17) to give
0
0
0 2
2 0
2 2 + 0
0 (Gs + g) = 2 2 s +
h
h
3 1
1 3
x
y
0
1
0
2
+ 0 + 1 (3x + y + h) + 1 (x + y) .
(6.20)
h
0
0
When we solve this, albeit just to some order of accuracy, then centre
manifold theory assures us that the long-term evolution will be described
by (6.19).
First, rearrange (6.20) into a form suitable for iteration. Do this by putting
all the dominant terms on the left-hand side:
0
0
0
1
0
1
= 0
0
0 g +
1 g +
h
h
h
h
h y x + x y
3/2 1/2
x
y
1
0
+ 1 (3x + y + h)2 + 1 (x + y) .
(6.21)
0
0
You can recognise which terms in g and h to put on the left-hand side as all
those terms which involve g and h and which are lowest order in s and .
h
This includes anything linear in g and h and anything of the form si s
j
because the derivative lowers the order by one but the multiplication raises
it again. However, other terms in g and h are put on the right-hand side
because they are either quadratic in h or multiplied by an si and so are of
order higher than the lowest.
Second, start the iteration by putting h = h(0) = 0 and g = g(0) = 0 in the
right-hand side of (6.21) to give:
0
1
0
1
0
1 g +
h
h
h y x + x y
3/2 1/2
Tony Roberts, 18 Sep 2009
395
1
0
= 1 (3x + y)2 + 1 (x + y) .
0
0
The first two components of this equation dictate that
(1)
1
=
+ y)2
.
(y x)
4 (3x
h
h
27
18
3
+x
= x y + x2 xy + y2 .
x
y
2
2
8
8
8
In general, such equations as these for the shape of a centre manifold are
solved in the space of multinomials in s. This is generic P
because we seek
a function h(s) whose shape is a multinomial in s, here m,n hmn xm yn ,
and so it is the coefficients of this multinomial, hmn , that must be found.
It is only in simple circumstances, such as that of the direct instability of
a pitchfork bifurcation, that we can overlook this feature of the solution
process; here we must be aware of it. Because the right-hand side is a
quadratic in x and y we therefore try a quadratic form for h, namely17
h = Ax + By + Cx2 + Dxy + Ey2 .
The equation then becomes
(A + B)x + (B A)y + (C + D)x2 + (2C + D 2E)xy + (D + E)y2
27
18
3
=
x y + x2 xy + y2 .
2
2
8
8
8
Equating coefficients of each term in the linearly independent xm yn we
16
The homological equation is hyperbolic with characteristic curves being trajectories
of the basic oscillation. This matches with the property that a centre manifold is any
union of suitable trajectories.
17
The functions x, y, x2 , xy and y2 are linearly independent multinomials that we use
as a basis.
396
1
1
0
0
equations
1 0
0
1 0
0
0 1
1
0 2 1
0 0 1
0
A
/2
0
B /2
0
C = 27/8
2 D
18/8
1
E
3/8
21
3
9
x + x2 + xy + y2 .
2
8
4
8
(6.22)
The homological equation is not too complicated. For a pitchfork bifurcation it is diagonal. For a simple Hopf bifurcation it is always effectively
tridiagonal. It is only for higher-order bifurcations that it need become
more complicated.
To conclude this subsection, observe that a first approximation to the nonlinear shape of the centre manifold is
x
.
x + y
u=
3
1
21 2
3
9 2
2 x + 2 y + 2 x + 8 x + 4 xy + 8 y
On this manifold the evolution is approximately described by18
x = y + 14 (3x + y)2 ,
y = x + (y x) .
(6.23)
However, as is generally the case for Hopf bifurcations, this quadratic approximation is almost always unreliable. For positive the origin of the
above model is linearly unstable. But the quadratic nonlinear terms may
or may not stabilise the dynamics. Here the dynamics are stabilised, but
it is just as easy for all model solutions eventually evolve to infinity. The
quadratic model is deficient. I emphasise that this is generally true for
Hopf bifurcations. To obtain a reliable model, you must perform one more
iteration to obtain the correct cubic nonlinear terms in the model.
18
6.2.3
397
Further iteration
(s) + g 0 (s) .
such that s = g
u
+ g0 +
+ g 0 = Lu
+ Lu 0 + f u
+ u0 .
g
g
s
s
and u
u 0
u
+ Lu
+ f(u
) .
Gs + Eg 0 = g
s
s
(6.24)
398
In short, the left-hand side is precisely the same operator as appeared before,
in equation (6.21), and the right-hand side is just the governing differential
equation.
In reduce we perform the iteration as follows.19
COMMENT Use iteration to form the centre manifold model of the eleme
Hopf bifurcation problem. The model is u=u(x,y) such that d(x,y)/dt
Tony Roberts, April 1999;
% formating for printed output
on div; off allfac; on revpri;
% useful matrices
ll:=mat((-1,-1,0),(2,1,0),(1,2,-1));
zt:=mat((1,0,0),(1,1,0));
e1:=mat((1),(0),(0));
e2:=mat((0),(1),(0));
% form of the centre manifold
u:=mat((x),(-x+y),(-3*x/2+y/2+h));
% linear approximation
depend x,t; let df(x,t) => g(1,1);
depend y,t; let df(y,t) => g(2,1);
h:=0;
g:=mat((-y),(x));
399
, eps^2=>0 };
% look for corrections hd as a multinomial in x and y
operator c;
hd:=for m:=0:o sum for n:=0:o-m sum c(m,n)*x^m*y^n$
clist:=for m:=0:o join for n:=0:o-m collect c(m,n)$
% iteration
repeat begin
dudt:=mat((df(u(1,1),t)),(df(u(2,1),t)),(df(u(3,1),t)));
write res:= -dudt +ll*u +(e1-e2)*u(3,1)^2 +eps*e2*u(2,1) ;
% solve first two components for evolution g
gd:=zt*res;
g:=g+gd;
% form and solve homological equation
heqn:= hd-y*df(hd,x)+x*df(hd,y) -(res(3,1) -(-3*gd(1,1)+gd(2,1))/
elist:=for m:=0:o join for n:=0:o-m collect coeffn(coeffn(heqn,x,
h:=h+sub(solve(elist,clist),hd);
showtime;
end until res=mat((0),(0),(0));
end;
The lines of code perform the following functions.
`?? These lines tell reduce to discard all terms higher than cubic in x and y,
and to only retain linear variations in the bifurcation parameter .
`?? This first tells reduce that x and y depend upon t and their time
derivatives are components of g. Then the linear solution is prescribed.
`?? These lines set up some quantities for solving the homological equation
in the space of multi-nomials in x and y up to the specified order.
`?? Performs the iteration.
Construct the current approximate to u.
Tony Roberts, 18 Sep 2009
400
The output of this program shows the centre manifold has nonlinear shape
21
3
9
3
311x3 + 319x2 y + 217xy2 + 177y3 +O 2 +s4 .
h = x+ x2 + xy+ y2
2
8
2
8
80
The corresponding model on this centre manifold for the evolution is20
x = y + 94 x2 32 xy + 14 y2
18 63x3 3x2 y + 21xy2 9y3 + O 2 + s4 ,
y = x + (y x) + O 2 + s4 .
(6.25)
401
manifold model is stable. Hence, the relevance theorem applies strictly, and
all trajectories close enough to the centre manifold approach a solution of
the centre manifold model.
6.2.4
Networks of near identical oscillators model a wide variety of pattern forming systems, such as neural networks, fluid convection, interacting lasers
and coupled biochemical systems. These networks exhibit rich collective
behaviour, including synchrony, travelling waves, spatiotemporal chaos and
incoherence (Abrams & Strogatz 2006, Moon et al. 2005). In particular,
much interest continues in the synchronisation of nonlinear oscillators (Sun
et al. 2009, e.g.).
Example 6.3: Coupled Van der Pol Oscillators
This example is a
simple case: the system is just a pair of oscillators with one forcing
the other, but not vice versa.
A lone van der Pol oscillator has governing equation
x 1 = y1 ,
y 1 = x1 + (1 x21 )y1 ,
(6.26)
(6.27)
where characterises the natural difference between the two oscillators, and the constant 2 parametrises the strength of the influence of
the first oscillator upon the second. In this simple system introduced
by Sun et al. (2009), the second oscillator is not coupled back to the
first.
Tony Roberts, 18 Sep 2009
402
The synchronised manifold emerges This centre subspace of linear synchronisation is linearly attractive like exp(t). Consequently,
in the full nonlinear dynamics we expect that nonlinear synchronised
oscillations emerge from among transients with similar such decay.
6.3
Sorry:
Modulation of oscillations
this section is incomplete as it is slowly written.
However, the low dimensional model (6.25) is in an important sense a failure. True it is low dimensional and of guaranteed accuracy and relevance.
However:
Tony Roberts, 24 Apr 2009
403
Algorithm 6.2 Constructs centre manifold of two coupled van der Pol
oscillators as a simple example of a synchronous manifold. Here for small
and .
on div; off allfac; on revpri;
factor z,eps,mu;
% Use z to control truncation in order (mu,eps).
let z^2=>0;
% Form homological equation to high enough order.
n:=3*deg((1+z)^9,z);
operator c;
cs:=for each k in {x,y} join
for i:=1:n join for j:=0:i collect c(i-j,j,k)$
hx:=for i:=1:n sum for j:=0:i sum c(i-j,j,x)*x1^(i-j)*y1^j$
hy:=for i:=1:n sum for j:=0:i sum c(i-j,j,y)*x1^(i-j)*y1^j$
homx:=-hy +2*hx+y1*df(hx,x1)-x1*df(hx,y1)$
homy:=+hx
+y1*df(hy,x1)-x1*df(hy,y1)$
% dependencies and initial centre subspace
depend x1,t; depend y1,t;
let { df(x1,t)=> y1 , df(y1,t)=> -x1+z*mu*(1-x1^2)*y1 };
x2:=x1; y2:=y1;
% iterate to centre manifold
repeat begin
write resx:=-df(x2,t)+y2+2*(x1-x2);
write resy:=-df(y2,t)-x2+z*(mu+eps)*(1-x2^2)*y2;
eqns:=for i:=0:n join append(
coeff(coeffn(homx-resx,x1,i),y1)
,coeff(coeffn(homy-resy,x1,i),y1));
soln:=solve(eqns,cs);
x2:=x2+sub(soln,hx);
y2:=y2+sub(soln,hy);
showtime;
end until {resx,resy}={0,0};
end;
Tony Roberts, 24 Apr 2009
404
Here we show how to derive a model for the modulation of the fast oscillations. The result remedies both the above problems. Further, the process
simplifies the homological equation and its solution process! Start by revisiting Example 6.2.
Example 6.4:
Deduce the centre manifold dynamics of the system (6.10),
but now express the centre manifold model in terms of complex amplitudes of the oscillations on the centre manifold. (Exercise 6.5 correspondingly asks you to find the centre manifold in terms of amplitudeangle variables.)
z 0,
(6.28)
??
Algorithm 6.3 finds
eit + 18 (1 2i)a3 ei3t + (1 + 2i)
x = aeit + a
a3 ei3t + O 2 ,
2
11
it
y = iaeit i
aeit + (1 + 11
a2 eit
2 i)a ae + (1 2 i)a
+ 18 (2 + i)a3 ei3t + (2 i)
a3 ei3t + O 2 ,
z = 10a
a + (1 2i)a2 ei2t + (1 + 2i)
a2 ei2t + O 2 ,
Tony Roberts, 24 Apr 2009
405
a = (1 + 11
2 i)a a + O .
(6.29)
6.3.1
To motivate the subsequent analysis, observe that the basic critical oscillation is sinusoidal. Thus to a leading approximation the solution21
x(t) r(t) cos (t) ,
where r(t) is the slowly varying amplitude of the oscillations, and the frequency = d/dt 1 . Writing (t) = t + (t) we note that d/dt
must be small. Now we can describe the evolution on the centre manifold in
terms of the evolution of r and . Because both dr/dt and d/dt are small,
the homological equation simplifiesit was previously complicated by the
presence of the terms yh/x + xh/y which came from the fact that
dx/dt and dy/dt are not small, being approximately y and x respectively.
In most cases it is more convenient, and very traditional, to model the
evolution in terms of the complex amplitude
1
a(t) = r(t)ei(t) .
2
For example,
x(t) r
ei + ei
2
21
See numerical simulations via osceg2.m for the critical = 0 for a slow algebraic
decay of the fast oscillations.
Tony Roberts, 24 Apr 2009
406
Algorithm 6.3 iteratively construct the centre manifold of the system (6.10) using complex amplitudes. Obtain code fragments from the
normal form Algorithm 1.15.
factor alf;
operator cis;
let { df(cis(~u),t) => i*df(u,t)*cis(u)
, cis(~u)*cis(~v) => cis(u+v)
, cis(~u)^~p => cis(p*u)
, cis(0)=>1 };
operator linv; linear linv;
depend a,t; depend b,t;
let { df(a,t)=>ga, df(b,t)=>gb };
ga:=gb:=0;
x:=a*cis(t)+b*cis(-t);
y:=df(x,t);
z:=0;
let alf^3=>0;
repeat begin
write resz:=-df(z,t)-z+5*alf*x^2;
z:=z+(linv(resz,cis) where { linv(1,cis)=>1,
linv(cis(~n*t),cis)=>cis(n*t)/(i*n+1) });
write res:=df(y,t)+x+x*z;
ga:=ga+i/2*(ca:=coeffn(res,cis(+t),1));
gb:=gb-i/2*(cb:=coeffn(res,cis(-t),1));
x:=x+(linv(res-ca*cis(t)-cb*cis(-t),cis)
where { linv(1,cis)=>-1,
linv(cis(~n*t),cis)=>cis(n*t)/(n^2-1) });
y:=df(x,t);
showtime;
end until resz=0 and res=0;
407
1 it+i
r e
+ eiti
2
1 i it 1 i it
=
re e + re e
2
2
it
it
e ,
= ae + a
where the overbar denotes complex conjugation.22 Observe that |a| = r/2
just measures the amplitude of the oscillations, whereas arg a = neatly
encapsulates any phase shift or slip of the oscillations. For example, oscillations with a small frequency shift from 1 to 1 + 0 are represented by a
0
complex amplitude a = ei t . In general, the complex amplitude a(t) is
itself slowly-varying in time because both r and are.
For dynamical systems in general,
u = Lu + f(, u) ,
we just substitute the ansatz
u = v(, a, t) ,
for some parameters measuring departure from the critical Hopf bifurcation. The new feature here is the presence of the explicit t dependence
is the shape of the centre manifold, v(, a, t). It is there to resolve the
basic oscillations whose amplitude and frequency shifts we wish to describe.
However, it introduces the extra complication that23
du
v v
=
+
g,
dt
t a
where the partial derivative /t is done keeping a constant, whereas /a is
done keeping t constant. This meaning of the partial derivatives is an important feature in the ensuing analysis. Thus we solve equations of the
form
v v
+
g = Lv + f(, v) ,
(6.30)
t a
.
for the centre manifold shape v(, a) and the slow evolution thereon a = g
22
Similar tricks to this allow one to analyse the evolution of the spatial modulation of
periodic patterns in space as seen in convection and Taylor vortices.
23
This is akin to the material derivative in fluid dynamics.
Tony Roberts, 24 Apr 2009
408
6.3.2
Linear approximation
1
1
eit
v 1 i aeit + 1 + i a
(6.31)
3i
2
3+i
2
6.3.3
Iterative refinement
Now we improve the above linear approximation: first by hand, then via
computer algebra.
Given any approximation to the centre manifold and the evolution thereon:
(a, t) such that a g
(a) . Seek a better approximation
uu
(a, t) + u 0 (a, t) ,
u=u
Tony Roberts, 24 Apr 2009
(a) + g 0 (a) .
such that a = g
409
u 0
u
u
g 0 eit =
+ f(u
) . (6.32)
+ Lu
Lu 0 +
+ wg 0 eit + w
g
t
t
a
Again the right-hand side is just the residual of the dynamical system, u +
Lu + f(u) , evaluated for the current approximation. In contrast to the
0
previous approach, the left-hand side no longer has the terms us Gs which
complicates the solution of the homological equation; consequently here the
solution process is simpler. However, an increase in complication occurs due
to the presence of the u 0 /t term.
First iteration & principles
We now explore the algebraic techniques needed to perform the iterative
improvement of the centre manifold model.
First, substitute the linear approximation (6.31) into the iterative equation (6.32) to
u 0
g 0 eit
Lu 0 +
+ wg 0 eit + w
t
1
4 3i 2 i2t
4 + 3i 2 i2t
e
1
a e + 5a
a+
a
=
2
2
0
0 h
i
+ 1 (1 i)aeit + (1 + i)
aeit .
(6.33)
0
The right-hand side is of the form of a sum of terms each involving a complex
exponential eint for n = 2, . . . , 2 . For each such component, say reint ,
pose that they give rise to a component in u 0 of the form veint , then we
obtain a linear equation for the vector coefficient v of the form
[L + inI] v = r .
Tony Roberts, 24 Apr 2009
410
1 + n2
1 + in
0
1
2(1 + in)
(n i)2
0 r.
v=
(1 + in)(n2 1)
2
(3 + in) (1 + 2in) n 1
(1 43 i)a2 ei2t
( 5 + 5 i)a2 ei2t + 5a
a + c.c. .
3
6
1
2
i2t
a
( 3 + i)a e + 10a
However, the n = 1 components are not so straightforward. This is seen
in the matrix inverse above where a division by n2 1 occurs; for n = 1
this is a division by 0. As is always the case for the construction of a centre
manifold, the operator on the left-hand side, here L + /t , is singular. It
has to be because that is the only way a non-trivial set of eigenvectors can
span a centre eigenspace. Here the centre eigenspace involves oscillations
in eit and hence we have difficulty solving for components with n = 1 .
As is always the case,24 we have to choose terms in the evolution, g 0 , so that
the right-hand side is put in the range of the operator L + /t .
We need to define an appropriate inner product such as
Z
1
T dt ,
h, i =
2
which is the average over one period of the normal vector inner product.
Upon finding the adjoint (left-) eigenvector z corresponding to the critical
oscillation, tz = LT z , computing hz, equationi will give an equation for g 0
as follows:
u 0
g 0 eit i = hz, rhsi
i + hz, wg 0 eit + w
t
z 0
it i
hLT z
, u i + hz, weit ig 0 + hz, we
g 0 = hz, rhsi .
t
hz, Lu 0 +
24
411
z = 2i eit ,
0
. Any linear combination of these would suffice,
and its complex conjugate z
it i = 0 and
but these have the advantage that hz, weit i = 1 and hz, we
25
similarly for the complex conjugate. Thus
g 0 = hz, rhsi .
Apply this solvability condition to (6.33) to find
g0 =
1i
a ,
2
The complex conjugate and average in h, i mean that z just extracts the component
in eit in the rhs.
Tony Roberts, 24 Apr 2009
412
Having chosen g 0 , the right-hand side is now in the range of the singular
linear operator L + /t . However, the right-hand side will still contain
components in eit , it is just that a u 0 can be found to suit. However, there
will be two degrees of freedom the solution, a freedom corresponding to an
arbitrary linear combination of the critical eigenvectors. We must pose an
extra condition upon u 0 in order to make the solution unique. For example,
seek solutions (with u10 = 0 in these components) containing the form
0
0
it
u = ve
= c2 eit .
c3
In essence this assumption26 defines the oscillation amplitude a to be the
is the comcomponent of eit in u1 (t) of the original variables. Similarly, a
ponent of eit in u1 (t). Seeking solutions in this form, and ignoring the first
component of the equation which is allowed because we know the right-hand
side is in the range, we find that components in the right-hand side of the
form reit give
0
0
0
1
0 r.
v = 0 1i
1
0 1 1i
Here the components in eit give rise to components in u 0 of
0
0
1+i aeit + 1i
aeit .
2
2
1+i
4
1i
4
Gathering all components together gives that the first nonlinear approximation to the shape of the centre manifold and the rapid oscillations thereon
is approximately described by
1
(1 34 i)
0
aeit + ( 5 + 5 i) a2 ei2t + 5a
a + c.c. .
u 1 i + 1+i
2
3
6
3i
1+i
1
10a
a
( 3 + i)
2 + 4
26
Must always define the amplitudes, whether implicit (as before) or explicitly.
413
414
r + ir =
r 15 + i r3 .
2
3
The real part of this equation gives
1
1
r = r 15r3 = r 30r2 ,
2
2
and consequently if < 0 , r evolves to zero which represents
the fixed
p
point at the origin; or if > 0 , r evolves to r = /30 which gives
the amplitude of the stable limit limit cycle. Observe that, similar to
the pitchfork bifurcation, the amplitude of this nonlinear state grows
2
3
At the origin r = 0 , this frequency shift is /2 as observed before in
the linear analysis. Whereas on the limit cycle, r = r ,
= 20 r2 = 13 .
2
3
18
This frequency shift is different to that for the origin because the
nonlinear limit cycle is distinctly different.
6.3.4
415
After studying these two sorts of bifurcations, the pitchfork and the Hopf,
we now have the basic analytical and computer algebra tools to rationally
analyse almost any dynamical system to extract a finite dimensional model.
This is because nomatter how many modes make up the centre manifold,
they all have eigenvalues which are either zero or pair up with a complex
conjugate. These are the two cases we have examined. The general analysis
and construction is then just as we have investigated, there are no new
features, just greater complexity of detail.
However, the case of infinite dimensional centre manifold models is rather
more interesting and is dealt with in the next chapter.
Exercise 6.5:
Deduce the centre manifold dynamics of the system (6.10),
but now express the centre manifold model in terms of the real ampliTony Roberts, 24 Apr 2009
416
+ 2 sin 3) + O 2 , y = r sin +
1
3
sin + 6 cos 3 44 sin 3) + O 2 , z = 12 r2 (5 + cos 2 +
32 r (8 cos 3
2
2 sin 2) + O 2 such that r = 14 r3 73
2 r5 + O 3 and = 1 + 11
80
8 r
2199 2 4
3
.
1280 r + O
Answer: x = r cos +
Consider
x = x 2y 2z + xy ,
y = x 3y 2z ,
z = x + 3y + 2z .
Explore and report all about the long-term dynamics of solutions to
this dynamical system. Use numerical solutions, centre manifold models, and so on.
6.4
Sorry:
417
Our task now is to return to double diffusive convection and model the fluid
flow that occurs when the static state loses stability via a Hopf bifurcation.
Do the oscillations settle down to a stable limit cycle? or do the nonlinear
interactions enhance the linear instability? Are there any special features of
the nonlinear dynamics?
6.4.1
Critical dynamics
Recall that in 6.1 we showed that the static state loses stability as the
Rayleigh number, a measure of the temperature difference between the
plates, increases past the critical value
Rac =
10
274
Rs +
,
11
4
iA =
20
10
274
Rs +
11
4
2
3
iB = kA
B,
2
iC = kA ,
kB + Rs kC +
94
A,
4
418
(6.35)
where a(t) is the complex amplitude of the oscillations. Our aim is to derive
an equation governing its variation, which in turn describes the modulation
and growth or decay of the oscillations.
6.4.2
+
= Ra
Rs
+ 4 .
Pr
t
z
x
x
x
Introducing the vorticity
= 2 =
w u
,
x
z
=
+ q = Ra
Rs
+ 2 .
Pr dt
Pr t
x
x
Tony Roberts, 1 Mar 2008
(6.36)
(6.37)
419
Now the vorticity is the amount of rotation in the fluid flow at each point.
This last equation just says that such rotation:
is carried by the fluid, d/dt;
+ q = w + 2 ,
t
+ q = w .
t
6.4.3
(6.38)
(6.39)
Nonlinear analysis
x, z, t) ,
= (a,
x, z, t) , = (a,
(a, x, z, t) , s.t. a = g
(a) ,
=
Tony Roberts, 1 Mar 2008
420
x, z, t) + 0 (a, x, z, t) ,
= (a,
x, z, t) + 0 (a, x, z, t) , = (a,
(a, x, z, t) + 0 (a, x, z, t) , s.t. a = g
(a) + g 0 (a) .
=
Then, in the same manner as in the previous section, equations are derived
for the corrections 0 , 0 , 0 and g 0 :
4 0 Rac
0
0
1 2 0
1 2 a 0
,
+ Rs
+
g =
x
x
Pr
t
Pr
c
0
a 0
0
(6.40)
2 0 +
+
g =
,
x
t
c
a 0
0 0
,
+
+
g =
x
t
c
and ,
are the residuals of the vorticity,
where the right-hand sides, ,
heat and salinity equations respectively.
The following reduce code to find such corrections is rather long. However,
much of it involves getting reduce to compute various inverse operators
rather than calculate them ourselveswith more effort I could have eliminated many steps at the expense of introducing my algebraic errors. After
the listing, I describe the various steps.
% Use iteration to form the centre manifold model of the
% double diffusive Hopf bifurcation problem. Here the
% description is in terms of the complex amplitude to describe
% the modulation of the oscillations.
% a is the complex amplitude and b is its complex conjugate.
% (c) Tony Roberts, October 1995
o:=3;
% only retain terms up to order o in a and b
let { a^~m => 0 when m>o
, b^~n => 0 when n>o
, a^~m*b^~n => 0 when m+n>o
, a*b^~n => 0 when 1+n>o
Tony Roberts, 1 Mar 2008
421
422
423
)$
ilop:=1/lop$
% Third, extract columns of inverse for later use
psiop:= ee(1)*ilop(1,1)*cist(l)*sinp(m,kx)*sinp(n,z)
+ee(2)*ilop(2,1)*cist(l)*cosp(m,kx)*sinp(n,z)
+ee(3)*ilop(3,1)*cist(l)*cosp(m,kx)*sinp(n,z)$
theop:= ee(1)*ilop(1,2)*cist(l)*sinp(m,kx)*sinp(n,z)
+ee(2)*ilop(2,2)*cist(l)*cosp(m,kx)*sinp(n,z)
+ee(3)*ilop(3,2)*cist(l)*cosp(m,kx)*sinp(n,z)$
sigop:= ee(1)*ilop(1,3)*cist(l)*sinp(m,kx)*sinp(n,z)
+ee(2)*ilop(2,3)*cist(l)*cosp(m,kx)*sinp(n,z)
+ee(3)*ilop(3,3)*cist(l)*cosp(m,kx)*sinp(n,z)$
% Fourth, deal with the special case of l=\pm 1, m=n=1
loq:=sub({l=1,m=1,n=1},lop)$
loq(1,2):=df(vortc ,a)/cist(1)/sinp(1,kx)/sinp(1,z)$
loq(2,2):=df(thetac,a)/cist(1)/cosp(1,kx)/sinp(1,z)$
loq(3,2):=df(sigmac,a)/cist(1)/cosp(1,kx)/sinp(1,z)$
iloq:=sub(om=l*om,1/loq)$
psioq:= ee(1)*iloq(1,1)*cist(l)*sinp(1,kx)*sinp(1,z)
+ee(3)*iloq(3,1)*cist(l)*cosp(1,kx)*sinp(1,z)$
theoq:= ee(1)*iloq(1,2)*cist(l)*sinp(1,kx)*sinp(1,z)
+ee(3)*iloq(3,2)*cist(l)*cosp(1,kx)*sinp(1,z)$
sigoq:= ee(1)*iloq(1,3)*cist(l)*sinp(1,kx)*sinp(1,z)
+ee(3)*iloq(3,3)*cist(l)*cosp(1,kx)*sinp(1,z)$
zpsi:=4*iloq(2,1)*sinp(1,kx)*sinp(1,z)*cist(-l)$
zthe:=4*iloq(2,2)*cosp(1,kx)*sinp(1,z)*cist(-l)$
zsig:=4*iloq(2,3)*cosp(1,kx)*sinp(1,z)*cist(-l)$
% Lastly, define inverse terms from each of the component eqns
operator psiinv; linear psiinv;
operator theinv; linear theinv;
operator siginv; linear siginv;
depend x,xzt; depend z,xzt; depend t,xzt;
let { psiinv(cist(~al)*sinp(~am,kx)*sinp(~an,z),xzt)
=> sub({m=am,n=an},sub(l=al,psiop)) when abs(al*am*an) neq 1
, theinv(cist(~al)*cosp(~am,kx)*sinp(~an,z),xzt)
Tony Roberts, 1 Mar 2008
424
%
% initial approximation
% uncomment the following line for numerical solution
%on rounded;k:=1/sqrt(2);rs:=1;om:=sqrt(10*rs/33);precision 6;
on complex; on list; factor cist,cosp,sinp;
ra:=rac+eps;
psi:=psic;
theta:=thetac;
sigma:=sigmac;
ga:=0;
gb:=0;
showtime;
%
% iteration
for iter:=1:o do begin
u:=-df(psi,z)$ w:=df(psi,x)$ vort:=df(psi,x,x)+df(psi,z,z)$
salteq:=-df(sigma,t)-df(sigma,a)*ga-df(sigma,b)*gb
-u*df(sigma,x)-w*df(sigma,z)+w;
heateq:=-df(theta,t)-df(theta,a)*ga-df(theta,b)*gb
-u*df(theta,x)-w*df(theta,z)+w+df(theta,x,x)
+df(theta,z,z);
vorteq:=(-df(vort,t)-df(vort,a)*ga-df(vort,b)*gb
-u*df(vort,x)-w*df(vort,z))/pr +ra*df(theta,x)
-rs*df(sigma,x)+df(vort,x,x)+df(vort,z,z);
% solve for evolution ga and gb
Tony Roberts, 1 Mar 2008
425
gad:=mean(sub(l=+1,zpsi*vorteq+zthe*heateq+zsig*salteq));
gbd:=mean(sub(l=-1,zpsi*vorteq+zthe*heateq+zsig*salteq));
write ga:=ga+gad;
write gb:=gb+gbd;
% update and solve homological equation
ud:= psiinv(vorteq-sub({a=gad,b=gbd},vortc ),xzt)
+theinv(heateq-sub({a=gad,b=gbd},thetac),xzt)
+siginv((salteq-sub({a=gad,b=gbd},sigmac)
where cist(0)*cosp(0,kx)=>0) ,xzt);
write psi:=psi+ud _1;
write theta:=theta+ud _2;
write sigma:=sigma+ud _3;
showtime;
end;
%
% check equations
u:=-df(psi,z)$ w:=df(psi,x)$ vort:=df(psi,x,x)+df(psi,z,z)$
salteq:=-df(sigma,t)-df(sigma,a)*ga-df(sigma,b)*gb
-u*df(sigma,x)-w*df(sigma,z)+w;
heateq:=-df(theta,t)-df(theta,a)*ga-df(theta,b)*gb
-u*df(theta,x)-w*df(theta,z)+w+df(theta,x,x)
+df(theta,z,z);
vorteq:=(-df(vort,t)-df(vort,a)*ga-df(vort,b)*gb
-u*df(vort,x)-w*df(vort,z))/pr +ra*df(theta,x)
-rs*df(sigma,x)+df(vort,x,x)+df(vort,z,z);
;end;
`738 As in the toy problem of the previous section these: set up the terms
to be dropped in the expressions for the model; some basis vectors for
the three components of the flow field, , and ; and the operator
cist to describe the oscillations.
`3950 However, in this problem we also have spatial structure.28 Because
of the nature of the governing equations and the boundary conditions
28
In essence, cosp and sinp act as basis vectors for the spatial structure in the fields ,
and .
Tony Roberts, 1 Mar 2008
426
`5264 The critical parameters and corresponding linear oscillations, equation (6.35), are encoded in these lines. Note that kx is used to denote kx in the argument of trigonometric functions.
`66128 This big chunk derives the inverse operators needed to later find
the corrections to the evolution and the centre manifold.
1. The solution fields are composed of sums of terms of the form
eilt cos mkx sin nz (except for for which the x dependence
is sin). Hence, we first find the effect of the linear operator on
fields of this form.
2. Then extract the coefficients, form into a matrix and invert. The
ith row of the matrix ilop describes the influence that residuals in the three equations have upon the ith field (, and
respectively).
3. Conversely,29 the jth column of ilop describes the influence that
the residual of the jth equation has upon the three solution fields.
Thus, third, extract these columns for later use and group together with the basis vectors ee.
4. However, as is always the case, the linear operator on the lefthand side of (6.40) is singular. This singularity affects the computation of the inverse only if l = 1 and n = m = 1 corresponding
to the critical mode. For these values of l, m and n we have to
choose g 0 so that the right-hand side is in the range of the operator. This is done by finding g 0 through a solvability condition.
29
427
!
ilt
clmn e
l,m,n
clmn psiinv eilt sin mkx sin nz .
l,m,n
The operators thus work on each term in turn, producing a result depending upon the particular value of l, m and n for each
particular term. The variable xzt is used to tell reduce to keep
any factor depending upon x, z and t within the argument of the
operator.
One curiosity is that we have to substitute for the time harmonic,
l, before substituting for the spatial wavenumbers m and n. the
reason is that if this is not done a division by 0 occurs! But how
could this be since we have properly dealt with the singularity
associated with the critical mode? It actually signifies that there
Tony Roberts, 1 Mar 2008
428
`129139 These just assign the critical linear solution as a first approximation to the centre manifold model.
`141166 The iteration has the same structure as before.
The residuals of each of the governing differential equations are
computed. These should be directly readable.
The solvability condition is applied to determine corrections to
the evolution in the model.
The new terms arising from the solvability are incorporated into
the residuals, then the inverse operators established earlier are
used to compute corrections to the centre manifold description.
Tony Roberts, 1 Mar 2008
429
6.4.4
The model
36302 i220
a ,
40 Rs +359374
(6.41)
p
where = 10 Rs /33 and = Ra Rac . This describes the linear dynamics of the pure conduction state:
2
430
0t
leads to
r = 0.008135 .
431
432
the fluid equations (6.376.39) and discover that it is indeed a critical mode
with eigenvalue = 0. We could avoid including the bulk flow mode in the
analysis because there the problem was 0/0 and we could arrange to evaluate
the numerator before the denominatorphysically there is no interaction
between the bulk flow and the convection rolls. Here this is not possible
because the numerator is definitely non-zero, here 36.56 for Rs = 1there is
interaction. Thus to do the centre manifold analysis properly we should have
included this mode in the description, say proportional to an amplitude c.
That is we should seek
= (a,
c, x, z, t) ,
c, x, z, t) ,
= (a,
(a, c, x, z, t) ,
=
(a, c) ,
s.t. a = g
c = h(a,
c) ,
and modify the reduce program accordingly. This will fix the problem.
Or will it? Why is = sin 2z a critical mode? Physically it corresponds
a change in the vertical structure of the mean salinity field. Horizontal
variations in salinity cause horizontal density gradients that drive flow, but
vertical variations apparently have no dynamic effect. The physical reason is that we set , the relative diffusivity of salt, to zero because it was
very small, in water 0.01. However, this removes diffusion of salt and
so nonlinear vertical structure of the salt field is consequently long-lived.
It is not just the sin 2z mode which is neutrally stable, all the modes
(, , ) = (0, 0, sin nz) are also critical modes! Really we should include
an infinite number of amplitudes cn for the centre manifold and include
P
n=1 cn sin nz into the linear description of the salinity field. Fortunately,
most of these modes do not interact with low-order approximations to the
flow dynamics on the centre manifold and so may be omitted from a model
of the oscillatory flow.
Another possible way around the problem is to re-introduce salt diffusion
by setting to be non-zero. Then when these vertical structure components
appear in the residual, they induce corresponding finite components into the
Tony Roberts, 1 Mar 2008
433
salt field. For example, here would have an additional term proportional
to a
a sin 2z. This would interact with the fluid flow to induce corrections
; there could not be a
to the modulation equation proportional to a2 a
correction proportional to a
a due to symmetry considerations. Thus taking
into account the finite diffusivity would not alter our earlier conclusions on
the existence of limit cycles.
However, the coefficient of the additional terms are proportional to 1/
and since is genuinely small in many applications we can expect that
the additional terms to be quite large. In the evolution of the complex
terms would become significant quickly due to
amplitude a, the order a2 a
their large coefficient. Physically, these effects are likely to be destabilising
because the advection of the background salinity field is likely to reduce the
stabilising salinity gradient in the bulk of the fluid domain, and hence assist
the development of more vigorous flow.
Either of the above approaches will fix the problem, which is best depends
upon the use to which the model is put.
6.5
dd??
Low-dimensional models of Hopf bifurcations may be constructed using centre manifold theory. A direct application of the theory generates a model in an abstract phase space (the xy model) through a
nontrivial form of the homological equation.
However, a simpler and usually more useful model is developed in
terms of the complex amplitude; it describes the modulation of the
basic oscillations involved in a Hopf bifurcation. From the modulation
equations we easily predict the amplitude and frequency of the stable
limit cycles that may occur for parameters past critical.
In this chapter we have introduced some of the dynamics of the advectiondiffusion of temperature and salinity. In the natural environment, variTony Roberts, 1 Mar 2008
434
Chapter 7
437
7.2
Dispersion in pipes . . . . . . . . . . . . . . . . .
440
7.3
7.4
7.2.1
7.2.2
7.2.3
7.2.4
7.2.5
Significance . . . . . . . . . . . . . . . . . . . . . . 453
7.2.6
457
7.3.1
7.3.2
7.3.3
7.3.4
7.3.5
Summary . . . . . . . . . . . . . . . . . . . . . . .
435
479
436
Physically, a thin band of wavenumbers, width proportional to , corresponds to a pattern of Taylor vortices/convective rolls which in any physical
locale looks like vortices/rolls with a local wavenumber and amplitude which
varies slowly from locale to locale. The pattern is slowly varying in the large
spatial dimension.
poise??
To greatly simplify the analysis we investigate a much simpler situation, that
of the dispersion of material carried in the flow of a viscous fluid along a
pipe. Physically, pipes are long and thin: we model the large-scale variations
Tony Roberts, 1 Mar 2008
437
r
#
-
6
I
@
- u(r)
"!
-x
"!
length L, radius a
Figure 7.1: schematic diagram of pipe flow and the coordinate system.
in concentration of the material along the pipe, seeking to ignore small scale
dynamics across the small cross-section of the pipe. Thus first we look at
the flow in a pipe.
Such models of dispersion in pipes are important in chemical engineering
where many reactions takes place in pipes. Similar models are important in
the modelling of dispersion in rivers and estuaries. It was, I believe, Henry
Stommel who dumped a truck load of turnips into a river to experimentally
test models of dispersion: turnips were chosen because they barely float and
so will be carried by the river without influence by the wind, are white so
are easily visible, and are cheap and biodegradable.
film??
7.1
Sorry:
Consider the flow of a viscous fluid in a pipe and we solve the NavierStokes
and continuity equations to determine the velocity field in the pipe.
We use cylindrical coordinates (x, r, ), where x is the distance along the
pipe, r is a radius, and is an angle around the pipe. The fluid velocity has
components (u, v, w) for convenience, that is q = ui + ver + we . Then the
Tony Roberts, 1 Mar 2008
438
1 p
+ 2 u ,
x
1 p
v
2 w
2
=
+ v 2 2
,
r
r
r
1 p
2 v w
=
+ 2 w + 2
,
r
r r2
=
= 0,
where
w
q=u
+v +
,
x
r
r
2
1
= 2+
x
r r
2
1 2
r
+ 2 2.
r
r
dependence,
= 0, and no longitudinal dependence, x
= 0, in the
flow, except that the pressure may vary linearly in x.
Reflection symmetry, about any plane through the pipes axis, shows
that w = 0there is no reason to favour helical flow in any direction
so no angular velocity is a possibility. w = 0 ensures that the angular
component of the NavierStokes equation is satisfied.
The continuity equation then shows that rv =constant. But since the
fluid cannot cross the pipes surface, v, the radial velocity must be
zero at r = a, hence the constant is zero, and thus v = 0 everywhere.
The radial component of the NavierStokes equation is automatically
satisfied provided the pressure is independent of r, that is the pressure
is constant across any given cross-section.
The axial/longitudinal component of the NavierStokes equation then
show that
1 p
u
u
=
+
r
.
t
x
r r
r
Tony Roberts, 1 Mar 2008
439
p 2
r + A log r + B .
4L
(7.1)
a4 p
=
.
8L
The interesting feature here is that for a given pressure gradient the flux
is proportional to the fourth power of the radius. For example, the flux
in 19 mm (3/4) water pipe is some 5 times as much as the flux in 13 mm
(1/2) water pipe for the same pressure gradient. Alternatively, 19 mm pipe
can be 5 times the length of 13 mm pipe and still deliver the same flow of
water for a given pressure drop.
1
The minus sign in front of p shows that fluid flows from high pressure to low.
Tony Roberts, 1 Mar 2008
440
7.2
Dispersion in pipes
Warning:
on r = a .
441
A simple but misleading model of the long-term spread of material along the
pipe is obtained by integrating (7.2) over a cross-section. Let C(x, t)
RR be the
1
cross-sectional average of the concentration c, namely C(x, t) = a2 c rd dr .
In a long and thin pipe, the cross-pipe diffusion has a long time to act and
so we may expect that c C . Now averaging (7.2) over a cross-section and
using the boundary conditions then leads to
C
1
+
t
a2
ZZ
u(r)
c
2 C
r d dr = 2 .
x
x
c
If c C then x
C
x which is independent of r and and so may be taken
outside the remaining integral to give
C
C
2 C
+U
= 2 ,
t
x
x
where U is the average velocity down the pipe. This is a simple model for
the transport of material along the pipe by the flow. It asserts that the
material is carried at the average velocity U and spreads out along the pipe
by molecular diffusion . Now, whereas the bulk transport at the average
velocity is a reasonable first approximation, the spread or dispersion along
the pipe is typically immensely faster than that due to molecular diffusion.2
To discover the correct dispersion and lesser effects, we need to use a better
modelling approachthat provided by centre manifold theory.
Before undertaking further analysis, it is convenient to non-dimensionalise
this problem in the following way:
scale the radial coordinate with respect to the pipe radius, r = r/a ;3
scale time with respect to a cross-pipe diffusion time, = a2 / , t =
t/a2 ;
and scale longitudinal distances with respect to the distance travelled
downstream, = U = Ua2 / , by a particle travelling at the average
velocity U in a cross-pipe diffusion time, x = x/(a2 U) .
2
3
442
Upon this scaling and upon omitting the stars that indicate non-dimensional
quantities we solve4
c
2 c 1
c
1 2 c
2 c
+ 2(1 r )
= 2 +
r
+ 2 2.
(7.3)
t
x
x
r r
r
r
in a pipe of radius 1 with boundary condition
c
=0
r
on r = 1 .
(7.4)
Ua
1000Re .
7.2.1
The spectrum
443
35
30
25
20
15
10
0
35
30
25
20
15
10
<
Figure 7.2: the spectrum of complex eigenvalues for dispersion in a pipe.
Each branch is plotted for longitudinal wavenumbers 0 k < 60 ; negative k
is just the complex conjugate. The angular mode is indicated by a symbol
plotted at intervals k = 14 : + for n = 0 , for n = 1 , for n = 2 , square
for n = 3 , and for n = 4 .
444
the very right-hand branch in the figure observe that there is a continuum
of modes with decay rate near 0. It is these modes which will dominate
the long-term dynamics of solutions to the advection-diffusion problem. It
is these modes which will form the basis of a low-dimensional model of the
dispersion.
Unfortunately, there is no clear spectral gap in this problem: there is a
continuum of eigenvalues ranging from 0 to large negative values. Thus we
cannot simply assert that a certain number of modes will form the basis
for the slow manifold model, and the rest of the modes are exponentially
decaying: there cannot be a clear dividing line between the two. However,
observe that a characteristic of the modes of long term importance, small
decay rate, is that they are of low wavenumber k, say k < 5 or so. Thus by
focussing attention on low wavenumbers k, large longitudinal length-scales,
we make progress.
7.2.2
The application of centre manifold theory appears rigorous Mercer & Roberts
(1990) if we take the Fourier transform: c c^ and /x ik. The
advection-diffusion equation (7.3) transforms to
^
c
1
=
t
r r
|
^
c
1 2 c^
r
+ 2 2 2i(1 r2 )k^
c k2 c^ .
|
{z
}
r
r
{z
}
non-linear
linear
(7.6)
445
n=0
0
-14.68
-49.22
n=1
-3.390
-28.42
-72.87
n=2
-9.328
-44.97
-99.39
n=3
-17.65
-64.24
-128.3
Table 7.1: spectrum of the linearised dynamics at zero wavenumber; that is,
the spectrum of cross-stream diffusion when there is no longitudinal variations in concentration.
to (7.6) the trivial dynamic equation
k
= 0.
t
(7.7)
446
problems each with a given value of wavenumber k. Then for small k we have
identified the existence of a low dimensional slow manifold. Upon constructing a slow manifold model for each k, we would have made a model of the
long-term evolution for each wavenumber k. Then the inverse Fourier transform would put all these models together, albeit only accurate for small k,
to create a model in physical space for the long-term evolution of dispersion
in a pipe, see Mercer & Roberts (1990) for example. The high-wavenumber
mode are not relevant to the long-term evolution because from the spectrum they can be seen to decay exponentially. In this way we justify the
construction of a slow manifold model for the long-term evolution of dispersion which, because it is based on small wavenumbers k, is a model for
solutions which vary slowly along the pipe.
7.2.3
Direct construction
However, although the previous approach puts the slow manifold analysis
on a firm foundation, the Fourier transforms and their inverse are unwieldy
and unnecessary. Precisely the same results may be obtained by a direct
and much cleaner approach.
The essential action of the Fourier transform is to split the terms on the
right-hand side of (7.6) into two groups: linear and non-linear terms.
In the ensuing slow manifold analysis the non-linear terms are treated as
providing small perturbations to the linear dynamics. We can do this
directly, simply by observing that it is the presence of /x which flags
whether a term is to be treated as a perturbation or not. Thus we rewrite
the physical equation (7.3) as
c
1
=
t
r r
|
c
1 2 c
c
2 c
r
+ 2 2 2(1 r2 )
+ 2 .
r
r |
x }
{zx
{z
}
non-linear
linear Lc
(7.8)
Then we analyse the dynamics based on this split and seek solutions which
are slowly-varying in x, /x is small; this is equivalent to low wavenumber k.
Tony Roberts, 16 Mar 2009
447
C
0.
t
(7.9)
Now seek to iteratively refine this model. Suppose that we have an approximate description
C
(C) ,
c v(C) s.t.
g
t
and we seek to find corrections v 0 (C) and g 0 (C) where
c v(C) + v 0 (C)
s.t.
C
(C) + g 0 (C) ,
g
t
is a more refined model. Substituting into the governing equation (7.8) and
neglecting products of primed quantities gives
C 0 C
v 0
2 v 0
C
v
2 v
u
+u
g 0g
+ Lv 0 + 2 =
g
Lv 2 .
v
v
x
x
v
x
x
The existence of the critical mode is a direct consequence of the overall conservation
of material.
Tony Roberts, 16 Mar 2009
448
(7.10)
0, requires us to solve
The first iteration, to improve v C and g
C
2 C
2 .
x
x
As always, L is singular (from the critical mode of zero eigenvalue), and so
we cannot find v 0 without first determining g 0 to put the right-hand side
in the range of L. There are two equivalent ways of phrasing the same
solvability condition:
Lv 0 g 0 = u
one
RR could observe that by cross-sectionally integrating this equation,
0
. . . rd dr, and using the conservative boundary conditions, v
r = 0 ,
the Lv 0 term disappears to give
ZZ
C
2 C
0
g = u rd dr
2 ,
x
x
which simplifies to
2 C
C
+ 2 ,
x
x
as u(r) = 2(1 r2 ) has cross-sectional average of 1 by the nondimensionalisation;
g0 =
449
RR
or one could note that L is self-adjoint under the inner product hv, wi =
vw rd dr and so an adjoint eigenvector corresponding to the critical mode is just z(r, ) = constant , so that hz, equationi is the same
solvability condition as just used above.
Using this g 0 , the problem for v 0 becomes
Lv 0 = (1 2r2 )
C
.
x
In the absence of any variations in the right-hand side, the solution to this
equation is found by integration in r to be
1 2 1 4 C
r r
+ A,
v0 =
4
8
x
where A is an arbitrary integration constant (although it need only be constant with respect to r and ). Such arbitrary constants are determined by
the definition we choose for the amplitude of the model. Here we want
to write the model in terms of C, the cross-sectional average of c. Since
the initial approximation (7.9) does this we therefore require that the crosssectional average of any correction v 0 has to be zero. Here this requires that
A = C
x /12 , thus
v0 =
C
1
2 + 6r2 3r4
.
24
x
Thus correcting the initial model (7.9) with g 0 and v 0 we deduce the more
refined model6
C
1
C
C
2 C
cC+
2 + 6r2 3r4
s.t.
+ 2 .
(7.11)
24
x
t
x
x
2
C
C
The evolution predicted by this model, C
t x + x2 , is the same as
that obtained in the introduction: advection at the mean velocity of the
fluid, and along pipe molecular diffusion. However, there is one important
6
450
C
2 C
C
2 C
0
0
Lv g = 1 + w
+ 2 +u
+w 2
x
x
x
x
x
2
C
x
C
(1 2r2 )
2 w 3C
x
x
n
2 C
= (u 1)w 2 .
x
Note that the derivative C
1 2 C
.
48 x2
+
+
.
(7.12)
t
x
48
x2
This is a simple advection-diffusion equation with an effective diffusion co1
efficient of 48
+ . Typical solutions are Gaussian distributions with mean
location travelling at the mean velocity of the flow and spreading according to the effective diffusion coefficient. In most applications the along-pipe
molecular diffusion, = 1/ Pe2 , is negligible by many orders of magnitude
when compared with the 1/48 contribution. Where does this extra dispersion come from? It is called shear dispersion because it arises from the
shear in the basic Poiseuille pipe flow. Imagine for a moment that there is
no molecular diffusion at all. Then a small slug of material released into
Tony Roberts, 16 Mar 2009
451
the flow across some section of the pipe will be differentially carried along
according to the velocities at the different radii inside the pipe: the part of
the slug released next to the pipe-wall will barely move, while the part of
the slug in the centre of the pipe will be carried a long way down the pipe.
Thus the spread along the pipe, as measured by the standard deviation in
the x-direction for example, must increase with the mean advection velocity U. Such shear induced spread is ameliorated by the cross-pipe diffusion
which causes individual contaminant particles to eventually sample the entire velocity distribution. These effects can be seen in the dimensional form
of the model (7.12):7
C
C
U
+
t
x
U 2 a2
+
48
2 C
,
x2
2 2
a
where the shear dispersion coefficient U48
is proportional to the square of
the velocity and inversely proportional to the cross-pipe diffusivity .
7.2.4
High-order convergence
Computer algebra easily verifies the above results and computes higher order
corrections. For example running the reduce program below informs us
that a better model for the long-term evolution is
C
C
+
t
x
2
1
C
41
x
1 x
+
3C
4C .
48
x2
2880 n
2580480 n
(7.13)
Physically these corrections describe how the skewness and the kurtosis of
the along-pipe distribution evolve in the long-term. They show, for example,
that a contaminant pulse is generally not a Gaussian but will be skewed
somewhat. The skewness does decay, but only algebraically slowly and so
is captured in the slow manifold model. This issue is discussed fully by
Chatwin (1970).
7
Observe the apparent paradox that the smaller the molecular diffusion, the larger the
shear dispersion!
Tony Roberts, 16 Mar 2009
452
453
Note that because of the linearity in c, really the only small parameter in
this problem is /x; that is, the wavenumber k. Substituting solutions
proportional to eikx+t find
1
i 3
41
ik
+ k2 +
k
k4 .
(7.14)
48
2880
2580480
This is simply the Taylors series of the m = n = 0 branch of the spectrum
of the full problem shown in Figure 7.2: the first two terms are a parabolic
approximation leading to an advection-diffusion model; higher-order approximations lead to higher-order models. In essence the slow manifold model
neglects all branches of the spectrum except the dominant m = n = 0
branch, and then only accurately approximates that part of the branch (near
critical) corresponding to small eigenvalue and small wavenumber as seen in
Figure 7.3.
Thus the slow manifold model is limited by its dependence upon:
the neglected transients of the neglected branches in the spectrum
the most significant of these decays approximately like e3.39t showing
that the model is valid on dimensional time scales t a2 //3.39 ;
and the necessarily slowly-varying dependence along the pipe from the
effective Taylor expansion in k.
In this particular problem we may use the computer discover quantitative
bounds on the spatial resolution.8 Computing to 24th order one may show
Mercer & Roberts (1994) that the above expansion for in wavenumber k
actually converges for wavenumber k < 13.8 . Thus we should only attempt
to use such a model to resolve along-pipe structures on a dimensional length
2
2
scale greater than 13.8
= 0.45 Ua
. Because of the smallness of molecular
diffusion this length can be surprisingly large in applications.
7.2.5
Significance
This sharp result for the spatial resolution is some 10 times better than that estimated
by Taylor Taylor (1954)
Tony Roberts, 16 Mar 2009
454
20
18
16
14
12
10
8
6
4
2
0
10
<
Figure 7.3: the spectrum of complex eigenvalues for dispersion in a pipe
compared with that of the slowly varying, slow manifold model (solid line).
Each branch is plotted for longitudinal wavenumbers 0 k < 30 ; the
angular mode is indicated by a symbol plotted at intervals k = 10 : + for
n = 0 , and for n = 1 . Observe the good fit between the model and the
leading branch up to wavenumber 10 or so.
455
The non-trivial models (7.12) and (7.13) are quite easy to understand
and to use to predict dispersion in a pipe.
The models are putatively low dimensional, and yet, being phrased in
terms of the time evolution of a function of x, are actually infinite
dimensional ! Nonetheless, they are considerably simpler than the
original mathematical description of the advection and diffusion in
a pipe because we have systematically eliminated all the cross pipe
dynamics to result in a description of the along pipe dispersion.
The slowly varying, low wavenumber, long wavelength approximation,
as used here for pipe dispersion, occurs in many applications. Here
we see that the essence of the approximation is to get the spectrum
of the low wavenumber, long wavelength modes correct, and to ignore
the dynamics on all other branches of exponentially decaying modes.
Although this is straightforward to see in a linear problem such as pipe
dispersion, there is no correspondingly simple statement that can be
made in a nonlinear problem. For truely nonlinear dynamics we need
the full power of centre manifold techniques.
The convergence that we see in the slow manifold model here is rare.
Most slow manifold models are only asymptotic. Here, the convergence
allows us to make a sharp estimates on the applicability of the model.
Exercise 7.1:
In flow in porous media such as soil or underground aquifers,
the fluid flows through the highly tortuous pipes of the pores between
granules of material. In such flow there is no regular physical structure on which to base analysis. Instead one may posit a microscopic
modelling parametrising the cross-stream direction by the advection
velocity v rather than the physical cross-stream direction. One argues
the following.
The concentration of interest f(x, v, t) is the concentration of particles at position x at time t that is travelling at velocity v through
the micropores.
The micropores occupy a fraction of the physical space, and
Tony Roberts, 16 Mar 2009
456
c(x, t) = f dv . Interpret.
3. Explore the slow manifold, slowly varying models for other equilibrium velocity distributions (v).
7.2.6
457
ck
t .
Assuming odes are weakly coupled implies there exists a slow manifold
parametrised by, say, countably infinite amplitudes Ck (t; X) = ck (y, t; X).
Somehow, consistency requires that if
some such??
7.3
Sorry:
C0
t
= g0 (C) then
Ck
t
= kX g0 (C), or
Our aim is to create relatively simple models of thin layers of fluid moving
on a solid bed. Examples include the flow of rainwater on a road or a
windscreen, paint, coating flows, and the flow of many protective biological
fluids.
When the fluid layer is thin then, as in dispersion in a pipe, the only important dynamics occur along the thin film of fluid. Across the fluid film,
viscosity acts quickly to damp almost all cross-film structure. This distinction between the longitudinal and cross fluid dynamics is exactly analogous
to that of dispersion in a pipe, and we proceed with a similar analysis.
However, the equations are very different (see Figure 7.4). In particular this
problem has many nonlinearities: not only is the advection in the Navier
Stokes equation described by a nonlinear term, but also the thickness of the
fluid film is to be found as part of the solution and its unknown location is
another source of nonlinearity.
Assuming no longitudinal variations, a linear analysis of the equations shows
that there is one critical mode in the cross-fluid dynamics, all others decay
Tony Roberts, 24 Apr 2009
458
y
6
atmospheric
pressure
hhh
((((
h
h
hhhh
y = (x,(
t)
h
(
((( 6
Navier--Stokes equation
h
u(x, y, t)
p(x, y, t)
?
x
solid, u = v = 0
Figure 7.4: schematic diagram of a thin fluid film flowing down a solid bed.
due to viscosity. This critical mode is associated with conservation of the
fluid. Consequently it is natural to express the low dimensional model in
terms of the film thickness (x, t). Seeking solutions which vary slowly
along the film, a centre manifold analysis creates an effective model of the
dynamics.
An interesting diversion is attributable to the fact that although this is a
nonlinear problem, conservation of fluid applies no matter how thick the
fluid layer. Thus the analysis, when properly done, is valid for arbitrarily
large variations in the thickness of the film! just so long as the variations
are sufficiently slow.
7.3.1
Governing equations
q = 0.
For simplicity we restrict attention to two-dimensional flow taking place in
the xy-plane; that is assume there is no dependence upon the other horizontal coordinate z. The x-axis is aligned along the solid bed of the flow;
Tony Roberts, 24 Apr 2009
459
the y-axis is perpendicular. The viscous flow must stick to the solid bed to
give the boundary condition
q = 0 on y = 0 .
The surface of the fluid, y = (x, t) , evolves with the flow. Because the
free-surface is unknown we not only need two boundary conditions for the
NavierStokes equations, we also need an extra boundary equation in order
to be able to find . One condition is that the fluid flow as given by the
velocity, q, must follow the free-surface y = (x, t) . A straightforward
way to derive such an equation9 is to note that y must be constant,
specifically 0, for the fluid particles on the free-surface. Thus the material
derivative, ddt (y ) , must be zero on the free-surface. Hence
(y ) + q (y ) =
+vu
t
t
x
= vu
on y = .
x
0 =
(7.16)
460
p
^ = ( 0 i + j) / 1 + 02
Now a normal vector across the free-surface y = is n
where 0 =
x . Thus the fluid stress across the free-surface is
^=p
T =n
1
1 + 02
0
p xx 0 + xy
.
0 yx p + yy
i
h
1
0
02
= 0.
)
+
1
xy
yy
xx
1 + 02
Thus
2
v u
y x
+ 1
02
u
v
+
y x
=0
on y = .
(7.17)
Since
a , is very low, then Bernoullis equation, pa +
the density,
1
2
a t + 2 |q| = constant, asserts that any air movement has little
effect on the air pressure, and hence the fluid stress exerted normally
across the surface has to be constant, say Tn = pa , equal and opposite to air pressure.
However, surface tension is an important force on the thin films we address. The effect of surface tension is like that of an elastic membrane,
it causes a pressure jump if the surface is curved: positive if the fluid
surface is convex; negative if it is concave.10 The jump in pressure
3/2
is proportional to the curvature R1 = 00 / 1 + 02
. With as
the coefficient of surface tension, then Tn + R = pa . For a clean
water/air interface = 72.8 dyn / cm (Batchelor 1979, p597), though
this coefficient does vary markedly with temperature and surface contamination.
10
Experimentally, surface tension is difficult to deal with as it is sensitive to surface contamination. For example, small gradients of contaminants can supply tangential stresses
which drive fluid flow.
Tony Roberts, 24 Apr 2009
461
For the normal stress and surface tension to oppose atmospheric pressure11
00
1 + 02 (p pa ) = yy 2 0 xy + 02 xx p
.
1 + 02
Without loss of generality we may take pa = 0 (since it is only gradients of pressure that are important), thus
v
v
02 u
0 u
02
+
+
p = 2
1+
y
x
y x
00
p
on y = .
1 + 02
Now non-dimensionalise these equations. For a reference length, suppose
that h is a characteristic thickness of the thin film as shown schematically
in Figure 7.4. Then non-dimensionalise by writing the equations with respect to: the reference length h; the reference time h2 / being a cross-film
diffusion time; the reference velocity /h, and the reference pressure 2 /h2 .
With these choices, and in non-dimensional quantities, we solve the Navier
Stokes and continuity equations
q
+ q q = p + 2 q ,
t
q = 0,
(7.18)
(7.19)
(7.20)
and on the surface the the kinematic condition (7.16), and two dynamic
conditions (7.17) and:
v
v
W 00
02
02 u
0 u
1+
p=2
+
+
p
on y = ,
y
x
y x
1 + 02
(7.21)
11
If the fluid is inviscid, as would be the case for large scale flows, then this condition
reduces to p = pa on the free-surface.
Tony Roberts, 24 Apr 2009
462
h
where W =
2 is a form of Weber number characterising the importance
of surface tension.
In the NavierStokes equation I have neglected body forces. In the presence of gravity were to be acknowledged then it would appear in the nondimensional combination gh3 /2 . This may be neglected if gh3 /2 is genuinely small (h 0.01 cm for water). In fact, including gravity is not a
great complication and I leave its inclusion as an exercise for the reader.
7.3.2
Linear picture
= v on y = 1 ,
t
u v
+
= 0 on y = 1 ,
y x
v
p = 2
W 00 on y = 1 .
y
(7.22)
To investigate the dynamics of long waves on this thin film we will treat
is actually an
463
464
0
-2
-4
-6
-8
real()
-10
-12
-14
-16
-18
-20
-22
0
0.5
1.5
2
k
2.5
3.5
7.3.3
Restricted model
465
u = v = p = 0,
and free to vary according to
0.
s.t.
() + q 0 () ,
q = q
() + p 0 () ,
p = p
() + g 0 () ;
= g
t
466
y2
y
v 0
y
=
=
p
,
+q
u+
2 u
t
x
v
p
v +
+q
2 v ,
t
y
(7.24)
(7.25)
,
= q
(7.26)
u v
+
on y = 1 ,
y x
v
=
p+2
W 00 on y = 1 ,
y
= 0 on y = 0 ,
= v + v 0 on y = 1 .
(7.27)
(7.28)
(7.29)
(7.30)
467
Pressure gradients induce a parabolic shear flow, just as in Poiseuille flow in a pipe.
Tony Roberts, 24 Apr 2009
468
W
[v]y=1 = iv .
t
3
(7.31)
This is a model for the long-term evolution of thin films under the action of
surface tension. At this level of approximation it is linear; however, further
iterations will generate nonlinear terms. But for now if we seek solutions
4
proportional to et+ikx , then observe that = W
3 k in agreement with the
small k behaviour of the leading eigenvalue branch shown in Figure 7.5. This
shows that all variations in thickness of a fluid film will eventually decay, to
result in a film of uniform thickness, although the decay, driven by surface
tension, is very slow for long-wavelength disturbances.
Another way of saying the same thing is that the model (7.31) is rather like
a diffusion equation, but with a much slower (4th order) diffusion than the
normal Fickian diffusion. A hump of fluid in the film slowly spreads out
due to surface tension, as shown in Figure 7.6, but the spread becomes very
slow over large distances.
The centre manifold theorems reasonably assure us that this model is indeed
relevant to the long-term dynamics of thin films. However, this assurance
is not yet rigorous because of deficiencies in the preconditions of current
theorems.
With a little more craft the above iteration may be restructured so that new
information about the structure of the solution fields is used as soon as it is
discovered. Such restructuring is akin to GaussSeidel iteration, and results
in fewer iterations being needed.
Instead we instruct a computer to carry out the algebra using the following
reduce code. The structure of the algorithm is exactly the same as always.
One innovation is that we need to count the number of spatial derivatives
Tony Roberts, 24 Apr 2009
469
initial
Figure 7.6: schematic diagram of a hump in a fluid film spreading out longitudinally under surface tension.
470
so that terms with many x-derivatives, here 9 or more, are removed from
the computation to restrict the algebra to the significant terms.
% Construct slowly-varying centre manifold of thin film fluids
% BUT only linearised boundary conditions on the free surface.
% (c) Tony Roberts, November, 1997
%
% Use d to count the number of derivatives of x,
% and throw away this order or higher in d/dx
let d^9=>0;
on div; off allfac; factor d,df,w;
% solves -df(p,y)=rhs s.t. sub(y=1,p)=0
operator psolv; linear psolv;
let {psolv(y^~n,y) => (1-y^(n+1))/(n+1)
,psolv(y,y) => (1-y^2)/2
,psolv(1,y) => (1-y) };
% solves df(u,y,2)=rhs s.t. sub(y=0,u)=0 and sub(y=1,df(u,y))=0
operator usolv; linear usolv;
let {usolv(y^~n,y) => (y^(n+2)/(n+2)-y)/(n+1)
,usolv(y,y) => (y^3/3-y)/2
,usolv(1,y) => (y^2/2-y) };
% linear solution
depend h,x,t; let df(h,t) => g;
u:=0; v:=0; p:=0; g:=0;
%
% iteration (Gauss-Seidel like)
repeat begin
% use vertical mom to determine the pressure
veq:=re*(df(v,t)+u*df(v,x)*d+v*df(v,y))+df(p,y)
-df(v,x,2)*d^2-df(v,y,2);
tn:= sub(y=1,-p+2*df(v,y)-df(h,x,2)*d^2);
write p:=p+psolv(veq,y)+tn;
% use horiz mom to determine horiz velocity
ueq:=re*(df(u,t)+u*df(u,x)*d+v*df(u,y))+df(p,x)*d
-df(u,x,2)*d^2-df(u,y,2);
Tony Roberts, 24 Apr 2009
471
tt:=-sub(y=1,df(u,y)+df(v,x)*d);
write u:=u+usolv(ueq,y)+tt*y;
% use continuity to find vertical velocity
ceq:=-df(u,x)*d-df(v,y);
write v:=v+int(ceq,y);
% find evolution
write g:=sub(y=1,v);
showtime;
end until {veq,tn,ueq,tt,ceq}={0,0,0,0,0};
end;
Upon running this code observe that a high-order model is
W iv 3W vi
23W 2W 2
37W 2 000 iv 0
+
viii +
. (7.32)
t
3
5
21
45
840
The linear terms correct the linear spectrum; the nonlinear term comes from
the horizontal advection term, u u
x , as this is the main nonlinearity in a film
with linearised boundary conditions.
Note that if we truncate the linearised version of the model (7.32) to the
first two terms, then for modes et+ikx we observe
1 3 2
4
= Wk
k
.
3 5
In use then we would observe modes with wavenumbers k > 5/3 growing
exponentially! Such growth is definitely unphysical. This is always a risk
with a long-wavelength, slowly-varying model: at finite wavenumber the
approximation may be so inaccurate that high wavenumber modes grow
instead of decay. Two things may be done: either higher-order terms are
included in the model to see if they fix the problem, or we can recast
the linear dynamics. Consider an example of the latter here. Instead of
approximating the linear dynamics by a Taylor series in wavenumber k, as
done up to now, we may approximate by a rational function, viz
=
Wk4
,
3 1 + 59 k2
472
=
.
2
t
5 x t
3 x4
Including nonlinear terms into this equation we then would derive a model
with the correct low-wavenumber and nonlinear behaviour, and with stable
high wavenumber dynamics so that useful solutions could be attained.
7.3.4
x
.
1+x
473
M
M0
474
Substituting this into the first of the pair of governing equations gives that
x evolves over ordinary time-scales according to
x (1 c)
x
.
1+x
We now show how centre manifold theory can put this heuristic argument
on a sound basis.
The first task17 is to stretch time so that the time-scales fit within the
framework of the theory as it has been presented. We do this by letting
t = so that ordinary times for t correspond to long times in and the
very rapid transients occurring in a t-time of order occur on a -time of
d
d
order 1; note that dt
= 1 d
. The dynamical equations then become
dx
d
dz
d
= [x + (x + c)z] ,
= x (x + 1)z ,
x
When = 0 , the curve z = 1+x
is a manifold of fixed points, M0 , with
exponential attraction in the z direction, at an x dependent rate = (1 +
x) . For small but non-zero, formally done by adjoining the trivial evolution
equation d
d = 0 and considering the dynamics local to M0 in xz-space,
the manifold of quasi-equilibrium is perturbed to some manifold, M, and
on M the evolution is slow. By basing the analysis on the whole of M0 we
obtain an approximate model which is globally valid along the whole of M,
just provided is small enough.
+ (x + c)y .
d
(1 + x)2
1+x
17
What is important is the relative separation between the two time-scales of the ignorable transients and the interesting dynamics.
Tony Roberts, 24 Apr 2009
475
(1 c)x
+ O 2 ,
4
(1 + x)
7.3.5
Thin films
Now we set up a good analysis of thin fluid film dynamics: good because
we cater for large variations in the the film thickness, just so long as the
variations are slow in x, that is /x is small.
The unknown location of the free surface of the film is still a major technical
difficulty. However, one way to proceed is to scale the vertical coordinate,
= y/ , so that the free surface corresponds to = 1 precisely. Unfortunately, because varies with x and t, this scaling of y affects space-time
derivatives and so plays havoc with details of the governing equations. Fortunately I have based much of our analysis on computer algebra and we
relegate most such details to the computer.
Under the change of coordinates from (x, y, t) to
= x,
= y/(x, t) ,
= t,
Tony Roberts, 24 Apr 2009
476
=
=
1
.
477
478
showtime;
end until {dseq,veq,tn,ueq,tt,ceq}={0,0,0,0,0,0};
end;
Note that for brevity of the printed results I used h(m) to denote m /xm .
This is implemented in lines 911, along with the interchange of time and
space derivatives in line 31.
Lastly, I have rewritten the iteration loop to use new information as it
becomes available. First, the pressure correction is found from the vertical
momentum equation, then the u correction is found from the horizontal
momentum equation, then v from continuity, and lastly the latest version of
the model evolution
t = g from the kinematic free surface condition.
Running this program we find the long-term evolution of long-wavelength
modes is approximately described by18
W 3 000
(7.33)
3 x
11 3 0 2 000
3 5 v
4 0 iv
4 00 000
3 0 00 2
W
+ 3 + +
.
x 5
6
I need to say more about the dynamics under this model, and what it
means physically. However, throwing out gravitational effects seem to
have severely limited the range of comments I can make.
Although this model describes the long term dynamics of thin films, it is
limited in its usefulness (even with gravitational effects included). For example, in the linearised problem at finite wavenumber, (7.22), the leading
branch of the spectrum, shown in Figure 7.5, merges with the next branch,
the gravest shear mode u = sin /2 , at a wavenumber k 2 to become
oscillatory decaying modes. Such oscillations are the remnants under the
strong viscosity in thin films of the waves which surface tension can support. In applications, such decaying waves seem important, for example see
the review by ChangChang (1994). However, the model (7.33) cannot describe the necessary oscillations because it has only one component and is
only first-order in time. This appears at first sight to be a strong limitation
18
The right-hand side can always be written as a gradient due to conservation of fluid.
7.4. Summary
479
7.4
Summary
W
3 3
t 3 x ( x3 ) in the absence of gravity.
These models, being infinite dimensional pdes, are rich enough to
describe a wide range of dynamics of interest. In contrast, models of
the pitchfork and Hopf bifurcation are just pale shadows of the full
range of dynamics inherent in the respective fluid problems.
Tony Roberts, 24 Apr 2009
480
Chapter 8
8.2
8.3
Sorry:
482
8.1.1
Sound . . . . . . . . . . . . . . . . . . . . . . . . . 483
8.1.2
8.1.3
. . . . . . . . . .
492
8.2.1
8.2.2
8.2.3
Summary . . . . . . . . . . . . . . . . . . . . . . .
508
482
relevant to the original high dimensional problem. However, in many practical situations we also make low dimensional approximations where there is
no clear and strong dissipative mechanism to cause such a dimensional collapse. Some examples introduced in the next section are the incompressible
approximation, rigid body dynamics, surface waves on water, atmospheric
dynamics, and irrotational flow. In these examples, the low dimensional
approximation is a relatively simple dynamical system which in some sense
acts as a guiding centre for the high dimensional dynamics; typically, the
high dimensional dynamics consists of relatively fast oscillations superimposed upon the slow evolution of the model. Hence such models are typically
based on what is called the slow manifold.
A little theory developed by Sijbrand (1985) may help justify such models.
However, there are severe limitations to the applicability of such theory,
primarily due to resonance. Nonetheless the crucial ideas are illustrated
by applying them to a toy model of atmospheric dynamics. Normal form
transformations justify the modelling and display the limitations of such
guiding centre slow manifolds.
scene??
subc??
8.1
Warning:
8.1.1
483
Sound
+ (q) = 0 ,
t
q
1
+ (q )q = p .
t
(8.1)
(8.2)
(8.3)
where and A are characteristics of the fluid and its temperature. For
example, in air 7/5 as appropriate for a gas of diatomic molecules. The
polytropic equation of state implies that
p =
p
= A1 = c2 ,
where c2 = p
in which c turns out to be the speed of sound. This is used
to eliminate the pressure from the Euler equation (8.2).
A stationary fluid, q = 0, of constant density, = 0 , is an equilibrium or
fixed point of the inviscid Euler equations above.
Consider small amplitude fluctuations, otherwise known as sound, to such
a density and velocity field. That is, seek perturbations
q = 0 + q0 ,
= 0 + 0 ,
484
where dashed quantities are small. The continuity equation (8.1) gives
0 =
=
0
t
+ (q)
t
0
+ (0 + 0 )q 0
t
neglecting products of perturbations
0
+ 0 q 0
t
= 0 q 0 .
A2
0 = 0 0 .
0
t
0
Then consider
2 0
t2
2 0
t2
q 0
by continuity
= 0
t
2
c0
0
by Euler
= 0
0
= c20 2 0
(a wave equation).
That is, fluctuations of density, sound, travels through the fluid according
to the above wave equation with wave speed c0 .
As seen in most undergraduate courses, solutions of the wave equation in
one spatial dimension may take the form 0 = f(x c0 t) + g(x + c0 t)a superposition of right and left travelling waves. This solution describes waves
travelling without change in form or amplitudeexactly our experience with
sound and exactly what makes it practical to communicate with sound over
long distances.
Tony Roberts, 1 Mar 2008
8.1.2
485
Slow flows
Now consider the flow of a fluid in the presence of fast sound waves.2 That
is, the speed of sound, c0 , is much faster than the relatively slow speed
of the fluid flow. Because the sound oscillates very fast, we show that it
has negligible effect on the slow bulk flow. Further, the slow flow evolves
according to the incompressible equations of fluid dynamics.
Now a large parameter such as c0 , the speed of sound, is a little inconvenient
to deal with in the following derivation. Instead I use the Mach number
M = U/c0 where U is a typical speed of the fluid flow. Here imagine we
have chosen units so that the fluid flows with velocities of size 1. Then the
Mach number, M = 1/c0 , is a small parameter.
To investigate flow in the presence of superposed sound, decompose the
velocity and density fields as
= 0 + M2 + M3 0 + ,
+ M2 q 0 + ,
q =
q
(8.4)
(8.5)
where 0 will be shown to be constant, quantities with tildes are those of the
slow bulk flow, and where primed quantities describe the sound field. One
oddity of the approximation process is a consequence of the fast oscillations
of sound: namely, time derivatives of any sound quantity is large. Specifically, the time derivatives of a primed quantity is taken to be of size c0 ,
or 1/M, times the quantity. Thus I write
0
=
+ M2
+ M2 .M
+ ,
t
t
t
t
0
2
where the M3
t term is of order M because M t of a primed quantity is
0
of size 1, order M , in the limit of large sound speed.
0
+ M2 0 q 0 + q
= O M3 .
+ M2
+ M2 .M
+ 0 q
t
t
t
2
486
) .
+ 0 q 0 =
(
q
t
t
The left-hand side is precisely the terms derived in the previous subsection for the propagation of sound. The right-hand side represents
the generation of sound by the slow flow, wind noise for example.
Eulers equation for a polytropic fluid,
q
c2
+ q q = ,
t
becomes
q 0
1 c2
q
q
= 2 2 0 + M2 + M3 0 +O M2 .
+M.M
+q
t
t
M c0
The largest term in this equation is of order 1/M2 and asserts (note
that c c0 to leading order)
0=
1
0 ,
q
1
q
=
+q
.
t
487
q 0
1
= 0 ,
t
8.1.3
488
sound
fast oscillations
compressible fluid
state space
q0
+ q0
uq = q
489
Kampen (1985), in his review of dynamical modelling, calls such slow manifold approximations as being based upon the principle of a guiding centre.
Consider some other examples, though the first two are not from fluid dynamics.
Rigid body dynamics Think of a ball thrown, hit or kicked through the
air. For most practical purposes we treat the ball as a rigid body, just
using equations for its translational and rotational degrees of freedom.
At its most complicated we would only consider it as a dynamical
system with a 12-dimensional state space: 3 for spatial position, 3 for
velocity, 3 for angular orientation, and 3 for angular velocity. However,
a real ball is an elastic or visco-elastic bodyit has infinitely many
internal dynamical modes of vibration. Thus in practise, we ignore the
infinitely many fast vibrational modes, and just consider the slow
translational and rotational dynamics of a rigid body. Such a model is
justified (Muncaster 1983, e.g.) because the internal vibrational modes
appear to have no effect on the flight of a ball.
Beams Beams are used all around us in constructions. To predict the
load a beam may safely carry, and how a beam deforms under load,
engineers do not solve the 3D elasticity equations for the beam as an
elastic solid. Instead they simply solve a beam equation, of the form3
2
AE 4
=
,
t2
x4
for the sideways deflection, , of the centreline of a beam with x measuring distance along the beam. Similar to the rigid body approximation, such a beam model neglects the fast vibrational modes in
the cross-section of the beam; instead it just resolves the much slower
dynamics of bending (Roberts 1993). Compare this with shear dispersion in a pipe where cross-pipe diffusional modes are neglected because
they are exponentially damped, here we neglect cross-beam vibration
because they appear as rapid oscillations centred upon the shape of
the bent beam.
3
Check
Tony Roberts, 1 Mar 2008
490
Tides Tidal flow around the world occurs on a 12 hour time scale. Computational models of tides need to resolve the dynamics on this time
scale and on a height scale of the order of meters. Such models are
used to not only predict tidal heights, but also the flushing of effluent, or the lack thereof, and the migration of sea organisms. Also
on the ocean there exist short waves with a period of the order of
10 secondswaves of interest in their own right for swimmers, sailors,
and surfers. However, how can tidal models neglect the short period
waves? Short period waves are neither exponentially decaying, nor
particularly small. The only reasonable rationale is to consider that
models of tidal dynamics act as a centre for actual realisations in
that short period waves are just oscillations about tidal flow.
Quasi-geostrophy In atmospheric dynamics, weather forecasters need to
predict the evolution of the high and low pressure patterns that we
see on the daily weather maps. These patterns, called Rossby waves,
have a characteristic time scale of one week as they drift from west
to east. However, in the dynamics of a rotating and vertically stratified atmosphere there also exist so-called gravity waves with a much
shorter time scale of the order of tens of minutes. In the large, these
gravity waves do not exist in the atmosphere;4 instead the atmosphere
is in a familiar balance characterised by winds travelling parallel to the
isobars seen on weather maps, called geostrophic balance. However,
numerical models of the atmosphere are plagued by the fast gravity waves when all meteorologists want to resolve is the slow Rossby
wave dynamicscalled quasi-geostrophic because nonlinearities have
a perceptible and important influence on the balance. Thus meteorologists want to construct and use models of the slow manifold of
quasi-geostrophic dynamics in an atmosphere which also supports the
much faster dynamics of gravity waves.
Irrotational flow Consider 2D fluid flow.5 The Navier-Stokes equations
4
Gravity waves may be important in local scale phenomena such the patchiness of
rainfall. If they occur, they angle upwards and eventually break and dissipate in the low
density of the stratosphere.
5
I may be on thin ice with this example.
491
+
(x 0 , y 0 ) dx 0 dy 0 ,
|{z}
2
(x x 0 )2 + (y y 0 )2
{z
}
irrotational |
rotational part
where is a potential, satisfying 2 = 0, and is the vorticity field;
time dependence is implicit in and . Now, the vorticity is carried
by the fluid, thus if we imagine it P
is made up of a large number of
discrete lumps called vortices, = i i (x xi ),6 then the location
of these vortices evolve according to the following adaptation of the
earlier integral:
1 X (yi yj )i + (xi xj )j
dxi
= +
j .
dt
2
(xi xj )2 + (yi yj )2
j6=i
This is a Hamiltonian system. Its dynamics may be incredibly complex, chaotic for example, but it still has a large number of conserved
6
Such a vortex decomposition is the basis for some high Reynolds number numerical
simulations.
Tony Roberts, 1 Mar 2008
492
With these examples I hope to have convinced you of the need to study the
modelling of slow dynamics in the presence of fast dynamics.
8.2
Warning:
There is a little theory7 that helps justify the construction of slow manifolds. Unfortunately, the theory I am aware of is extremely limited and is
mentioned more for its conceptual support than in any hope of rigour in an
interesting application. Perhaps the best support comes from a normal form
transformation which I introduce in a toy model of the quasi-geostrophic approximation.
493
complex
i2 v
i1 v
-
i1 v
i2
<()
i3 v
8.2.1
Building upon earlier work by Lyapunov, Sijbrand (1985) [7] looked briefly
at dynamical systems with purely oscillatory dynamics. As usual, consider
a dynamical system in the form
u = Lu + f(u) ,
(8.6)
This is viewed as being composed, indeed as it written, of a number of vibra m eim t , sometimes called normal modes. The
tional modes am eim t + a
oscillations of such a normal mode all take place in an invariant subspace
Tony Roberts, 3 Apr 2009
494
Em = span {<(am ), =(am )}, and all take the form of limit cycles. That is,
associated with each complex conjugate eigenvalue pair is an invariant subspace. The question is: how much of this manifold and limit cycle structure
is maintained in the actual nonlinear problem?
The essence of the existence theorem is as follows. Provided (Sijbrand 1985,
Thm7.1):
the spectrum is not degenerate (no multiple eigenvalues);
the linear frequencies are non-resonant, that is jm 6= kn for any
j, k, m, n (roughly); and
there exists a non-trivial integral of the motion, such as a conserved
energy integral;
then for each m there exists an invariant manifold, Mm ,8 of (8.6) tangent
to Em at the origin and composed of limit cycles. Such an invariant manifold
is termed a sub-centre manifold because it is the nonlinear analogue of a
subspace of the centre eigenspace. Such a theorem more-or-less justifies the
construction of nonlinear normal modes for vibrational systems as recently
done by Shaw & Pierre (1993), Shaw (1994a,b).
However, does the theory give any support for using sub-centre manifolds
as the basis for low-dimensional models of high-dimensional dynamical systems? The answer is certainly not directly. For a sobering example, consider the system (Sijbrand 1985, Eq.(7.23)) which in bi-polar coordinates,
u1 = r cos , u2 = r sin , u3 = cos and u4 = sin , is
dr
dt
d
dt
d
dt
d
dt
8
= r r4 10r2 2 + 54 ,
= 5r4 10r2 2 + 4 ,
= 1 ,
= 2 .
(8.7)
495
496
8.2.2
The present state of theory being unsatisfactory, I propose that one can
justify slow manifold models of dynamical systems via the normal form
transformation. Here I introduce the ideas and results through a specific
example.
Lorenz (1986) proposed the following system of 5 coupled equations in order
to model some important characteristics of atmospheric dynamics.
u = vw + bvz ,
v =
uw buz ,
= uv ,
w
x = z ,
z =
x + buv .
(8.8)
(8.9)
(8.10)
(8.11)
(8.12)
497
and I = v2 + w2 + x2 + z2 .
and z = Z(u, v, w) ,
(8.13)
on which the evolution is given by the first three equations of the Lorenz
system (8.88.10) with Z substituted for z. Then
x = z
from (8.11)
= Z by (8.13)
X
X
X
by chain rule
=
u +
v +
w
u
v
w
X
X
X
= (vw bvZ) +
(uw buZ)
uv ,
u
v
w
9
498
and similarly
z = x + buv
from (8.12)
= X + buv by (8.13)
Z
Z
Z
by chain rule
u +
v +
w
=
u
v
w
Z
Z
Z
= (vw bvZ) +
(uw buZ)
uv ,
u
v
w
Rearranging the second and last lines of these two equations gives
Z
Z
Z
(vw bvZ) +
(uw buZ)
uv ,
u
v
w
X
X
X
(vw bvZ)
(uw buZ) +
uv .
u
v
w
X = buv
(8.14)
Z =
(8.15)
and Z(1) = 0 .
and Z(2) = b u2 v2 w ,
499
= uv .
w
This model of the dynamics is simpler because it is low-dimensional, 3dimensional instead of the original 5-dimensions. In many applications this
is where an enormous simplification occurs. However, in some applications,
and atmospheric dynamics is one such, the gain in the dimensional reduction
is rather small: as here, one reduces the dimensionality by simply a factor
of two or three rather than an infinite amount as in other applications.
However, in such applications a further and important gain is attained on
the slow manifold. In atmospheric dynamics, the evolution on the slow
manifold is of the order of a week, whereas that of the fast dynamics is
of the order of tens of minutes; thus in the absence of fast waves, explicit
numerical simulations can take a time step some one hundred times larger
than would otherwise be the case. An important characteristic of the slow
manifold is the long time scale of all its dynamics.
I cannot resist giving the following simple reduce program to compute
the slow manifold of Lorenzs five equation toy of the atmosphere.
500
8.2.3
y = z ,
z = y .
(8.18)
= y,
= z.
(8.19)
= ,
= .
(8.20)
The significant generic feature of this example is that the presence of fast
waves causes a drift that is not described by the slow manifold model. The
drift depends upon the amplitude of the waves. Consequently, there can not
be a strong relevance theorem for slow manifold models when fast oscillatory
modes are eliminated. In this example we see that the evolution on and off
the slow manifold are fundamentally incompatible.
Tony Roberts, 3 Apr 2009
501
We quickly explore how to derive the normal form (8.20) to introduce the
general procedure developed later. Seek a near identity coordinate transformation of (8.18), but since the oscillations are not coupled to the slow mode
we only seek to transform the slow mode:
x = + X(, , ) ,
y = ,
z = ,
(8.21)
??fill in details??
In Chapter ?? devoted to normal forms we will explore in depth many more
of these issues and their implications in practise.
Tony Roberts, 3 Apr 2009
502
(8.23)
x = X(s) ,
z = Z(s) .
To the given linear approximation: is identified with u, with v, with w,
with x and with z. Thus and will encapsulate the fast dynamics
in the transformed variables, whereas , and encapsulate the slow. In
essence, with such a transformation we seek to warp our view of the state
space to see if we can find a simpler view of the dynamics.
Since u, v, w, x and z evolve in time, then so must , , , and ; they
do so according to some prescription such as s = g(s) which written in
components is
= A(s) ,
= B(s) ,
= C(s) ,
= F(s) ,
(8.24)
= G(s) ,
where A, B and C are to be nonlinear functions of s. Note that the last
two equations incorporate the dominant behaviour of and , that of fast
oscillation, through the given linear approximation. Our aim is to choose
the transformation (8.23) so that A, B, C, F and G are as simple as possible.
Tony Roberts, 3 Apr 2009
503
then substituting
X 0 0 X 0 0 X 0 0
=
A +
B +
C
X 0
X 0
+ 1+
( + F 0 ) +
( + G 0 )
Rearranging the equality between the first line and the last line leads
to
X 0
X 0
F0 + Z0
+
= 0.
(8.25)
then substituting
Tony Roberts, 3 Apr 2009
504
Z 0
Z 0
0
+
( + F ) + 1 +
( + G 0 )
Rearranging the equality between the third and last lines leads to
G0 X0
Z 0
Z 0
+
= b .
(8.26)
Now solve the two coupled equations (8.25) and (8.26).10 Clearly a
solution keeping F 0 and G 0 as simple as possible is F 0 = G 0 = Z 0 = 0
and X 0 = b. Thus, we remove the quadratic coupling term in
the z-equation, coupling the fast to the slow dynamics, by choosing
the transformation x + X 0 = b. Of course there may well
be induced cubic or higher order coupling between the dynamics, but
such coupling is asymptotically weaker.11
Second, consider the slow modes. Similar arguments to the above
lead to the following equations for the corrections, upon substituting,
using the chain rule, and neglecting products of primed quantities or
products of primed quantities and small variables:
U 0
U 0
+
0
V
V 0
B0
+
0
W
W 0
C0
+
A0
= + b ,
=
b ,
= .
505
506
z 0,
U 0
U 0 X
+
=
fmn (, , )m n ,
m,n
where fmn is some multinomial expression in , and . Typically, nonlinear terms will generate all possible such terms. Specifically, suppose that
terms quadratic12 in and are generated so that we need to solve
A0
U 0
U 0
+
= f20 2 + f11 + f02 2 .
U
The homological operator U
+ maps quadratics in and into
quadratics, cubics into cubics, etc. Thus it is sufficient to consider a quadratic
form for U 0 in an attempt to simplify A 0 . Trying the general form U 0 =
a2 + b + c2 , the equation becomes
(8.27)
507
508
8.3
Summary
scene??
subc??
Chapter 9
Contents
9.1
9.2
One-dimensional introduction . . . . . . . . . . .
510
9.1.1
9.1.2
9.1.3
Summary . . . . . . . . . . . . . . . . . . . . . . .
519
How does the zebra get its stripes? How does the leopard get its spots?
Gently heat a thin layer of oil and see a shimmering of the surface. These
are examples of pattern forming systems. Use the SwiftHohenberg equation
as a prototypical system:
u
= (1 + 2 )2 u + ru u3 .
t
(9.1)
510
perhaps, that positive u represents hot spots in some layer of fluid, and
negative u represents cold spots.
oneDintro??
9.1
Sorry:
One-dimensional introduction
this section is incomplete as it is slowly written.
Almost all of the richness of pattern forming systems is lost in one spatial
dimension. However, we choose to start simply. the one-dimensional Swift
Hohenberg equation for a field u(x, t) is
u
= (1 + 2x )2 u + ru u3 .
t
(9.2)
See the rather complicated dissipation represented by (1+2x )2 u, the destabilising forcing ru controlled by the forcing parameter r (analogous to the
strength of heating in a convective system for example), and the nonlinear
stabilisation represented by u3 .
First we see the classic pitchfork bifurcation in a small box. Then second we
discover that the freedom in large boxes provides exciting pattern evolution.
9.1.1
511
(9.4)
has constant coefficients we expect trigonometric and exponential spatiotemporal structures. Without justifying fully we seek solutions u et cos kx
for some spatial wavenumber k and some growth-rate (eigenvalue) . Why
not sine? Because sin kx does not satisfy the boundary condition (9.3) at
x = 0. To satisfy the boundary condition (9.3) at x = 2 with cos kx we must
have k = 21 , 1, 32 , 2, . . . . To satisfy the linearised SwiftHohenberg equation
the growth-rate
= r (1 k2 )2 .
(9.5)
Thus for the allowed discrete values for the wavenumber, the growth-rate
=r
9
16
, r, r
25
16
, r 9, ...
for k = 12 , 1, 32 , 2, . . . .
(9.6)
u = a(t) cos x +
such that
a = ra 3a3 + .
512
9.1.2
Now explore the start of the complexity that arises when we look at patterns
in large boxes.
513
See that as for dynamics in a small box, all modes decay when the forcing
parameter r < 0. When the forcing parameter r crosses zero, the trivial
solution u = 0 loses stability and we must explore the nonlinear dynamics.
Nonlinear modulation
The huge difference now is that the allowed wavenumbers are not discrete,
but are continuously variable. See that for even small forcing parameter r >
0, there are an infinite number of modes near wavenumber k 1 which are
unstable.
Suppose we try to justify a slow manifold model as for bifurcations. At
the critical value of the forcing parameter r, namely zero, there are two
critical modes, eix with wavenumbers k = 1 . But there is a continuous
spectrum of modes with growth rates arbitrarily close to zero. Thus there
is no spectral gap, and the centre manifold theory currently available to us
cannot be applied.1 However, we make progress by adapting the techniques
we used to model dispersion in pipes and channels.
We introduce the idea of modulation. Recall that broadcast radio waves are
a carrier wave with frequency or amplitude modulated. The modulations
carry the information we hear. We adopt the same principle to study the
patterns: the carrier wave here is the basic spatial pattern of wavelength 2
formed from eix ; the modulation contains the information about how that
pattern evolves in the long-term due to variations from regular spacing. For
example, suppose the pattern is set up with a slightly different wavenumber
k = 1 + , where is small: then
cos[(1 + )x] =
=
1 i(1+)x
+ 12 ei(1+)x
2e
1 ix ix
e + 12 eix eix
2e
ix
ix
= a(x)e + b(x)e
1
Nonetheless, for some systems, researchers have been able to justify the results which
we obtain, namely the GinzburgLandau equation. The issue is that the proofs are specific
to the particular system under study, and not as yet general.
Tony Roberts, 1 Mar 2008
514
where a = 21 eix and b = 12 eix are modulations to the carrier waves eix .
The modulations a and b have large scale x structure, small derivatives in x,
and are slowly varying in x. All these are equivalent properties and the same
as those we used for shear dispersion.
Thus our modelling task is to pose that the solution field is
u(x, t) = a(x, t)eix + b(x, t)eix +
(9.7)
where modulations a and b vary slowly in space x; they also vary slowly in
time as this is a pitchfork-like bifurcation rather than a Hopf-like bifurcation.
We eventually find the modulations evolve according to a pde of the form
a
2 a
= ra 3a2 b + 4 2 + ,
t
x
and similarly for b. For real solutions a and b are complex conjugates, which
these equations maintain, and so the model is of the form of a Ginzburg
Landau equation
a
2 a
ra 3a|a|2 + 4 2 .
(9.8)
t
x
Now interpret this equation. As forcing r crosses zero, the trivial a =
515
(9.9)
geix + heix +
a3 ei3x 3a2 beix 3ab2 eix b3 ei3x + [aeix + beix ]2 u + O u2 .
Omit for this first approximation the products of small quantities and the
third and fourth derivatives of a and b, as these are small compared to the
dominant second derivatives, to leave
Lu + geix + heix r[aeix + beix ] + 4axx eix + 4bxx eix
a3 ei3x 3a2 beix 3ab2 eix b3 ei3x ,
2
Perhaps see Cross & Hohenberg (1993) [IV.A.1.a.(i)] to read more about this topic.
Tony Roberts, 1 Mar 2008
516
1 3 i3x
64 a e
1 3 i3x
.
64 b e
(9.11)
and
b
rb 3ab2 + 4bxx .
t
(9.12)
9.1.3
517
cold spots, with wavenumber k = 1 which is precisely the critical wavenumber. Here test its stability to perturbations of some long wavelength, small
wavenumber.
Seek solutions
r
r
ipx+t ,
+ eipx+t + e
3
r
r
eipx+t ,
b =
+ eipx+t +
3
a =
where and are the small amplitudes of the perturbations of wavenumber p. The perturbation wavenumber p must be small for the Ginzburg
Landau equation (9.8) to be valid. This form ensures that a and b are
complex conjugates and so the stability analysis applies to real fields u.
Substitute into the GinzburgLandau equation pair (9.12), omit products
of the small perturbation amplitudes, equate coefficients of like wavenumbers, and arrive at the eigenproblem
+ r + 4p2
r
= 0,
r
+ r + 4p2
and similarly for the complex conjugate components. The eigenvalues are
thus = 2r 4p2 and = 4p2 . These are always negative, as parameter
r > 0 past the bifurcation, sopthe finite amplitude mode of wavenumber
precisely k = 1 and amplitude r/3 is stable to such perturbations.
Nearby modes may also grow
But other modes are also linearly unstable and so may grow. Try a eikx+t
518
p
These modes of wavenumber k = 1+k grow to amplitude |a| = (r 4k2 )/3 .
But when they reach this finite amplitude state, is the pattern stable? Can
all these nonlinear patterns be observed? The answer is no as we see next.
Some growing modes are nonlinearly unstable
Seek solutions
r
r
ipx+t ,
+ eipx+t + e
3
r
r
ipx+t
ipx+t
ikx
e
+ e
+
,
b = e
3
ikx
a = e
where and are the small amplitudes of the perturbations of wavenumber p, but now relative to the wavenumber k = 1 + k . The wavenumbers
k and p must be small for the GinzburgLandau equation (9.8) to be valid.
This form again ensures that a and b are complex conjugates and so the
stability analysis applies to real fields u. As before, substitute into the
GinzburgLandau equation pair (9.12), omit products of the small perturbation amplitudes, equate coefficients of like wavenumbers, and arrive at the
eigenproblem
+ r 4k2 + 4p2 + 8kp
r 4k2
= 0,
r 4k2
+ r 4k2 + 4p2 8kp
The characteristic equation for nontrivial solutions is thus
2 + (2r 8k2 + 8p2 ) + 8p2 (r 12k2 + 2p2 ) = 0 .
The coefficients of 2 and are both positive, for r4k2 > 0 that we need for
linear instability. Hence the pattern can only be stable when the constant
term is positive; that is, the pattern can be stable when r > 12k2 2p2 . But
the perturbation wavenumber p is arbitrary, so forcing parameter r must be
greater than all of them, thus the pattern may be stable when r > 12k2 .
Conversely, at any given forcing parameter r, a band of wavenumbers |k| <
1
2 r, wavenumbers relative to the critical wavenumber k = 1 , may grow
Tony Roberts, 1 Mar 2008
9.2. Summary
519
9.2
Summary
oneDintro??
520
Bibliography
Abraham, R. H. & Shaw, C. D. (1983), Dynamics-the Geometry Of Behaviour. Part 1: Periodic Behaviour, Aerial Press.
Abraham, R. H. & Shaw, C. D. (1988), Dynamicsthe geometry of behaviour. Part 4: Bifurcation behaviour, Aeriel Press.
Abramowitz, M. & Stegun, I. A., eds (1965), Handbook of mathematical
functions, Dover.
Abrams, D. M. & Strogatz, S. H. (2006), Chimera states in a ring of nonlocally coupled oscillators, Int. J. of Bifurcation and Chaos 16(1), 2137.
Arneodo, A., Coullet, P. H. & Spiegel, E. A. (1985a), The dynamics of
triple convection, Geophys. Astro. Fluid Dyn. 31, 148.
Arneodo, A., Coullet, P. H., Spiegel, E. A. & Tresser, C. (1985b), Asymptotic chaos, Physica D 14, 327347.
Batchelor, G. K. (1979), An Introduction to Fluid Dynamics, CUP.
Bender, C. M. & Orszag, S. A. (1981), Advanced mathematical methods for
scientists and engineers, McGrawHill.
Bensoussan, A., Lions, J. L. & Papanicolaou, G. (1978), Asymptotic Analysis
For Periodic Structures, Vol. 5, Stud Appl Maths And Applications.
Carr, J. (1981), Applications of centre manifold theory, Vol. 35 of Applied
Math. Sci., SpringerVerlag.
521
522
Bibliography
Bibliography
523
524
Bibliography
Bibliography
525
Index
:=, 33
begin-end, 34
bye, 33
df, 33
end, 34
factorial, 33
factor, 38, 46
for, 33, 36
int, 33, 36, 46
in, 34
let, 34
linear, 34
operator, 34
quit, 33
repeat-until, 34, 45
until, 34
write, 33, 38
accuracy, 272
action, 105
adiabatic approximation, 190
adiabatic manifold, 190
adjoint, 306
Airys equation, 65
algebraic equation, 3, 13, 14, 18
assembly, 266, 272, 277
asymptotic approximation, 62
asymptotic completeness, 168
Index
dispersion, 273
Duffings ode, 7383, 94
element integral, 265, 272, 275
finite element method, 263, 271
527
Landau equation, 177, 195
Laurent series, 62
Legendres equation, 43, 44, 66
linear combination, 93
linearisation, 50
local analysis, 154
local manifolds, 350
logarithm, 51
528
Index
singular problem, 56
slow manifold, 161, 165, 171, 171,
quadratic equation, 4, 6, 8, 12
190, 206, 208, 247
quasi-stationary probability, 149
slow subspace, 161
solution, 267, 273, 278
Rayleigh number, 374
solvability condition, 194, 306, 307
RayleighBenard convection, 209
solvability conditions, 203
Reduce, 3, 9, 11, 1821, 23, 25 stable manifold, 171, 348, 349, 350,
34, 3642, 4547, 49, 53
351, 359
55, 57, 6567, 80, 81, 89
stationary probability, 149
91, 9396, 99, 100, 102
steady flow, 376
104, 184, 185, 189, 195
Stieltjes, 56, 57
197, 208, 210, 214, 216, 219,
Stieltjes integral, 5860, 68
220, 298, 300, 304, 354, 360,
Stieltjes series, 57, 59, 60, 62, 68
393, 394, 396, 416, 417, 423,
stiffness matrix, 266, 267
424, 426429, 449, 466, 474,
strain energy, 266
497, 509
stream-function, 376
regular, 15
stress, 268
regular perturbation, 22
structurally stable, 397
rescale, 18, 22
supplementary features, 268, 273
residual, 68, 1114, 16, 44, 44, 46,
SwiftHohenberg equation, 209, 507
48, 50, 52, 54, 171
510, 513515
residuals, 279
resonance, 357
Taylor series, 58, 11, 16, 35, 36, 38,
river, 273
42, 43, 46, 50, 60, 62, 6567
Rossler system, 413
testing, 50
row vector, 149
tetrahedral element, 272
Prandtl number, 375
Index
529