3. Non-equilibrium Field Theory
《非平衡态系统中的集体过程 (Collective processes in non-equilibrium systems)》是位于德累斯顿的马克思普朗克复杂物理研究所 (Max Planck Institute for the Physics of Complex Systems) Steffen Rulands 研究员的一门课程。
课程主页链接在此,网页上有课程的课件,录像发布于 YouTube。
YouTube 把视频中讲者说的话从语音转化成了文字,我把这些转录复制了下来,进行了简单的断句,并且推测了各段文字对应的课件的内容。
The slide 8 of examples was talked about at the end of the lecture rather than between neighboring slides.
random
(00:00)
so i i have to set this sound to you so
(00:02)
you hear them as well so you’re using
(00:04)
headphones okay right that goes
(00:08)
okay so then let’s start uh so welcome
(00:11)
uh to our third lecture in collective
(00:14)
processes and this will be
(00:17)
the final lecture that is more
(00:18)
methodology methodological
(00:21)
and technical so today we’ll talk about
(00:24)
a field theory representations of
(00:28)
the kind of processes that we’ve been
(00:29)
looking at last time
(00:32)
let me just share the screen
(00:42)
okay
slide 1
(00:49)
here we go so if you remember the
(00:52)
lecture so you can see that
(00:54)
great uh so you remember if you remember
(00:56)
lecture that we had last time
(00:59)
uh i gave a little introduction to the
(01:02)
mathematics that is
(01:04)
behind the description of the stochastic
(01:07)
processes
(01:09)
and we discussed two kinds
(01:12)
of stochastic processes alternatives not
(01:15)
two kinds of stochastic processes but
(01:17)
two ways
(01:18)
of describing sarcastic processes the
(01:21)
first way
(01:22)
that was associated with the name
(01:24)
nonzero
(01:25)
that we already featured in the very
(01:26)
first lecture
(01:28)
relied on deriving
(01:32)
a stochastic differential equation that
(01:34)
describes
(01:36)
the time evolution of a single
(01:38)
realization
(01:39)
of such a stochastic process and the
(01:41)
alternative approach that einstein
(01:44)
applied for the description of broad
(01:46)
emotion
(01:47)
was to derive an equation
(01:52)
for the time evolution of the
(01:54)
probability density
(01:56)
itself yeah and that was the master
(01:58)
equation this master equation is not a
(02:00)
stochastic
(02:01)
differential equation is a deterministic
(02:04)
equation
(02:05)
but it’s typically high dimensional and
(02:09)
relies as you saw last time on some
(02:13)
integrations over the possible states
(02:16)
that you can jump into
(02:18)
and then for those of you who remained
(02:21)
for the example
(02:23)
section last time you would have seen
(02:26)
that these master equations although
(02:29)
they look
(02:30)
pretty complicated can be derived
(02:32)
phenologically
(02:34)
phenologically actually
(02:37)
in most cases in a very simple way yeah
(02:41)
so these are the two kinds of
(02:42)
description and why are we not
(02:44)
completely happy with that
(02:48)
so there exists approximately
(02:52)
suppose both in general conditions both
(02:54)
of these different kinds of equations
(02:55)
cannot be solved
(02:57)
there exists approximative methods that
(03:00)
we discussed last time for example the
(03:01)
focal plunk equation
(03:03)
that was an approximation for the master
(03:06)
equation
(03:07)
and these appropriate approximative
(03:09)
methods that work in certain special
(03:11)
cases
(03:13)
now the focal planck equation as you
(03:15)
remember the
(03:16)
derivation relied on making strong
(03:19)
assumptions
(03:20)
on how these jumps in states state space
(03:24)
look like they’re very nicely behave
(03:26)
these jumps are very small
(03:28)
and that they effectively rely on very
(03:32)
large
(03:32)
system size and
(03:36)
in other cases there are no
(03:39)
approximative
(03:40)
methods at all so what field theory
(03:44)
does for us is it gives us a flexible
(03:47)
framework a general
(03:48)
and flexible framework that allows us to
(03:51)
describe a large class of stochastic
(03:54)
systems
(03:55)
and it is even
(03:59)
maybe if you remember from statistical
(04:01)
physics or quantum mechanics
(04:03)
it’s uh so general that you can apply it
(04:06)
to
(04:06)
a very diverse set of systems that’s
(04:10)
one thing is for example spatially
(04:12)
extended systems
(04:13)
yes you can see on the bottom uh right
(04:16)
yeah so this is a system where
(04:20)
the evolution over the spatial
(04:22)
information
(04:23)
itself is very important not because you
(04:26)
see here there’s a chemical
(04:28)
system you see that here structures form
(04:31)
it’s an example of what’s called binodal
(04:33)
decomposition
(04:35)
now so here the spatial component is
(04:36)
very important
(04:38)
and this has to be taken into account
(04:41)
when you want to understand of course
(04:43)
collective processes
(04:45)
another example are non-linearities in
(04:48)
stochastic differential equations
(04:51)
now these become especially if they
(04:53)
affect an oyster and become pretty hard
(04:55)
very quickly
(04:56)
and another example is that i would have
(04:58)
showed you last time
(05:00)
that field theory that typical
(05:02)
approximative methods like the focal
(05:04)
planck
(05:06)
method or other kinds of expansions
(05:10)
uh are not very suitable for rare events
(05:14)
for the tales of probability
(05:16)
distributions
(05:17)
so in fields here we have a general
(05:19)
framework
(05:20)
of understanding these variables
(05:24)
and so there has been a lot of work on
(05:26)
how to make approximations to these
(05:28)
field theories
(05:30)
that allow us to understand the tales of
(05:32)
probability distributions and very often
(05:34)
these tales
(05:36)
are pretty important so they are not
(05:39)
they’re rare
(05:40)
but uh if you think about a nuclear
(05:44)
reactor or something like this yes these
(05:46)
rare events are rare
(05:47)
but you want to know how often they hear
(05:49)
and there are a few theoretic methods
(05:51)
that allow you
(05:52)
to calculate the tails of probability
(05:54)
distributions
(05:56)
very nicely so
(06:00)
this is what we’ll do and in this
(06:02)
lecture
slide 2
(06:03)
i want to show you how we can derive
(06:07)
a field theory description um
(06:11)
for the longitudinal equation for
(06:13)
stochastic differential equations
(06:15)
we’ll be looking at very simple a
(06:20)
very simple automatic equation that
(06:22)
doesn’t have any
(06:23)
explicit time derivative that doesn’t
(06:25)
have
(06:26)
multiplicative noise and we can
(06:30)
nevertheless in the framework of field
(06:31)
theory we can
(06:33)
straightforwardly uh extend
(06:36)
this methodology to more complicated
(06:38)
systems so we restrict ourselves
(06:40)
to this very simple larger equation uh
(06:43)
with gaussian white noise
(06:45)
that is as last time uncorrelated
(06:50)
so if you
(06:54)
so the first step here is to discretize
(06:58)
time yeah and uh so we take
(07:02)
the logical equation and we write it
(07:03)
down in discrete time
(07:06)
and uh as you saw uh last time
(07:09)
uh or in the first of what was the first
(07:12)
lecture already i saw
(07:13)
two lectures ago oh wait whatever yeah
(07:16)
so so no last time last time was the
(07:18)
catholic processes
(07:19)
yeah so as you saw last time the way we
(07:22)
discretize
(07:24)
stochastic differential equations and i
(07:27)
showed you that for uh stochastic
(07:28)
integrals is very important
(07:32)
yeah and in this case we also have to
(07:34)
discretize
(07:35)
our stochastic differential equation in
(07:37)
a certain way
(07:39)
which is called the ito discretization
(07:43)
and but if we do that if we discretize
(07:47)
our stochastic differential equation in
(07:49)
this way
(07:50)
the idea is that in principle we can
(07:53)
write down
(07:55)
averages over any
(07:58)
observable you know that’s here on the
(08:01)
left left hand side by some complicated
(08:04)
thing that is on the right hand side
(08:07)
so this what is on the right hand side
(08:09)
and this equation looks pretty
(08:10)
complicated but it’s not that
(08:12)
complicated so let’s have a look
(08:14)
at these different terms in the first
(08:18)
term here
(08:19)
on the left right hand side the red term
(08:22)
we just integrate over all possible
(08:25)
realizations
(08:27)
that a stochastic trajectory can take
(08:32)
on the right hand side
(08:35)
now if you look at the right hand side
(08:37)
oh there’s a delta function missing here
(08:40)
right at here delta here
(08:43)
on the right hand side you have this
(08:45)
delta function
(08:49)
here you have this delta function and
(08:52)
this delta function
(08:54)
just makes sure that whatever we
(08:56)
integrate here whatever trajectories we
(08:59)
integrate over
(09:01)
that they fulfill the discretized
(09:04)
version
(09:05)
of the larger equation
(09:08)
yeah so the delta function is just the
(09:11)
left-hand side
(09:12)
minus the right-hand side of this launch
(09:15)
of our equation
(09:16)
and if the left-hand side is equal to
(09:18)
the right-hand side
(09:20)
then we take that trajectory into
(09:22)
account
(09:23)
you know and then sandwiched between
(09:26)
that we have
(09:27)
our observable o that is some functional
(09:31)
of our trajectory x
(09:34)
now as i say functional so i don’t know
(09:37)
how much uh
(09:37)
functional analysis all of you had so
(09:40)
most
(09:41)
so i would expect that most of you would
(09:44)
have that
(09:44)
until the first fourth term
(09:48)
or so also but just just to make sure a
(09:51)
functional
(09:52)
is basically a function that maps um
(09:56)
that maps this that maps
(09:59)
this function x uh yes that map
(10:03)
a function to a real number yeah
(10:06)
and this is our how we define these
(10:08)
observables
(10:09)
now you take a trajectory and you map it
(10:12)
to them to a number
(10:16)
okay so this looks very complicated it
(10:18)
doesn’t help us
(10:19)
anything at all and uh
(10:22)
one the other step that we need to make
(10:24)
on the slide is we that we
(10:26)
introduce some notations here
(10:29)
and uh of course we don’t always want to
(10:32)
write
(10:33)
all of these integrals here on the left
(10:36)
hand side
(10:37)
we just say we write that in this
(10:39)
functional form here
(10:41)
and of course many of you will know that
(10:43)
this is just
(10:44)
a functional integral or a path into
(10:47)
yeah that’s how we define it here
slide 3
(10:49)
and now for uh notational convenience
(10:54)
now we just go to a continuum locate
(10:56)
notation now we forget that we
(10:58)
discretize time in the previous step
(11:01)
and we write the equation back in
(11:03)
continuous form
(11:05)
just for the sake of simplicity and we
(11:07)
define
(11:08)
some delta functional
(11:11)
that is just equal to the product over
(11:14)
all data functions that we had on the
(11:16)
previous slide
(11:17)
now if you look here are the product of
(11:20)
our data functions
(11:21)
just make sure that you fulfill the
(11:23)
larger equation really at each time
(11:25)
point
(11:27)
and we just define like a super delta
(11:29)
function or delta functional
(11:31)
that makes sure that we really satisfy
(11:33)
the larger equation for each time point
(11:41)
so now we make a little trick
(11:45)
now we say we have this delta function
(11:47)
or this product of delta functions
(11:50)
and what we say is that in fourier space
(11:54)
the delta function is represented by a
(11:58)
plane
(11:58)
wave yeah so we fully transform the
(12:02)
delta function
(12:15)
or a functional is now and
(12:18)
the fourier transform of the delta
(12:20)
function
(12:21)
just becomes delta x dot
(12:25)
minus f of x minus
(12:29)
sine because our delta our function
(12:31)
equation
(12:33)
is equal to a fury space
(12:38)
d x tilde
(12:43)
e to the minus i x to the
(12:48)
x dot minus
(12:52)
f of x minus c
(12:56)
now we now get this variable x tilde
(13:00)
so this is nothing there’s nothing
(13:01)
happening it’s just the definition of
(13:03)
the forage in the form
(13:04)
of a delta function and because we have
(13:08)
this
(13:08)
product for many delta functions here
(13:12)
we have the integral here that the path
(13:15)
integral here will also get a path
(13:17)
integral
(13:18)
delta x total you know that’s that just
(13:22)
the definition of the fury transform
(13:24)
and now we plug this
(13:28)
in again so
(13:32)
we obtain from that
(13:35)
that the expectation value of our
(13:39)
observable yeah
(13:43)
taken to the average now this average is
(13:45)
over the distance
(13:47)
over over a different voice realizations
(13:52)
yeah it is equal now we plug that in
(13:55)
an integral now over
(13:58)
x and x tilde so it’s a path integral
(14:03)
over x and x tilde so we there’s an
(14:06)
integral
(14:07)
over all realization of x and all
(14:10)
realization
(14:11)
of x2 now so now this x still that pops
(14:14)
up yeah so we don’t really know what it
(14:15)
is
(14:16)
but it will hang around and it will show
(14:18)
you later what it actually mean
(14:21)
means so we have this fourth integral
(14:25)
so our observable of x and then
(14:33)
plugging in e to the minus
(14:37)
i x tilde
(14:40)
um sorry
(14:44)
we’ve got an integral so we have here a
(14:46)
little
(14:47)
integral dt
(14:52)
and this integral
(14:56)
we get because we had this
(15:00)
product over here now so here the
(15:03)
product
(15:04)
we had here gives us an integral so our
(15:07)
sum
(15:08)
in the exponential so this was
(15:10)
originally like a product of many
(15:12)
exponential
(15:14)
and uh so this is this and then we have
(15:18)
our
(15:18)
x tilde x dot minus
(15:23)
f of x minus x psi
(15:28)
and then we close the average
(15:32)
and then we just plug this in yeah we
(15:34)
can move out
(15:36)
everything that does not depend on psi
(15:39)
or the noise
(15:41)
out of the average because average is
(15:43)
over the noise
(15:46)
b x x to the
(15:51)
power of x right now comes the stuff
(15:54)
that does not depend
(15:56)
on x under x i integral
(16:00)
e t x to the
(16:04)
x dot minus f of x
(16:08)
and now we have some
(16:12)
average over minus i
(16:15)
dt x tilde
(16:19)
times psi
slide 4
(16:23)
you know so
(16:26)
now the question is what is this here
(16:32)
can we calculate this at the moment we
(16:34)
cannot do anything with this equation
(16:36)
with this expression can you can we
(16:38)
calculate
(16:40)
this last term that involves a noise
(16:44)
average
(16:45)
over e to the minus dt
(16:49)
and psi now and there’s hope that we can
(16:52)
calculate this because we know what
(16:54)
psi is now we said that x i
(16:58)
is a gaussian random variable
(17:01)
yeah it has follows a normal
(17:03)
distribution
(17:04)
it’s uncorrelated so we know a lot of
(17:07)
things about this
(17:08)
sign and what we do right now
(17:12)
right because psi is gaussian there’s
(17:14)
also hope that we can actually
(17:17)
solve or integrate this integral that is
(17:20)
this average here
(17:21)
now i’ll show you now how this works in
(17:24)
detail
(17:25)
yeah so we make use of the definition of
(17:28)
phi
(17:30)
so let’s see what this second average
(17:32)
looks like
(17:34)
now that’s o of x
(17:38)
uh sorry it’s not o of x it is
(17:42)
e to the minus i dt
(17:47)
integral x of
(17:50)
sine
(17:54)
now and this is
(17:58)
by definition by the definition of the
(18:00)
average
(18:03)
is equal to the integral over dxi
(18:08)
times the probability over the
(18:11)
probability distribution of
(18:13)
psi which is a gaussian or a normal
(18:15)
distribution
(18:18)
2 pi a a is the strength of our noise
(18:23)
e to the minus psi
(18:26)
squared over 2a
(18:29)
yeah so this is the probability
(18:31)
distribution
(18:33)
and we know that psi is normally
(18:36)
distributed
(18:37)
yeah and that’s why we have this normal
(18:40)
distribution here
(18:42)
at now we multiply this by the thing
(18:45)
we’re averaging over
(18:48)
yeah e to the minus i
(18:51)
d t x to the sine
(18:57)
now so this excuse me yes isn’t it
(19:00)
supposed to be exponential
(19:02)
plus psi integral delta t whatever
(19:08)
integral over let me see
(19:12)
you mean the second integral this one
(19:13)
here
(19:16)
yeah okay so this one is that minus or
(19:18)
plus
(19:20)
let me just check
(19:23)
let me check my notes
(19:27)
okay
(19:31)
so here we go
(19:35)
so no there’s a minus
(19:39)
i i wouldn’t be able to find him i
(19:42)
wouldn’t be able to find it
(19:43)
[Music]
(19:45)
on the previous slide you can go and
(19:47)
just
(19:48)
okay so let me see maybe there’s a here
(19:51)
we have the minus
(19:54)
here we have the minus u of the minus
(19:57)
here we have the minus and another minus
(20:00)
yeah you’re right
(20:01)
let me see what’s wrong here uh
(20:04)
okay oh yes okay so here
(20:12)
i think this minus here is not right
(20:17)
now that’s just the definition here wait
(20:20)
minus my i don’t know
(20:22)
okay minus minus is plus
(20:28)
minus minus with plus i think i think
(20:29)
you’re right
(20:31)
but then here should be a minus so where
(20:33)
does the
(20:34)
is there an i squared somewhere
(20:44)
excuse me i think there should be a
(20:46)
minus sign
slide 4
(20:47)
um in the first exponential as well in
(20:50)
the last step on this particular slide
(20:52)
there are two exponential studies yes
(20:54)
yes yes yes yes
(20:56)
that’s what i’m wondering about um let
(20:58)
me see if it’s
(20:59)
actually the final result
(21:02)
um i x minus minus
(21:08)
[Music]
(21:14)
okay so so mike
(21:17)
let me let me just see
(21:24)
let me just see
(21:28)
so let’s let’s put this minus here in
(21:30)
brackets
(21:31)
now so this one minus this should have
(21:34)
the opposite
(21:35)
i don’t wait with me okay this is the
(21:37)
minus then
(21:38)
this should be plus you say
(21:42)
you know i think i think it i think it
(21:45)
will come out correctly later
(21:46)
now let’s see how it goes um let’s see
(21:50)
how it goes
(21:51)
now it makes sense
(21:56)
now we have to have a gaussian integral
(22:00)
and that will determine whether what
(22:04)
we’re doing is right
(22:05)
okay so let’s let’s go on so this here
(22:08)
is the integral and now we just
(22:12)
put everything together and say
(22:15)
that the sign 1 over
(22:20)
2 a e
(22:24)
minus dt x to the
(22:28)
psi
(22:32)
now that’s the first one so that would
(22:34)
be a plus then
(22:36)
minus i dt x
(22:40)
to the sine
(22:43)
now i just i just uh i just combined the
(22:46)
exponentials
(22:47)
now thanks for thanks for paying so much
(22:49)
attention
(22:51)
um so yeah and this is here
(22:54)
a gaussian integral and if you don’t
(22:57)
know remember
(22:58)
the gaussian integrals and you can look
(23:00)
at the bottom here
(23:01)
these gaussian intervals are pretty easy
(23:04)
to solve
(23:05)
probably most of you have heard of that
(23:08)
and we can also solve
(23:10)
this gaussian integral here and what we
(23:13)
get
(23:13)
is that this is just e to the a over 2
(23:18)
dt x to the squared
(23:23)
yeah because we know yeah and here
(23:27)
one thing you have to make uh you need
(23:29)
to remember
(23:30)
is that this here has eyes in it
(23:33)
now if you look at this formula at the
(23:35)
bottom you need to take into account
(23:36)
that these are complex
(23:38)
integrals so we can calculate
(23:42)
this quantity here
(23:45)
we can calculate this quant this
(23:46)
quantity here because we know
(23:49)
how psi looks like and that we integrate
(23:52)
over the distribution of
(23:53)
sine of the noise organization and we
(23:56)
get
(23:56)
the term that we have here yeah
(23:59)
and now we already have arrived at the
(24:02)
famous
(24:03)
no at least famous for a very small
(24:05)
number of people
slide 5
(24:06)
and this is the so-called martin citra
(24:09)
rose johnson they dominicus
(24:12)
uh functional integral yeah that’s what
(24:14)
you see here
(24:16)
in the red box now we just now put
(24:18)
everything together
(24:19)
here we have our noise here we have
(24:23)
uh what we had before and here
(24:28)
is so to say that it’s a deterministic
(24:30)
part that came from the larger equation
(24:33)
yeah and uh so what this tells us now
(24:37)
here
(24:38)
and maybe it looks familiar to some of
(24:40)
you yet quantum field theory
(24:42)
this looks very familiar so what we do
(24:45)
now is
(24:46)
we want to calculate the average over
(24:49)
some observable
(24:52)
we integrate over all possible
(24:56)
realizations and over all possible
(24:58)
realizations of some weird quantity
(25:00)
x tilde so we got a second field here
(25:05)
and weight the contributions of
(25:08)
different trajectories
(25:11)
by this exponential factor here
(25:14)
now and this exponential factor looks
(25:16)
very much like what you know from
(25:19)
other field theories like quantum field
(25:21)
theory and this is why this is very
(25:23)
often called
(25:24)
an action
(25:28)
now that’s the martin sutra rose or ms
(25:31)
rjd functional
(25:34)
integral that allows us to really write
(25:37)
a field theory
(25:39)
for stochastic processes
(25:42)
now so here our axes are not fields yet
(25:45)
now they’re trajectories they don’t have
(25:47)
a space component now they don’t have
(25:49)
spatial dimensions
(25:51)
but as you can see later the structure
(25:53)
of a real spatial
(25:55)
the the integrals they will look very
(25:58)
similar
(26:00)
now we can also rewrite this a little
(26:04)
bit here and look at specific
(26:05)
trajectories and then we can look for
(26:09)
example at a special case
(26:11)
where it is observable is just
(26:14)
the propagator here or this will
(26:16)
propagate as the probability
(26:18)
that we end up at some state x at a time
(26:22)
t
(26:23)
if we start it add some x naught
(26:27)
at a time t naught you know and we
(26:30)
obtain this
(26:31)
not by just requiring that this
(26:33)
observable is a delta function
(26:36)
where uh we say that the x the specific
(26:39)
time
(26:40)
t at a time t needs to be equal
(26:43)
to the x that we give to the probability
(26:46)
distribution here
(26:50)
and then we plug that in and
(26:53)
what we now need to say is that we only
(26:56)
integrate
(26:57)
over trajectories that actually started
(27:00)
it’s not an end
(27:01)
and x at a given times now that’s that’s
(27:04)
what we have to take to account for in
(27:06)
the boundaries and the bounds of the
(27:07)
integrals
(27:08)
yeah and then we get this form here
(27:13)
now that looks very similar and that’s a
(27:16)
different representation if you remember
(27:18)
last time we had the
(27:19)
koi maguro of uh chap and komogorov
(27:21)
equation
(27:22)
it was an equation for the same quantity
(27:25)
and the idea was a little bit similar
(27:27)
in this technical mcgovern
(27:30)
equation we had the same spirit
(27:33)
and we looked at different sums of
(27:35)
different paths
(27:37)
um a process could take to go from x
(27:40)
naught to x
(27:41)
and here we do the same thing in a more
(27:43)
fancy way
(27:49)
just to reflect on this a little bit
(27:51)
more
(27:52)
so what happened now so we started
(27:55)
here with
(27:59)
a longer equation we discretized it
(28:03)
and if you remember uh quantum fields
(28:06)
theory that’s also the step that you do
(28:08)
there
(28:08)
you discretize time into small intervals
(28:12)
as the first step if you derive
(28:13)
a quantum fluid theory and then
(28:17)
we wrote expectation values of some
(28:19)
observables
(28:20)
formally in a way that involved
(28:23)
integrals of
(28:23)
all possible trajectories and the
(28:25)
resultant trajectories
(28:27)
filtered four trajectories that solve
(28:30)
the launch of a creation
(28:32)
now that was to say self uh
(28:35)
circular starting point and in the next
(28:39)
uh step we then got this field
(28:42)
side tilde now that we still don’t know
(28:44)
what it is exactly about
(28:46)
now we got that from the fourier
(28:47)
transform of the delta function
(28:52)
and now now we have two fields we
(28:54)
integrate over two fields
(28:56)
x and x x tilde and
(28:59)
in the end we managed to integrate out
(29:03)
the noise so what we have here
(29:07)
now is something that does not depend on
(29:09)
the sign anymore
(29:11)
it’s a deterministic
(29:14)
equation and deterministic integral so
(29:17)
somehow
(29:18)
our noise was observed observed absorbed
(29:23)
into a new field a new fluctuating field
(29:28)
x tilde yeah and that’s how
(29:31)
it very often goes now that you
(29:34)
make a field theory and what you gain is
(29:37)
you get a nice nice integral but you
(29:39)
have to pay for
(29:40)
it by having additional fields conjugate
(29:43)
fields
(29:44)
that you have to integrate over and the
(29:47)
same is true here
(29:48)
and i’ll show you later what these x
(29:50)
tilde actually mean before i do that let
slide 6
(29:55)
me just mention so so now we have a few
(29:56)
theory i have a path integral and these
(29:59)
path integrals are very useful
(30:01)
because we can make use of a lot of
(30:03)
tools
(30:04)
from other field fields from quantum
(30:06)
field theory
(30:08)
renormalization perturbation theory and
(30:10)
so we can
(30:11)
make use of these tools very powerful
(30:14)
frameworks developed in the last 70
(30:18)
years or so
(30:20)
we can make use of these frameworks and
(30:21)
apply them to these
(30:23)
few theories for stochastic processes
(30:27)
and one of the things that we can do is
(30:29)
we can
(30:30)
define a so-called generating functional
(30:35)
that’s maybe something that you already
(30:36)
know from other field theories
(30:38)
so what you do is you add some auxiliary
(30:42)
fields
(30:43)
external fields h and h
(30:46)
tilde and these fields cuddle
(30:51)
to x and x tilde the accelerations
(30:55)
respectively
(30:57)
and what is this is then the generating
(31:00)
function
(31:01)
and of course we know that these fields
(31:04)
don’t really exist
(31:05)
now we just added them and the reason
(31:08)
why we added them
(31:10)
is that if we take derivatives
(31:13)
with respect to these fields h or this
(31:16)
external
(31:17)
forces or external fields h here
(31:21)
and here what will happen is that each
(31:24)
time
(31:25)
because this is an in exponential
(31:28)
the x
(31:31)
let or just draw that now the x
(31:36)
or the x tilde
(31:40)
will go down here
(31:44)
yeah and if we do that if we take these
(31:46)
derivatives
(31:48)
you know so we can therefore get the
(31:52)
expectation values of combinations of
(31:56)
the x
(31:56)
and x tildes just by differentiating
(31:59)
this
(32:01)
generating functional with respect to
(32:04)
these
(32:05)
weird virtual fields
(32:08)
yeah and when we’ve done that we have to
(32:12)
remove these fields again so we have to
(32:14)
set them back to zero
(32:15)
now for example if you want to have the
(32:17)
correlation the autocorrelation function
(32:19)
so how much
(32:21)
is the process at a time t correlated to
(32:24)
the state at a time t
(32:25)
prime yeah then we
(32:28)
differentiate and we
(32:32)
take the derivative first with respect
(32:34)
to
(32:35)
with respect to h at a certain time t
(32:39)
yeah and then we get one of these axes
(32:42)
here and then we take the derivative
(32:44)
with uh to h at
(32:48)
a time a different time t prime and then
(32:50)
we get the field again
(32:52)
at a different time here
(32:55)
yeah and these are of course functional
(32:57)
derivatives now that was
(32:59)
uh functional calculus if you haven’t
(33:01)
done that
(33:02)
it’s uh for these what you what you do
(33:04)
is you
(33:05)
look at the change of a functional
(33:08)
now for example this here is a
(33:10)
functional
(33:12)
we look at the change of a functional
(33:15)
with respect to small
(33:16)
changes of its argument also you have a
(33:20)
look at
(33:20)
perturbations in age and h tilde
(33:24)
around some value yeah and then you
(33:28)
see how your function changes that’s
(33:30)
called a functional derivative
(33:32)
and if you do that you get very
(33:34)
conveniently these pre-factors here
(33:39)
right here in front of the action and if
(33:42)
you look at
(33:43)
the definition here this is just what
(33:46)
gives us our observable
(33:48)
as for example if our observable is x
(33:52)
that we just take the derivative once
(33:55)
we get an x here yeah
(33:58)
and if we have the x here we have the
(34:01)
first moment so the mean
(34:03)
of x the average of x if we take the
(34:06)
derivative twice
(34:07)
at different times then we get a
(34:09)
correlation here on the left-hand side
(34:13)
so this is a very convenient tool and as
(34:15)
i said
(34:16)
if you want to have a correlation
(34:18)
function for example we take the
(34:19)
derivative
(34:20)
twice and you must always remember to
(34:23)
set
(34:23)
these fields to zero again it’s actually
(34:26)
the same approach as you do in the
(34:28)
quantity
(34:29)
and classical field theory the
(34:31)
equilibrium equilibrium field t
(34:37)
okay so now what is this x
(34:40)
of t that’s just a remark so i’m not
(34:42)
doing the calculations like
slide 7
(34:44)
what is this x tilde of t that we get
(34:48)
got in this process
(34:52)
there are different uh
(34:55)
c theories for stochastic processes and
(34:57)
for master equations
(34:58)
and you always get some kind of
(35:01)
auxiliary
(35:02)
field some some conjugate field that you
(35:05)
have to pay for
(35:07)
and uh this x tilde from here from here
(35:10)
you can
(35:12)
get an intuition about that i’m not
(35:14)
super rigorous but you can get
(35:16)
your intuition about this if you write
(35:18)
down the master equation
(35:20)
sorry the larger equation with respect
(35:24)
and add some external source
(35:27)
capital h of t
(35:32)
you know and if you add this external
(35:34)
force capital h
(35:36)
of t now some temporally fluctuating
(35:39)
force
(35:40)
that doesn’t depend on x itself
(35:43)
and you plug that in into this martin
(35:46)
sergio rose
(35:47)
functional integral or martin citra ruse
(35:50)
johnson did you limit it to minikit
(35:54)
then you see a formal analogy that if
(35:57)
you
(35:57)
that this you get a term that looks like
(36:00)
h tilde
(36:04)
if you define this to be minus i this
(36:07)
external field
(36:09)
yeah and now we can see what happens
(36:14)
to x what is the effect of this external
(36:17)
field
(36:19)
h of t on x so there’s a little bit in
(36:22)
an
(36:23)
analogy already here now so here on the
(36:25)
left hand side it has something to do
(36:27)
with this uh external field from the
(36:31)
generating functional that couples to x
(36:33)
tilde
(36:34)
now let’s see what this does to x to the
(36:37)
actual stochastic process
(36:39)
now to this end we calculate a response
(36:43)
function
(36:43)
and this response function is just the
(36:47)
change
(36:48)
in the average of f x with respect
(36:52)
to changes in this external field
(36:55)
that’s called the response so how does
(36:56)
the system respond
(36:58)
to changes in this external field
(37:01)
yeah and so as you remember the average
(37:05)
here we just get by
(37:08)
integrating uh by by taking this for
(37:12)
this this generating functional
(37:15)
and taking the derivative with respect
(37:17)
to h
(37:18)
once so that’s the first moment
(37:22)
and because of this equality here
(37:27)
now we see that this year
(37:30)
this response function is we also get
(37:33)
that
(37:33)
if we take the derivative with respect
(37:36)
to h
(37:37)
and then h tilde at a certain time
(37:41)
t let me see let me just say
(37:45)
here that’s the tilde
(37:51)
yeah and this here what this is
(37:59)
as on the last slide is the correlation
(38:02)
between
(38:02)
x of t and x tilde of t
(38:06)
now somehow the response of the system
(38:10)
with respect to an external force
(38:13)
or an infinitely miserable external
(38:16)
force
(38:17)
is given by how the x total
(38:21)
couples to x
(38:24)
so it describes the activity or is
(38:27)
related
(38:28)
to an infinitesimal response
(38:32)
of x of the field x with respect
(38:35)
to a small perturbation and that’s why
(38:39)
this field x tilde is also called the
(38:41)
response field
(38:44)
now there’s just a little bit of
(38:46)
intuition and very often this these
(38:48)
conjugate variables somehow in some way
(38:51)
encode the noise
(38:56)
now i have an example uh let me just
(38:59)
check the time whether we do it now or
(39:01)
at the end of the lecture
(39:05)
let’s do it at the end of the lecture
(39:06)
again not this example for those of you
(39:09)
or who
(39:10)
heard a few theory course last year uh
(39:13)
there were no
(39:14)
examples so i’ll leave that to the to
(39:16)
the end of the lecture that
(39:17)
people can dial out uh if they’re tired
slide 9
(39:24)
now i want to just give you a second
(39:26)
remark
(39:28)
so we are now in the framework of
(39:29)
physicians
(39:31)
and in field theory now if you remember
(39:34)
we can have different formulations of
(39:36)
field theories otherwise the lagrangian
(39:39)
field theory and the hamiltonian field
(39:42)
and we can translate these two into each
(39:45)
other
(39:45)
and we can do the same things here for
(39:49)
the non-equilibrium for the stochastic
(39:51)
process
(39:53)
yeah and it’s also just a remark no it’s
(39:56)
not
(39:57)
so important for the rest of the lecture
(40:00)
but you can formally define
(40:03)
new variables q and
(40:07)
p by these relations here
(40:10)
and then this probability will take the
(40:14)
form
(40:14)
that i wrote down here now so this
(40:18)
has formally the form of a hamiltonian
(40:22)
action oh or having a hamiltonian theory
(40:25)
so we have
(40:26)
p times q dot minus
(40:29)
some hamiltonian and this hamiltonian
(40:32)
is given by this term here this looks a
(40:35)
little bit like a kinetic energy so it’s
(40:38)
just
(40:38)
just just saying that we can write by
(40:41)
variable transformation we can write
(40:42)
these
(40:43)
field theories uh then quite analogously
(40:47)
to feed theories that we already know
(40:51)
and we can even go one step further
(40:55)
now we can take this here and integrate
(40:58)
out the piece
(40:59)
now because it’s just gradually
(41:01)
quadratic in p
(41:03)
so these are just essentially gaussian
(41:05)
integrals that you can integrate over
(41:06)
them
(41:07)
and just to tell you the results that we
(41:09)
then get
(41:10)
a field theory that looks like a
(41:13)
lagrangian field theory
(41:16)
now we have a lagrangian that depends on
(41:17)
q and the
(41:19)
derivative of q with respect to time
(41:22)
and if you then look at the analogy
(41:26)
at this hammer at this land range here
(41:28)
then
(41:29)
this looks like it describes some kind
(41:31)
of particle
(41:32)
with that has some mass one over a uh
(41:35)
that lays
(41:36)
like in a potential that happens that’s
(41:38)
coupled to some
(41:39)
q dot here and uh we have now here
(41:43)
the quadratic potential of laughter now
(41:46)
so now these analogies they’re not very
(41:48)
helpful yeah they don’t tell you
(41:49)
anything
(41:50)
uh these hamiltonians that you get here
(41:53)
uh
(41:53)
they’re not comparable to hamiltonians
(41:55)
that you get important systems
(41:57)
for example these hamiltonians they’re
(42:01)
not emission quantities
(42:05)
for quantum physicists
(42:10)
and just to say that they’re different
(42:11)
ways of formulating these
(42:13)
theories that you get by variable
(42:16)
transformations
(42:17)
uh this second equation here has a name
(42:19)
that’s the own sagar
(42:21)
mac look functional you know and like
(42:24)
you will
(42:25)
pop up these these these kind of
(42:27)
functional
(42:28)
um pop up in papers if you read papers
(42:31)
that
(42:32)
they they pop up in different ways but
(42:34)
in the end
(42:35)
the same uh um
(42:40)
implementations as a few implementations
(42:43)
of the same
(42:44)
stochastic differential equation but
(42:46)
just reformulations of the same thing
slide 10
(42:52)
okay now
(42:55)
i told you in the beginning of the
(42:57)
lecture that um
(43:01)
i actually what most of the cases
(43:03)
interested in
(43:04)
now and what field theories are also
(43:07)
good for
(43:08)
are spatially extended systems
(43:13)
now how do you write down a larger
(43:16)
equation or a spatially extended system
(43:19)
now it gets of course a little bit more
(43:20)
complicated but there’s actually a
(43:22)
classification
(43:23)
theme for a long general equation that
(43:27)
describes systems
(43:28)
that have spatial degrees of freedom
(43:32)
and this classification scheme you can
(43:34)
see on the slide
(43:36)
and so we’re looking here at some time
(43:40)
evolution of some field phi
(43:43)
at a position x at a time t
(43:47)
and this time evolution is given by
(43:49)
different components
(43:51)
yeah that’s r on the right hand sides on
(43:54)
the very right we have the noise
(43:56)
as always yeah and this noise
(44:01)
now has the usual properties the average
(44:04)
of the noise is zero
(44:06)
and it also has correlations so how is
(44:09)
a fluctuation a fluctuating force of
(44:12)
that xi represents how is that
(44:15)
correlated between different time points
(44:18)
and how they correlated between
(44:20)
different positions
(44:21)
in space so while the assumption that
(44:24)
different
(44:25)
correlations that this noise is
(44:28)
uncorrelated in time so that
(44:30)
you don’t have memory is something that
(44:32)
is
(44:33)
very reasonable and that we that we
(44:35)
typically make
(44:37)
it’s not so clear that the noise is also
(44:40)
independent at each position in space
(44:44)
you know so generally there will be some
(44:47)
some
(44:48)
length you perturb the system and then
(44:49)
you don’t perturb a single atom
(44:51)
yeah but you deter like a small region
(44:54)
for example
(44:55)
and then these fluctuating forces are
(44:58)
correlated
(45:00)
over small regions and these
(45:02)
correlations in the spatial
(45:06)
uh in these spatial systems and the
(45:09)
fluctuating force that act on these
(45:11)
systems
(45:13)
are given by so-called spatial
(45:15)
correlation
(45:16)
now that just says how are these
(45:19)
fluctuating forces
(45:21)
how are they correlated between two
(45:24)
given positions
(45:25)
in space yeah and
(45:28)
now we have the rest here
(45:32)
the black parts here
(45:35)
yeah it’s just a fancy way of writing
(45:38)
down
(45:39)
the deterministic part of what’s
(45:42)
actually happening
(45:44)
yeah and it’s just a fancy way of
(45:47)
writing down uh the actual
(45:50)
uh for example chemical reactions
(45:53)
that change the value of the field at a
(45:56)
given position
(45:58)
yeah and we can formally write down this
(46:01)
uh this term by saying okay so we have
(46:04)
some kind of potential
(46:06)
and the dynamics will go so to some
(46:09)
minimum
(46:10)
of this potential not just some some
(46:13)
equilibrium state
(46:15)
yeah and then uh we can formally write
(46:18)
this
(46:18)
this way here that we have some
(46:21)
functions and free energy functional
(46:23)
f that depends on the fields and this is
(46:26)
given by some input
(46:28)
space integral uh where we have this
(46:31)
term here that will flatten
(46:33)
the field that something like can become
(46:36)
a diffusion term
(46:37)
and on the right hand side we have
(46:40)
something
(46:41)
that is a potential that describes
(46:43)
what’s actually which where we’re going
(46:44)
to with the field
(46:46)
now if you have to have heard
(46:48)
statistical physics
(46:50)
yeah then uh these will look familiar to
(46:53)
them
(46:53)
to you like fight to the force theory
(46:55)
ginsburg londo and so on
(46:58)
now there’s this red stuff here yeah and
(47:01)
this red stuff is just a way
(47:03)
of classifying different kinds of
(47:06)
systems
(47:07)
and people distinguish between system
(47:10)
where the order parameter so our phi
(47:13)
is not conserved for example in chemical
(47:16)
systems where you can
(47:17)
convert one chemical species to another
(47:20)
chemical species
(47:21)
yeah and then another chemical species
(47:23)
and so on then the con
(47:25)
concentration of one chemical species is
(47:28)
not
(47:28)
conserved in the entire system
(47:32)
yeah and this is what you get if you set
(47:34)
this exponent to zero
(47:36)
that means you don’t have this laplacian
(47:38)
the second derivative here
(47:42)
the other case is when you have
(47:45)
set this to one here set this n to one
(47:49)
then this will here be part of
(47:52)
something like a diffusion term yeah
(47:55)
this will go into a diffusion term
(47:57)
and what you uh will then get is uh what
(48:00)
is
(48:00)
what what these kind of systems that
(48:02)
describe are
(48:04)
situations where the field supply is
(48:06)
conserved
(48:07)
now so you remove stuff at one point of
(48:09)
the system if you remove stuff at one
(48:11)
point of the system
(48:12)
it has to pop up in another point
(48:16)
an example is for example uh if you have
(48:18)
something like hydrodynamics also where
(48:20)
you just
(48:20)
move mars around and uh but it doesn’t
(48:24)
really disappear
(48:25)
and you just move things around but
(48:27)
things don’t disappear
(48:28)
that’s an example for these kind of
(48:30)
model b
(48:32)
systems yeah and
(48:35)
uh so if you plug these things in so
(48:37)
typical examples very famous example
(48:39)
systems also called reaction diffusion
(48:43)
systems
(48:44)
now so and the typical reaction
(48:45)
diffusion equation
(48:47)
is seen here on the right hand side that
(48:50)
describes a situation
(48:52)
where uh the phi if you look here at the
(48:55)
phi
(48:56)
now then what happens what this term
(48:59)
here does
(49:00)
is a 5 is just a little bit larger than
(49:03)
0
(49:04)
yeah then this term will be positive
(49:07)
yeah and this this
(49:10)
this will give to rise an increase in
(49:12)
the field
(49:14)
now this term is a diffusion term uh
(49:17)
that just
(49:17)
transports information to the set
(49:19)
between different positions
(49:21)
in the system and then we have our noise
(49:24)
and this noise typically has
(49:26)
prefactors it’s multiplicative uh if you
(49:29)
look for example at biological systems
slide 11
(49:34)
so at the take home message for these
(49:36)
spatial systems
(49:38)
now we you can think about okay we just
(49:40)
add an x variable well what i just said
(49:42)
is just a complicated way of
(49:44)
saying okay so we have some x variable
(49:47)
and some diffusion yeah but everything
(49:50)
else will look very similar as before
(49:53)
now and indeed we can follow exactly the
(49:56)
same steps now
(49:57)
now we take the long-term equation
(49:59)
general launch of my equation here
(50:02)
you know which is this one i do exactly
(50:05)
the same steps as before
(50:07)
and now this here would be the first
(50:10)
step
(50:13)
now where we have this delta functions
(50:16)
now now we don’t only have a product
(50:18)
over i
(50:19)
that is already in the over time that is
(50:21)
already in this delta function here
(50:23)
but also over x yeah and then we just
(50:26)
plug in this
(50:27)
generalized launch of our equation and
(50:29)
do the same step
(50:31)
and then we get the field theory the
(50:33)
martin citra rose functional
(50:35)
for the spatially extended system so
(50:38)
this looks pretty complicated
(50:40)
but it is actually the same as before
(50:43)
you know so we have we integrate again
(50:46)
over the field fine and the response
(50:50)
field
(50:50)
file tilde and
(50:54)
then we sum up different contributions
(50:59)
to our variable to our observable like
(51:02)
the second term
(51:03)
and then as before we wait that
(51:06)
with some exponential that essentially
(51:10)
quantifies
(51:10)
how far we are away from a realistic
(51:14)
realization of the launching equation
(51:17)
yeah and here we have
(51:19)
exactly the same structure as before
(51:22)
here you have the launch of the equation
(51:25)
now you see you say that you need to
(51:27)
solve the deterministic part of the
(51:29)
moisture equation
(51:31)
and this was previously just
(51:34)
x tilde squared i said previously
(51:40)
this was this term here
(51:43)
was something like a
(51:46)
over 2 x to the squared
(51:50)
well now that looks more complicated
(51:52)
yeah but it has the same form as if you
(51:54)
look at this
(51:56)
you have here your x tilde or phi tilde
(52:00)
here you have another one that just
(52:03)
coupled by the correlations in the noise
(52:07)
uh that is described by this voice
(52:10)
kernel
(52:11)
now but the structure is the same as
(52:14)
before
(52:15)
yeah and uh so so uh so this is the
(52:18)
martial citra rose
(52:20)
functional for the noise for the
(52:22)
spatially extended system
(52:25)
yeah and with this uh i’m done with the
(52:29)
definitions
(52:29)
and with the actual
(52:33)
derivation of the field theory now
slide 8
(52:36)
let’s see if we can apply it and now i
(52:39)
go back
(52:40)
to the example and just see how we’re
(52:43)
doing in time
(52:47)
okay great i think almost exactly an
(52:49)
hour
(52:50)
so if you don’t already know all of this
(52:52)
maybe from last year’s lecture
(52:53)
then uh feel free to feel free for her
(52:56)
to drop out
(52:56)
and for the rest i’ll go through one
(52:58)
example
(53:00)
and i’ll double check so if i upload
(53:02)
when i upload the uh
(53:04)
the lecture notes i’ll try to make sure
(53:06)
that actually these minus signs
(53:09)
are correct yeah so so if there was a
(53:11)
mistake i’ll correct it
(53:12)
in the uploaded version that you find on
(53:15)
the website
(53:17)
excuse me yes so
(53:20)
when you wrote the response function
(53:23)
there was this i guess heavyside
(53:28)
function multiplied with
(53:29)
your autocorrelation so
(53:33)
does that come up because uh you have
(53:36)
a kind of the etho formalism
(53:39)
into your system the point
(53:44)
the response function if you have a
(53:46)
perturbation here
(53:48)
if you perturb at t prime and you look
(53:50)
at the response at t
(53:52)
now that’s how this response function is
(53:55)
defined
(53:56)
answer you should have here that’s a d
(53:58)
prime
(53:59)
here as you perturb at t prime and you
(54:02)
look
(54:03)
at the time t yeah then
(54:06)
oh the other way around actually the
(54:08)
answer so then then this
(54:12)
herbicide function just ensures
(54:15)
causality
(54:16)
that you cannot observe the response
(54:19)
that hap that happens before you could
(54:20)
do the perturbation before you apply the
(54:22)
force
(54:25)
now so that’s that’s that’s the reason
(54:27)
why you get these tata functions
(54:30)
okay okay thanks for the question
(54:35)
let’s go then to the
(54:42)
first example
(54:46)
here we go yeah and
(54:49)
uh so let’s let’s have a look at the
(54:52)
fields here for a very
(54:54)
simple example and all like in the
(54:56)
stochastic processes
(54:58)
also simple examples quickly become
(55:04)
complicated
(55:06)
so let me find the notes
(55:11)
here we go
(55:18)
okay we start with a simple
(55:22)
with probably one of the simplest larger
(55:24)
way equation uh you can
(55:26)
imagine and uh because it’s so simple
(55:29)
uh it has a name because it has been
(55:32)
extensively
(55:33)
studied and it’s called the einstein
(55:35)
ruling
(55:36)
process and this process is just given
(55:40)
by the time
(55:41)
by the time derivative of x
(55:45)
dot and this is equal to
(55:48)
some restoring force all right so plus
(55:51)
sometimes
(55:52)
uh some external fluctuating force
(55:56)
side and as always we have the usual
(55:59)
conditions that are outside
(56:00)
very nicely behaved doesn’t have a mean
(56:04)
and it’s uncorrelated in time
(56:08)
so now writing down the margin citra
(56:12)
rose
(56:13)
function integral easy because we just
(56:16)
have to plug in this laundromate
(56:18)
equation
(56:19)
so martin
(56:23)
martin cedar rose johnson the dominicus
(56:28)
some people say martin said rose i did
(56:30)
that once and i happened to be in munich
(56:32)
and then i they were very angry that i
(56:34)
forgot johnson
(56:36)
my parents apparently had some
(56:37)
connections to munich
(56:40)
and uh now so that’s the full
(56:43)
that that that contains i think
(56:45)
everybody who contributed
(56:47)
uh
(56:52)
reads now so what is the action now just
(56:56)
i don’t write the full integral just
(56:57)
write the action
(56:59)
so the action is
(57:03)
dt i x tilde
(57:12)
del t x plus alpha x
(57:19)
now let’s make sure that we fulfill the
(57:21)
login equation
(57:23)
plus a over 2
(57:27)
dt
(57:31)
x tilde now that would be squared
(57:37)
now we can write down the generating
(57:39)
function
(57:41)
generating
(57:47)
functional yeah and this is just as
(57:50)
before
(57:52)
that of h is tilde
(57:57)
is equal to the integral
(58:00)
over our two fields function interval of
(58:02)
the word two fields
(58:04)
field response field uh
(58:07)
e to the minus s
(58:11)
plus dt
(58:15)
h x so that the external auxiliary field
(58:19)
that comes to x
(58:21)
plus an external field that couples
(58:24)
to x tilde
(58:28)
yeah and uh just
(58:31)
just to make sure everybody understands
(58:33)
now so why do i say that this field
(58:35)
couples to x
(58:36)
and x tilde here so these and why are
(58:38)
these external fields
(58:40)
provided like this this just follows if
(58:42)
you write down
(58:43)
um something like for example the
(58:45)
ginsburg landlord theory also
(58:48)
or you look at the icing model now then
(58:50)
terms like this here
(58:55)
will tilt the potential in one direction
(58:59)
yeah and here this tilts the potential
(59:04)
in the x-direction and this tills the
(59:06)
potential in the axillary
(59:07)
direction this tilts it in the x
(59:10)
direction
(59:12)
and this tilts it in the x tilde
(59:14)
direction
(59:16)
now so this is why we call these
(59:17)
external fields
(59:19)
and that they couple to uh
(59:22)
these fields x and x that’s the analogy
(59:26)
to uh the ginsberg london theory and
(59:29)
other fields
(59:32)
okay so let’s go this is the
(59:35)
um functional generating functional
(59:40)
and now we use that
(59:44)
integral over dx e to the iq x
(59:48)
is just delta of q so we use the
(59:52)
definition of the fourier transform of
(59:54)
the delta function
(59:56)
and by doing that we get that
(60:00)
this functional generative function
(60:04)
is equal to d of
(60:07)
x to the e
(60:11)
to the a over two
(60:16)
integral dt x to the squared
(60:22)
plus dt
(60:26)
our two fields h x
(60:29)
plus h tilde x
(60:33)
yeah and now we’ve got our delta
(60:37)
function
(60:39)
i del t minus alpha
(60:44)
x to the plus h
(60:53)
now when is this delta function here the
(60:56)
delta function
(60:58)
actually non-zero now so the delta
(61:02)
delta functional
(61:08)
is non zero
(61:11)
if our x tilde
(61:15)
solves this ordinary differential
(61:18)
equation
(61:20)
you know that we have here in the delta
(61:22)
function now we can just write it down
(61:26)
i times t to infinity
(61:30)
dt prime e to the minus
(61:35)
alpha t minus p prime
(61:39)
h of t prime yeah
(61:45)
now we substitute that so we saw that we
(61:48)
already got rid of
(61:49)
one field here by doing this reverse
(61:53)
fury transform now we substitute that we
(61:56)
get
(61:57)
rid of our second field now we
(61:59)
substitute
(62:05)
into that
(62:11)
and what we get
(62:14)
is age h tilde
(62:17)
is equal to and now comes a large
(62:20)
exponential
(62:21)
to see how to write that on the screen
(62:26)
minus integral dt
(62:30)
integral dt prime the second it comes
(62:34)
from this x tilde
(62:35)
and then we have
(62:39)
a over 2 e to the minus
(62:42)
alpha t minus t prime
(62:48)
times h of t
(62:53)
h of t prime plus
(62:57)
i theta
(63:00)
t minus t prime
(63:03)
e to the minus alpha t
(63:07)
minus t prime
(63:10)
times i’m sorry i said it like
(63:14)
even for the simplest process it gets
(63:16)
lengthy
(63:17)
h of t h of t prime
(63:21)
yeah but what’s happening here is
(63:24)
nothing magical it’s just a calculation
(63:26)
such
(63:26)
integrals well we substituted that x
(63:29)
total
(63:31)
into our generating functional and then
(63:34)
we collected the terms
(63:35)
and made sure
(63:40)
that the right time order
(63:43)
is given and now we have this generating
(63:46)
function it looks a little bit
(63:47)
complicated
(63:48)
but we can deal with that now so we can
(63:51)
take
(63:52)
functional derivatives with this now for
(63:54)
example to get a correlation function
(64:02)
and so we have x
(64:06)
of the x of t prime
(64:10)
now we take the second
(64:14)
derivative
(64:19)
once at time t
(64:24)
and once at time t prime
(64:30)
and then we must not forget
(64:33)
to set the fields equal to zero again
(64:40)
now and if you look at this equation if
(64:43)
we do that
(64:44)
then and do a little bit of calculations
(64:48)
don’t
(64:48)
get that actually this correlation
(64:50)
function is
(64:51)
a over alpha e to the minus alpha
(64:56)
t minus t prime
(65:02)
so that’s just uh an example yeah so
(65:05)
these actual calculations
(65:06)
are a little bit messy but that’s not
(65:09)
just how you use these field theories
(65:12)
that you have the field theory you write
(65:14)
down the general functional
(65:16)
yeah and then you look that you are able
(65:19)
to
(65:19)
take these functional derivatives with
(65:22)
respect to these external fields
(65:24)
and the rest is then say
(65:28)
mathematics yeah
(65:32)
okay great so this was an example and
(65:35)
from next week
(65:35)
we’ll be a little bit more intuitive
(65:37)
again now we’ve covered the technical
(65:39)
stuff
(65:40)
uh and we can look into some real uh
(65:43)
physics
(65:44)
and physics problems okay see you all
(65:47)
next week
(65:48)
bye