slide 1
00:00
there we go
00:07
okay so now you can see the screen i
00:09
hope
00:10
with a little overview so what where are
00:13
we actually
00:14
uh at the moment yeah so we had uh
00:18
two lectures ago we started thinking
00:20
about how order can emerge
00:22
you know and we said that this somehow
00:25
um very often uh relies on a balance
00:29
between
00:30
fluctuations you know that favor
00:31
disorder and
00:33
interactions that favor order
00:37
and then we went on next week last week
00:40
to study how we can have transitions
00:44
between different non-equilibrium states
00:48
for example between a
00:49
disordered and an order state or between
00:52
different kinds of ordered states
00:54
and we started that in a situation where
00:56
we neglected
00:58
moles we pretended that our
01:00
non-equilibrium system
01:01
or our general system could also be an
01:04
equilibrium system is very very large
01:06
and then we were basically back in the
01:08
framework
01:09
of non-linear differential equations and
01:12
non-linear partial differential
01:14
equations
01:15
and to understand then what kind of
01:17
order we had
01:19
we looked at specific situations here
01:21
where this order
01:23
or these patterns arose continuously
01:26
or not abruptly but they first were
01:28
small and then we can larger larger
01:31
and in these situations were allowed and
01:33
to linearize
01:34
and to ignore these difficult
01:38
non-linearities in these equations
01:42
so today we're now in a state where we
01:44
want to
01:46
join these two lectures now we're
01:49
looking at transitions between
01:51
non-equilibrium
01:53
steady states where
01:57
we have noise yeah and i wouldn't tell
02:01
you that i wouldn't have a lecture on
02:02
this if
02:03
the noise these fluctuations in these
02:06
transitions weren't super important
02:09
and maybe yeah then
02:12
once we've understood that we are in a
02:15
position
02:16
uh just to understand how does actually
02:19
order look like
02:20
how can we identify order and after that
02:23
we'll then
02:23
go to data and actually learn some tools
02:26
from data science
02:27
on how to extract such order
02:31
from complex and big data sets
02:34
and then at the end of this lecture of
02:37
this uh this term
02:38
uh we'll have a specific example from
02:40
research where we bring that all
02:42
together and we see
02:43
how this works together uh in the
02:45
context of
02:46
current research
slide 2
02:50
okay so to start
02:54
let's remind ourselves about
02:57
how we actually transit from order to
03:00
from disorder to order in equilibrium
03:03
yeah and we'll continue we'll
03:06
continue looking at continuous phase
03:09
transition and
03:10
in equilibrium these continuous phase
03:12
positions are characterized
03:14
by a critical point and this is just the
03:16
point where this balance
03:18
between fluctuations and interactions
03:21
this ordering and these disordering
03:24
forces
03:25
are just equal now the system doesn't
03:27
really know
03:28
exactly where to go whether to create an
03:30
ordered state or to a completely
03:32
disintegrated
03:33
order state and it's somewhere in
03:34
between and
03:36
uh so i i suppose
03:39
uh you've you you'll have had that in
03:42
your
03:42
statistical physics lecture normally
03:44
yeah but at the end of the first
03:46
statistical physics lectures but i'll
03:49
just give you
03:51
a little reminder of the most important
03:54
concepts
03:55
so here you see like this example from
03:58
equilibrium
03:59
this very powerful and intuitive model
04:02
system which is called the izing model
04:04
which is essentially modeling a
04:05
ferromagnet magnet
04:07
and on the left hand side you can see
04:10
simulations
04:12
for three different temperatures it's
04:15
just
04:16
such an icing model and what you see
04:18
here if the temperature is very low
04:21
you get an ordered state now everything
04:23
is black
04:24
all spins are putting the same
04:26
directions
04:27
in the same direction and if the
04:29
temperature is high
04:32
then you get a disordered state and this
04:35
disorder state as you see as a kind of
04:37
salt and pepper state now the average
04:40
magnetization is zero
04:43
but and locally uh you will always find
04:46
spin that goes up and down
04:48
go up and down and then you have that
04:50
thing in between these states
04:52
yeah when the temperature is exactly
04:56
equal to a critical temperature and this
04:59
is where
04:59
there is a balance between energy and
05:02
entropy of
05:03
fluctuations and order and
05:06
what you see here is this state
05:09
where you have these domains here
05:11
domains of these black domains
05:14
and if you zoom in to such a system
05:18
what you'll see at the critical point
05:19
you'll see that it looks exactly the
05:21
same
05:21
now you zoom in and you would wouldn't
05:24
be able to say whether this is a
05:25
snapshot a zoomed in version of the one
05:27
on the left hand side
05:29
or whether this is an entirely different
05:32
simulation
05:34
and then you can zoom in further and
05:35
further and
05:37
you again all the time get the same
05:39
impression that you can't really make
05:41
out
05:42
uh what is now the typical length how
05:44
large are these classes
05:46
you have here these domains of all sizes
05:49
equally now because you have domains
05:53
these white things or black things of
05:55
all
05:56
size is equally represented you can't
05:59
make out a single sound you can't make a
06:01
typical
06:02
length scale here because you can't say
06:04
that
06:06
the typical size of such a cluster here
06:10
is for example that large
06:13
as you always find clusters that are
06:15
much smaller you find glasses that are
06:17
much
06:17
larger here's a cluster now that has the
06:20
size of the entire system that's
06:22
infinitely large
06:23
and then you have all a spectrum of
06:26
sizes in between that
06:28
yeah well here in this example you kind
06:31
of get an idea
06:33
that these clusters are typically very
06:35
small
06:36
these domains whereas pin points up or
06:38
so are very small they have a
06:40
typical size but you can't see it on the
06:42
left hand side
06:44
yeah same as if i would zoom in with
06:46
this camera you know so
06:49
if we do like this now you immediately
06:52
see
06:53
that i zoomed in now because i have a
06:56
characteristic size i'm
06:57
one meter 80 or something yeah and now
07:00
you
07:00
see that you that have zoomed in the
07:02
camera and the picture is not the same
07:04
as before
07:06
so i'm not in a critical state yeah but
07:09
the icing system is in a critical state
07:12
and this critical state is characterized
07:15
by self-similarity
07:17
so they have structures of all sizes
07:19
represented
07:21
in this system now this is the
07:23
self-similarity
07:24
and the mathematical representation of
07:27
the self-similarity
07:28
the fact that you don't have an average
07:30
cluster size
07:32
is that the correlation length diverges
07:35
now so the correlation length is
07:37
infinity
07:38
that means the correlation function you
07:41
know
07:41
so how that represents how large
07:45
these clusters are is scale in variance
07:48
that means
07:49
that really classes of different sizes
07:51
are equally represented
07:53
and such if you ask what is the
07:55
probability that i
07:56
am now in a white cluster that i'm in a
07:59
black cluster a certain distance away
08:01
then the answer to this doesn't depend
08:03
on any specific distance
08:05
you know it's the same this probability
08:07
is the same this
08:08
correlation function here it's the same
08:11
when we
08:12
calculated in the original version of
08:15
the simulation or a zoomed in
08:17
uh fraction of this this is the
08:19
self-similarity
08:21
the self-similarity at a critical point
08:23
goes along with power laws
08:26
you know so the power laws have the um
08:30
for example like this here the critical
08:31
length as you go the critical
08:34
correlation length as you go closer to
08:37
the
08:37
critical point diverges to infinity it
08:40
goes to infinity
08:42
and it does so with an exponent that's
08:44
to be
08:45
called new when this gets zero here
08:49
this term will be infinity and these
08:51
exponents
08:52
capture how fast you go to infinity
08:56
and these exponents are very the fact
08:58
that you have an exponent but
08:59
that you have such a power rule means
09:02
that yourself similar you have a power
09:04
law if you have
09:06
something like this you can zoom in and
09:08
you still have the same exponent here
09:10
and you can't do that with an
09:11
exponential function or so
09:14
now and it also tells you that this is
09:17
here
09:18
uh if you have power laws from some
09:20
distribution that goes over
09:22
that has a power law that has this long
09:25
tail some exponent you cannot typically
09:29
calculate averages or moments because
09:31
these integrals diverge
09:34
now you have the power law of the
09:35
correlation function and you have a
09:38
power law
09:38
now if you have a power of the
09:39
correlation function you also have the
09:41
power laws
09:42
for example in the density and the
09:44
magnetization
09:46
uh near the critical point and all kinds
09:48
of other thermodynamic
09:50
quantities and i just wanted to briefly
slide 3
09:53
show you why this is the case
09:55
now it's actually the background of this
09:58
the background of this is actually an
09:59
assumption
10:01
that you say you have a free energy
10:05
and with this free energy uh
10:08
this free energy as you go to the
10:10
critical pound point uh
10:11
gets in finite it has singularity
10:15
and then you say that you assume
10:19
that the free energy here
10:23
this free energy has
10:26
one part that has all the physics and
10:29
all the details a regular part
10:32
but you say that the part of the free
10:33
energy that
10:35
diverges at a critical point
10:39
this one here so now it's is a
10:44
t t
10:46
is something like t minus
10:50
t c over t you know and h
10:53
is the external field if you have
10:56
something like this as a free energy as
10:58
the function of these parameters
11:00
yeah then the singular part the one that
11:03
goes to infinity
11:05
is a homogeneous function
11:08
homogeneous function just uh tells you
11:12
that if you have f of
11:15
lambda x that this is equal to
11:18
lambda to the power of some alpha
11:22
f on x yeah and this represents that
11:25
just this these gains zooming in you
11:28
have the same function you zoom in where
11:29
you rescale your variable
11:31
you zoom in and you get the same
11:33
function back
11:34
now this is a homogeneous function and
11:37
you still assume that this free energy
11:39
density in this case is a homogeneous
11:42
function
11:43
and this homogeneous function can only
11:45
depend
11:46
on dimensionless quantities now for
11:49
example
11:50
you wouldn't expect this divergence to
11:53
infinity
11:54
to depend on how you measure a length
11:58
now whether you measure the length in
12:01
units of centimeters or meters
12:04
whether you measure temperature in
12:07
kelvin or in units of one kelvin or two
12:10
kelvins or so
12:12
so you can see you these kind of things
12:15
these units
12:15
dimensions should be irrelevant for how
12:18
this quantity goes to infinity
12:22
yeah and if we say that then we say okay
12:24
so we have
12:26
here so-called scaling function that
12:29
depends on dimensional parameter
12:31
a dimensional dimensionless
12:34
combinations of our parameters so the
12:37
external field
12:38
divided to the temperature and then we
12:40
have to
12:41
take the temperature to some power of
12:43
something
12:45
to make everything dimensionless
12:48
so that it has no units no and there's
12:51
something that's not
12:52
something we don't know yeah and
12:56
this has the free energy has some units
12:59
therefore the whole thing gets doesn't
13:01
have the units you need a pre-factor
13:04
that gives you the right units you know
13:06
and then
13:08
you have this alpha which we don't know
13:11
yeah
13:11
this is some exponent that depends on
13:13
the specific model
13:14
for example for the eisenmann zero
13:17
you know and then you have these uh this
13:20
is the
13:21
consequence of how you translate this
13:23
homogeneity
13:25
here of the free energy to something
13:28
that you give names
13:29
as you have here this part that diverges
13:32
yeah that has the units
13:34
and this part here is the so-called
13:37
scaling function
13:38
that only depends on dimensionless
13:40
parameters
13:41
and it turns out that these exponents
13:43
and this scaling function are universal
13:45
so if you know it for one model then you
13:47
know it will know it
13:49
you will know it for a very large class
13:52
and we'll see that once we do
13:53
renormalization later today
13:57
so uh so what does it mean yeah so if we
13:59
make this
14:00
assumption that's really an assumption
14:02
about
14:04
homogeneity of the free energy then we
14:07
can calculate for example
14:09
the magnetization m of th
14:12
now this is in thermodynamics something
14:14
like
14:16
del f 2 del h
14:19
you know and then we just plug this in
14:22
and we get something like temperature
14:25
this reduced temperature
14:27
t to the power of minus 2 minus alpha
14:30
means
14:30
minus delta some other function that we
14:34
don't know
14:35
that again depends on a dimensionless
14:40
parameter and then
14:44
this scales like some better that's the
14:47
definition
14:48
of this exponent better of the
14:50
magnetization
14:52
yeah and uh so in thermodynamics this is
14:55
called
14:56
rhythm scaling basically in any textbook
14:59
on statistical physics
15:00
and it's just just to show you how the
15:04
assumption
15:05
of homogeneity near the critical point
15:09
leads to power laws in other
15:11
thermodynamic quantities
15:13
i've shown you here the magnetization
15:17
this was the magnetization
15:23
but the same holds true for example for
15:26
susceptibility
15:27
and other thermodynamic quantities that
15:29
you can get by taking derivatives of
15:31
your
15:32
energy so
15:35
this homogeneity or this self-similarity
15:39
that i showed you here that is a
15:41
reflection that is one of the hallmarks
15:44
of uh critical behavior and that's what
15:47
we're looking for when we look for
15:49
critical behavior
15:50
and now the question is can we see
15:53
something like this
15:55
also a non-equilibrium system
slide 4
15:58
before before i start with that
16:01
uh let's just have a look at one
16:02
specific how this scaling
16:04
behaves if you look at these equations
16:08
here
16:10
what does it mean it means that
16:13
actually the curves that you get you
16:16
know so if you just divide so once you
16:19
make a measurement for example with a
16:21
known temperature
16:23
and a known magnetic field you measure
16:26
this function here
16:28
then you know that it doesn't depend
16:29
separately
16:31
on the age and the temperature
16:35
so you can rescale your axis so this is
16:37
what is your y
16:38
x axis you can use that your x axis and
16:41
your y
16:42
axis to make all of these curves
16:46
collapse onto each other yeah and this
16:49
is this uh
16:50
scaling form that we see in equilibrium
16:54
physics this is for the icing model
16:56
so on the x-axis we have this scaled
16:58
temperature
16:59
that would be t on the previous slide
17:02
lowercase t
17:04
uh times something so this is v scale
17:08
and then these uh people in these
17:11
experiments
17:12
for uh for
17:15
for a ferro ferromagnet measured
17:18
the magnetization for different values
17:20
of the
17:22
of different experimental parameter
17:24
values
17:25
like magnetic field external magnetic
17:27
field temperature
17:30
and by making use of this formula here
17:33
you see that uh the scaling
17:37
behavior what is where is my
17:40
is it going down here yeah that's the
17:42
scaling behavior here
17:45
yeah that the only thing
17:49
that you don't know is this g of m that
17:52
you have to measure
17:53
yeah once you know the h and the
17:56
temperature
17:58
you can make all of these different g of
18:00
m's the gms
18:02
this scaling function you can rescale
18:05
these axes
18:06
to make them collapse onto each other
18:09
yeah and this is this observation this
18:11
is how you observe
18:12
scaling an experiment so you manage to
18:15
collapse
18:16
your experimental curves by multiplying
18:19
this x-axis and the y-axis with certain
18:23
values
18:23
of offense you have to guess you can
18:27
collapse all of these curves on the same
18:30
uh universal so-called scaling form
18:34
now this is the manifestation of scaling
18:36
and that's of course
18:37
also something we'll be looking at and
18:39
non-equilibrium systems
18:41
but also in data
18:44
now scaling is a whole mark of critical
18:48
behavior and today
slide 5
18:52
we want to see whether these concepts
18:56
of scaling and criticality yeah
18:59
where and these these continuous
19:02
phase transitions actually also extend
19:05
to non-equilibrium systems
19:07
yeah and it turns out so now we first we
19:10
need to find a non-equilibrium system
19:12
that is as intuitive
19:15
as the ising model and the icing model
19:18
is very intuitive
19:19
you have that in your lectures when
19:21
you're a student and
19:22
in your later life as a scientist you
19:24
always refer to that because it's so
19:26
simple and intuitive that uh you can
19:28
explain a lot of things a lot of
19:30
things about continuous phase
19:32
transitions in equilibrium
19:33
just based on this very simple model
19:35
like i did in the beginning of this
19:37
lecture
19:38
and it turns out now that the uh
19:41
icing model of non-equilibrium physics
19:44
of course is
19:46
that so people would dig a disagree but
19:48
one of the simplest models in
19:49
mono-equilibrium physics that shows
19:52
critical behavior
19:53
is an epidemic model and this epidemic
19:56
model we knew already
19:57
from the previous lectures here we have
20:02
our good old si model again
20:06
now so this epidemic model is this is
20:08
the simplest model
20:10
that you can think about so you have
20:11
infected individuals
20:14
i and susceptible individuals or
20:17
healthier people's less
20:19
now if an i meets an s
20:22
then the s gets infected with the rate
20:25
say lambda half
20:27
and turns into an affected infidel and
20:30
in the end you have two of them
20:33
then you have the other process that we
20:34
recover
20:36
and we set this rate to one now we can
20:38
just set that to one
20:40
without any loss of generality and uh
20:43
so that infected we measure units
20:46
time in units of this recovery rate
20:50
you know service infected individuals
20:52
can also
20:54
then recover and become
20:57
healthy again we have these two kinds of
21:00
individuals
21:01
and now we put them in the real world so
21:04
last time we were only looking at some
21:06
well-mixed
21:07
average quantities but now we put them
21:10
into the real world like the city of
21:11
brisbane also
21:13
where they actually can where actually
21:16
space
21:17
matters yeah so i'm more likely to
21:19
infect somebody else working at the
21:22
mp rpk pks than somebody looking at
21:25
another max planck institute for example
21:28
yeah so
21:29
so here uh these spatial structures
21:32
these special degrees of freedom
21:34
uh are taken into account and the
21:36
simplest way of you
21:38
how you can think about this is at the
21:40
bottom here
21:42
now that you look at letters
21:45
you have a letters each site
21:48
carries either an affected individual or
21:51
a recovered individual and
21:55
you know an infected or a recovered
21:56
individual and
21:58
when an infected individual
22:02
is next to a recovered a healthy one
22:05
then
22:05
the healthy one can turn into an
22:07
infected one
22:09
with a certain probability of with a
22:10
certain rate lambda over two
22:13
yeah and also there's another process
22:16
here if i saw if the
22:18
individual on a certain position is
22:21
infected
22:21
it can turn into a healthy one at a rate
22:24
lambda
22:26
so this is this simple spatial version
22:29
that you can think about for this and
22:32
simple epidemic model and it's also the
22:35
literature is often called the contact
22:38
process
22:40
so of course real epidemic model models
22:43
have
22:44
typically one more component namely the
22:47
um
22:49
[Music]
22:51
the infected recovered uh wait
22:54
is this also here okay so so the third
22:57
component that you normally have
22:59
and these models are the recovered one
23:01
the immune people
23:02
now you have the disease yeah and then
23:04
you are fine for the rest of your life
23:06
and you're immune to this disease
23:08
so you can only forget in fact one then
23:11
you have a third
23:12
species here a third kinds of particles
23:15
which would be the recovered ones
23:17
or the immune ones and they cannot be a
23:21
faculty again
23:22
but this slightly more complicated model
23:25
uh is shows very similar behavior to the
23:28
model that we're studying here
23:31
for the things that we're interested in
23:32
so here we're interested in infinities
23:34
in singularities so once you
23:38
once you look at these kind of things
23:40
then these models will qualitatively the
23:42
same
23:43
although also the exponents will be
23:45
different
23:47
but once you look of course into
23:49
non-singularities it's more critical
23:51
behavior
23:52
than the messiness of how wide your
23:55
roads are
23:57
how often the tram goes uh between the
24:00
blasphemy institutes and so on these
24:02
things will matter
24:05
yeah but close to the critical point uh
24:07
we'll be fine
slide 6
24:10
so this is a stochastic simulation of
24:13
such a system
24:14
and we can just see what happens on the
24:17
left hand side
24:18
you see a simulation of such a lattice
24:20
system
24:21
where you initially have random random
24:25
random initialization so every site
24:28
is either with the probability of one
24:30
half
24:32
a certain probability infected or
24:35
not infected and what you see here
24:39
now is in blue infected
24:42
individuals now if this lambda
24:46
now this infection rate is smaller
24:49
than a certain critical value
24:52
then what you will see is that this
24:54
infraction
24:55
this infection can spread for a while
24:58
but most of the time with a certain
24:59
probability
25:00
it will uh disappear
25:04
now so in this regime here in this phase
25:08
the recovery rate outweighs
25:11
the infection rate you know and
25:14
uh so that's what we're supposed to be
25:17
on in this regime we're supposed to be
25:19
investing
25:19
starting next week and then on the right
25:23
hand side
25:24
that's the regime that we're currently
25:26
in now then the infection rate
25:28
is larger than the uh
25:31
than the recovery rate now so the
25:33
infection probability is higher
25:36
and what you will then end up is is a
25:39
state where most of the individuals will
25:42
carry the disease will be infected
25:45
so you will you will reach a steady
25:47
state
25:49
not everybody is all the time in fact
25:50
that you will reach some steady state
25:52
with a certain percentage of infected
25:55
people
25:57
and now we have this situation in
26:00
between
26:02
that's this one here and this is
26:05
where the um where
26:08
the infection rate is more or less
26:11
balanced
26:12
with the recovery rate it's not exactly
26:15
equal to one
26:16
so that's these things are complicated
26:18
yeah you might think that
26:20
okay if this this lambda is equal to one
26:23
or one half or so
26:25
yeah then uh then that's the critical
26:27
point of these systems i'll show you are
26:29
more complicated than you might think
26:32
because the noise is so important here
26:35
and here
26:36
what you see is that you have domains
26:39
that become larger and larger over time
26:41
so from
26:42
when we go from top to bottom we have
26:43
time now so we go we start here at the
26:46
top
26:47
and then this domain goes large and
26:49
larger you have merging
26:51
of domains with infected individuals
26:54
and then we have branches that they die
26:56
out
26:57
like this one here and uh
27:00
it looks a little bit like a
27:02
self-similar state
27:04
now we have domains of all sizes for
27:07
example
27:08
in the time domain or from top to bottom
27:11
you have some branches that die out
27:13
quite quickly here
27:15
but then you have other branches like
27:17
the big one in the middle
27:19
where that just go on for uh forever
27:23
without really occupying the whole
27:24
system
27:26
and then if you look take a slice in
27:28
this direction here
27:30
these simulations are very small also
27:33
it's not like a design
27:34
you can't see that well you'll also see
27:36
that here you have
27:37
structures of all different sizes
27:41
there are small ones like this one here
27:44
yeah and the larger ones like this one
27:46
and so you have these structures of all
27:48
different sizes
27:51
and this is again reminiscent of cell
27:53
similarity
27:54
and the critical point so it turns out
27:57
our little empiric model has a critical
28:00
point
28:00
can i ask a question um i i think i
28:04
roughly understand the model
28:06
but this simulation is the simulation of
28:08
what
28:09
so is it a hamiltonian system where the
28:12
um just yes so what is this basically
28:15
nothing
28:16
yeah so uh you just take what it is here
28:19
or you take that just this year the way
28:21
here that you write these simulations
28:23
the different ways to write them you
28:25
have a lattice you know you have a
28:26
numerical simulation you have a vector
28:28
an array and you either have like
28:31
one or zero and then you pick a side
28:35
randomly
28:36
and perform these reactions here
28:39
you know so so the one way to do that is
28:41
to pick a side randomly
28:44
and check if your neighbors overcome if
28:45
you pick this side
28:47
and if your neighbor is susceptible or
28:49
it does not is not infected
28:52
then you infect the neighbor with the
28:53
probability
28:55
lambda over two so like a monte carlo
28:58
simulation it's a monte carlo simulation
29:00
the different ways you can also think of
29:02
a cellular automaton
29:04
yeah but uh the typical way to simulate
29:06
these things are multicolored
29:08
simulations
29:09
okay but there's no hamiltonian there's
29:10
no deeper insight you can just take
29:12
these rules
29:13
and simulate them on a lattice and the
29:15
only thing you have to do
29:16
is to take into account that this is not
29:19
a deterministic process here
29:21
but it's a random process with a
29:23
probability one-half
29:25
you turn this one here lambda over two
29:27
you turn this one here
29:29
into an infected blue one can i then
29:32
properly
29:33
um make a make a statement about the
29:36
time scale how
29:37
some how this thing spreads because it's
29:40
it's random right so i can
29:44
that's what we'll be trying to do today
29:47
okay uh but we'll
29:48
uh only be managing to do that tomorrow
29:50
uh let me just go on
29:52
so here you get some kind of idea
29:55
already
29:56
in this slide here you can get us some
29:58
kind of idea here so that
30:00
that you have here a time scale yeah
30:03
it's not
30:04
where things uh where things uh
30:06
disappear
30:08
yeah so you say that for example here
30:10
the typically the
30:12
the number of infected individuals will
30:14
go down with an exponential function
30:17
yeah and then this has a typical time
30:18
and then you get rid of most of the
30:20
infected ones
30:22
this has some certain time at this
30:25
typical
30:25
this certain time where you say okay
30:29
at this time i'm i have lost most of my
30:31
infected
30:32
people now that they're healthy again
30:35
this is then called
30:36
psi parallel
30:39
yeah so this is like a correlation
30:41
length so i'm actually actually getting
30:43
a hat a little bit too far so so this
30:45
you have here a correlation length in
30:47
time
30:48
but that tells you exactly that how f
30:51
how quickly
30:52
does this disease disappear
30:55
yes you have a correlation length in
30:57
space like this
30:59
so for example this one here
31:03
now our analysis is it depends on how
31:05
you define it it can relate to that
31:07
uh this is typically called psi
31:10
perpendicular
31:12
you have a correlation length in space
31:13
now that tells you how large are your
31:15
clusters
31:16
but you also have a correlation length
31:18
and time how long
31:20
lived are your clusters how long does it
31:23
take for them to disappear
31:25
and it turns out that both of these
31:27
things at the critical point are
31:29
infinite
31:30
so the system is not only self-similar
31:32
in space but also in time
31:37
yeah but first before we before we do
31:39
that um
31:40
uh before we do that formally so what i
31:43
see here
31:44
is is the stochastic simulation uh we'll
31:46
later
31:48
motivate some larger equation that we
31:50
actually will be studying
31:52
but for now now the system is as simple
31:54
as it gets now you have a neighbor
31:56
if this neighbor is not infected you
31:58
infect it with a certain probability
32:00
yeah it's the simplest there's like five
32:03
lines of code also
32:04
in matlab now there's nothing there's
32:08
nothing
32:09
uh in terms of the simulation the modal
32:11
definition is nothing
32:12
that is nothing deep in there but of
32:15
course the consequences as we see on the
32:16
slide
32:17
are rather non-trivial
slide 7
32:22
so now we want to go one step
32:26
ahead and try to formalize this
32:29
mathematically
32:32
and um to formalize this we first need
32:36
to
32:36
have something to put into our larger
32:39
equation
32:40
yeah and that something that we put into
32:42
our laundry equation
32:44
is the density of this or this order
32:47
parameter
32:49
is the density of infected
32:52
individuals all right to get this we uh
32:58
um we we do a double average
33:02
so this average here is over the lattice
33:06
now we sum over the letters and we count
33:10
now with this si variable like a spin
33:14
how many infected individuals we have
33:17
and divide it by the total number of
33:19
lattice sites
33:20
the system size and then we average
33:23
again
33:23
over the ensemble now this is our order
33:26
parameter and this parameter this order
33:28
parameter
33:29
tells us whether we have order or not
33:33
you know if this is one then everybody
33:35
is infected
33:36
yeah it's not or not order or not if
33:38
this is one
33:39
everybody's infected and if this is zero
33:42
then everybody is healthy
33:45
so this is our like our magnetization
33:48
and now
33:49
we want to do the same thing as an
33:52
equilibrium
33:53
i also want to ask what are we actually
33:56
looking at yeah so
34:00
what we say is we don't know
34:04
but we make the assumption
34:07
that in this non-equilibrium critical
34:10
point
34:11
we also have scaling behavior and we
34:14
also have self-similarity
34:16
now and of course you can test this
34:18
assumption if you do large enough
34:20
computer simulations
34:22
so one thing is that this
34:25
if our system obtains a steady state
34:28
with some density
34:30
you know so that's a row
34:33
stationary density something like the
34:36
magnetization you know the process of 90
34:39
of the people
34:40
are infected uh
34:43
this goes with
34:48
lambda minus lambda c to the power
34:51
of beta it's like the magnetization we
34:54
don't know what beta is
34:56
but there is some better that we want to
35:00
know
35:02
now as i've discussed already before we
35:04
have now not just
35:05
one correlation length but two so one is
35:08
the spatial
35:13
correlation length
35:18
and this is typically denoted by psi
35:21
perpendicular
35:22
because it's perpendicular to time and
35:26
perpendicular so if you look at these
35:28
pictures here
35:30
you can kind of get an idea
35:33
why this is called perpendicular and
35:35
parallel
35:37
suppose that this is here whether this
35:39
actually an
35:40
equivalent model is the one of water
35:43
pouring
35:44
into soil you know so you have little
35:48
channels
35:48
it's a rough thing yeah and then for
35:51
example here
35:52
the water flows down
35:56
but at some point now the density of the
35:59
soil is too large
36:00
and the water stops this is
36:03
this is an example where the soil is
36:05
like the soil is like
36:06
it's very open it's not very dense you
36:09
put water in it
36:10
and it flows all the way to the bottom
36:13
so that's
36:14
what is what is sorry yes um you said
36:17
that
36:19
in case of critical systems the
36:20
correlation length in space can be
36:22
divergent
36:24
yes and also the correlation length in
36:27
time could be divergent
36:28
yes so if the correlation length in time
36:31
is divergent then in this specific
36:33
example
36:35
the it'll the number of infected
36:38
clusters will always be present
36:40
right yes exactly you will not you will
36:43
never get rid of this
36:44
disease but of course this infinities
36:47
when i talk about infinity
36:49
these infinities are not defined really
36:52
in this small simulation where we maybe
36:54
have
36:54
100 individuals also now these
36:57
infinities are defined for
36:58
systems that don't really have an
37:00
infinite infinitely large size
37:03
now this here what you see the
37:05
simulation in the middle
37:07
can just by chance disappear
37:11
and it will disappear and i can tell you
37:15
even that the disease in this case here
37:18
the right hand side will disappear with
37:21
a very small probability
37:23
right so it's a very nice feature of
37:25
this model that will turn out to be very
37:28
important
37:29
what happens if all individuals
37:32
are healthy what happens if all
37:36
individuals are healthy
37:39
then there's no process here
37:44
there are only s's there's no process
37:46
here that can give you the disease back
37:49
once the disease is extinct
37:52
it will never come back and because this
37:55
is a stochastic system
37:57
you just have to wait long enough and
37:59
just by chance
38:01
even this is casey on the right hand
38:03
side will turn into the
38:05
case just because it's stochastic just
38:08
by chance
38:09
maybe you have to wait 100 billion years
38:11
or so for this to happen but you know
38:13
that at some point you will end up in
38:15
this state
38:17
where the disease went extinct by chance
38:20
you have to wait extremely long for that
38:22
but you know that it will happen
38:24
and these states here now like in this
38:27
system here
38:28
you go to zero and then there's no way
38:30
it can come back
38:32
yeah in reality you will have to wait
38:34
for evolution
38:36
to create another virus that has the
38:39
same properties
38:40
now to come back so that texas goes
38:42
extremely long
38:43
yeah it's a much it's much longer than
38:45
the spreading of the disease itself it
38:47
happens in one or two years
38:50
you know and so these are called
38:52
absorbing states you can go in there
38:54
but you can never go out again
38:58
so in other words this means that in
39:00
these absorbing states
39:02
they're very important for not only for
39:03
virus spreading but in any ecological
39:05
model
39:06
now we have extinction and this
39:09
absorbing states
39:10
uh you can get in but you will never be
39:12
able to get out
39:14
now once you're in there you're trapped
39:16
and these absorbing states they don't
39:18
have
39:19
fluctuations they don't have any noise
39:21
and we'll see that in the larger
39:22
equation you know so these absorbing
39:25
states don't have any noise
39:28
and by this you can already see that
39:30
this whole system
39:31
is a non-equilibrium system because if
39:34
you have a state that has no noise
39:36
this is not a thermal system where you
39:38
have a temperature
39:40
now so this here is a system where you
39:41
have noise now for example here you have
39:44
noise but once you reach the state
39:47
where there's no disease no virus left
39:51
you don't have any noise anymore you
39:54
know and that cannot happen in a
39:55
thermodynamic system that is an
39:57
equilibrium
39:58
that you always have your temperature
40:00
and this will always give you noise
40:01
regardless of how many
40:03
particles you have or whatever yeah so
40:06
this already tells you that this is a
40:07
non-equilibrium system
40:09
and it's a very interesting system and
40:11
the system is actually one of the
40:13
universality classes non-equilibrium
40:15
physics
40:16
so that's once you are getting a little
40:19
bit ahead
40:20
once you know that your system has one
40:22
absorbing state
40:24
many ecological systems for example one
40:26
absorbing state
40:28
then it's quite likely that what i'll
40:29
show you in these
40:31
uh renewalization calculations today and
40:34
next week
40:35
will also apply to these systems is a
40:37
very powerful
40:38
you know universality class and
40:40
universal system
40:42
for non-equilibrium systems
40:45
yeah but let's let's i was here talking
40:47
about did i actually answer your
40:49
question so i got a little bit
40:50
uh distracted uh i i distracted myself
40:55
a little bit yeah did i do that
40:58
okay okay i i forgot it at the end i
41:01
forgot this question but i hope i
41:02
answered it at some point
41:04
okay so you have the two correlations
41:06
just spatial correlation length
41:07
you know that's uh sigma
41:11
uh side perpendicular and actually
41:14
what i want to say here is actually
41:16
that's that's what i would say
41:18
yeah so you have here you have soul and
41:20
you have water
41:21
flowing through this then the parallel
41:23
length here
41:26
this one it's called parallel because
41:28
it's parallel to the direction of
41:29
gravitation
41:31
and the other length here is
41:34
perpendicular because it's perpendicular
41:36
to the
41:37
direction of gravitation now that's
41:38
where these names come from
41:40
because these models called direct or
41:42
the directed percolation
41:44
is that you have something flowing
41:47
through a rough
41:48
medium like soil and then you have a
41:51
direct gravitation force that pulls the
41:54
fluid into one direction
41:55
but not in the other direction and
41:57
that's where these parallel and
41:58
perpendicular
42:00
yeah so and then we give that some
42:02
exponent
42:04
lambda minus lambda c to the power of
42:08
minus
42:09
mu perpendicular
42:12
now that we have the temporal or dynamic
42:22
correlation length
42:25
side parallel and this
42:28
we call and very surprisingly minus
42:32
new parallel
42:35
and this is now as you said so
42:38
our temporal correlation can become
42:41
infinity
42:42
what does it mean yeah so i have a
42:45
perturbation
42:46
to the system so what so if you have a
42:49
the spatial correlation and infinity
42:52
like an isomorph
42:53
you make a perturbation and this
42:56
perturbation will in principle
42:58
affect all parts of the magnet
43:01
yeah you will have a very very long
43:03
range correlation you flip a spin
43:05
somewhere
43:05
and it has an effect somewhere
43:07
completely somewhere else
43:10
now we have an infinite correlation
43:13
length
43:13
in time what does that mean so that
43:16
means that if we perturb the system we
43:18
are at a critical point
43:19
we will perturb the system and the time
43:23
that it takes the system to go back to
43:26
forget this perturbation
43:27
is infinitely long yeah
43:30
so so you have again now processes in
43:33
all time scales and parallel
43:35
very long very slow process and also
43:37
infinitely long processes
43:39
yeah it's like like the space in the
43:41
isis mode you have classes of all
43:43
different sizes now you have also
43:44
processes
43:45
of all different length scales at the
43:48
same time
43:49
now and this is what this criticality
43:51
does to time
43:52
the time domain you make it perturbation
43:55
and it never just disappears again
43:57
that the effects of this particular
43:59
perturbation you will see in this system
44:01
infinitely long now so that's the cool
44:04
thing about critical systems that does
44:06
uh
44:08
it does uh they do very straight things
44:11
and then now we define another uh
44:15
i've got a question yes sorry is are we
44:18
still dealing with a mean field model
44:21
uh i i didn't tell you yet uh but we'll
44:24
we won't be dealing with the mean field
44:26
of model
44:27
yeah so mean field is not very good for
44:30
these kind of things
44:32
yeah so we're not uh we're not dealing
44:34
with the mean fit model
44:35
last last time last week we were dealing
44:37
with mean field models
44:39
but this time we have to take propaganda
44:41
fluctuations properly into account
44:44
and we will have also to take into
44:46
confluctuations on all
44:48
different temporal and spatial scales
44:51
you know so that's that's what we will
44:53
have to do and that's what we will uh do
44:55
with the renovation group
44:58
so mean field theory is typically pretty
45:00
bad for these things
45:03
even just for the getting what is this
45:05
lambda c i'm going to show you what this
45:06
lambda c is but this is
45:08
in the mean field version we would say
45:10
okay this is just one half or something
45:12
like this
45:12
yeah where you write down some
45:14
differential equation like you did last
45:16
time
45:16
you write you guess some differential
45:18
equation you motivate it
45:20
and then you get some lambda c but i'll
45:22
show you today
45:23
now that this is actually not how it
45:26
works if you have these
45:27
strong fluctuations to get different
45:29
results
45:31
okay so then we have a third exponent
45:35
the dynamic critical exponents
45:39
and that means that near lambda c near
45:42
the critical point
45:45
we say that uh these correlation lengths
45:49
now they they both follow power rules
45:52
you know and we say that they are
45:54
connected
45:55
by this exponent called z
45:59
and this is called
46:02
a dynamical
46:07
critical
46:11
exponent yeah so we what we did now
46:14
is okay we said okay these systems look
46:17
self-similar
46:18
uh like in equilibrium and we assume
46:21
that the same concepts of scaling
46:23
and power laws also apply to
46:26
non-equilibrium systems
slide 8
46:30
and now i have a slide here that we
46:31
already talked about
46:33
this is just what are these correlation
46:35
lengths now so what are these
46:37
correlation uh
46:40
what are what are these correlation
46:41
lengths these two and we discussed that
46:44
basically already
46:45
and so if you look here on the right
46:49
hand side for example
46:51
so on the left hand side i have two
46:53
simulations
46:55
that were started from an initial seat
46:58
from this just one
46:59
infected individual and on the right
47:02
hand side
47:03
these figures uh they are from
47:06
assimilation
47:06
were maybe 50 percent of the letters
47:10
was infected you know other
47:13
simulations i didn't show you uh there's
47:16
a nice review
47:17
article by hindi uh
47:24
about a non-equilibrium criticality and
47:26
face transitions
47:28
and that's where it took this pair this
47:30
uh it's a very nice review
47:32
uh about the kind of things that we're
47:34
doing this week and next week
47:37
so here now we this is just an intuition
47:40
about what these correlation lengths are
47:42
uh i told you already you know that this
47:45
um that this temporal correlation length
47:49
side parallel gives you to say the time
47:52
that the disease
47:54
dies out as you can see that here
47:57
in this simulation here that the
47:59
parallel correlation length
48:01
that's the time for such a droplet here
48:04
to go away again now we know that
48:08
if we're below the critical point that
48:09
at some point it will disappear
48:12
and this has a typical time and this is
48:14
just the excite
48:16
parallel and then you have also a
48:18
typical size of such a droplet here
48:21
that's psi perpendicular that's just how
48:24
large
48:25
do these domains get and you can also
48:28
see that here on the right hand side of
48:29
course
48:30
how long does how large does the domain
48:33
of uninfected
48:34
or infected individuals get now there's
48:37
this one here and how long does it
48:39
survive
48:41
now that's how these correlation lengths
48:44
are to be
48:45
interpreted intuitively
slide 9
48:49
okay now
48:53
i'm going to do a little step
48:57
uh that where i'm trying to avoid a long
49:02
calculation
49:05
what we usually would do now is we would
49:08
take this lattice model and we would
49:11
try to derive a master equation
49:15
and also this probability that the
49:17
lattice
49:18
has a certain configuration and then we
49:21
would be looking
49:22
at this master equation uh
49:25
this very complex master equation that
49:27
tells us the time evolution of this
49:29
vector and we try to get the rates
49:33
and then we would uh do approximations
49:36
uh like the system size expansion or the
49:38
chromosomes
49:39
expansion and so on and then we would
49:41
try to derive this large
49:43
which now that's a lengthy business and
49:46
it's not the subject of our
49:48
lecture what i just want to show you is
49:52
that why does this larger equation that
49:55
i'll show you
49:56
later not look like what you naively
49:59
would expect
50:01
to this end you can just show you that
50:03
such a letter
50:04
system you can interpret in different
50:06
ways
50:08
one way is to say that it won't give you
50:11
like
50:12
the mathematical rigorous form of the
50:14
larger equation but it shows you why
50:16
it looks not like you expect it to look
50:19
like
50:20
so the first way we can interpret
50:24
this lattice system is to say what
50:27
happens
50:28
at t and what is the state at some d
50:32
plus dt some like a very short time
50:35
interval after that
50:38
and now i'm taking like in this master
50:40
equation i'm taking the perspective
50:43
of the state in a certain site
50:48
and then ask what are the previous
50:51
states
50:52
that give rise to me being infected
50:57
you know what gives rises yeah and
51:00
suppose that i was not infected before
51:03
yeah then my left neighbor could have
51:05
been infected
51:07
my right neighbor could have been
51:09
infected or both of them could have been
51:12
infected
51:12
and have affected me these are the three
51:15
processes that
51:16
lead to me being infected and then
51:19
if it was there's another process
51:23
and i should use here different colors
51:25
it's not the recovery
51:35
i switched the colors not red is
51:39
healthy sorry
51:43
red is healthy and blue is infected
51:48
okay yeah if i in this recovery process
51:53
it doesn't matter what my neighbors were
51:54
i would just i just know that was
51:56
previously infected
51:57
if i'm now in the process of recovery
52:01
now this is this master equation picture
52:03
where we ask the word which state do i
52:04
come from
52:06
and now we can take an equivalent
52:09
description
52:10
and take more a lattice picture now that
52:13
corresponds more a little bit left to
52:14
the longer
52:15
picture yeah where update where i say
52:18
what is the state of the lattice now
52:20
and what is the state of the lattice the
52:21
next step
52:23
now that corresponds to this
52:24
differential equation picture
52:28
yeah and then how can i update it so
52:32
then i need at least two lattice points
52:35
at the same time to update to define
52:37
these updating rules
52:39
and then i have different possibilities
52:41
here
52:42
yeah if my two lattice points are
52:45
infected
52:46
that previously either the left one or
52:49
the right one
52:50
was infected
52:54
the recovery process is no more
52:56
complicated
52:57
yeah if i have known that in this two
53:00
side picture i have one
53:03
infected and one one infected and one
53:07
not affected once this white is down
53:09
here
53:10
this is
53:13
healthy this is
53:17
infected
53:21
now if i have one infected and one
53:24
healthy one
53:25
previously my system could have been
53:29
must have been in a state now if i'm
53:31
looking at the
53:32
recovery process where both of them were
53:35
infected
53:37
now so with the probability one half i
53:39
have either this
53:41
or this one here and then
53:44
if both of the states in the second time
53:47
step
53:48
are healthy then one of them was in fact
53:52
before now so here i update two sides at
53:55
the same time
53:57
or i can also say i updated the whole
53:59
letters at the same time
54:00
in parallel and now
54:04
the thing is what is this here
54:09
what is this here these two processes
54:11
here
54:12
suddenly i have a process a process for
54:15
recovery
54:17
that involves two individuals
54:20
yeah formally that involves two
54:22
individuals so here have two individuals
54:25
to infect it
54:26
and after that only one of them is
54:28
infected
slide 10
54:30
now suddenly i have two individuals and
54:32
if i ask how
54:33
such a term here if i take this
54:37
description as the basis for my larger
54:40
equation
54:41
how will this term pop up in my larger
54:44
equation
54:45
then it's this term is proportional to
54:49
one-half
54:50
times
54:55
the probability that one of them is
54:57
infected and the probability that the
54:59
other one is
55:00
also infected so we will have something
55:03
like rho
55:04
of x t squared
55:08
and we'll get a minus because we
55:10
decrease the number
55:11
of infected individuals
55:15
now this is just to show you yeah by
55:16
this uh
55:18
magnesium inside the exotic lattice
55:20
representation
55:22
that you suddenly get a term you write
55:25
down the larger equation here
55:27
this is the larger equation
55:30
and this is here what you expect the
55:32
first thing is what you expect is an
55:34
infection term
55:35
you have the density of infected people
55:37
at a certain position x
55:40
and this depends on how many infected
55:43
people i already have
55:44
that is this typical exponential
55:47
increase of the infection rate
55:51
now we get another term here
55:55
this term that describes the recovery
55:58
process now this describes
56:02
the recovery process and
56:05
it suddenly has this quadratic term
56:08
although this recovery process like this
56:10
picture looked completely linear because
56:12
every individual was doing it
56:14
individually now we have now the second
56:18
degree term here and that looks like an
56:21
interaction
56:22
and this comes just because we're
56:23
updating all the letters in parallel in
56:25
this
56:26
lingerie equation and that's why we get
56:29
this
56:30
second degree term here we have a
56:33
diffusion term
56:34
this one here that was also not in our
56:36
model description
56:38
you know that we never said that these
56:41
particles are actually moving around
56:44
but there is some spread of spatial
56:46
information because you
56:47
interact with the nearest neighbor yeah
56:50
and this
56:51
is models if you zoom out and go to a
56:54
continuous picture
56:56
it's a spread diffusive spread of
56:58
infection information
57:00
and that's why you effectively get a
57:02
term here you need to have a term
57:05
that involves spatial derivatives now
57:07
where you actually spread something over
57:09
space you spread the infection of
57:11
space and that's why you get this term
57:13
here
57:14
and you have of course again a noise
57:16
term
57:18
now and i'll give these here parameters
57:20
and also new names the combinations of
57:22
the old parameters
57:24
and we want to keep this we want to have
57:26
different parameters at each of these
57:28
rates because we in the next step want
57:30
to renormalize these parameters that we
57:33
we need them and the noise here
57:36
on the right hand side is our good old
57:39
gaussian noise now that has
57:41
zero mean and
57:44
correlations in space and time
57:47
that are luckily delta distributors so
57:50
they're
57:50
memory less they don't have memory in
57:52
space or in time
57:54
but they depend on this density here
58:00
they depend on this density and what
58:03
this means
58:04
is that this noise
58:08
the correlation
58:11
in the noise or the strength of this
58:13
noise that's the strength of
58:15
noise
58:19
is zero
58:24
if the disease
58:28
is extinct that's called
58:31
multiplicative noise because now the
58:33
density
58:35
rho of x and t is a pre-factor in the
58:39
noise term that
58:40
if is it is contributing
58:43
or defining the strength of the nodes
58:45
and once we have
58:46
zero infected individuals left then the
58:49
noise disappeared and we can never have
58:51
get away out of this term out of this
58:53
point
58:54
where the infection is lost
slide 11
58:59
so maybe
59:02
quite late in time as a next step
59:10
um we get a little bit more formal so
59:12
now we have the larger
59:13
equation now and what you see here
59:16
already
59:17
is the martensite rose functional
59:20
integral that we derived
59:22
and this functional martin citra rose
59:24
function integral
59:25
or martensite rules johnson the dominic
59:27
is functional integral
59:29
you can divide very easily we remember
59:32
the equation that we had
59:33
a few lectures ago first
59:36
part of this function integral
59:42
what was it here
59:45
it's just the launch of a equation
59:48
yeah that that makes sure that you
59:50
actually solve the launcher equation
59:52
here you have the laundry equation and
59:55
then you have these terms here
59:57
on the right hand side there are higher
60:00
order
60:01
here we have this phi squared
60:04
that's just this one here and here
60:08
you have a noise term that we also had
60:10
before now that was this
60:12
gaussian noise term that we came from
60:14
integrating out the psi
60:16
in the martensitic rows formula now so
60:19
we have this term here
60:21
is quite noise but in contrast to the
60:23
previous case
60:25
where the noise didn't depend on the
60:27
density itself
60:28
we now here that's the only difference
60:31
have
60:31
another phi here
60:35
now we have another phi here
60:38
that's the only difference that we get
60:40
for multiplicative noise
60:42
and because we have multiplicative noise
60:46
you can see that somehow now the noise
60:50
term
60:51
here looks a little bit like another
60:55
term that is
60:55
actually as an interaction term where we
60:58
couple the two fields
61:00
one linearly to each other this term and
61:03
suddenly the noise term is not
61:04
simple a gaussian it's not simple a
61:07
gaussian that the inti can integrate
61:08
over
61:09
suddenly it couples to the other fields
61:11
now the strength of this noise curve
61:13
is proportional to phi that's this
61:16
multiplicative noise that makes life
61:18
complicated
61:20
now we now do a simple step now we
61:23
re-scale we now
61:24
we now we take this martensitic rows
61:27
generating function that i wrote down
61:29
here we take this
61:31
yeah and we just make our lives easier
61:33
life easier for later
61:35
yeah and by doing this to do this
61:39
we rescale some of these
61:42
fields and parameters so rescale
61:47
the fields so what we do
61:50
is we want to
61:53
get rid of
61:59
we have had two terms this one and this
62:02
one
62:03
they look kind of similar and the idea
62:06
is now
62:06
that if we have a proper transformation
62:09
or fields
62:10
that we can make them exactly equal up
62:13
to some pre-factors
62:15
yeah so that's what we want to do we
62:16
want to simplify this action
62:19
and uh to symmetrize it now so that we
62:22
can treat these two terms
62:23
and e equally now that's how we may we
62:26
want to make the prefactor here
62:28
this and this prefactor equal
62:32
now we can summarize these two trends
62:35
now so really scared the fields but just
62:37
tell you how to do that you have phi
62:40
goes over to 2 lambda
62:43
over gamma times phi
62:46
if i tilde the response field goes over
62:50
to
62:52
gamma to lambda
62:56
and gamma goes over to
63:00
two gamma lambda
63:05
yeah so we rescale this we are allowed
63:07
to do that
63:08
and then we get our new generating
63:11
functional
63:13
and this generating functional is again
63:16
of this form
63:17
d phi d phi tilde
63:22
e to the some action as naught and we
63:26
find that
63:26
what this is phi
63:30
by children
63:33
plus now this is has a naught the others
63:36
the non-interacting term
63:38
and now we have a term that of course
63:40
that describes interactions
63:43
now that's where the fun is happening
63:46
by artillery and just write that down
63:50
this as not phi
63:54
phi tilde this action is the integral
63:57
over dx
63:59
dt
64:03
so here we should have
64:06
dt as well
64:10
yeah so at the x dt
64:13
and now we um have
64:17
5 x t
64:21
tau sorry what we need that
64:25
del t minus d naught
64:28
squared minus kappa
64:32
phi of x t
64:37
and then we have another part as
64:40
interaction phi phi to the
64:45
gamma over two integral dx
64:49
dt phi tilde
64:53
x t minus i sorry that looks
64:56
not so nice
65:01
all right tilde of x and t
65:05
times phi of x
65:09
t minus phi to the
65:13
of x and t
65:17
phi of x
65:20
t so this is this interaction term
65:24
and it's called interaction term because
65:26
we have here
65:27
higher orders of the field coupling to
65:29
each other
65:31
yeah so
65:34
this is now our martin's intervals
65:37
integral
65:38
yeah and this is what we'll be dealing
65:41
with and this is what we'll define
65:43
the result renormalization group on and
65:46
because i knew that we would be doing
65:48
renormalization
65:50
i have already introduced this little
65:54
towel here now because in
65:57
renormalization
65:58
we want to cause grain we want to
66:00
transform this action
66:02
and we want to see how our action and
66:04
how the parameters are
66:06
of our action change in this procedure
66:09
and that's why all our terms here need
66:11
to have a prefactor
66:14
yeah and that's why i introduced this
66:15
tau that's why i have this d
66:18
you know and this kappa here i have them
66:20
all
66:21
giving them different names although
66:22
they're not independent of each other
66:25
because our original model just had one
66:27
parameter that was the lambda
66:29
yeah so that's the uh that's the martin
66:32
sutra rose
66:34
functional integral and um
66:38
next so we're already quite late now
66:41
next time
66:42
we start right away with uh first with
66:47
introducing renormalization intuitively
66:51
and then as a second step you'll then
66:53
apply that
66:54
to this epidemic model and re-normalize
66:57
this margin central rows
66:58
functional integral now apparently i was
67:01
always
67:01
very optimistic and it was trying to
67:04
introduce already
67:05
the renomination today but we can do
67:07
that just next week
67:08
now because it's already quite late
67:11
today