slide 1

00:00 there we go 00:07 okay so now you can see the screen i 00:09 hope 00:10 with a little overview so what where are 00:13 we actually 00:14 uh at the moment yeah so we had uh 00:18 two lectures ago we started thinking 00:20 about how order can emerge 00:22 you know and we said that this somehow 00:25 um very often uh relies on a balance 00:29 between 00:30 fluctuations you know that favor 00:31 disorder and 00:33 interactions that favor order 00:37 and then we went on next week last week 00:40 to study how we can have transitions 00:44 between different non-equilibrium states 00:48 for example between a 00:49 disordered and an order state or between 00:52 different kinds of ordered states 00:54 and we started that in a situation where 00:56 we neglected 00:58 moles we pretended that our 01:00 non-equilibrium system 01:01 or our general system could also be an 01:04 equilibrium system is very very large 01:06 and then we were basically back in the 01:08 framework 01:09 of non-linear differential equations and 01:12 non-linear partial differential 01:14 equations 01:15 and to understand then what kind of 01:17 order we had 01:19 we looked at specific situations here 01:21 where this order 01:23 or these patterns arose continuously 01:26 or not abruptly but they first were 01:28 small and then we can larger larger 01:31 and in these situations were allowed and 01:33 to linearize 01:34 and to ignore these difficult 01:38 non-linearities in these equations 01:42 so today we're now in a state where we 01:44 want to 01:46 join these two lectures now we're 01:49 looking at transitions between 01:51 non-equilibrium 01:53 steady states where 01:57 we have noise yeah and i wouldn't tell 02:01 you that i wouldn't have a lecture on 02:02 this if 02:03 the noise these fluctuations in these 02:06 transitions weren't super important 02:09 and maybe yeah then 02:12 once we've understood that we are in a 02:15 position 02:16 uh just to understand how does actually 02:19 order look like 02:20 how can we identify order and after that 02:23 we'll then 02:23 go to data and actually learn some tools 02:26 from data science 02:27 on how to extract such order 02:31 from complex and big data sets 02:34 and then at the end of this lecture of 02:37 this uh this term 02:38 uh we'll have a specific example from 02:40 research where we bring that all 02:42 together and we see 02:43 how this works together uh in the 02:45 context of 02:46 current research

slide 2

02:50 okay so to start 02:54 let's remind ourselves about 02:57 how we actually transit from order to 03:00 from disorder to order in equilibrium 03:03 yeah and we'll continue we'll 03:06 continue looking at continuous phase 03:09 transition and 03:10 in equilibrium these continuous phase 03:12 positions are characterized 03:14 by a critical point and this is just the 03:16 point where this balance 03:18 between fluctuations and interactions 03:21 this ordering and these disordering 03:24 forces 03:25 are just equal now the system doesn't 03:27 really know 03:28 exactly where to go whether to create an 03:30 ordered state or to a completely 03:32 disintegrated 03:33 order state and it's somewhere in 03:34 between and 03:36 uh so i i suppose 03:39 uh you've you you'll have had that in 03:42 your 03:42 statistical physics lecture normally 03:44 yeah but at the end of the first 03:46 statistical physics lectures but i'll 03:49 just give you 03:51 a little reminder of the most important 03:54 concepts 03:55 so here you see like this example from 03:58 equilibrium 03:59 this very powerful and intuitive model 04:02 system which is called the izing model 04:04 which is essentially modeling a 04:05 ferromagnet magnet 04:07 and on the left hand side you can see 04:10 simulations 04:12 for three different temperatures it's 04:15 just 04:16 such an icing model and what you see 04:18 here if the temperature is very low 04:21 you get an ordered state now everything 04:23 is black 04:24 all spins are putting the same 04:26 directions 04:27 in the same direction and if the 04:29 temperature is high 04:32 then you get a disordered state and this 04:35 disorder state as you see as a kind of 04:37 salt and pepper state now the average 04:40 magnetization is zero 04:43 but and locally uh you will always find 04:46 spin that goes up and down 04:48 go up and down and then you have that 04:50 thing in between these states 04:52 yeah when the temperature is exactly 04:56 equal to a critical temperature and this 04:59 is where 04:59 there is a balance between energy and 05:02 entropy of 05:03 fluctuations and order and 05:06 what you see here is this state 05:09 where you have these domains here 05:11 domains of these black domains 05:14 and if you zoom in to such a system 05:18 what you'll see at the critical point 05:19 you'll see that it looks exactly the 05:21 same 05:21 now you zoom in and you would wouldn't 05:24 be able to say whether this is a 05:25 snapshot a zoomed in version of the one 05:27 on the left hand side 05:29 or whether this is an entirely different 05:32 simulation 05:34 and then you can zoom in further and 05:35 further and 05:37 you again all the time get the same 05:39 impression that you can't really make 05:41 out 05:42 uh what is now the typical length how 05:44 large are these classes 05:46 you have here these domains of all sizes 05:49 equally now because you have domains 05:53 these white things or black things of 05:55 all 05:56 size is equally represented you can't 05:59 make out a single sound you can't make a 06:01 typical 06:02 length scale here because you can't say 06:04 that 06:06 the typical size of such a cluster here 06:10 is for example that large 06:13 as you always find clusters that are 06:15 much smaller you find glasses that are 06:17 much 06:17 larger here's a cluster now that has the 06:20 size of the entire system that's 06:22 infinitely large 06:23 and then you have all a spectrum of 06:26 sizes in between that 06:28 yeah well here in this example you kind 06:31 of get an idea 06:33 that these clusters are typically very 06:35 small 06:36 these domains whereas pin points up or 06:38 so are very small they have a 06:40 typical size but you can't see it on the 06:42 left hand side 06:44 yeah same as if i would zoom in with 06:46 this camera you know so 06:49 if we do like this now you immediately 06:52 see 06:53 that i zoomed in now because i have a 06:56 characteristic size i'm 06:57 one meter 80 or something yeah and now 07:00 you 07:00 see that you that have zoomed in the 07:02 camera and the picture is not the same 07:04 as before 07:06 so i'm not in a critical state yeah but 07:09 the icing system is in a critical state 07:12 and this critical state is characterized 07:15 by self-similarity 07:17 so they have structures of all sizes 07:19 represented 07:21 in this system now this is the 07:23 self-similarity 07:24 and the mathematical representation of 07:27 the self-similarity 07:28 the fact that you don't have an average 07:30 cluster size 07:32 is that the correlation length diverges 07:35 now so the correlation length is 07:37 infinity 07:38 that means the correlation function you 07:41 know 07:41 so how that represents how large 07:45 these clusters are is scale in variance 07:48 that means 07:49 that really classes of different sizes 07:51 are equally represented 07:53 and such if you ask what is the 07:55 probability that i 07:56 am now in a white cluster that i'm in a 07:59 black cluster a certain distance away 08:01 then the answer to this doesn't depend 08:03 on any specific distance 08:05 you know it's the same this probability 08:07 is the same this 08:08 correlation function here it's the same 08:11 when we 08:12 calculated in the original version of 08:15 the simulation or a zoomed in 08:17 uh fraction of this this is the 08:19 self-similarity 08:21 the self-similarity at a critical point 08:23 goes along with power laws 08:26 you know so the power laws have the um 08:30 for example like this here the critical 08:31 length as you go the critical 08:34 correlation length as you go closer to 08:37 the 08:37 critical point diverges to infinity it 08:40 goes to infinity 08:42 and it does so with an exponent that's 08:44 to be 08:45 called new when this gets zero here 08:49 this term will be infinity and these 08:51 exponents 08:52 capture how fast you go to infinity 08:56 and these exponents are very the fact 08:58 that you have an exponent but 08:59 that you have such a power rule means 09:02 that yourself similar you have a power 09:04 law if you have 09:06 something like this you can zoom in and 09:08 you still have the same exponent here 09:10 and you can't do that with an 09:11 exponential function or so 09:14 now and it also tells you that this is 09:17 here 09:18 uh if you have power laws from some 09:20 distribution that goes over 09:22 that has a power law that has this long 09:25 tail some exponent you cannot typically 09:29 calculate averages or moments because 09:31 these integrals diverge 09:34 now you have the power law of the 09:35 correlation function and you have a 09:38 power law 09:38 now if you have a power of the 09:39 correlation function you also have the 09:41 power laws 09:42 for example in the density and the 09:44 magnetization 09:46 uh near the critical point and all kinds 09:48 of other thermodynamic 09:50 quantities and i just wanted to briefly

slide 3

09:53 show you why this is the case 09:55 now it's actually the background of this 09:58 the background of this is actually an 09:59 assumption 10:01 that you say you have a free energy 10:05 and with this free energy uh 10:08 this free energy as you go to the 10:10 critical pound point uh 10:11 gets in finite it has singularity 10:15 and then you say that you assume 10:19 that the free energy here 10:23 this free energy has 10:26 one part that has all the physics and 10:29 all the details a regular part 10:32 but you say that the part of the free 10:33 energy that 10:35 diverges at a critical point 10:39 this one here so now it's is a 10:44 t t 10:46 is something like t minus 10:50 t c over t you know and h 10:53 is the external field if you have 10:56 something like this as a free energy as 10:58 the function of these parameters 11:00 yeah then the singular part the one that 11:03 goes to infinity 11:05 is a homogeneous function 11:08 homogeneous function just uh tells you 11:12 that if you have f of 11:15 lambda x that this is equal to 11:18 lambda to the power of some alpha 11:22 f on x yeah and this represents that 11:25 just this these gains zooming in you 11:28 have the same function you zoom in where 11:29 you rescale your variable 11:31 you zoom in and you get the same 11:33 function back 11:34 now this is a homogeneous function and 11:37 you still assume that this free energy 11:39 density in this case is a homogeneous 11:42 function 11:43 and this homogeneous function can only 11:45 depend 11:46 on dimensionless quantities now for 11:49 example 11:50 you wouldn't expect this divergence to 11:53 infinity 11:54 to depend on how you measure a length 11:58 now whether you measure the length in 12:01 units of centimeters or meters 12:04 whether you measure temperature in 12:07 kelvin or in units of one kelvin or two 12:10 kelvins or so 12:12 so you can see you these kind of things 12:15 these units 12:15 dimensions should be irrelevant for how 12:18 this quantity goes to infinity 12:22 yeah and if we say that then we say okay 12:24 so we have 12:26 here so-called scaling function that 12:29 depends on dimensional parameter 12:31 a dimensional dimensionless 12:34 combinations of our parameters so the 12:37 external field 12:38 divided to the temperature and then we 12:40 have to 12:41 take the temperature to some power of 12:43 something 12:45 to make everything dimensionless 12:48 so that it has no units no and there's 12:51 something that's not 12:52 something we don't know yeah and 12:56 this has the free energy has some units 12:59 therefore the whole thing gets doesn't 13:01 have the units you need a pre-factor 13:04 that gives you the right units you know 13:06 and then 13:08 you have this alpha which we don't know 13:11 yeah 13:11 this is some exponent that depends on 13:13 the specific model 13:14 for example for the eisenmann zero 13:17 you know and then you have these uh this 13:20 is the 13:21 consequence of how you translate this 13:23 homogeneity 13:25 here of the free energy to something 13:28 that you give names 13:29 as you have here this part that diverges 13:32 yeah that has the units 13:34 and this part here is the so-called 13:37 scaling function 13:38 that only depends on dimensionless 13:40 parameters 13:41 and it turns out that these exponents 13:43 and this scaling function are universal 13:45 so if you know it for one model then you 13:47 know it will know it 13:49 you will know it for a very large class 13:52 and we'll see that once we do 13:53 renormalization later today 13:57 so uh so what does it mean yeah so if we 13:59 make this 14:00 assumption that's really an assumption 14:02 about 14:04 homogeneity of the free energy then we 14:07 can calculate for example 14:09 the magnetization m of th 14:12 now this is in thermodynamics something 14:14 like 14:16 del f 2 del h 14:19 you know and then we just plug this in 14:22 and we get something like temperature 14:25 this reduced temperature 14:27 t to the power of minus 2 minus alpha 14:30 means 14:30 minus delta some other function that we 14:34 don't know 14:35 that again depends on a dimensionless 14:40 parameter and then 14:44 this scales like some better that's the 14:47 definition 14:48 of this exponent better of the 14:50 magnetization 14:52 yeah and uh so in thermodynamics this is 14:55 called 14:56 rhythm scaling basically in any textbook 14:59 on statistical physics 15:00 and it's just just to show you how the 15:04 assumption 15:05 of homogeneity near the critical point 15:09 leads to power laws in other 15:11 thermodynamic quantities 15:13 i've shown you here the magnetization 15:17 this was the magnetization 15:23 but the same holds true for example for 15:26 susceptibility 15:27 and other thermodynamic quantities that 15:29 you can get by taking derivatives of 15:31 your 15:32 energy so 15:35 this homogeneity or this self-similarity 15:39 that i showed you here that is a 15:41 reflection that is one of the hallmarks 15:44 of uh critical behavior and that's what 15:47 we're looking for when we look for 15:49 critical behavior 15:50 and now the question is can we see 15:53 something like this 15:55 also a non-equilibrium system

slide 4

15:58 before before i start with that 16:01 uh let's just have a look at one 16:02 specific how this scaling 16:04 behaves if you look at these equations 16:08 here 16:10 what does it mean it means that 16:13 actually the curves that you get you 16:16 know so if you just divide so once you 16:19 make a measurement for example with a 16:21 known temperature 16:23 and a known magnetic field you measure 16:26 this function here 16:28 then you know that it doesn't depend 16:29 separately 16:31 on the age and the temperature 16:35 so you can rescale your axis so this is 16:37 what is your y 16:38 x axis you can use that your x axis and 16:41 your y 16:42 axis to make all of these curves 16:46 collapse onto each other yeah and this 16:49 is this uh 16:50 scaling form that we see in equilibrium 16:54 physics this is for the icing model 16:56 so on the x-axis we have this scaled 16:58 temperature 16:59 that would be t on the previous slide 17:02 lowercase t 17:04 uh times something so this is v scale 17:08 and then these uh people in these 17:11 experiments 17:12 for uh for 17:15 for a ferro ferromagnet measured 17:18 the magnetization for different values 17:20 of the 17:22 of different experimental parameter 17:24 values 17:25 like magnetic field external magnetic 17:27 field temperature 17:30 and by making use of this formula here 17:33 you see that uh the scaling 17:37 behavior what is where is my 17:40 is it going down here yeah that's the 17:42 scaling behavior here 17:45 yeah that the only thing 17:49 that you don't know is this g of m that 17:52 you have to measure 17:53 yeah once you know the h and the 17:56 temperature 17:58 you can make all of these different g of 18:00 m's the gms 18:02 this scaling function you can rescale 18:05 these axes 18:06 to make them collapse onto each other 18:09 yeah and this is this observation this 18:11 is how you observe 18:12 scaling an experiment so you manage to 18:15 collapse 18:16 your experimental curves by multiplying 18:19 this x-axis and the y-axis with certain 18:23 values 18:23 of offense you have to guess you can 18:27 collapse all of these curves on the same 18:30 uh universal so-called scaling form 18:34 now this is the manifestation of scaling 18:36 and that's of course 18:37 also something we'll be looking at and 18:39 non-equilibrium systems 18:41 but also in data 18:44 now scaling is a whole mark of critical 18:48 behavior and today

slide 5

18:52 we want to see whether these concepts 18:56 of scaling and criticality yeah 18:59 where and these these continuous 19:02 phase transitions actually also extend 19:05 to non-equilibrium systems 19:07 yeah and it turns out so now we first we 19:10 need to find a non-equilibrium system 19:12 that is as intuitive 19:15 as the ising model and the icing model 19:18 is very intuitive 19:19 you have that in your lectures when 19:21 you're a student and 19:22 in your later life as a scientist you 19:24 always refer to that because it's so 19:26 simple and intuitive that uh you can 19:28 explain a lot of things a lot of 19:30 things about continuous phase 19:32 transitions in equilibrium 19:33 just based on this very simple model 19:35 like i did in the beginning of this 19:37 lecture 19:38 and it turns out now that the uh 19:41 icing model of non-equilibrium physics 19:44 of course is 19:46 that so people would dig a disagree but 19:48 one of the simplest models in 19:49 mono-equilibrium physics that shows 19:52 critical behavior 19:53 is an epidemic model and this epidemic 19:56 model we knew already 19:57 from the previous lectures here we have 20:02 our good old si model again 20:06 now so this epidemic model is this is 20:08 the simplest model 20:10 that you can think about so you have 20:11 infected individuals 20:14 i and susceptible individuals or 20:17 healthier people's less 20:19 now if an i meets an s 20:22 then the s gets infected with the rate 20:25 say lambda half 20:27 and turns into an affected infidel and 20:30 in the end you have two of them 20:33 then you have the other process that we 20:34 recover 20:36 and we set this rate to one now we can 20:38 just set that to one 20:40 without any loss of generality and uh 20:43 so that infected we measure units 20:46 time in units of this recovery rate 20:50 you know service infected individuals 20:52 can also 20:54 then recover and become 20:57 healthy again we have these two kinds of 21:00 individuals 21:01 and now we put them in the real world so 21:04 last time we were only looking at some 21:06 well-mixed 21:07 average quantities but now we put them 21:10 into the real world like the city of 21:11 brisbane also 21:13 where they actually can where actually 21:16 space 21:17 matters yeah so i'm more likely to 21:19 infect somebody else working at the 21:22 mp rpk pks than somebody looking at 21:25 another max planck institute for example 21:28 yeah so 21:29 so here uh these spatial structures 21:32 these special degrees of freedom 21:34 uh are taken into account and the 21:36 simplest way of you 21:38 how you can think about this is at the 21:40 bottom here 21:42 now that you look at letters 21:45 you have a letters each site 21:48 carries either an affected individual or 21:51 a recovered individual and 21:55 you know an infected or a recovered 21:56 individual and 21:58 when an infected individual 22:02 is next to a recovered a healthy one 22:05 then 22:05 the healthy one can turn into an 22:07 infected one 22:09 with a certain probability of with a 22:10 certain rate lambda over two 22:13 yeah and also there's another process 22:16 here if i saw if the 22:18 individual on a certain position is 22:21 infected 22:21 it can turn into a healthy one at a rate 22:24 lambda 22:26 so this is this simple spatial version 22:29 that you can think about for this and 22:32 simple epidemic model and it's also the 22:35 literature is often called the contact 22:38 process 22:40 so of course real epidemic model models 22:43 have 22:44 typically one more component namely the 22:47 um 22:49 [Music] 22:51 the infected recovered uh wait 22:54 is this also here okay so so the third 22:57 component that you normally have 22:59 and these models are the recovered one 23:01 the immune people 23:02 now you have the disease yeah and then 23:04 you are fine for the rest of your life 23:06 and you're immune to this disease 23:08 so you can only forget in fact one then 23:11 you have a third 23:12 species here a third kinds of particles 23:15 which would be the recovered ones 23:17 or the immune ones and they cannot be a 23:21 faculty again 23:22 but this slightly more complicated model 23:25 uh is shows very similar behavior to the 23:28 model that we're studying here 23:31 for the things that we're interested in 23:32 so here we're interested in infinities 23:34 in singularities so once you 23:38 once you look at these kind of things 23:40 then these models will qualitatively the 23:42 same 23:43 although also the exponents will be 23:45 different 23:47 but once you look of course into 23:49 non-singularities it's more critical 23:51 behavior 23:52 than the messiness of how wide your 23:55 roads are 23:57 how often the tram goes uh between the 24:00 blasphemy institutes and so on these 24:02 things will matter 24:05 yeah but close to the critical point uh 24:07 we'll be fine

slide 6

24:10 so this is a stochastic simulation of 24:13 such a system 24:14 and we can just see what happens on the 24:17 left hand side 24:18 you see a simulation of such a lattice 24:20 system 24:21 where you initially have random random 24:25 random initialization so every site 24:28 is either with the probability of one 24:30 half 24:32 a certain probability infected or 24:35 not infected and what you see here 24:39 now is in blue infected 24:42 individuals now if this lambda 24:46 now this infection rate is smaller 24:49 than a certain critical value 24:52 then what you will see is that this 24:54 infraction 24:55 this infection can spread for a while 24:58 but most of the time with a certain 24:59 probability 25:00 it will uh disappear 25:04 now so in this regime here in this phase 25:08 the recovery rate outweighs 25:11 the infection rate you know and 25:14 uh so that's what we're supposed to be 25:17 on in this regime we're supposed to be 25:19 investing 25:19 starting next week and then on the right 25:23 hand side 25:24 that's the regime that we're currently 25:26 in now then the infection rate 25:28 is larger than the uh 25:31 than the recovery rate now so the 25:33 infection probability is higher 25:36 and what you will then end up is is a 25:39 state where most of the individuals will 25:42 carry the disease will be infected 25:45 so you will you will reach a steady 25:47 state 25:49 not everybody is all the time in fact 25:50 that you will reach some steady state 25:52 with a certain percentage of infected 25:55 people 25:57 and now we have this situation in 26:00 between 26:02 that's this one here and this is 26:05 where the um where 26:08 the infection rate is more or less 26:11 balanced 26:12 with the recovery rate it's not exactly 26:15 equal to one 26:16 so that's these things are complicated 26:18 yeah you might think that 26:20 okay if this this lambda is equal to one 26:23 or one half or so 26:25 yeah then uh then that's the critical 26:27 point of these systems i'll show you are 26:29 more complicated than you might think 26:32 because the noise is so important here 26:35 and here 26:36 what you see is that you have domains 26:39 that become larger and larger over time 26:41 so from 26:42 when we go from top to bottom we have 26:43 time now so we go we start here at the 26:46 top 26:47 and then this domain goes large and 26:49 larger you have merging 26:51 of domains with infected individuals 26:54 and then we have branches that they die 26:56 out 26:57 like this one here and uh 27:00 it looks a little bit like a 27:02 self-similar state 27:04 now we have domains of all sizes for 27:07 example 27:08 in the time domain or from top to bottom 27:11 you have some branches that die out 27:13 quite quickly here 27:15 but then you have other branches like 27:17 the big one in the middle 27:19 where that just go on for uh forever 27:23 without really occupying the whole 27:24 system 27:26 and then if you look take a slice in 27:28 this direction here 27:30 these simulations are very small also 27:33 it's not like a design 27:34 you can't see that well you'll also see 27:36 that here you have 27:37 structures of all different sizes 27:41 there are small ones like this one here 27:44 yeah and the larger ones like this one 27:46 and so you have these structures of all 27:48 different sizes 27:51 and this is again reminiscent of cell 27:53 similarity 27:54 and the critical point so it turns out 27:57 our little empiric model has a critical 28:00 point 28:00 can i ask a question um i i think i 28:04 roughly understand the model 28:06 but this simulation is the simulation of 28:08 what 28:09 so is it a hamiltonian system where the 28:12 um just yes so what is this basically 28:15 nothing 28:16 yeah so uh you just take what it is here 28:19 or you take that just this year the way 28:21 here that you write these simulations 28:23 the different ways to write them you 28:25 have a lattice you know you have a 28:26 numerical simulation you have a vector 28:28 an array and you either have like 28:31 one or zero and then you pick a side 28:35 randomly 28:36 and perform these reactions here 28:39 you know so so the one way to do that is 28:41 to pick a side randomly 28:44 and check if your neighbors overcome if 28:45 you pick this side 28:47 and if your neighbor is susceptible or 28:49 it does not is not infected 28:52 then you infect the neighbor with the 28:53 probability 28:55 lambda over two so like a monte carlo 28:58 simulation it's a monte carlo simulation 29:00 the different ways you can also think of 29:02 a cellular automaton 29:04 yeah but uh the typical way to simulate 29:06 these things are multicolored 29:08 simulations 29:09 okay but there's no hamiltonian there's 29:10 no deeper insight you can just take 29:12 these rules 29:13 and simulate them on a lattice and the 29:15 only thing you have to do 29:16 is to take into account that this is not 29:19 a deterministic process here 29:21 but it's a random process with a 29:23 probability one-half 29:25 you turn this one here lambda over two 29:27 you turn this one here 29:29 into an infected blue one can i then 29:32 properly 29:33 um make a make a statement about the 29:36 time scale how 29:37 some how this thing spreads because it's 29:40 it's random right so i can 29:44 that's what we'll be trying to do today 29:47 okay uh but we'll 29:48 uh only be managing to do that tomorrow 29:50 uh let me just go on 29:52 so here you get some kind of idea 29:55 already 29:56 in this slide here you can get us some 29:58 kind of idea here so that 30:00 that you have here a time scale yeah 30:03 it's not 30:04 where things uh where things uh 30:06 disappear 30:08 yeah so you say that for example here 30:10 the typically the 30:12 the number of infected individuals will 30:14 go down with an exponential function 30:17 yeah and then this has a typical time 30:18 and then you get rid of most of the 30:20 infected ones 30:22 this has some certain time at this 30:25 typical 30:25 this certain time where you say okay 30:29 at this time i'm i have lost most of my 30:31 infected 30:32 people now that they're healthy again 30:35 this is then called 30:36 psi parallel 30:39 yeah so this is like a correlation 30:41 length so i'm actually actually getting 30:43 a hat a little bit too far so so this 30:45 you have here a correlation length in 30:47 time 30:48 but that tells you exactly that how f 30:51 how quickly 30:52 does this disease disappear 30:55 yes you have a correlation length in 30:57 space like this 30:59 so for example this one here 31:03 now our analysis is it depends on how 31:05 you define it it can relate to that 31:07 uh this is typically called psi 31:10 perpendicular 31:12 you have a correlation length in space 31:13 now that tells you how large are your 31:15 clusters 31:16 but you also have a correlation length 31:18 and time how long 31:20 lived are your clusters how long does it 31:23 take for them to disappear 31:25 and it turns out that both of these 31:27 things at the critical point are 31:29 infinite 31:30 so the system is not only self-similar 31:32 in space but also in time 31:37 yeah but first before we before we do 31:39 that um 31:40 uh before we do that formally so what i 31:43 see here 31:44 is is the stochastic simulation uh we'll 31:46 later 31:48 motivate some larger equation that we 31:50 actually will be studying 31:52 but for now now the system is as simple 31:54 as it gets now you have a neighbor 31:56 if this neighbor is not infected you 31:58 infect it with a certain probability 32:00 yeah it's the simplest there's like five 32:03 lines of code also 32:04 in matlab now there's nothing there's 32:08 nothing 32:09 uh in terms of the simulation the modal 32:11 definition is nothing 32:12 that is nothing deep in there but of 32:15 course the consequences as we see on the 32:16 slide 32:17 are rather non-trivial

slide 7

32:22 so now we want to go one step 32:26 ahead and try to formalize this 32:29 mathematically 32:32 and um to formalize this we first need 32:36 to 32:36 have something to put into our larger 32:39 equation 32:40 yeah and that something that we put into 32:42 our laundry equation 32:44 is the density of this or this order 32:47 parameter 32:49 is the density of infected 32:52 individuals all right to get this we uh 32:58 um we we do a double average 33:02 so this average here is over the lattice 33:06 now we sum over the letters and we count 33:10 now with this si variable like a spin 33:14 how many infected individuals we have 33:17 and divide it by the total number of 33:19 lattice sites 33:20 the system size and then we average 33:23 again 33:23 over the ensemble now this is our order 33:26 parameter and this parameter this order 33:28 parameter 33:29 tells us whether we have order or not 33:33 you know if this is one then everybody 33:35 is infected 33:36 yeah it's not or not order or not if 33:38 this is one 33:39 everybody's infected and if this is zero 33:42 then everybody is healthy 33:45 so this is our like our magnetization 33:48 and now 33:49 we want to do the same thing as an 33:52 equilibrium 33:53 i also want to ask what are we actually 33:56 looking at yeah so 34:00 what we say is we don't know 34:04 but we make the assumption 34:07 that in this non-equilibrium critical 34:10 point 34:11 we also have scaling behavior and we 34:14 also have self-similarity 34:16 now and of course you can test this 34:18 assumption if you do large enough 34:20 computer simulations 34:22 so one thing is that this 34:25 if our system obtains a steady state 34:28 with some density 34:30 you know so that's a row 34:33 stationary density something like the 34:36 magnetization you know the process of 90 34:39 of the people 34:40 are infected uh 34:43 this goes with 34:48 lambda minus lambda c to the power 34:51 of beta it's like the magnetization we 34:54 don't know what beta is 34:56 but there is some better that we want to 35:00 know 35:02 now as i've discussed already before we 35:04 have now not just 35:05 one correlation length but two so one is 35:08 the spatial 35:13 correlation length 35:18 and this is typically denoted by psi 35:21 perpendicular 35:22 because it's perpendicular to time and 35:26 perpendicular so if you look at these 35:28 pictures here 35:30 you can kind of get an idea 35:33 why this is called perpendicular and 35:35 parallel 35:37 suppose that this is here whether this 35:39 actually an 35:40 equivalent model is the one of water 35:43 pouring 35:44 into soil you know so you have little 35:48 channels 35:48 it's a rough thing yeah and then for 35:51 example here 35:52 the water flows down 35:56 but at some point now the density of the 35:59 soil is too large 36:00 and the water stops this is 36:03 this is an example where the soil is 36:05 like the soil is like 36:06 it's very open it's not very dense you 36:09 put water in it 36:10 and it flows all the way to the bottom 36:13 so that's 36:14 what is what is sorry yes um you said 36:17 that 36:19 in case of critical systems the 36:20 correlation length in space can be 36:22 divergent 36:24 yes and also the correlation length in 36:27 time could be divergent 36:28 yes so if the correlation length in time 36:31 is divergent then in this specific 36:33 example 36:35 the it'll the number of infected 36:38 clusters will always be present 36:40 right yes exactly you will not you will 36:43 never get rid of this 36:44 disease but of course this infinities 36:47 when i talk about infinity 36:49 these infinities are not defined really 36:52 in this small simulation where we maybe 36:54 have 36:54 100 individuals also now these 36:57 infinities are defined for 36:58 systems that don't really have an 37:00 infinite infinitely large size 37:03 now this here what you see the 37:05 simulation in the middle 37:07 can just by chance disappear 37:11 and it will disappear and i can tell you 37:15 even that the disease in this case here 37:18 the right hand side will disappear with 37:21 a very small probability 37:23 right so it's a very nice feature of 37:25 this model that will turn out to be very 37:28 important 37:29 what happens if all individuals 37:32 are healthy what happens if all 37:36 individuals are healthy 37:39 then there's no process here 37:44 there are only s's there's no process 37:46 here that can give you the disease back 37:49 once the disease is extinct 37:52 it will never come back and because this 37:55 is a stochastic system 37:57 you just have to wait long enough and 37:59 just by chance 38:01 even this is casey on the right hand 38:03 side will turn into the 38:05 case just because it's stochastic just 38:08 by chance 38:09 maybe you have to wait 100 billion years 38:11 or so for this to happen but you know 38:13 that at some point you will end up in 38:15 this state 38:17 where the disease went extinct by chance 38:20 you have to wait extremely long for that 38:22 but you know that it will happen 38:24 and these states here now like in this 38:27 system here 38:28 you go to zero and then there's no way 38:30 it can come back 38:32 yeah in reality you will have to wait 38:34 for evolution 38:36 to create another virus that has the 38:39 same properties 38:40 now to come back so that texas goes 38:42 extremely long 38:43 yeah it's a much it's much longer than 38:45 the spreading of the disease itself it 38:47 happens in one or two years 38:50 you know and so these are called 38:52 absorbing states you can go in there 38:54 but you can never go out again 38:58 so in other words this means that in 39:00 these absorbing states 39:02 they're very important for not only for 39:03 virus spreading but in any ecological 39:05 model 39:06 now we have extinction and this 39:09 absorbing states 39:10 uh you can get in but you will never be 39:12 able to get out 39:14 now once you're in there you're trapped 39:16 and these absorbing states they don't 39:18 have 39:19 fluctuations they don't have any noise 39:21 and we'll see that in the larger 39:22 equation you know so these absorbing 39:25 states don't have any noise 39:28 and by this you can already see that 39:30 this whole system 39:31 is a non-equilibrium system because if 39:34 you have a state that has no noise 39:36 this is not a thermal system where you 39:38 have a temperature 39:40 now so this here is a system where you 39:41 have noise now for example here you have 39:44 noise but once you reach the state 39:47 where there's no disease no virus left 39:51 you don't have any noise anymore you 39:54 know and that cannot happen in a 39:55 thermodynamic system that is an 39:57 equilibrium 39:58 that you always have your temperature 40:00 and this will always give you noise 40:01 regardless of how many 40:03 particles you have or whatever yeah so 40:06 this already tells you that this is a 40:07 non-equilibrium system 40:09 and it's a very interesting system and 40:11 the system is actually one of the 40:13 universality classes non-equilibrium 40:15 physics 40:16 so that's once you are getting a little 40:19 bit ahead 40:20 once you know that your system has one 40:22 absorbing state 40:24 many ecological systems for example one 40:26 absorbing state 40:28 then it's quite likely that what i'll 40:29 show you in these 40:31 uh renewalization calculations today and 40:34 next week 40:35 will also apply to these systems is a 40:37 very powerful 40:38 you know universality class and 40:40 universal system 40:42 for non-equilibrium systems 40:45 yeah but let's let's i was here talking 40:47 about did i actually answer your 40:49 question so i got a little bit 40:50 uh distracted uh i i distracted myself 40:55 a little bit yeah did i do that 40:58 okay okay i i forgot it at the end i 41:01 forgot this question but i hope i 41:02 answered it at some point 41:04 okay so you have the two correlations 41:06 just spatial correlation length 41:07 you know that's uh sigma 41:11 uh side perpendicular and actually 41:14 what i want to say here is actually 41:16 that's that's what i would say 41:18 yeah so you have here you have soul and 41:20 you have water 41:21 flowing through this then the parallel 41:23 length here 41:26 this one it's called parallel because 41:28 it's parallel to the direction of 41:29 gravitation 41:31 and the other length here is 41:34 perpendicular because it's perpendicular 41:36 to the 41:37 direction of gravitation now that's 41:38 where these names come from 41:40 because these models called direct or 41:42 the directed percolation 41:44 is that you have something flowing 41:47 through a rough 41:48 medium like soil and then you have a 41:51 direct gravitation force that pulls the 41:54 fluid into one direction 41:55 but not in the other direction and 41:57 that's where these parallel and 41:58 perpendicular 42:00 yeah so and then we give that some 42:02 exponent 42:04 lambda minus lambda c to the power of 42:08 minus 42:09 mu perpendicular 42:12 now that we have the temporal or dynamic 42:22 correlation length 42:25 side parallel and this 42:28 we call and very surprisingly minus 42:32 new parallel 42:35 and this is now as you said so 42:38 our temporal correlation can become 42:41 infinity 42:42 what does it mean yeah so i have a 42:45 perturbation 42:46 to the system so what so if you have a 42:49 the spatial correlation and infinity 42:52 like an isomorph 42:53 you make a perturbation and this 42:56 perturbation will in principle 42:58 affect all parts of the magnet 43:01 yeah you will have a very very long 43:03 range correlation you flip a spin 43:05 somewhere 43:05 and it has an effect somewhere 43:07 completely somewhere else 43:10 now we have an infinite correlation 43:13 length 43:13 in time what does that mean so that 43:16 means that if we perturb the system we 43:18 are at a critical point 43:19 we will perturb the system and the time 43:23 that it takes the system to go back to 43:26 forget this perturbation 43:27 is infinitely long yeah 43:30 so so you have again now processes in 43:33 all time scales and parallel 43:35 very long very slow process and also 43:37 infinitely long processes 43:39 yeah it's like like the space in the 43:41 isis mode you have classes of all 43:43 different sizes now you have also 43:44 processes 43:45 of all different length scales at the 43:48 same time 43:49 now and this is what this criticality 43:51 does to time 43:52 the time domain you make it perturbation 43:55 and it never just disappears again 43:57 that the effects of this particular 43:59 perturbation you will see in this system 44:01 infinitely long now so that's the cool 44:04 thing about critical systems that does 44:06 uh 44:08 it does uh they do very straight things 44:11 and then now we define another uh 44:15 i've got a question yes sorry is are we 44:18 still dealing with a mean field model 44:21 uh i i didn't tell you yet uh but we'll 44:24 we won't be dealing with the mean field 44:26 of model 44:27 yeah so mean field is not very good for 44:30 these kind of things 44:32 yeah so we're not uh we're not dealing 44:34 with the mean fit model 44:35 last last time last week we were dealing 44:37 with mean field models 44:39 but this time we have to take propaganda 44:41 fluctuations properly into account 44:44 and we will have also to take into 44:46 confluctuations on all 44:48 different temporal and spatial scales 44:51 you know so that's that's what we will 44:53 have to do and that's what we will uh do 44:55 with the renovation group 44:58 so mean field theory is typically pretty 45:00 bad for these things 45:03 even just for the getting what is this 45:05 lambda c i'm going to show you what this 45:06 lambda c is but this is 45:08 in the mean field version we would say 45:10 okay this is just one half or something 45:12 like this 45:12 yeah where you write down some 45:14 differential equation like you did last 45:16 time 45:16 you write you guess some differential 45:18 equation you motivate it 45:20 and then you get some lambda c but i'll 45:22 show you today 45:23 now that this is actually not how it 45:26 works if you have these 45:27 strong fluctuations to get different 45:29 results 45:31 okay so then we have a third exponent 45:35 the dynamic critical exponents 45:39 and that means that near lambda c near 45:42 the critical point 45:45 we say that uh these correlation lengths 45:49 now they they both follow power rules 45:52 you know and we say that they are 45:54 connected 45:55 by this exponent called z 45:59 and this is called 46:02 a dynamical 46:07 critical 46:11 exponent yeah so we what we did now 46:14 is okay we said okay these systems look 46:17 self-similar 46:18 uh like in equilibrium and we assume 46:21 that the same concepts of scaling 46:23 and power laws also apply to 46:26 non-equilibrium systems

slide 8

46:30 and now i have a slide here that we 46:31 already talked about 46:33 this is just what are these correlation 46:35 lengths now so what are these 46:37 correlation uh 46:40 what are what are these correlation 46:41 lengths these two and we discussed that 46:44 basically already 46:45 and so if you look here on the right 46:49 hand side for example 46:51 so on the left hand side i have two 46:53 simulations 46:55 that were started from an initial seat 46:58 from this just one 46:59 infected individual and on the right 47:02 hand side 47:03 these figures uh they are from 47:06 assimilation 47:06 were maybe 50 percent of the letters 47:10 was infected you know other 47:13 simulations i didn't show you uh there's 47:16 a nice review 47:17 article by hindi uh 47:24 about a non-equilibrium criticality and 47:26 face transitions 47:28 and that's where it took this pair this 47:30 uh it's a very nice review 47:32 uh about the kind of things that we're 47:34 doing this week and next week 47:37 so here now we this is just an intuition 47:40 about what these correlation lengths are 47:42 uh i told you already you know that this 47:45 um that this temporal correlation length 47:49 side parallel gives you to say the time 47:52 that the disease 47:54 dies out as you can see that here 47:57 in this simulation here that the 47:59 parallel correlation length 48:01 that's the time for such a droplet here 48:04 to go away again now we know that 48:08 if we're below the critical point that 48:09 at some point it will disappear 48:12 and this has a typical time and this is 48:14 just the excite 48:16 parallel and then you have also a 48:18 typical size of such a droplet here 48:21 that's psi perpendicular that's just how 48:24 large 48:25 do these domains get and you can also 48:28 see that here on the right hand side of 48:29 course 48:30 how long does how large does the domain 48:33 of uninfected 48:34 or infected individuals get now there's 48:37 this one here and how long does it 48:39 survive 48:41 now that's how these correlation lengths 48:44 are to be 48:45 interpreted intuitively

slide 9

48:49 okay now 48:53 i'm going to do a little step 48:57 uh that where i'm trying to avoid a long 49:02 calculation 49:05 what we usually would do now is we would 49:08 take this lattice model and we would 49:11 try to derive a master equation 49:15 and also this probability that the 49:17 lattice 49:18 has a certain configuration and then we 49:21 would be looking 49:22 at this master equation uh 49:25 this very complex master equation that 49:27 tells us the time evolution of this 49:29 vector and we try to get the rates 49:33 and then we would uh do approximations 49:36 uh like the system size expansion or the 49:38 chromosomes 49:39 expansion and so on and then we would 49:41 try to derive this large 49:43 which now that's a lengthy business and 49:46 it's not the subject of our 49:48 lecture what i just want to show you is 49:52 that why does this larger equation that 49:55 i'll show you 49:56 later not look like what you naively 49:59 would expect 50:01 to this end you can just show you that 50:03 such a letter 50:04 system you can interpret in different 50:06 ways 50:08 one way is to say that it won't give you 50:11 like 50:12 the mathematical rigorous form of the 50:14 larger equation but it shows you why 50:16 it looks not like you expect it to look 50:19 like 50:20 so the first way we can interpret 50:24 this lattice system is to say what 50:27 happens 50:28 at t and what is the state at some d 50:32 plus dt some like a very short time 50:35 interval after that 50:38 and now i'm taking like in this master 50:40 equation i'm taking the perspective 50:43 of the state in a certain site 50:48 and then ask what are the previous 50:51 states 50:52 that give rise to me being infected 50:57 you know what gives rises yeah and 51:00 suppose that i was not infected before 51:03 yeah then my left neighbor could have 51:05 been infected 51:07 my right neighbor could have been 51:09 infected or both of them could have been 51:12 infected 51:12 and have affected me these are the three 51:15 processes that 51:16 lead to me being infected and then 51:19 if it was there's another process 51:23 and i should use here different colors 51:25 it's not the recovery 51:35 i switched the colors not red is 51:39 healthy sorry 51:43 red is healthy and blue is infected 51:48 okay yeah if i in this recovery process 51:53 it doesn't matter what my neighbors were 51:54 i would just i just know that was 51:56 previously infected 51:57 if i'm now in the process of recovery 52:01 now this is this master equation picture 52:03 where we ask the word which state do i 52:04 come from 52:06 and now we can take an equivalent 52:09 description 52:10 and take more a lattice picture now that 52:13 corresponds more a little bit left to 52:14 the longer 52:15 picture yeah where update where i say 52:18 what is the state of the lattice now 52:20 and what is the state of the lattice the 52:21 next step 52:23 now that corresponds to this 52:24 differential equation picture 52:28 yeah and then how can i update it so 52:32 then i need at least two lattice points 52:35 at the same time to update to define 52:37 these updating rules 52:39 and then i have different possibilities 52:41 here 52:42 yeah if my two lattice points are 52:45 infected 52:46 that previously either the left one or 52:49 the right one 52:50 was infected 52:54 the recovery process is no more 52:56 complicated 52:57 yeah if i have known that in this two 53:00 side picture i have one 53:03 infected and one one infected and one 53:07 not affected once this white is down 53:09 here 53:10 this is 53:13 healthy this is 53:17 infected 53:21 now if i have one infected and one 53:24 healthy one 53:25 previously my system could have been 53:29 must have been in a state now if i'm 53:31 looking at the 53:32 recovery process where both of them were 53:35 infected 53:37 now so with the probability one half i 53:39 have either this 53:41 or this one here and then 53:44 if both of the states in the second time 53:47 step 53:48 are healthy then one of them was in fact 53:52 before now so here i update two sides at 53:55 the same time 53:57 or i can also say i updated the whole 53:59 letters at the same time 54:00 in parallel and now 54:04 the thing is what is this here 54:09 what is this here these two processes 54:11 here 54:12 suddenly i have a process a process for 54:15 recovery 54:17 that involves two individuals 54:20 yeah formally that involves two 54:22 individuals so here have two individuals 54:25 to infect it 54:26 and after that only one of them is 54:28 infected

slide 10

54:30 now suddenly i have two individuals and 54:32 if i ask how 54:33 such a term here if i take this 54:37 description as the basis for my larger 54:40 equation 54:41 how will this term pop up in my larger 54:44 equation 54:45 then it's this term is proportional to 54:49 one-half 54:50 times 54:55 the probability that one of them is 54:57 infected and the probability that the 54:59 other one is 55:00 also infected so we will have something 55:03 like rho 55:04 of x t squared 55:08 and we'll get a minus because we 55:10 decrease the number 55:11 of infected individuals 55:15 now this is just to show you yeah by 55:16 this uh 55:18 magnesium inside the exotic lattice 55:20 representation 55:22 that you suddenly get a term you write 55:25 down the larger equation here 55:27 this is the larger equation 55:30 and this is here what you expect the 55:32 first thing is what you expect is an 55:34 infection term 55:35 you have the density of infected people 55:37 at a certain position x 55:40 and this depends on how many infected 55:43 people i already have 55:44 that is this typical exponential 55:47 increase of the infection rate 55:51 now we get another term here 55:55 this term that describes the recovery 55:58 process now this describes 56:02 the recovery process and 56:05 it suddenly has this quadratic term 56:08 although this recovery process like this 56:10 picture looked completely linear because 56:12 every individual was doing it 56:14 individually now we have now the second 56:18 degree term here and that looks like an 56:21 interaction 56:22 and this comes just because we're 56:23 updating all the letters in parallel in 56:25 this 56:26 lingerie equation and that's why we get 56:29 this 56:30 second degree term here we have a 56:33 diffusion term 56:34 this one here that was also not in our 56:36 model description 56:38 you know that we never said that these 56:41 particles are actually moving around 56:44 but there is some spread of spatial 56:46 information because you 56:47 interact with the nearest neighbor yeah 56:50 and this 56:51 is models if you zoom out and go to a 56:54 continuous picture 56:56 it's a spread diffusive spread of 56:58 infection information 57:00 and that's why you effectively get a 57:02 term here you need to have a term 57:05 that involves spatial derivatives now 57:07 where you actually spread something over 57:09 space you spread the infection of 57:11 space and that's why you get this term 57:13 here 57:14 and you have of course again a noise 57:16 term 57:18 now and i'll give these here parameters 57:20 and also new names the combinations of 57:22 the old parameters 57:24 and we want to keep this we want to have 57:26 different parameters at each of these 57:28 rates because we in the next step want 57:30 to renormalize these parameters that we 57:33 we need them and the noise here 57:36 on the right hand side is our good old 57:39 gaussian noise now that has 57:41 zero mean and 57:44 correlations in space and time 57:47 that are luckily delta distributors so 57:50 they're 57:50 memory less they don't have memory in 57:52 space or in time 57:54 but they depend on this density here 58:00 they depend on this density and what 58:03 this means 58:04 is that this noise 58:08 the correlation 58:11 in the noise or the strength of this 58:13 noise that's the strength of 58:15 noise 58:19 is zero 58:24 if the disease 58:28 is extinct that's called 58:31 multiplicative noise because now the 58:33 density 58:35 rho of x and t is a pre-factor in the 58:39 noise term that 58:40 if is it is contributing 58:43 or defining the strength of the nodes 58:45 and once we have 58:46 zero infected individuals left then the 58:49 noise disappeared and we can never have 58:51 get away out of this term out of this 58:53 point 58:54 where the infection is lost

slide 11

58:59 so maybe 59:02 quite late in time as a next step 59:10 um we get a little bit more formal so 59:12 now we have the larger 59:13 equation now and what you see here 59:16 already 59:17 is the martensite rose functional 59:20 integral that we derived 59:22 and this functional martin citra rose 59:24 function integral 59:25 or martensite rules johnson the dominic 59:27 is functional integral 59:29 you can divide very easily we remember 59:32 the equation that we had 59:33 a few lectures ago first 59:36 part of this function integral 59:42 what was it here 59:45 it's just the launch of a equation 59:48 yeah that that makes sure that you 59:50 actually solve the launcher equation 59:52 here you have the laundry equation and 59:55 then you have these terms here 59:57 on the right hand side there are higher 60:00 order 60:01 here we have this phi squared 60:04 that's just this one here and here 60:08 you have a noise term that we also had 60:10 before now that was this 60:12 gaussian noise term that we came from 60:14 integrating out the psi 60:16 in the martensitic rows formula now so 60:19 we have this term here 60:21 is quite noise but in contrast to the 60:23 previous case 60:25 where the noise didn't depend on the 60:27 density itself 60:28 we now here that's the only difference 60:31 have 60:31 another phi here 60:35 now we have another phi here 60:38 that's the only difference that we get 60:40 for multiplicative noise 60:42 and because we have multiplicative noise 60:46 you can see that somehow now the noise 60:50 term 60:51 here looks a little bit like another 60:55 term that is 60:55 actually as an interaction term where we 60:58 couple the two fields 61:00 one linearly to each other this term and 61:03 suddenly the noise term is not 61:04 simple a gaussian it's not simple a 61:07 gaussian that the inti can integrate 61:08 over 61:09 suddenly it couples to the other fields 61:11 now the strength of this noise curve 61:13 is proportional to phi that's this 61:16 multiplicative noise that makes life 61:18 complicated 61:20 now we now do a simple step now we 61:23 re-scale we now 61:24 we now we take this martensitic rows 61:27 generating function that i wrote down 61:29 here we take this 61:31 yeah and we just make our lives easier 61:33 life easier for later 61:35 yeah and by doing this to do this 61:39 we rescale some of these 61:42 fields and parameters so rescale 61:47 the fields so what we do 61:50 is we want to 61:53 get rid of 61:59 we have had two terms this one and this 62:02 one 62:03 they look kind of similar and the idea 62:06 is now 62:06 that if we have a proper transformation 62:09 or fields 62:10 that we can make them exactly equal up 62:13 to some pre-factors 62:15 yeah so that's what we want to do we 62:16 want to simplify this action 62:19 and uh to symmetrize it now so that we 62:22 can treat these two terms 62:23 and e equally now that's how we may we 62:26 want to make the prefactor here 62:28 this and this prefactor equal 62:32 now we can summarize these two trends 62:35 now so really scared the fields but just 62:37 tell you how to do that you have phi 62:40 goes over to 2 lambda 62:43 over gamma times phi 62:46 if i tilde the response field goes over 62:50 to 62:52 gamma to lambda 62:56 and gamma goes over to 63:00 two gamma lambda 63:05 yeah so we rescale this we are allowed 63:07 to do that 63:08 and then we get our new generating 63:11 functional 63:13 and this generating functional is again 63:16 of this form 63:17 d phi d phi tilde 63:22 e to the some action as naught and we 63:26 find that 63:26 what this is phi 63:30 by children 63:33 plus now this is has a naught the others 63:36 the non-interacting term 63:38 and now we have a term that of course 63:40 that describes interactions 63:43 now that's where the fun is happening 63:46 by artillery and just write that down 63:50 this as not phi 63:54 phi tilde this action is the integral 63:57 over dx 63:59 dt 64:03 so here we should have 64:06 dt as well 64:10 yeah so at the x dt 64:13 and now we um have 64:17 5 x t 64:21 tau sorry what we need that 64:25 del t minus d naught 64:28 squared minus kappa 64:32 phi of x t 64:37 and then we have another part as 64:40 interaction phi phi to the 64:45 gamma over two integral dx 64:49 dt phi tilde 64:53 x t minus i sorry that looks 64:56 not so nice 65:01 all right tilde of x and t 65:05 times phi of x 65:09 t minus phi to the 65:13 of x and t 65:17 phi of x 65:20 t so this is this interaction term 65:24 and it's called interaction term because 65:26 we have here 65:27 higher orders of the field coupling to 65:29 each other 65:31 yeah so 65:34 this is now our martin's intervals 65:37 integral 65:38 yeah and this is what we'll be dealing 65:41 with and this is what we'll define 65:43 the result renormalization group on and 65:46 because i knew that we would be doing 65:48 renormalization 65:50 i have already introduced this little 65:54 towel here now because in 65:57 renormalization 65:58 we want to cause grain we want to 66:00 transform this action 66:02 and we want to see how our action and 66:04 how the parameters are 66:06 of our action change in this procedure 66:09 and that's why all our terms here need 66:11 to have a prefactor 66:14 yeah and that's why i introduced this 66:15 tau that's why i have this d 66:18 you know and this kappa here i have them 66:20 all 66:21 giving them different names although 66:22 they're not independent of each other 66:25 because our original model just had one 66:27 parameter that was the lambda 66:29 yeah so that's the uh that's the martin 66:32 sutra rose 66:34 functional integral and um 66:38 next so we're already quite late now 66:41 next time 66:42 we start right away with uh first with 66:47 introducing renormalization intuitively 66:51 and then as a second step you'll then 66:53 apply that 66:54 to this epidemic model and re-normalize 66:57 this margin central rows 66:58 functional integral now apparently i was 67:01 always 67:01 very optimistic and it was trying to 67:04 introduce already 67:05 the renomination today but we can do 67:07 that just next week 67:08 now because it's already quite late 67:11 today

本文收录于以下合集: