introduction

00:00 to our little uh lecture 00:03 so meanwhile uh most of my group 00:07 are in current time and i'm one of the 00:09 few people left 00:11 and um and ethan detang who used to join 00:15 me here in the lecture hall 00:17 uh has escaped to a safer place 00:20 and so now uh buddy tang is watching the 00:24 chat and 00:25 the participantsness and we'll let you 00:26 in and 00:29 moderate a little bit the chat and um 00:34 so it's a good time actually to be 00:35 working on some epidemic models that we 00:37 start 00:38 that we started working on last time and 00:42 uh in particular last time

slide 1

00:46 we were looking at the simplest possible 00:49 epidemic model you might think about 00:52 so let me share the screen just to give 00:54 you a little bit of an 00:56 um 01:01 of reminder 01:10 okay 01:32 okay 01:39 okay here we go so we introduced this 01:43 little epidemic model where you had 01:46 two kinds of people infected one 01:50 and the susceptible ones and the model 01:53 was very simple so these infected ones 01:55 could uh meet up with an affected one uh 01:58 within with a susceptible one 02:00 that's that maybe they weren't uh to a 02:01 party or so it just 02:03 used the same train yeah and then the 02:05 susceptible one uh 02:07 got infected and then you have the 02:11 second process 02:12 where if you are infected with a certain 02:15 rate also probability 02:17 per unit of time you get rid of your 02:20 disease 02:21 and you recover and go back to 02:23 susceptible 02:25 you know so you're a recovered person 02:28 so and then we went one step further and 02:31 we 02:31 put this very simple model on a lattice 02:34 in a spatial context 02:36 the simplest spatial context you can 02:38 think about is just having 02:40 lattice in one dimension yeah 02:43 so suppose that this letter looks a 02:45 little like the train like a train 02:47 and uh so on this little letters you 02:50 react 02:51 basically with your nearest neighbors if 02:53 you if you infect somebody 02:55 you're in fact your nearest neighbor but 02:57 uh not 02:58 anybody else yeah and then

slide 2

03:02 we saw that such a simple model gave 03:04 rise 03:05 uh to rich phenomenology depending on 03:08 the relation between the rate of 03:10 infection 03:11 and the rate of recovery which was set 03:14 to one 03:15 now if the infection rate was very large 03:18 then 03:18 we got a state where we had a large 03:22 fraction of individuals constantly 03:24 carrying the disease 03:26 and when the rate of infection was very 03:29 low 03:30 then we ended up in a stainless 03:31 so-called absorbing state 03:33 where the disease got extinct 03:37 now in between that we found that there 03:39 is a value of this infection rate 03:41 where infection and recovery just 03:43 balance 03:44 and this state was characterized uh 03:47 very in a very similar way to the ising 03:50 model 03:51 in equilibrium was characterized by 03:54 scale and variance 03:55 that means that these correlation 03:57 lengths that we looked at 03:59 along the space axis 04:03 and along the time axis we had two 04:05 correlation lengths 04:06 they both diverged that means we had 04:09 self-similarity 04:10 in this critical point

slide 3

04:15 yeah and then we went on and we derived 04:17 the launch of the equation 04:19 and from this launch of an equation we

slide 4

04:21 then just wrote down 04:22 the martin citra rose functional 04:26 integral and i'll show you that later 04:28 when we actually used it 04:30 when we'll actually use it for the 04:31 realization procedure 04:33 and for once we once we had written down 04:36 these actions at the margin et cetera 04:38 rose function integral 04:40 of course we can go to free of space and 04:42 write down 04:43 this action also in fourier space

slide 5

04:47 now that's the formalities yeah 04:50 that looks pretty complicated now 04:55 we have a problem right so this looks 04:57 pretty complicated 04:59 and uh we have a situation 05:02 where we have divergences where we our 05:06 correlation length is infinite infinite 05:10 and we don't really know how to deal 05:12 with this infinite correlation

slide 6

05:16 and this is exactly the situation where 05:19 we have developed in the last 05:21 70 years or so mathematical techniques 05:25 originally in 05:26 quantum field theory that allow us to 05:29 deal with these 05:30 divergences with these infinities 05:33 now we've made two observations here so 05:36 one is 05:37 scale scale invariance that i just 05:38 mentioned yeah scale variance 05:40 at this critical point there's no 05:43 distinguished length scale we zoom 05:45 in and the picture we get is 05:48 statistically 05:49 still the same as the original picture 05:54 the second observation is 05:57 that so and as a result of the scale 05:59 invariance we saw that 06:00 in equilibrium all of these 06:03 thermodynamic quantities 06:05 i like the like the free energy 06:08 the magnetization and icing model obey 06:10 power laws 06:12 as you approach the critical point so 06:14 they diverge with certain power laws 06:17 and the second observation is 06:20 that at this critical point 06:23 our empirical observation is for example 06:26 in the equilibrium but also in 06:27 non-equilibrium 06:29 that there are a lot of different 06:32 real world systems that are described by 06:36 the same 06:38 theory the same simple theory so for 06:42 example 06:43 you have the icing more you know the 06:44 icing model is super simple it's much 06:46 simpler than actually a real magnet 06:49 nevertheless the icing motor can predict 06:52 exponents 06:54 near the critical point of real magnets 06:58 also of different materials very nicely 07:03 and it even the icing model even can 07:05 predict 07:07 critical phenomena in completely 07:09 unrelated systems 07:12 like phase transitions in water if you 07:14 put water in a high pressure then you 07:15 have a phase where you have 07:17 like a coexistence of water and steam 07:20 and this you can predict these exponents 07:22 you can predict with the icing water 07:24 and this power of such simple models to 07:28 predict a large 07:29 range of critical phenomena is called 07:33 universality and both of these 07:36 observations 07:37 are the scale and variance the cell 07:39 similarity 07:40 and universality can be systematically 07:43 started with a renormalization group 07:46 the first thing you might actually think 07:47 about okay so why do we need actually 07:49 something complicated 07:51 now we just do a naive approach 07:54 uh similar to the one that we used 07:57 implicitly in these lectures about 07:59 pattern formation 08:00 and non-linear dynamics now what we 08:03 could do is okay we say okay so we just 08:06 instantaneously zoom out of the system 08:09 yeah go to 08:10 wave vector zero so in finite wavelength 08:13 you look only at the very large 08:15 properties 08:16 and just pretend that we can average 08:18 over everything we just look at average 08:20 quantities 08:21 that's that should be called mean field 08:23 theory where you pretend 08:25 that your system the state of the system 08:28 at a certain 08:28 site is governed by 08:32 a field that rises over 08:35 basically an average over the entire 08:37 system 08:38 yeah so this mean field theory where you 08:41 instantaneously 08:42 go to the macroscopic scale 08:46 this doesn't work in others it fails to 08:49 predict these exponents that we get at a 08:51 critical point 08:52 and it also doesn't give any uh reason 08:55 why we should have universality 08:59 now renewalization does something very 09:02 smart now that provides a systematic way 09:06 of connecting these microscopic degrees 09:09 of freedom that are described for 09:10 example by hamiltonian 09:13 to uh macroscopic descriptions 09:17 now it does so by taking into account 09:21 all the intermediary scales between 09:24 micro 09:24 the microscopic level and the 09:26 macroscopic level 09:28 and as you uh my guess from these 09:30 critical phenomena if you have scale 09:32 environments and all 09:33 scales are equal and equally important 09:36 this approach where you 09:38 take into account all of these scales 09:41 one at a time 09:43 now going from microscopic to 09:44 macroscopic which 09:46 makes much more sense than arbitrarily 09:48 focusing only on the largest scale as 09:50 you do in new york theory 09:54 now so rememberization group the 09:56 romanization 09:57 group allows us to go from the 09:59 microscopic 10:00 level now that is described by some 10:03 hamiltonian 10:04 by some symmetries all the way to the 10:07 macroscopic 10:08 level without ever forgetting what is 10:11 going on 10:12 in between now that's the power of the 10:14 renalization group and in this lecture i 10:16 will show you 10:18 how this is actually implemented 10:21 now how we can actually do that 10:30 okay so there's a little bit of a lot of 10:33 information i should have split this 10:34 slide

slide 7

10:35 um so 10:39 so so we are now going to follow this 10:41 program now going from microscopic 10:43 to macroscopic one scale 10:46 one length scale at a time now we're 10:49 starting with something microscopic that 10:51 is super complicated don't think about 10:53 the icing model 10:54 think about something that has 10 000 10:57 different couplings or so 11:00 now we take that model with 10 000 11:03 different parameters 11:05 on the microscopic scale so the real 11:07 physics 11:08 that's going on on the microscopic scale 11:10 that's complicated 11:12 yeah it's much more complicated than the 11:13 icing model and it just as the disease 11:16 is much more complicated as the ice i 11:18 model that i showed you 11:21 and then we go to larger and larger 11:24 scales 11:26 and hope to end up with something that 11:28 is simpler 11:30 and less correlated than on the 11:32 microscopic scale 11:34 yeah so how do we do that so realization 11:37 consists of three 11:38 steps uh the most important step 11:42 now the the actual uh what's actually 11:46 underlying renovation 11:48 is a course grading step now in this 11:51 course grading step 11:53 you can think about uh that you 11:57 unsharpen an image suppose you take a 11:59 photo 12:00 and then you can focus you can make the 12:02 picture 12:04 uh sharp or less sharp yeah by turning 12:07 like the lens 12:08 and the ring on the lens if you still 12:10 use a rear camera 12:11 yeah so you can make a job or let's try 12:13 or think about the microscope 12:15 you can have to turn some knobs to make 12:17 the image sharp or not sharp 12:19 and the first step we 12:22 cause grain so we cause grain 12:26 and that means literally that we make 12:28 the image that we're looking at in the 12:30 system 12:31 unsharp so that we that's what we do 12:35 yeah so we 12:36 mathematically that means if you uh if 12:39 you close grain if you make something 12:41 out sharp unsharp then that means that 12:43 you integrate out 12:46 fast fluctuations or short range 12:48 fluctuations 12:49 i'll show you on the next slide how this 12:51 looks like intuitively so first you get 12:53 rid of 12:53 all these uh fluctuations that happen on 12:56 very small length scales 12:59 and you have to perform an integral to 13:01 do that 13:02 and this integral of course is very 13:05 complicated 13:06 to calculate now the second state 13:10 we have a new field yeah 13:13 now let's say phi average 13:17 but because we have course grade this 13:19 new field 13:21 our new image is blurry but 13:25 it is also not on the same length scale 13:27 as before 13:29 yeah so it's just blurry but what we now 13:32 have to do 13:33 is to rescale length 13:36 to make the structures that we have in 13:38 this new image 13:40 similar to the structures that we had in 13:42 the original image 13:43 now so that means we need to rescale 13:45 length scales 13:48 by a factor of b 13:52 and the other thing that happens if you 13:53 make things blurry 13:55 is that you lose contrast in the image 13:58 so the image looks a bit dull so we 14:02 also need to now to increase the 14:03 contrast again 14:05 and we do that by rescaling our 14:09 fields as well so these are the three 14:12 steps of renormalization as if we first 14:14 integrate out very short 14:18 fluctuations happening on the smallest 14:20 length scales 14:22 and the second step and the third step 14:25 make sure that once we have integrated 14:28 out 14:29 these uh these short length skills 14:32 that our new theory that we get is 14:34 actually comparable 14:36 to the previous one so we have to reset 14:39 the length scales and we have to reset 14:41 the fields the contrast of the field by 14:44 multiplying them 14:45 with appropriate numbers

slide 8

14:50 so how does that look like so we start 14:52 with a field file 14:53 now that has short range fluctuations 14:56 you know fast fluctuations in space 14:58 now for example um 15:02 yeah here these wobbly things yeah these 15:04 are fast fluctuations 15:07 but it also has slow fluctuations that 15:09 would be 15:11 that would be here a slow fluctuation 15:15 now you have fluctuations on all 15:17 different length scales and now 15:19 we cause grain that means we integrate 15:21 out 15:22 these fast fluctuations and what we get 15:25 is a new field if i uh this phi 15:29 average here because phi r 15:33 that is smooth that doesn't have these 15:35 small 15:36 wiggles anymore but it is smooth here 15:41 now so we have course grade the fields 15:44 and now we have coarse grains the field 15:46 we reset the length 15:48 rescale the length scale 15:51 x prime to make fluctuations on the new 15:57 field comparable to typical fluctuations 16:00 on the 16:00 original field the second step 16:05 we then renormalize um 16:08 there's actually no three here because i 16:11 started 16:12 i know there is a three okay so uh 16:16 as a final step and we have to rescale 16:18 the fields now we have to 16:20 change the magnitude of the signal here 16:23 to make it comparable 16:24 to the original signal over here 16:28 now this sounds very intuitive 16:31 and very simple but of course in reality 16:34 it's quite difficult

answering a question

16:39 so let's see how this works uh 16:42 what's the result so can i have a 16:45 question here 16:50 um in the previous slide 16:54 when we re-normalize uh fi 16:57 prime we essentially doing it 17:01 because we want uh whatever abs 17:04 magnitude or value of phi prime is to 17:07 equal 17:08 uh the value of five the original field 17:11 yes 17:11 exactly so what we want to do is so so 17:14 we do this procedure not only once 17:16 but many times now the next step will do 17:18 that many times 17:20 and uh each time we do these three steps 17:24 we get a new hamiltonian or a new action 17:28 and this new action will be 17:31 different to the original action 17:35 it will be different for trivial reasons 17:38 namely because once we cause when 17:41 think about this course grading as an 17:43 average instead 17:44 now you have i'll show you later a 17:47 specific example suppose you average 17:49 uh in some area here and that's why 17:52 that's how you smooth 17:54 yeah when you average you know think 17:56 about the spin system you average 17:59 then you don't have plus minus plus one 18:01 or minus one 18:02 but the new average field will be plus 18:04 0.1 18:06 and minus 0.1 now just because you 18:08 averaged 18:09 over in the field yeah but you don't 18:11 want the new spins to live in this world 18:13 of plus 18:14 0.1 and minus 0.1 but you'd want them 18:18 also 18:18 live on the scale of plus minus 1 just 18:21 to make the 18:21 hamiltonians comparable now so this is a 18:24 trivial effect that you get by course 18:26 grading that you don't want to have 18:28 to dominate your results and we need to 18:31 get rid of that 18:32 this trivial rescalings of the fields 18:35 and of the 18:36 of the um of the length scales just by 18:39 explicitly taking the step and saying 18:41 okay 18:42 now i have co-strained my field now i 18:45 have to reset the length scales 18:46 and i have to reset the amplitude of my 18:49 fields 18:50 by multiplying these these quantities 18:53 with appropriate numbers 18:56 i just want to have comparable things at 18:58 each step 19:00 yes thank you so much uh but regarding 19:03 this particular point 19:04 you showed in the first slide that phi 19:07 prime 19:07 phi is divided by b uh in the 19:10 in introductory slide remember 19:12 normalization i couldn't understand that 19:15 yes here yeah five in the next one 19:19 five bar by b or um 19:23 so which which one third point 19:26 third point yeah that's just a number 19:28 yeah we don't know this number yet 19:31 we don't know it yet we can get it by 19:33 dimensional analysis 19:34 uh very often um so we don't know this 19:37 number yet 19:38 it doesn't the same b okay it's not 19:42 from one i think it's an s actually 19:47 so so it doesn't have to be b it depends 19:50 on the dimensions of your field 19:53 typically b to the power of something 19:55 also 19:56 that doesn't have to be b and it depends 19:58 on the on the dimensions of your fields 20:00 here 20:01 um okay just at this point we just say 20:03 okay so we have to do something with our 20:05 fields to make them comparable 20:08 now think about an average now you can't 20:10 get average and if you're doing average 20:12 all the time 20:13 yeah then your uh the central limited 20:16 theorem will be that the 20:17 average like in a disordered system will 20:20 get very very small 20:22 yes the variance of this average will 20:24 get very very small 20:26 and we don't want this effect to happen 20:28 because it destroys this basically 20:30 or if i what we actually want to look at 20:33 now what we actually want to look at is 20:34 how does the theory in itself the 20:36 structure of the theory 20:38 change as we go through the scale and we 20:41 don't want to have these effects that 20:42 come 20:43 by uh by uh 20:46 just that we that we can't compare 20:49 um suppose you compare like uh uh 20:53 you compare um the velocity of a car 20:56 you live in the uk also or in the united 20:57 states and you compare the velocity of 20:59 the car in miles per hour 21:01 or in kilometers per hour now you have 21:04 to you cannot compare that 21:05 but pure numbers you have to do 21:07 something you have to be scared by 1.6 21:09 to make them comparable and here also we 21:12 go through these scales look 21:13 like meters uh kilo kilometers miles and 21:17 so on 21:18 and to make these things comparable all 21:20 the different lengths that we have to 21:21 rescale 21:22 them all the time yeah just that way 21:24 that we're all the way talking about 21:26 um the same thing um 21:30 so in germany we say we uh you cannot 21:32 compare apples and oranges 21:34 different things you have to uh 21:37 if you compare apple and origins then um 21:40 then then you're doing something wrong 21:43 in other words to compare apples 21:45 to different kinds of apples so to say 21:48 yes but we always want to talk about 21:50 apples and not about miles and 21:52 kilometers now so that's that's the idea 21:55 about this rescaling step 21:59 i'll show you later an example where 22:00 this rescaling step is already implicit 22:02 of course you can choose this course 22:06 grading step in a way 22:07 that it doesn't change the magnitude of 22:10 the field 22:12 yeah so that you can choose the course 22:14 grading step in a way here this one 22:16 in a way that it doesn't change the 22:18 magnitude of this field and then you 22:19 don't have to do this renormalization 22:21 step 22:24 now but the principle you will have to 22:26 do that 22:28 okay so now we have these three points 22:31 now we rescale uh we of course grain

slide 9

22:35 rescale and renormalize and once we do 22:39 that 22:39 our action or our equilibrium or 22:41 hamiltonian 22:43 will become a new action 22:48 now s prime and say we 22:52 did this rescaling on a very small 22:54 length scale 22:55 dl now this s prime 23:00 is then given by some operator r 23:03 of s and if we do that repeatedly 23:08 you know so what we then get is a 23:11 remoralization group flow our g 23:13 flow and that's basically 23:17 the differential equation for the action 23:21 yeah the s over the l 23:26 is then something like this r 23:29 of s so we normalize one step further 23:33 minus the previous step 23:38 you know so we change we do these three 23:41 steps 23:41 just a little bit not just and we call 23:44 square 23:45 we integrate out a very small additional 23:48 scale 23:50 yeah and our action is then different on 23:53 the next scale 23:54 of course we assume that this is somehow 23:55 continuous and well behaved 23:58 and then we'll have a flow equation 24:01 of our action and of course in reality 24:04 this will not be a flow equation of our 24:06 action 24:06 but of the parameters that define our 24:09 action 24:10 now suppose so how does it look like 24:18 now so here is in some space of all 24:21 possible actions 24:23 of some parameters p1 24:27 p2 p3 24:32 now and this action now think about the 24:34 uh 24:35 think about our non-linear dynamics or 24:37 dynamical systems lecture 24:39 what we can this is this is a 24:41 differential equation here 24:44 it looks complicated but this is a 24:46 differential equation it tells you 24:47 how the parameters of our action 24:51 change as we cause gradients we do this 24:54 renormalization procedure 24:57 now and this gives us a differential 24:59 equation 25:01 and if you have a differential equation 25:02 that will be highly non-linear of course 25:05 now we can do the same thing as we did 25:07 in this lecture or nonlinear system we 25:09 can 25:10 derive a phase portrait now in the space 25:12 of all possible actions 25:14 of all possible models where does this 25:17 minimalization group 25:18 group flow carry us 25:22 so let's start with a very simple 25:27 line 25:31 that's the line in the space of all 25:33 possible models 25:34 now this is the line of models 25:37 that actually describe our physical 25:39 systems 25:40 now think about different combinations 25:42 of temperature 25:43 and magnetic fields in the icing model 25:48 now so this is this is where this model 25:50 lives in this space 25:52 now if we take the right parameter 25:55 combinations 25:58 we will be at some critical point 26:06 yeah so there's no flow yet now so this 26:09 is just 26:10 uh the the the the range of different 26:13 models that we can have for example in 26:15 eisenhower that correspond to some real 26:17 physical system 26:18 but there of course there are many other 26:20 model in the models in this space 26:22 that don't describe our physical system 26:24 now they don't describe a magnet but 26:26 something else or that there are not 26:28 given by a simple hamiltonian with 26:29 nearest neighbors 26:30 interactions but by something that has 26:33 long-range interactions or something 26:35 very weird 26:36 now there's a space of all possible 26:37 models is very large 26:40 now we have this critical point and 26:43 other models in this space also have 26:46 critical points 26:47 now and these critical points live on a 26:51 manifold 26:58 now they leave on a manifold you know 27:01 that is the critical manifold 27:09 you know so that's all the points of 27:11 this critical manifolds are critical 27:13 points 27:14 of some actions 27:17 and our action our critical point of our 27:20 action 27:21 is also on this manifold but all 27:24 other points of this manifolds are also 27:26 some critical points of some other 27:29 actions so 27:33 what happens now now 27:37 we are close to the critical point let's 27:39 say 27:40 we're here now very close to the 27:44 critical point 27:46 and now we really normalize 27:50 we go through we make this procedure of 27:53 renormalizing 27:54 the core straining going to larger and 27:56 larger scales and rescaling our fields 27:59 and lengths 28:00 and then our anatomy or our action will 28:03 change 28:04 so it will flow in this space 28:07 in some direction 28:11 so where does it flow to in the 28:13 non-dynamics lecture 28:16 we've seen that what determines such 28:18 dynamical systems 28:20 are fixed points and 28:24 what you typically have to assume in the 28:25 real normalization group theory 28:28 that there is some fixed point 28:31 on this critical manifold yeah 28:36 the fixed point of the realization would 28:39 flow 28:40 on this critical manifold 28:44 and now what happens with our flow 28:50 of course we will go to this fixed point 28:55 and then we might go away again 28:59 now the sixth point the fixed points 29:02 like in the dynamic resistance lectures 29:04 determine the flow of our dynamical 29:07 system 29:09 now what is this fixed point here this 29:11 fixed point is not the critical point 29:14 it's the critical point of some other 29:16 model 29:18 but this critical point here 29:23 has stability yeah just like in normal 29:25 domains so this is non-linear dynamics 29:27 here 29:28 so it's very often exactly what we did 29:30 in this lectures before 29:31 so we asked what is the flow now we ask 29:33 them to ask about the stability 29:35 of this fixed point now so this fixed 29:39 point 29:40 has stable directions now you perturb 29:44 and you get pushed out and unstable 29:46 directions 29:48 yeah typically or by definition 29:52 the directions on the critical manifolds 29:55 are stable now that's how this manifold 29:59 is actually defined 30:00 you can and there's a theory and only 30:02 resistance 30:04 that tells you that and there are also 30:07 other directions 30:09 that are not stable now let's have them 30:12 in green 30:14 for example the way i drew this these 30:17 are the ones 30:19 in this direction 30:26 so this determines the stability of the 30:28 flow of our fix 30:29 of our our system and if we have only 30:32 one fixed point this one fixed point 30:35 will tell us what happens to the flow of 30:37 our system just like in nonlinear 30:39 dynamics 30:40 lecture okay so 30:43 now we re-normalize 30:46 it because we didn't start exactly at 30:48 the critical point we stay in the 30:50 vicinity of this manifold here 30:52 here and this critic this fixed point of 30:55 the flow will do 30:56 something to us no it will push us away 30:59 or will attract us and now you can 31:04 do the same thing that you do in 31:05 learning your dynamics namely you linear 31:07 wise 31:08 around this fixed point so we say that 31:11 our action 31:13 [Music] 31:16 called linear stability 31:21 you linearize around this fixed point so 31:24 you say 31:25 that our action is equal to the action 31:29 at this fixed point 31:32 plus sum over different 31:35 directions i who operate is i 31:39 h i 31:45 times b to the power of lambda i 31:52 times 31:56 q i yeah so this here are 32:01 these two eyes are the eigen directions 32:08 so these are operators you can think 32:09 about this as operators so like eigen 32:12 direction of these operators and 32:16 the b is our course grading scale 32:30 the h tells us 32:34 how far we are away from the fixed point 32:42 and the s tells us uh is just the action 32:48 that we have at this fixed point 32:51 you know so we it's the same thing for 32:53 all of this was once we have once we 32:55 have this mineralization group flow 32:58 we're in the subject of non-linear 32:59 dynamics and we use the tools 33:02 of nonlinear dynamics renalization group 33:05 theory this language is slightly 33:09 different 33:10 you know so that's that's why you have 33:12 this b to the lambdas and so on 33:14 here and you separate the h i from the b 33:17 to the lambda 33:18 that's just the framework the how you 33:20 write it in the realization 33:23 theory for convenience reasons but what 33:26 we do here is 33:27 a simple linear stability analysis 33:30 of a non-linear system i will 33:34 re-linearize around the fixed point 33:36 we see how whether perturbations grow or 33:39 shrink 33:40 in different directions yeah and that 33:43 characterizes 33:44 then our nonlinear system and now we can 33:47 ask if we perturb around this fixed 33:50 point 33:51 in one direction i does it grow 33:54 this perturbation or does it shrink 33:57 lambda i 33:58 is larger than zero now this bi 34:03 is larger than one or b is larger than 34:07 one 34:08 now so lambda i is larger than zero then 34:11 our perturbation will grow 34:13 yeah perturbation 34:20 growth and then we say this direction or 34:23 this operator q 34:24 i you can also think about q i as one of 34:27 the 34:28 terms in the action now think about one 34:31 of the terms in the action 34:32 or one of the terms in the hamiltonian 34:36 this direction qi 34:40 is then called relevant 34:45 why is this relevant what we call when 34:46 do we call this relevance 34:48 you know if this perturbation grows we 34:51 are in this green direction here that 34:53 pushes us away from the critical point 34:56 so that means 34:56 that if we are an experimentalist 35:00 and if we want to tune our system 35:03 to get into the critical point 35:07 then 35:10 then we know that we have to turn 35:13 these relevant parameters qi 35:18 now this relevant parameter for which 35:19 this uh that are 35:21 unstable directions of this fixed point 35:25 now and if we have a relevant directions 35:28 we also have irrelevant directions 35:30 lambda is smaller than zero 35:32 and these directions are called 35:34 irrelevant 35:35 now perturbation 35:43 shrinks that means that 35:47 qi is irrelevant 35:53 that means the qi in this case is not a 35:56 parameter that drives us into this 35:58 critical point 36:01 and then we have the case that lambda i 36:03 is exactly equal to zero 36:05 then we don't really know what to do 36:08 then this q 36:09 i is 36:12 marginal and we cannot tell from this 36:15 linear stability analysis alone 36:17 from the linear realization around this 36:19 fixed bond we cannot tell alone 36:21 whether this perturbation will grow or 36:24 shrink and we have to use 36:25 other methods okay 36:28 so what happened now so we 36:31 started our theory close to the critical 36:34 point 36:35 we did this rememberization group 36:37 procedure 36:39 course grading rescaling renormalization 36:43 and then in this procedure 36:46 our action or our model will flow 36:49 through the space of all possible models 36:52 yeah 36:55 and then we ask where does it flow to 37:00 and we look at the non-linear dynamics 37:01 lecture and ask 37:03 so where does such a nonlinear system 37:05 drive us to 37:07 and our nonlinear dynamics lecture will 37:09 tell us look at the fixed points 37:12 right and in this realization flow you 37:14 also have fixed points you assume that 37:15 you have a fixed point 37:17 and this fixed point tells us about 37:21 what is going on on the macroscopic 37:24 scale now that determines this fixed 37:26 point 37:26 determines which is the end result of 37:28 our minimization group 37:30 procedure 37:33 so at this fixed point is characterized 37:36 by stability 37:38 it has a finite number of relevant 37:40 directions 37:43 now it's had a finite number of 37:44 parameters that are actually important 37:47 to change if you want to go to the 37:48 critical point 37:51 and because you only have a finite 37:53 number of directions that are relevant 37:56 you usually get away with models that 37:58 also have this very finite number 38:01 of parameters instead of models like a 38:04 bigger magnet or so that is a very 38:06 complicated geometry and everything 38:08 there are instead of model that has 15 38:10 000 parameters 38:12 now you get away with a finite number of 38:13 parameters that are given 38:15 by the relevant eigen directions 38:19 of this fixed point now so now what 38:22 happens if we have a different model so 38:24 now this is now the 38:25 magnet one that this is not a 38:29 magnet or material one we can also look 38:31 at a magnet at another of another 38:33 material 38:34 now so this magnet one 38:40 and now we have here 38:47 a magnitude 38:51 now each of them at the physical real 38:53 microscopic level is destroyed by 15 000 38:56 parameters or whatever something very 38:57 complicated 38:59 and for this magnitude we can do the 39:02 same procedure 39:04 we renormalize 39:07 and we while we renormalize we will end 39:10 up at the vicinity of this fixed point 39:14 and in the vicinity of the fixed point 39:17 the behavior of the flow is determined 39:20 by a finite number 39:22 of parameters again 39:26 and both of these magnets here are 39:29 under renormalization now on the large 39:31 scale repeat these 39:33 procedures determined by the same fixed 39:37 point 39:37 about the same point but this one here 39:41 and because they're determined by the 39:42 same fixed point with the same stability 39:46 and with the same properties of how they 39:48 go 39:49 uh of how the flow behaves around this 39:51 fixed point 39:53 that's why these two magnets here are 39:55 described macroscopically by the same 39:57 theory 39:58 and that's then the reason why we have 40:01 universality 40:03 so in this way in this very general way 40:05 so we look of course in our in 40:07 more detail the renal isolation group 40:10 theory gives us a justification for why 40:14 only a finite number of parameters 40:17 matter 40:17 on the or finite a limited level of 40:21 description is sufficient to describe 40:23 large scale properties of a large 40:26 number of very different systems and the 40:29 reason is 40:30 at this critical point they're 40:33 described macroscopically by the same 40:36 fixed point of the renewalization 40:39 workflow

slide 10

40:42 now how does this look in detail 40:47 so the very first or the 40:50 most uh simple way of doing 40:53 remoralization 40:55 is to take what i said initially about 40:58 this 40:59 defocusing about this course training 41:01 literally 41:02 and do the whole procedure in real space 41:06 now suppose you have here a lattice 41:09 system 41:11 and also suppose you have this lattice 41:12 system here and you have some 41:15 spins here and what you can do then to 41:19 cause grain 41:20 is to 41:24 you know what you can do to coarse grain 41:26 is to create 41:27 boxes or blocks now of a certain size 41:31 and then to calculate this cross here 41:34 that is a representation of all of these 41:37 microscopic spin for each block 41:40 yeah so we have this uh spins here 41:44 like what are 16 in each block and you 41:48 now transform them to a single number 41:50 you can do that by averaging over them 41:52 or you can take 41:54 you can say that i take the spin 41:58 that is the majority of these other 42:01 spins here 42:02 if the majority goes up then my new spin 42:06 that describes the entire block 42:07 will also go up and this would be a way 42:11 to get a rid of this renormalization 42:14 step if you say i take the majority 42:18 i take a majority rule so this new spin 42:20 here x 42:21 will take the value of the majority of 42:24 the original spins 42:27 then the new spin will also be plus or 42:29 minus one 42:31 if my new spin yeah and i don't have to 42:33 renormalize them because it has the same 42:35 values as the original splits 42:38 if i take the new spin as the average 42:43 over all of these spins here then 42:47 i typically get a very small number the 42:49 average won't be one or minus one but it 42:51 will be 42:52 0.3 or 0.1 or 0.5 or so but it will not 42:57 be 42:57 one or minus one in most cases yeah so 43:00 in this case if i perform this procedure 43:02 i would have to renormalize 43:04 you know and i have to would have to 43:05 rescale my fields 43:08 to make them comparable to the original 43:10 step 43:12 yeah so now we have these blocks 43:17 and we define some new spin that 43:20 describes each of these blocks 43:23 and now we write down a new model 43:26 a new hamiltonian for these new spins 43:29 here 43:31 and what we hope is that the spins that 43:35 we have here 43:38 in this new system this course-grade 43:40 system are described by a theory 43:43 that is structurally very similar to the 43:45 original theory 43:48 and this hope is actually justified by 43:52 the observation of 43:55 scale and variance now so if your system 43:58 is scaled in variance we can hope that 43:59 if we zoom out 44:01 and our system is statistically the same 44:04 then then our partition function 44:06 or our action will also be the same 44:09 just with some different parameters now 44:11 that is the hope that is underlying 44:13 renewalization group procedures with 44:16 these 44:16 block spins here what you typically get 44:19 is that you get 44:20 higher order turns all the time you know 44:23 so that's this hope is not 44:24 mathematically super precise uh but 44:28 that's what you have to assume in order 44:30 to achieve anything 44:33 okay yeah 44:36 okay so we call screen and then 44:40 we rescale the second step so that the 44:42 distance between these 44:44 spins here the new spins is the same 44:47 distance as we had 44:49 between the original spins now that's 44:52 what we have to do anyway 44:53 and that's why how we divide length 44:55 scales 44:57 by the same factor that corresponds to 45:00 the size of our boxes 45:02 yeah and now the lengths are the same as 45:04 before

slide 11

45:09 so let's do this procedure in a very 45:13 simple case which is the 1d 45:16 ising model 45:21 now so the one deising model now is 45:24 written the tradition function 45:26 it can be written in like a long form 45:30 in this way here that i sum up 45:33 all combinations of nearest neighbor 45:36 interactions 45:38 now that's just the hamiltonian here of 45:40 the ising model without an external 45:42 field 45:43 now and then i have to have a sum about 45:45 all possible values of the sigma i's 45:48 of the of the that my fee that my all 45:51 the possible values 45:52 that the sigma can take that gives me my 45:54 partition function 45:57 now what i do know is the first 46:01 course grading stuff now this first 46:04 coast grading step 46:06 means that these red spins here 46:09 like with the all these these black 46:11 spins are the white splits 46:13 now i'll integrate out the white spins 46:17 here in this picture i'll integrate 46:20 these ones out 46:23 now every second spin all even spins 46:26 and if i do that i will get a new theory 46:30 that is described by interactions 46:32 between these 46:33 uneven spins that are and these 46:36 interactions occurred to us are 46:37 depicted in these red with these red 46:41 lines 46:44 okay so and the stars is actually very 46:47 simple activated talks 46:49 so that's so for if we just do it for 46:56 for sigma 2 now we just sum 46:59 out we take the terms that correspond to 47:02 sigma 2 47:04 and we get that that is equal to 47:08 not many terms and then we have the 47:11 contribution from sigma 2 47:14 uh k sigma 1 47:17 plus sigma 3 now that's what what's left 47:20 plus e to the minus k sigma 1 47:25 sigma 3 47:28 and then all the rest e to the k 47:32 sigma 3 sigma 47:36 4 plus four sigma 47:40 five and all the other splits 47:44 so what i just that said is that sigma 47:46 two can have two values 47:48 minus one plus one yeah and i just 47:51 substituted i expected 47:53 explicitly now set sigma 2 to plus 1 and 47:56 minus 1 47:57 and perform the sum and that's what i 47:59 get then here for this first 48:01 term and now i can do that for all 48:12 even sigmas and then what i get is 48:16 exactly the same thing sum over 48:20 many terms e to the k 48:23 sigma 1 plus sigma 3 as before 48:27 plus e to the minus k sigma 1 48:31 sigma 3. now that was the original one 48:34 where we set sigma 48:36 2 to -1 and then we get the same thing 48:42 e to the k for the next 48:45 term now for the next interaction sigma 48:48 3 48:49 plus sigma 5 48:54 e to the minus k sigma 3 48:58 sigma 5 49:01 and so on yeah 49:07 for all the other terms yeah

slide 12

49:13 so now the idea is 49:18 that because when we are at a critical 49:21 state 49:23 that we expect our partition function to 49:25 be self-similar when we call screen 49:28 the statistics of the system remains the 49:31 same as we zoom out 49:33 and that's why we also expect 49:36 the quantity that gives us the statistic 49:40 statistics the partition function to be 49:42 self-similar as well 49:45 and what we now do is that we find 49:49 a new value of k prime 49:53 and some function f 49:56 of k that tells us that we 49:59 that these terms that we got 50:02 here sigma 1 sigma 3 we always got the 50:06 sum for each coupling 50:08 they should take the same form as the 50:11 original hamiltonian but with some 50:14 pre-factor here 50:17 and some new coupling here 50:20 but the form should be the same as 50:22 before 50:24 now as i said this is not usually 50:27 well justified but we have to do that in 50:29 order to do anything 50:31 and if you require that if you do some 50:33 algebra you will find that if you set 50:36 k prime to this one one half 50:39 logarithm hyperbolic cosine of 2k 50:44 and the function f of k to this year 50:48 then it fulfills this condition yeah 50:54 so now we can plug this in now so if we 50:59 if we use that 51:01 then our hammer tilt our partition 51:05 function 51:06 will read again we have many 51:10 terms f of k 51:14 e to the k prime 51:17 sigma 1 sigma 3 51:21 f of k e to the k 51:24 prime sigma 3 sigma 5 51:29 and so on 51:32 now and this is just the same we can 51:35 write this now 51:36 as a new partition function that has a 51:40 new prefactor 51:42 f of k now we pull this out this 51:45 prefactor 51:46 f of k and we have that n over two times 51:51 times a new partition function that 51:55 depends on the new system size 52:00 and a new coupling k prime 52:04 yeah so we have this down this one 52:06 renormalization step 52:08 we get a new partition function that 52:10 looks exactly the same as the old one 52:12 in structure but we have a new coupling 52:15 k prime 52:17 and a pre-factor here that is this 52:20 function 52:21 f and that also depends on the coupling 52:25 also what we did here is that we now 52:27 have 52:31 a relationship between the partition 52:34 functions 52:35 at different stages of the 52:38 renewalization procedure 52:45 yeah and now 52:50 what does it mean look at 52:53 this one here k prime 52:57 this is already a description 53:02 now this is already a description of how 53:04 our coupling 53:05 one parameter k prime 53:09 depends on the value of this okay now 53:12 how this parameter k 53:13 evolves in this course grading procedure 53:17 in this renewalization procedure 53:20 you know so this k that gets updated 53:23 now it's not in differential form yeah 53:26 like in 53:26 like uh like we did in the non-linear 53:28 dynamics 53:29 actually but it's in this uh other way 53:32 that you can describe non-linear 53:34 systems but by iterative updating now so 53:38 the new value of k 53:39 prime is given by the old value 53:43 is by this function here applied to the 53:45 old value 53:47 and now now is this is this updating 53:50 scheme here 53:52 and we can expand this term here the 53:54 logarithm 53:55 of the hyperbolic cosine and so for low 53:58 values of this coupling 54:00 this goes with k squared now so normally 54:03 i have an idea 54:05 about how this looks like i have already 54:08 prepared this 54:09 very nice and we can now solve this 54:11 equation here 54:13 graphically so we want to get the flow 54:15 and we can solve this graphically 54:17 and see where this linearization 54:19 procedure 54:20 carries our k our coupling k 54:23 now and because our partition function 54:27 remains invariant

slide 13

54:30 well our k the update of our k 54:35 describes actually the behavior of our 54:37 hamiltonian 54:38 under renormalization 54:41 okay so this is the plot here so what i 54:44 what you do 54:44 is that you plot the left hand side 54:48 of this equation k prime it's just 54:51 linear with slope one 54:54 and uh the right hand side 54:58 now this is uh this here 55:02 and where the left hand side is equal to 55:04 the right hand side there you have a 55:06 fixed point 55:07 now so this is here the way you need to 55:10 read this this is the next value of the 55:11 realization 55:12 this is the previous one if you start 55:14 here we'll go 55:15 here then here and here and here 55:19 and at some point these two lines meet 55:22 and that's 55:22 that's where your fixed point is so this 55:25 humanization procedure 55:27 will bring us to some fixed point which 55:29 happens 55:30 to be uh down here at zero 55:36 yeah and then we can look at how this 55:39 uh we can also then plot 55:42 a flow diagram as i did in a more 55:45 complicated way before 55:47 and how we also did it in the actual 55:49 nonlinear dynamics lecture 55:51 now we can plot it on this in this 55:53 one-dimensional line 55:55 then we have a stable fixed point at 55:57 zero 55:58 now that's here stable 56:02 and any value where we start with our 56:04 course grading procedure 56:06 will be driven to a value of k equals 56:09 zero 56:10 now we start with a very strong coupling 56:12 we renormalize 56:14 and we will be driven to zero 56:17 that means that this renormalization 56:21 procedure this course grading procedure 56:24 in this one-dimensional icing model 56:28 will always on the macroscopic skin on 56:31 large 56:31 length scales always lead to a model 56:35 that is effectively described by 56:40 a system that has zero coupling 56:43 yeah this coupling here vanishes and if 56:46 we have zero coupling that means that 56:48 our system 56:49 is above the critical point that's 56:52 non-critical 56:54 so we will normalize we go on and on 56:57 and we always end up on the system that 56:59 has very high temperature 57:01 or very low coupling yeah that's a it is 57:04 a disordered system 57:06 and that's just a reflection of the fact 57:07 that the one deicing model 57:10 doesn't have any order for a finite 57:12 temperature 57:15 now so you have to start with coupling 57:18 exactly equal to infinity 57:21 to get an order over the temperature 57:24 exactly 57:24 equal to zero only then you can have 57:26 order everything else 57:28 will drive you to this fixed point here 57:32 that corresponds to a system where you 57:35 have no 57:36 coupling at all now so the one that we 57:39 knew that already that the 1d 57:40 system doesn't show all in the ising 57:42 model doesn't have order 57:44 it doesn't have a really critical point 57:47 and that's why our flow 57:48 tells us that macroscopic scales this 57:51 system 57:51 goes to a system that doesn't have 57:55 any interaction so it's completely 57:56 disordered

slide 14

57:59 of course you can do the same procedure 58:01 for um 58:04 for the 2d system yeah and there is of 58:06 course 58:07 again much more complicated then this 2d 58:10 system 58:12 you get a flow diagram that looks like 58:15 this here 58:17 on the bottom so here you suddenly 58:21 have another fixed point an unstable 58:23 fixed point 58:25 in between these two extremes now this 58:28 unstable fixed point 58:30 here if you start to the left of this 58:33 unstable fixed point you were driven to 58:36 a state 58:37 without that that we had previously 58:40 where the coupling is very low 58:42 or that corresponds to the system at 58:43 very high temperature 58:45 if you start to the right of this you'll 58:47 be driven to a state 58:49 where you have order you know where your 58:51 coupling is basically infinity or your 58:53 temperature effective temperature 58:55 is zero and because you have now this 58:58 fixed point here this new fixed point 59:00 right you get 59:02 this singularity or this discontinuity 59:05 of the free energy because if you go a 59:06 little bit to the left 59:08 you go to another a different 59:10 macroscopic state 59:11 then if you go a little bit to the right 59:14 and of course you can test that with 59:15 numerical simulations 59:17 now so that's here from the book of 59:19 cardi which is a very very nice book um 59:22 scaling and renormalization and 59:23 statistical physics 59:26 and i have to say that 59:29 the class don't look very good on this 59:32 ipad 59:34 so what you see here are just 59:36 simulations of the two the eisenmann 59:39 and what they did is they performed one 59:42 block spin removalization procedure 59:45 that's that's what we did right now 59:47 that's the coarse grain so one step of 59:50 this coarse graining 59:52 and uh if you are right at the critical 59:55 point on the left hand side 59:57 you do the coarse graining step yeah 59:59 then 60:00 the system remains invariant if you 60:02 start in a fixed point you'll stay there 60:05 if you start a little bit this course 60:07 grading procedure 60:08 and here that's not there's nothing 60:10 fancier they just took the simulations 60:13 and they did one course grading step 60:15 they averaged you have maybe over a 60:17 block of spins or so 60:19 and if you are above this 60:23 critical temperature that means you 60:24 start here on the left of this fixed 60:26 point 60:27 and if you do this course grading step 60:29 then your system looks more disordered 60:31 than before 60:33 yeah so this year it looks like a higher 60:35 temperature than this year because you 60:37 have a lot of these small domains 60:39 now this is just a reflection of this 60:41 stat like an intuitive 60:43 picture of how you go in this 60:45 reunionization procedure 60:47 to the left to a state that has no 60:49 coupling at all 60:50 k equals zero and if you're coupling a 60:53 zero that's the same as when your 60:54 temperature 60:55 is very large and if you want to read 60:58 more about this 60:59 have a look at the book of john carty 61:01 and there's also of what i just showed 61:03 you 61:04 there is a there is a nice book by um 61:11 so there's a nice book there's a nice 61:13 article by karanov 61:15 um by teaching the immunization group 61:17 and he does these calculations also for 61:19 the 2d 61:20 model okay so now 61:23 we state in a real space yeah and in 61:26 real space 61:28 uh is very intuitive now 61:31 and it works for the one deising model 61:33 for the 2d eisen model gets already 61:34 complicated 61:36 and it's basically impractical to do 61:38 that 61:39 for general um 61:43 for for general physical models now it 61:46 gets very complicated to do that 61:47 procedure 61:48 in real space and the reason is that 61:50 there's no small parameter involved 61:53 that you can use for an expansion

slide 15

61:57 then there was another guy called wilson 61:59 who came up with another idea 62:01 now that was actually and that's called 62:03 the wilson 62:05 momentum shell idea 62:09 that's the world's in momentum shell 62:11 idea so what does it mean so what was 62:14 wilson's idea was that we caused grain 62:18 by integrating out fast degrees of 62:21 freedom 62:22 or degrees of freedom that have a very 62:24 short wavelength 62:26 in fourier space that's the way you do 62:29 that 62:30 is uh you look at free space also this 62:33 is our 62:34 free space let's say we have two 62:35 directions in free space 62:38 then we have here 62:44 a maximum wave vector that's the maximum 62:52 wave vector and this wave vector let's 62:55 call it 62:56 capital omega is the same 63:00 just given by that's the smallest 63:02 structure we can have in the system 63:04 that's a microscopic length scale 63:06 yeah and that's in these lattice systems 63:08 this typical 63:09 uh one over the the lattice the 63:12 microscopic level the lattice spacing 63:15 yeah so we cannot go any smaller than 63:17 that 63:19 now starting from the smallest length 63:21 scale now so a description of our system 63:24 on the smallest length scale 63:27 we now integrate out the blue stuff here 63:32 that's this one here that's the momentum 63:35 shell 63:42 and we integrate out this momentum shell 63:45 until we reach 63:48 a new wavelength a new y vector of a new 63:51 length scale 63:54 omega prime is equal to the original 63:57 omega 63:58 divided sum by some number lambda 64:01 yeah so 64:05 we integrate out one bit in momentum 64:08 space 64:10 at a time and that means that we perform 64:13 an integration 64:15 uh on a momentum shell on a tiny shell 64:18 in momentum space and also our new field 64:26 in the momentum shell 64:32 is then called typically something like 64:35 feel 64:36 find larger of q 64:40 and this is just undefined by 64:43 phi of q 64:46 with q in this interval 64:51 omega over lambda to omega 64:56 yeah so we integrate 65:00 out one step and now wilson's scheme is 65:03 actually very similar 65:08 to what we've done already yes the first 65:10 step will be 65:12 recall screen 65:18 now by rescaling oops sorry 65:29 no by rescaling 65:34 and that will give rise to some 65:38 change in the coefficients 65:46 in the action 65:49 and the second step is that we perform 65:56 this integration in the momentum shell 66:00 integrate out 66:04 short range 66:08 fluctuations 66:12 or momentum 66:19 a second step and we'll in the next 66:22 lecture 66:22 we do exactly this yeah we performed 66:24 this wilson's ruralization group 66:26 procedure that is much more practical 66:29 than the 66:29 block spin immunization group that we 66:32 had in the 66:32 beginning of this lecture and the good 66:36 thing about this wilson's 66:38 of wilson's idea is that it actually has 66:40 a small parameter 66:41 this momentum shell is very small 66:44 and uh yes so this 66:48 has a small parameter that means that we 66:50 can actually then 66:51 hope to get some approximative um 66:55 uh scheme out of this approximative 67:00 so that we can approximate our integrals 67:02 that we get from the course grade 67:04 yeah so we'll do that uh next week 67:07 for our little epidemic model and we'll 67:10 derive 67:11 the renalization group flow from our 67:13 equity epidemic model 67:15 and from this flow we'll then get the 67:17 exponents that derive that describe 67:20 the action of the the the behavior 67:23 of this epidemic model near the critical 67:26 point 67:27 yeah and um 67:30 exactly yeah so so that's what we'll do 67:33 i'll 67:33 just leave it for here today because and 67:35 then next week we do the calculation and 67:37 if you're 67:38 not interested in calculating that 67:41 because it's so 67:42 uh so uh 67:46 so uh short before christmas if you're 67:48 free to skip the next lecture 67:50 yeah and uh officially i think it's not 67:53 a lecture data 67:56 but i wanted to get that done before 67:57 christmas that we can after christmas at 68:00 january what is that fifth or so i can't 68:03 remember 68:03 um we'll start actually done with data 68:06 science and 68:07 to look at some real data yeah okay 68:10 great 68:11 so that was today only the intuitive 68:13 part about romanization so next week 68:15 we'll do 68:15 we'll see that in action and see how it 68:18 actually works in the non-equilibrium 68:19 system 68:22 bye

answering a question

68:28 excuse me yes 68:31 um could you please explain again why 68:33 were we trying to reach the fixed point 68:35 from the critical point 68:37 so so we're not trying to 68:41 it's just that you assume for this to 68:44 work that there is such a fixed point 68:47 that determines the flow of the 68:50 renormalization group yeah in this 68:53 generality you have to assume that that 68:55 there is 68:56 such a fixed point and from nonlinear 68:58 dynamics dynamical system 69:00 lecture that we have we know that once 69:02 we have such a fixed point 69:04 we basically know already how the system 69:07 behaves also in other parts 69:09 of uh of the faith yeah so that's why 69:12 the system these fixed points are so 69:14 important now we have to assume that 69:16 there exists some verb 69:17 uh but we have to basically for every 69:19 individual model that we look at we have 69:21 to show that they actually 69:23 that they that we have actually 69:25 meaningful 69:27 or moral more than one thing of course 69:31 yeah so but once you have the flow it's 69:33 a problem in normal dynamics 69:35 that's that once you have the flow you 69:36 do what you do in non-aerodynamic 69:38 dynamic dynamics so typically in these 69:40 books of linearization they use a 69:42 different language that's a little bit 69:43 disconnected from this dynamical systems 69:47 field 69:48 no but what you do is you have 69:50 non-linear 69:51 differential equations and then you just 69:53 want to see what happens to these 69:55 nonlinear differential equations 69:57 and then if you ask this question then 70:00 and non-in your system you need to ask 70:01 about the fixed points 70:03 and about the stability of yeah and 70:05 that's why this fixed point 70:07 in the randomization group flow is so 70:10 important 70:11 now if there wasn't any fixed point at 70:13 all yeah then 70:14 it would be a very good system so we 70:16 have to have such a big 70:19 point on the critical manifold for this 70:22 uh for this procedure to work 70:25 now we have to assume at this stage here 70:27 we have to assume 70:29 that it exists and if it exists then it 70:31 will determine 70:32 our flow yeah 70:36 but we don't want to we don't want to do 70:38 so once we have to fix the the system 70:40 the renewalization group will 70:41 automatically 70:42 carry us to the fixed point now we don't 70:44 we don't want to go the system to 70:47 the system to go there if the fixed 70:49 point exists 70:50 we'll have to uh we know that the flow 70:54 will determine we will be determined by 70:56 this 70:58 yeah so that's the that's the that's the 71:01 idea 71:02 but there's non-linear systems that in 71:04 principle has not much to do with 71:06 renormalization 71:07 yeah it is a general property of the 71:10 dynamic 71:14 okay thank you okay great 71:24 any other questions 71:28 um i have a question um so 71:32 does this does this um 71:35 require you to have a very 71:38 good understanding of the microscopic 71:42 um dynamics i guess 71:45 you have to have a model to start with i 71:48 mean 71:48 um but that's kind of what i mean is 71:51 if you have some model that's leaving 71:55 out certain things that are 71:57 that maybe 72:01 you know you thought wasn't so important 72:03 or something like that 72:04 but then as you core scream more and 72:08 more 72:08 like does it like um 72:11 will it give rise to like a different rg 72:14 flow than 72:15 if you had included you know other other 72:19 uh okay i think i think that that 72:21 depends on 72:22 what you leave away and what you leave 72:24 out and what you can use so in principle 72:28 for the usual model including this si 72:30 model that we have here 72:32 that we're looking at and also but also 72:34 for the icing model and 72:36 they they fall a little bit out of the 72:37 blue when somebody presents them to you 72:40 and for example in the lecture now they 72:43 completely make sense that they 72:45 destroyed these systems 72:47 but usually people have used 72:50 renormalization studies 72:52 to show that additional terms don't 72:55 don't matter for these yeah also for 72:57 this active 72:58 uh for these active systems that we 73:00 talked about 73:02 with these aligning uh with these 73:04 aligning directions 73:05 and so on now for these systems i showed 73:08 you briefly a larger way equation and 73:09 people did a 73:10 lot of work to show that this is 73:12 actually the simplest 73:14 uh description that you can have that 73:15 describes this system now because the 73:18 rigid rimmelization they found 73:20 that all other terms so this is the 73:22 minimal set of terms that i need 73:24 to describe still the same 73:26 renewalization 73:27 or still renumerization yeah 73:31 and uh whether you get new terms i 73:34 so i i would i wouldn't expect that if 73:35 you 73:37 i wouldn't expect to get new relevant 73:39 terms 73:40 out of nothing yeah in general you know 73:43 otherwise you could start with just 73:45 nothing at all and see what happens and 73:47 then you get like a theory of everything 73:49 i wouldn't expect these things to pop up 73:52 yeah but of course you get all kind of 73:53 messy things 73:55 now you can have if you start with the 73:57 icing more then in reality then you get 73:59 higher order 74:00 interactions and so on and 74:04 and you have to basically have to get 74:06 rid of them 74:08 okay i think let's uh okay great 74:18 okay so if there are no more questions 74:20 then uh 74:21 see you all next week so as as 74:25 as usual next week i'll uh 74:28 start with a repetition of this and 74:31 explain explain it again before we 74:33 actually do the 74:34 real calculation okay bye see you next 74:40 week 74:45 you

本文收录于以下合集: