3

.mpipks-transcript 07. Renormalization Group | 阿掖山:一个博客

 3 years ago
source link: https://mountaye.github.io/blog/articles/mpipks-non-equilibrium-physics-transcript-07-renormalization
Go to the source link to view the article. You can view the picture content, updated content and better typesetting reading experience. If the link is broken, please click the button below to view the snapshot at that time.
neoserver,ios ssh client

introduction

00:00 to our little uh lecture
00:03 so meanwhile uh most of my group
00:07 are in current time and i’m one of the
00:09 few people left
00:11 and um and ethan detang who used to join
00:15 me here in the lecture hall
00:17 uh has escaped to a safer place
00:20 and so now uh buddy tang is watching the
00:24 chat and
00:25 the participantsness and we’ll let you
00:26 in and
00:29 moderate a little bit the chat and um
00:34 so it’s a good time actually to be
00:35 working on some epidemic models that we
00:37 start
00:38 that we started working on last time and
00:42 uh in particular last time

slide 1

00:46 we were looking at the simplest possible
00:49 epidemic model you might think about
00:52 so let me share the screen just to give
00:54 you a little bit of an
00:56 um
01:01 of reminder
01:10 okay
01:32 okay
01:39 okay here we go so we introduced this
01:43 little epidemic model where you had
01:46 two kinds of people infected one
01:50 and the susceptible ones and the model
01:53 was very simple so these infected ones
01:55 could uh meet up with an affected one uh
01:58 within with a susceptible one
02:00 that’s that maybe they weren’t uh to a
02:01 party or so it just
02:03 used the same train yeah and then the
02:05 susceptible one uh
02:07 got infected and then you have the
02:11 second process
02:12 where if you are infected with a certain
02:15 rate also probability
02:17 per unit of time you get rid of your
02:20 disease
02:21 and you recover and go back to
02:23 susceptible
02:25 you know so you’re a recovered person
02:28 so and then we went one step further and
02:31 we
02:31 put this very simple model on a lattice
02:34 in a spatial context
02:36 the simplest spatial context you can
02:38 think about is just having
02:40 lattice in one dimension yeah
02:43 so suppose that this letter looks a
02:45 little like the train like a train
02:47 and uh so on this little letters you
02:50 react
02:51 basically with your nearest neighbors if
02:53 you if you infect somebody
02:55 you’re in fact your nearest neighbor but
02:57 uh not
02:58 anybody else yeah and then

slide 2

03:02 we saw that such a simple model gave
03:04 rise
03:05 uh to rich phenomenology depending on
03:08 the relation between the rate of
03:10 infection
03:11 and the rate of recovery which was set
03:14 to one
03:15 now if the infection rate was very large
03:18 then
03:18 we got a state where we had a large
03:22 fraction of individuals constantly
03:24 carrying the disease
03:26 and when the rate of infection was very
03:29 low
03:30 then we ended up in a stainless
03:31 so-called absorbing state
03:33 where the disease got extinct
03:37 now in between that we found that there
03:39 is a value of this infection rate
03:41 where infection and recovery just
03:43 balance
03:44 and this state was characterized uh
03:47 very in a very similar way to the ising
03:50 model
03:51 in equilibrium was characterized by
03:54 scale and variance
03:55 that means that these correlation
03:57 lengths that we looked at
03:59 along the space axis
04:03 and along the time axis we had two
04:05 correlation lengths
04:06 they both diverged that means we had
04:09 self-similarity
04:10 in this critical point

slide 3

04:15 yeah and then we went on and we derived
04:17 the launch of the equation
04:19 and from this launch of an equation we

slide 4

04:21 then just wrote down
04:22 the martin citra rose functional
04:26 integral and i’ll show you that later
04:28 when we actually used it
04:30 when we’ll actually use it for the
04:31 realization procedure
04:33 and for once we once we had written down
04:36 these actions at the margin et cetera
04:38 rose function integral
04:40 of course we can go to free of space and
04:42 write down
04:43 this action also in fourier space

slide 5

04:47 now that’s the formalities yeah
04:50 that looks pretty complicated now
04:55 we have a problem right so this looks
04:57 pretty complicated
04:59 and uh we have a situation
05:02 where we have divergences where we our
05:06 correlation length is infinite infinite
05:10 and we don’t really know how to deal
05:12 with this infinite correlation

slide 6

05:16 and this is exactly the situation where
05:19 we have developed in the last
05:21 70 years or so mathematical techniques
05:25 originally in
05:26 quantum field theory that allow us to
05:29 deal with these
05:30 divergences with these infinities
05:33 now we’ve made two observations here so
05:36 one is
05:37 scale scale invariance that i just
05:38 mentioned yeah scale variance
05:40 at this critical point there’s no
05:43 distinguished length scale we zoom
05:45 in and the picture we get is
05:48 statistically
05:49 still the same as the original picture
05:54 the second observation is
05:57 that so and as a result of the scale
05:59 invariance we saw that
06:00 in equilibrium all of these
06:03 thermodynamic quantities
06:05 i like the like the free energy
06:08 the magnetization and icing model obey
06:10 power laws
06:12 as you approach the critical point so
06:14 they diverge with certain power laws
06:17 and the second observation is
06:20 that at this critical point
06:23 our empirical observation is for example
06:26 in the equilibrium but also in
06:27 non-equilibrium
06:29 that there are a lot of different
06:32 real world systems that are described by
06:36 the same
06:38 theory the same simple theory so for
06:42 example
06:43 you have the icing more you know the
06:44 icing model is super simple it’s much
06:46 simpler than actually a real magnet
06:49 nevertheless the icing motor can predict
06:52 exponents
06:54 near the critical point of real magnets
06:58 also of different materials very nicely
07:03 and it even the icing model even can
07:05 predict
07:07 critical phenomena in completely
07:09 unrelated systems
07:12 like phase transitions in water if you
07:14 put water in a high pressure then you
07:15 have a phase where you have
07:17 like a coexistence of water and steam
07:20 and this you can predict these exponents
07:22 you can predict with the icing water
07:24 and this power of such simple models to
07:28 predict a large
07:29 range of critical phenomena is called
07:33 universality and both of these
07:36 observations
07:37 are the scale and variance the cell
07:39 similarity
07:40 and universality can be systematically
07:43 started with a renormalization group
07:46 the first thing you might actually think
07:47 about okay so why do we need actually
07:49 something complicated
07:51 now we just do a naive approach
07:54 uh similar to the one that we used
07:57 implicitly in these lectures about
07:59 pattern formation
08:00 and non-linear dynamics now what we
08:03 could do is okay we say okay so we just
08:06 instantaneously zoom out of the system
08:09 yeah go to
08:10 wave vector zero so in finite wavelength
08:13 you look only at the very large
08:15 properties
08:16 and just pretend that we can average
08:18 over everything we just look at average
08:20 quantities
08:21 that’s that should be called mean field
08:23 theory where you pretend
08:25 that your system the state of the system
08:28 at a certain
08:28 site is governed by
08:32 a field that rises over
08:35 basically an average over the entire
08:37 system
08:38 yeah so this mean field theory where you
08:41 instantaneously
08:42 go to the macroscopic scale
08:46 this doesn’t work in others it fails to
08:49 predict these exponents that we get at a
08:51 critical point
08:52 and it also doesn’t give any uh reason
08:55 why we should have universality
08:59 now renewalization does something very
09:02 smart now that provides a systematic way
09:06 of connecting these microscopic degrees
09:09 of freedom that are described for
09:10 example by hamiltonian
09:13 to uh macroscopic descriptions
09:17 now it does so by taking into account
09:21 all the intermediary scales between
09:24 micro
09:24 the microscopic level and the
09:26 macroscopic level
09:28 and as you uh my guess from these
09:30 critical phenomena if you have scale
09:32 environments and all
09:33 scales are equal and equally important
09:36 this approach where you
09:38 take into account all of these scales
09:41 one at a time
09:43 now going from microscopic to
09:44 macroscopic which
09:46 makes much more sense than arbitrarily
09:48 focusing only on the largest scale as
09:50 you do in new york theory
09:54 now so rememberization group the
09:56 romanization
09:57 group allows us to go from the
09:59 microscopic
10:00 level now that is described by some
10:03 hamiltonian
10:04 by some symmetries all the way to the
10:07 macroscopic
10:08 level without ever forgetting what is
10:11 going on
10:12 in between now that’s the power of the
10:14 renalization group and in this lecture i
10:16 will show you
10:18 how this is actually implemented
10:21 now how we can actually do that
10:30 okay so there’s a little bit of a lot of
10:33 information i should have split this
10:34 slide

slide 7

10:35 um so
10:39 so so we are now going to follow this
10:41 program now going from microscopic
10:43 to macroscopic one scale
10:46 one length scale at a time now we’re
10:49 starting with something microscopic that
10:51 is super complicated don’t think about
10:53 the icing model
10:54 think about something that has 10 000
10:57 different couplings or so
11:00 now we take that model with 10 000
11:03 different parameters
11:05 on the microscopic scale so the real
11:07 physics
11:08 that’s going on on the microscopic scale
11:10 that’s complicated
11:12 yeah it’s much more complicated than the
11:13 icing model and it just as the disease
11:16 is much more complicated as the ice i
11:18 model that i showed you
11:21 and then we go to larger and larger
11:24 scales
11:26 and hope to end up with something that
11:28 is simpler
11:30 and less correlated than on the
11:32 microscopic scale
11:34 yeah so how do we do that so realization
11:37 consists of three
11:38 steps uh the most important step
11:42 now the the actual uh what’s actually
11:46 underlying renovation
11:48 is a course grading step now in this
11:51 course grading step
11:53 you can think about uh that you
11:57 unsharpen an image suppose you take a
11:59 photo
12:00 and then you can focus you can make the
12:02 picture
12:04 uh sharp or less sharp yeah by turning
12:07 like the lens
12:08 and the ring on the lens if you still
12:10 use a rear camera
12:11 yeah so you can make a job or let’s try
12:13 or think about the microscope
12:15 you can have to turn some knobs to make
12:17 the image sharp or not sharp
12:19 and the first step we
12:22 cause grain so we cause grain
12:26 and that means literally that we make
12:28 the image that we’re looking at in the
12:30 system
12:31 unsharp so that we that’s what we do
12:35 yeah so we
12:36 mathematically that means if you uh if
12:39 you close grain if you make something
12:41 out sharp unsharp then that means that
12:43 you integrate out
12:46 fast fluctuations or short range
12:48 fluctuations
12:49 i’ll show you on the next slide how this
12:51 looks like intuitively so first you get
12:53 rid of
12:53 all these uh fluctuations that happen on
12:56 very small length scales
12:59 and you have to perform an integral to
13:01 do that
13:02 and this integral of course is very
13:05 complicated
13:06 to calculate now the second state
13:10 we have a new field yeah
13:13 now let’s say phi average
13:17 but because we have course grade this
13:19 new field
13:21 our new image is blurry but
13:25 it is also not on the same length scale
13:27 as before
13:29 yeah so it’s just blurry but what we now
13:32 have to do
13:33 is to rescale length
13:36 to make the structures that we have in
13:38 this new image
13:40 similar to the structures that we had in
13:42 the original image
13:43 now so that means we need to rescale
13:45 length scales
13:48 by a factor of b
13:52 and the other thing that happens if you
13:53 make things blurry
13:55 is that you lose contrast in the image
13:58 so the image looks a bit dull so we
14:02 also need to now to increase the
14:03 contrast again
14:05 and we do that by rescaling our
14:09 fields as well so these are the three
14:12 steps of renormalization as if we first
14:14 integrate out very short
14:18 fluctuations happening on the smallest
14:20 length scales
14:22 and the second step and the third step
14:25 make sure that once we have integrated
14:28 out
14:29 these uh these short length skills
14:32 that our new theory that we get is
14:34 actually comparable
14:36 to the previous one so we have to reset
14:39 the length scales and we have to reset
14:41 the fields the contrast of the field by
14:44 multiplying them
14:45 with appropriate numbers

slide 8

14:50 so how does that look like so we start
14:52 with a field file
14:53 now that has short range fluctuations
14:56 you know fast fluctuations in space
14:58 now for example um
15:02 yeah here these wobbly things yeah these
15:04 are fast fluctuations
15:07 but it also has slow fluctuations that
15:09 would be
15:11 that would be here a slow fluctuation
15:15 now you have fluctuations on all
15:17 different length scales and now
15:19 we cause grain that means we integrate
15:21 out
15:22 these fast fluctuations and what we get
15:25 is a new field if i uh this phi
15:29 average here because phi r
15:33 that is smooth that doesn’t have these
15:35 small
15:36 wiggles anymore but it is smooth here
15:41 now so we have course grade the fields
15:44 and now we have coarse grains the field
15:46 we reset the length
15:48 rescale the length scale
15:51 x prime to make fluctuations on the new
15:57 field comparable to typical fluctuations
16:00 on the
16:00 original field the second step
16:05 we then renormalize um
16:08 there’s actually no three here because i
16:11 started
16:12 i know there is a three okay so uh
16:16 as a final step and we have to rescale
16:18 the fields now we have to
16:20 change the magnitude of the signal here
16:23 to make it comparable
16:24 to the original signal over here
16:28 now this sounds very intuitive
16:31 and very simple but of course in reality
16:34 it’s quite difficult

answering a question

16:39 so let’s see how this works uh
16:42 what’s the result so can i have a
16:45 question here
16:50 um in the previous slide
16:54 when we re-normalize uh fi
16:57 prime we essentially doing it
17:01 because we want uh whatever abs
17:04 magnitude or value of phi prime is to
17:07 equal
17:08 uh the value of five the original field
17:11 yes
17:11 exactly so what we want to do is so so
17:14 we do this procedure not only once
17:16 but many times now the next step will do
17:18 that many times
17:20 and uh each time we do these three steps
17:24 we get a new hamiltonian or a new action
17:28 and this new action will be
17:31 different to the original action
17:35 it will be different for trivial reasons
17:38 namely because once we cause when
17:41 think about this course grading as an
17:43 average instead
17:44 now you have i’ll show you later a
17:47 specific example suppose you average
17:49 uh in some area here and that’s why
17:52 that’s how you smooth
17:54 yeah when you average you know think
17:56 about the spin system you average
17:59 then you don’t have plus minus plus one
18:01 or minus one
18:02 but the new average field will be plus
18:04 0.1
18:06 and minus 0.1 now just because you
18:08 averaged
18:09 over in the field yeah but you don’t
18:11 want the new spins to live in this world
18:13 of plus
18:14 0.1 and minus 0.1 but you’d want them
18:18 also
18:18 live on the scale of plus minus 1 just
18:21 to make the
18:21 hamiltonians comparable now so this is a
18:24 trivial effect that you get by course
18:26 grading that you don’t want to have
18:28 to dominate your results and we need to
18:31 get rid of that
18:32 this trivial rescalings of the fields
18:35 and of the
18:36 of the um of the length scales just by
18:39 explicitly taking the step and saying
18:41 okay
18:42 now i have co-strained my field now i
18:45 have to reset the length scales
18:46 and i have to reset the amplitude of my
18:49 fields
18:50 by multiplying these these quantities
18:53 with appropriate numbers
18:56 i just want to have comparable things at
18:58 each step
19:00 yes thank you so much uh but regarding
19:03 this particular point
19:04 you showed in the first slide that phi
19:07 prime
19:07 phi is divided by b uh in the
19:10 in introductory slide remember
19:12 normalization i couldn’t understand that
19:15 yes here yeah five in the next one
19:19 five bar by b or um
19:23 so which which one third point
19:26 third point yeah that’s just a number
19:28 yeah we don’t know this number yet
19:31 we don’t know it yet we can get it by
19:33 dimensional analysis
19:34 uh very often um so we don’t know this
19:37 number yet
19:38 it doesn’t the same b okay it’s not
19:42 from one i think it’s an s actually
19:47 so so it doesn’t have to be b it depends
19:50 on the dimensions of your field
19:53 typically b to the power of something
19:55 also
19:56 that doesn’t have to be b and it depends
19:58 on the on the dimensions of your fields
20:00 here
20:01 um okay just at this point we just say
20:03 okay so we have to do something with our
20:05 fields to make them comparable
20:08 now think about an average now you can’t
20:10 get average and if you’re doing average
20:12 all the time
20:13 yeah then your uh the central limited
20:16 theorem will be that the
20:17 average like in a disordered system will
20:20 get very very small
20:22 yes the variance of this average will
20:24 get very very small
20:26 and we don’t want this effect to happen
20:28 because it destroys this basically
20:30 or if i what we actually want to look at
20:33 now what we actually want to look at is
20:34 how does the theory in itself the
20:36 structure of the theory
20:38 change as we go through the scale and we
20:41 don’t want to have these effects that
20:42 come
20:43 by uh by uh
20:46 just that we that we can’t compare
20:49 um suppose you compare like uh uh
20:53 you compare um the velocity of a car
20:56 you live in the uk also or in the united
20:57 states and you compare the velocity of
20:59 the car in miles per hour
21:01 or in kilometers per hour now you have
21:04 to you cannot compare that
21:05 but pure numbers you have to do
21:07 something you have to be scared by 1.6
21:09 to make them comparable and here also we
21:12 go through these scales look
21:13 like meters uh kilo kilometers miles and
21:17 so on
21:18 and to make these things comparable all
21:20 the different lengths that we have to
21:21 rescale
21:22 them all the time yeah just that way
21:24 that we’re all the way talking about
21:26 um the same thing um
21:30 so in germany we say we uh you cannot
21:32 compare apples and oranges
21:34 different things you have to uh
21:37 if you compare apple and origins then um
21:40 then then you’re doing something wrong
21:43 in other words to compare apples
21:45 to different kinds of apples so to say
21:48 yes but we always want to talk about
21:50 apples and not about miles and
21:52 kilometers now so that’s that’s the idea
21:55 about this rescaling step
21:59 i’ll show you later an example where
22:00 this rescaling step is already implicit
22:02 of course you can choose this course
22:06 grading step in a way
22:07 that it doesn’t change the magnitude of
22:10 the field
22:12 yeah so that you can choose the course
22:14 grading step in a way here this one
22:16 in a way that it doesn’t change the
22:18 magnitude of this field and then you
22:19 don’t have to do this renormalization
22:21 step
22:24 now but the principle you will have to
22:26 do that
22:28 okay so now we have these three points
22:31 now we rescale uh we of course grain

slide 9

22:35 rescale and renormalize and once we do
22:39 that
22:39 our action or our equilibrium or
22:41 hamiltonian
22:43 will become a new action
22:48 now s prime and say we
22:52 did this rescaling on a very small
22:54 length scale
22:55 dl now this s prime
23:00 is then given by some operator r
23:03 of s and if we do that repeatedly
23:08 you know so what we then get is a
23:11 remoralization group flow our g
23:13 flow and that’s basically
23:17 the differential equation for the action
23:21 yeah the s over the l
23:26 is then something like this r
23:29 of s so we normalize one step further
23:33 minus the previous step
23:38 you know so we change we do these three
23:41 steps
23:41 just a little bit not just and we call
23:44 square
23:45 we integrate out a very small additional
23:48 scale
23:50 yeah and our action is then different on
23:53 the next scale
23:54 of course we assume that this is somehow
23:55 continuous and well behaved
23:58 and then we’ll have a flow equation
24:01 of our action and of course in reality
24:04 this will not be a flow equation of our
24:06 action
24:06 but of the parameters that define our
24:09 action
24:10 now suppose so how does it look like
24:18 now so here is in some space of all
24:21 possible actions
24:23 of some parameters p1
24:27 p2 p3
24:32 now and this action now think about the
24:34 uh
24:35 think about our non-linear dynamics or
24:37 dynamical systems lecture
24:39 what we can this is this is a
24:41 differential equation here
24:44 it looks complicated but this is a
24:46 differential equation it tells you
24:47 how the parameters of our action
24:51 change as we cause gradients we do this
24:54 renormalization procedure
24:57 now and this gives us a differential
24:59 equation
25:01 and if you have a differential equation
25:02 that will be highly non-linear of course
25:05 now we can do the same thing as we did
25:07 in this lecture or nonlinear system we
25:09 can
25:10 derive a phase portrait now in the space
25:12 of all possible actions
25:14 of all possible models where does this
25:17 minimalization group
25:18 group flow carry us
25:22 so let’s start with a very simple
25:27 line
25:31 that’s the line in the space of all
25:33 possible models
25:34 now this is the line of models
25:37 that actually describe our physical
25:39 systems
25:40 now think about different combinations
25:42 of temperature
25:43 and magnetic fields in the icing model
25:48 now so this is this is where this model
25:50 lives in this space
25:52 now if we take the right parameter
25:55 combinations
25:58 we will be at some critical point
26:06 yeah so there’s no flow yet now so this
26:09 is just
26:10 uh the the the the range of different
26:13 models that we can have for example in
26:15 eisenhower that correspond to some real
26:17 physical system
26:18 but there of course there are many other
26:20 model in the models in this space
26:22 that don’t describe our physical system
26:24 now they don’t describe a magnet but
26:26 something else or that there are not
26:28 given by a simple hamiltonian with
26:29 nearest neighbors
26:30 interactions but by something that has
26:33 long-range interactions or something
26:35 very weird
26:36 now there’s a space of all possible
26:37 models is very large
26:40 now we have this critical point and
26:43 other models in this space also have
26:46 critical points
26:47 now and these critical points live on a
26:51 manifold
26:58 now they leave on a manifold you know
27:01 that is the critical manifold
27:09 you know so that’s all the points of
27:11 this critical manifolds are critical
27:13 points
27:14 of some actions
27:17 and our action our critical point of our
27:20 action
27:21 is also on this manifold but all
27:24 other points of this manifolds are also
27:26 some critical points of some other
27:29 actions so
27:33 what happens now now
27:37 we are close to the critical point let’s
27:39 say
27:40 we’re here now very close to the
27:44 critical point
27:46 and now we really normalize
27:50 we go through we make this procedure of
27:53 renormalizing
27:54 the core straining going to larger and
27:56 larger scales and rescaling our fields
27:59 and lengths
28:00 and then our anatomy or our action will
28:03 change
28:04 so it will flow in this space
28:07 in some direction
28:11 so where does it flow to in the
28:13 non-dynamics lecture
28:16 we’ve seen that what determines such
28:18 dynamical systems
28:20 are fixed points and
28:24 what you typically have to assume in the
28:25 real normalization group theory
28:28 that there is some fixed point
28:31 on this critical manifold yeah
28:36 the fixed point of the realization would
28:39 flow
28:40 on this critical manifold
28:44 and now what happens with our flow
28:50 of course we will go to this fixed point
28:55 and then we might go away again
28:59 now the sixth point the fixed points
29:02 like in the dynamic resistance lectures
29:04 determine the flow of our dynamical
29:07 system
29:09 now what is this fixed point here this
29:11 fixed point is not the critical point
29:14 it’s the critical point of some other
29:16 model
29:18 but this critical point here
29:23 has stability yeah just like in normal
29:25 domains so this is non-linear dynamics
29:27 here
29:28 so it’s very often exactly what we did
29:30 in this lectures before
29:31 so we asked what is the flow now we ask
29:33 them to ask about the stability
29:35 of this fixed point now so this fixed
29:39 point
29:40 has stable directions now you perturb
29:44 and you get pushed out and unstable
29:46 directions
29:48 yeah typically or by definition
29:52 the directions on the critical manifolds
29:55 are stable now that’s how this manifold
29:59 is actually defined
30:00 you can and there’s a theory and only
30:02 resistance
30:04 that tells you that and there are also
30:07 other directions
30:09 that are not stable now let’s have them
30:12 in green
30:14 for example the way i drew this these
30:17 are the ones
30:19 in this direction
30:26 so this determines the stability of the
30:28 flow of our fix
30:29 of our our system and if we have only
30:32 one fixed point this one fixed point
30:35 will tell us what happens to the flow of
30:37 our system just like in nonlinear
30:39 dynamics
30:40 lecture okay so
30:43 now we re-normalize
30:46 it because we didn’t start exactly at
30:48 the critical point we stay in the
30:50 vicinity of this manifold here
30:52 here and this critic this fixed point of
30:55 the flow will do
30:56 something to us no it will push us away
30:59 or will attract us and now you can
31:04 do the same thing that you do in
31:05 learning your dynamics namely you linear
31:07 wise
31:08 around this fixed point so we say that
31:11 our action
31:13 [Music]
31:16 called linear stability
31:21 you linearize around this fixed point so
31:24 you say
31:25 that our action is equal to the action
31:29 at this fixed point
31:32 plus sum over different
31:35 directions i who operate is i
31:39 h i
31:45 times b to the power of lambda i
31:52 times
31:56 q i yeah so this here are
32:01 these two eyes are the eigen directions
32:08 so these are operators you can think
32:09 about this as operators so like eigen
32:12 direction of these operators and
32:16 the b is our course grading scale
32:30 the h tells us
32:34 how far we are away from the fixed point
32:42 and the s tells us uh is just the action
32:48 that we have at this fixed point
32:51 you know so we it’s the same thing for
32:53 all of this was once we have once we
32:55 have this mineralization group flow
32:58 we’re in the subject of non-linear
32:59 dynamics and we use the tools
33:02 of nonlinear dynamics renalization group
33:05 theory this language is slightly
33:09 different
33:10 you know so that’s that’s why you have
33:12 this b to the lambdas and so on
33:14 here and you separate the h i from the b
33:17 to the lambda
33:18 that’s just the framework the how you
33:20 write it in the realization
33:23 theory for convenience reasons but what
33:26 we do here is
33:27 a simple linear stability analysis
33:30 of a non-linear system i will
33:34 re-linearize around the fixed point
33:36 we see how whether perturbations grow or
33:39 shrink
33:40 in different directions yeah and that
33:43 characterizes
33:44 then our nonlinear system and now we can
33:47 ask if we perturb around this fixed
33:50 point
33:51 in one direction i does it grow
33:54 this perturbation or does it shrink
33:57 lambda i
33:58 is larger than zero now this bi
34:03 is larger than one or b is larger than
34:07 one
34:08 now so lambda i is larger than zero then
34:11 our perturbation will grow
34:13 yeah perturbation
34:20 growth and then we say this direction or
34:23 this operator q
34:24 i you can also think about q i as one of
34:27 the
34:28 terms in the action now think about one
34:31 of the terms in the action
34:32 or one of the terms in the hamiltonian
34:36 this direction qi
34:40 is then called relevant
34:45 why is this relevant what we call when
34:46 do we call this relevance
34:48 you know if this perturbation grows we
34:51 are in this green direction here that
34:53 pushes us away from the critical point
34:56 so that means
34:56 that if we are an experimentalist
35:00 and if we want to tune our system
35:03 to get into the critical point
35:07 then
35:10 then we know that we have to turn
35:13 these relevant parameters qi
35:18 now this relevant parameter for which
35:19 this uh that are
35:21 unstable directions of this fixed point
35:25 now and if we have a relevant directions
35:28 we also have irrelevant directions
35:30 lambda is smaller than zero
35:32 and these directions are called
35:34 irrelevant
35:35 now perturbation
35:43 shrinks that means that
35:47 qi is irrelevant
35:53 that means the qi in this case is not a
35:56 parameter that drives us into this
35:58 critical point
36:01 and then we have the case that lambda i
36:03 is exactly equal to zero
36:05 then we don’t really know what to do
36:08 then this q
36:09 i is
36:12 marginal and we cannot tell from this
36:15 linear stability analysis alone
36:17 from the linear realization around this
36:19 fixed bond we cannot tell alone
36:21 whether this perturbation will grow or
36:24 shrink and we have to use
36:25 other methods okay
36:28 so what happened now so we
36:31 started our theory close to the critical
36:34 point
36:35 we did this rememberization group
36:37 procedure
36:39 course grading rescaling renormalization
36:43 and then in this procedure
36:46 our action or our model will flow
36:49 through the space of all possible models
36:52 yeah
36:55 and then we ask where does it flow to
37:00 and we look at the non-linear dynamics
37:01 lecture and ask
37:03 so where does such a nonlinear system
37:05 drive us to
37:07 and our nonlinear dynamics lecture will
37:09 tell us look at the fixed points
37:12 right and in this realization flow you
37:14 also have fixed points you assume that
37:15 you have a fixed point
37:17 and this fixed point tells us about
37:21 what is going on on the macroscopic
37:24 scale now that determines this fixed
37:26 point
37:26 determines which is the end result of
37:28 our minimization group
37:30 procedure
37:33 so at this fixed point is characterized
37:36 by stability
37:38 it has a finite number of relevant
37:40 directions
37:43 now it’s had a finite number of
37:44 parameters that are actually important
37:47 to change if you want to go to the
37:48 critical point
37:51 and because you only have a finite
37:53 number of directions that are relevant
37:56 you usually get away with models that
37:58 also have this very finite number
38:01 of parameters instead of models like a
38:04 bigger magnet or so that is a very
38:06 complicated geometry and everything
38:08 there are instead of model that has 15
38:10 000 parameters
38:12 now you get away with a finite number of
38:13 parameters that are given
38:15 by the relevant eigen directions
38:19 of this fixed point now so now what
38:22 happens if we have a different model so
38:24 now this is now the
38:25 magnet one that this is not a
38:29 magnet or material one we can also look
38:31 at a magnet at another of another
38:33 material
38:34 now so this magnet one
38:40 and now we have here
38:47 a magnitude
38:51 now each of them at the physical real
38:53 microscopic level is destroyed by 15 000
38:56 parameters or whatever something very
38:57 complicated
38:59 and for this magnitude we can do the
39:02 same procedure
39:04 we renormalize
39:07 and we while we renormalize we will end
39:10 up at the vicinity of this fixed point
39:14 and in the vicinity of the fixed point
39:17 the behavior of the flow is determined
39:20 by a finite number
39:22 of parameters again
39:26 and both of these magnets here are
39:29 under renormalization now on the large
39:31 scale repeat these
39:33 procedures determined by the same fixed
39:37 point
39:37 about the same point but this one here
39:41 and because they’re determined by the
39:42 same fixed point with the same stability
39:46 and with the same properties of how they
39:48 go
39:49 uh of how the flow behaves around this
39:51 fixed point
39:53 that’s why these two magnets here are
39:55 described macroscopically by the same
39:57 theory
39:58 and that’s then the reason why we have
40:01 universality
40:03 so in this way in this very general way
40:05 so we look of course in our in
40:07 more detail the renal isolation group
40:10 theory gives us a justification for why
40:14 only a finite number of parameters
40:17 matter
40:17 on the or finite a limited level of
40:21 description is sufficient to describe
40:23 large scale properties of a large
40:26 number of very different systems and the
40:29 reason is
40:30 at this critical point they’re
40:33 described macroscopically by the same
40:36 fixed point of the renewalization
40:39 workflow

slide 10

40:42 now how does this look in detail
40:47 so the very first or the
40:50 most uh simple way of doing
40:53 remoralization
40:55 is to take what i said initially about
40:58 this
40:59 defocusing about this course training
41:01 literally
41:02 and do the whole procedure in real space
41:06 now suppose you have here a lattice
41:09 system
41:11 and also suppose you have this lattice
41:12 system here and you have some
41:15 spins here and what you can do then to
41:19 cause grain
41:20 is to
41:24 you know what you can do to coarse grain
41:26 is to create
41:27 boxes or blocks now of a certain size
41:31 and then to calculate this cross here
41:34 that is a representation of all of these
41:37 microscopic spin for each block
41:40 yeah so we have this uh spins here
41:44 like what are 16 in each block and you
41:48 now transform them to a single number
41:50 you can do that by averaging over them
41:52 or you can take
41:54 you can say that i take the spin
41:58 that is the majority of these other
42:01 spins here
42:02 if the majority goes up then my new spin
42:06 that describes the entire block
42:07 will also go up and this would be a way
42:11 to get a rid of this renormalization
42:14 step if you say i take the majority
42:18 i take a majority rule so this new spin
42:20 here x
42:21 will take the value of the majority of
42:24 the original spins
42:27 then the new spin will also be plus or
42:29 minus one
42:31 if my new spin yeah and i don’t have to
42:33 renormalize them because it has the same
42:35 values as the original splits
42:38 if i take the new spin as the average
42:43 over all of these spins here then
42:47 i typically get a very small number the
42:49 average won’t be one or minus one but it
42:51 will be
42:52 0.3 or 0.1 or 0.5 or so but it will not
42:57 be
42:57 one or minus one in most cases yeah so
43:00 in this case if i perform this procedure
43:02 i would have to renormalize
43:04 you know and i have to would have to
43:05 rescale my fields
43:08 to make them comparable to the original
43:10 step
43:12 yeah so now we have these blocks
43:17 and we define some new spin that
43:20 describes each of these blocks
43:23 and now we write down a new model
43:26 a new hamiltonian for these new spins
43:29 here
43:31 and what we hope is that the spins that
43:35 we have here
43:38 in this new system this course-grade
43:40 system are described by a theory
43:43 that is structurally very similar to the
43:45 original theory
43:48 and this hope is actually justified by
43:52 the observation of
43:55 scale and variance now so if your system
43:58 is scaled in variance we can hope that
43:59 if we zoom out
44:01 and our system is statistically the same
44:04 then then our partition function
44:06 or our action will also be the same
44:09 just with some different parameters now
44:11 that is the hope that is underlying
44:13 renewalization group procedures with
44:16 these
44:16 block spins here what you typically get
44:19 is that you get
44:20 higher order turns all the time you know
44:23 so that’s this hope is not
44:24 mathematically super precise uh but
44:28 that’s what you have to assume in order
44:30 to achieve anything
44:33 okay yeah
44:36 okay so we call screen and then
44:40 we rescale the second step so that the
44:42 distance between these
44:44 spins here the new spins is the same
44:47 distance as we had
44:49 between the original spins now that’s
44:52 what we have to do anyway
44:53 and that’s why how we divide length
44:55 scales
44:57 by the same factor that corresponds to
45:00 the size of our boxes
45:02 yeah and now the lengths are the same as
45:04 before

slide 11

45:09 so let’s do this procedure in a very
45:13 simple case which is the 1d
45:16 ising model
45:21 now so the one deising model now is
45:24 written the tradition function
45:26 it can be written in like a long form
45:30 in this way here that i sum up
45:33 all combinations of nearest neighbor
45:36 interactions
45:38 now that’s just the hamiltonian here of
45:40 the ising model without an external
45:42 field
45:43 now and then i have to have a sum about
45:45 all possible values of the sigma i’s
45:48 of the of the that my fee that my all
45:51 the possible values
45:52 that the sigma can take that gives me my
45:54 partition function
45:57 now what i do know is the first
46:01 course grading stuff now this first
46:04 coast grading step
46:06 means that these red spins here
46:09 like with the all these these black
46:11 spins are the white splits
46:13 now i’ll integrate out the white spins
46:17 here in this picture i’ll integrate
46:20 these ones out
46:23 now every second spin all even spins
46:26 and if i do that i will get a new theory
46:30 that is described by interactions
46:32 between these
46:33 uneven spins that are and these
46:36 interactions occurred to us are
46:37 depicted in these red with these red
46:41 lines
46:44 okay so and the stars is actually very
46:47 simple activated talks
46:49 so that’s so for if we just do it for
46:56 for sigma 2 now we just sum
46:59 out we take the terms that correspond to
47:02 sigma 2
47:04 and we get that that is equal to
47:08 not many terms and then we have the
47:11 contribution from sigma 2
47:14 uh k sigma 1
47:17 plus sigma 3 now that’s what what’s left
47:20 plus e to the minus k sigma 1
47:25 sigma 3
47:28 and then all the rest e to the k
47:32 sigma 3 sigma
47:36 4 plus four sigma
47:40 five and all the other splits
47:44 so what i just that said is that sigma
47:46 two can have two values
47:48 minus one plus one yeah and i just
47:51 substituted i expected
47:53 explicitly now set sigma 2 to plus 1 and
47:56 minus 1
47:57 and perform the sum and that’s what i
47:59 get then here for this first
48:01 term and now i can do that for all
48:12 even sigmas and then what i get is
48:16 exactly the same thing sum over
48:20 many terms e to the k
48:23 sigma 1 plus sigma 3 as before
48:27 plus e to the minus k sigma 1
48:31 sigma 3. now that was the original one
48:34 where we set sigma
48:36 2 to -1 and then we get the same thing
48:42 e to the k for the next
48:45 term now for the next interaction sigma
48:48 3
48:49 plus sigma 5
48:54 e to the minus k sigma 3
48:58 sigma 5
49:01 and so on yeah
49:07 for all the other terms yeah

slide 12

49:13 so now the idea is
49:18 that because when we are at a critical
49:21 state
49:23 that we expect our partition function to
49:25 be self-similar when we call screen
49:28 the statistics of the system remains the
49:31 same as we zoom out
49:33 and that’s why we also expect
49:36 the quantity that gives us the statistic
49:40 statistics the partition function to be
49:42 self-similar as well
49:45 and what we now do is that we find
49:49 a new value of k prime
49:53 and some function f
49:56 of k that tells us that we
49:59 that these terms that we got
50:02 here sigma 1 sigma 3 we always got the
50:06 sum for each coupling
50:08 they should take the same form as the
50:11 original hamiltonian but with some
50:14 pre-factor here
50:17 and some new coupling here
50:20 but the form should be the same as
50:22 before
50:24 now as i said this is not usually
50:27 well justified but we have to do that in
50:29 order to do anything
50:31 and if you require that if you do some
50:33 algebra you will find that if you set
50:36 k prime to this one one half
50:39 logarithm hyperbolic cosine of 2k
50:44 and the function f of k to this year
50:48 then it fulfills this condition yeah
50:54 so now we can plug this in now so if we
50:59 if we use that
51:01 then our hammer tilt our partition
51:05 function
51:06 will read again we have many
51:10 terms f of k
51:14 e to the k prime
51:17 sigma 1 sigma 3
51:21 f of k e to the k
51:24 prime sigma 3 sigma 5
51:29 and so on
51:32 now and this is just the same we can
51:35 write this now
51:36 as a new partition function that has a
51:40 new prefactor
51:42 f of k now we pull this out this
51:45 prefactor
51:46 f of k and we have that n over two times
51:51 times a new partition function that
51:55 depends on the new system size
52:00 and a new coupling k prime
52:04 yeah so we have this down this one
52:06 renormalization step
52:08 we get a new partition function that
52:10 looks exactly the same as the old one
52:12 in structure but we have a new coupling
52:15 k prime
52:17 and a pre-factor here that is this
52:20 function
52:21 f and that also depends on the coupling
52:25 also what we did here is that we now
52:27 have
52:31 a relationship between the partition
52:34 functions
52:35 at different stages of the
52:38 renewalization procedure
52:45 yeah and now
52:50 what does it mean look at
52:53 this one here k prime
52:57 this is already a description
53:02 now this is already a description of how
53:04 our coupling
53:05 one parameter k prime
53:09 depends on the value of this okay now
53:12 how this parameter k
53:13 evolves in this course grading procedure
53:17 in this renewalization procedure
53:20 you know so this k that gets updated
53:23 now it’s not in differential form yeah
53:26 like in
53:26 like uh like we did in the non-linear
53:28 dynamics
53:29 actually but it’s in this uh other way
53:32 that you can describe non-linear
53:34 systems but by iterative updating now so
53:38 the new value of k
53:39 prime is given by the old value
53:43 is by this function here applied to the
53:45 old value
53:47 and now now is this is this updating
53:50 scheme here
53:52 and we can expand this term here the
53:54 logarithm
53:55 of the hyperbolic cosine and so for low
53:58 values of this coupling
54:00 this goes with k squared now so normally
54:03 i have an idea
54:05 about how this looks like i have already
54:08 prepared this
54:09 very nice and we can now solve this
54:11 equation here
54:13 graphically so we want to get the flow
54:15 and we can solve this graphically
54:17 and see where this linearization
54:19 procedure
54:20 carries our k our coupling k
54:23 now and because our partition function
54:27 remains invariant

slide 13

54:30 well our k the update of our k
54:35 describes actually the behavior of our
54:37 hamiltonian
54:38 under renormalization
54:41 okay so this is the plot here so what i
54:44 what you do
54:44 is that you plot the left hand side
54:48 of this equation k prime it’s just
54:51 linear with slope one
54:54 and uh the right hand side
54:58 now this is uh this here
55:02 and where the left hand side is equal to
55:04 the right hand side there you have a
55:06 fixed point
55:07 now so this is here the way you need to
55:10 read this this is the next value of the
55:11 realization
55:12 this is the previous one if you start
55:14 here we’ll go
55:15 here then here and here and here
55:19 and at some point these two lines meet
55:22 and that’s
55:22 that’s where your fixed point is so this
55:25 humanization procedure
55:27 will bring us to some fixed point which
55:29 happens
55:30 to be uh down here at zero
55:36 yeah and then we can look at how this
55:39 uh we can also then plot
55:42 a flow diagram as i did in a more
55:45 complicated way before
55:47 and how we also did it in the actual
55:49 nonlinear dynamics lecture
55:51 now we can plot it on this in this
55:53 one-dimensional line
55:55 then we have a stable fixed point at
55:57 zero
55:58 now that’s here stable
56:02 and any value where we start with our
56:04 course grading procedure
56:06 will be driven to a value of k equals
56:09 zero
56:10 now we start with a very strong coupling
56:12 we renormalize
56:14 and we will be driven to zero
56:17 that means that this renormalization
56:21 procedure this course grading procedure
56:24 in this one-dimensional icing model
56:28 will always on the macroscopic skin on
56:31 large
56:31 length scales always lead to a model
56:35 that is effectively described by
56:40 a system that has zero coupling
56:43 yeah this coupling here vanishes and if
56:46 we have zero coupling that means that
56:48 our system
56:49 is above the critical point that’s
56:52 non-critical
56:54 so we will normalize we go on and on
56:57 and we always end up on the system that
56:59 has very high temperature
57:01 or very low coupling yeah that’s a it is
57:04 a disordered system
57:06 and that’s just a reflection of the fact
57:07 that the one deicing model
57:10 doesn’t have any order for a finite
57:12 temperature
57:15 now so you have to start with coupling
57:18 exactly equal to infinity
57:21 to get an order over the temperature
57:24 exactly
57:24 equal to zero only then you can have
57:26 order everything else
57:28 will drive you to this fixed point here
57:32 that corresponds to a system where you
57:35 have no
57:36 coupling at all now so the one that we
57:39 knew that already that the 1d
57:40 system doesn’t show all in the ising
57:42 model doesn’t have order
57:44 it doesn’t have a really critical point
57:47 and that’s why our flow
57:48 tells us that macroscopic scales this
57:51 system
57:51 goes to a system that doesn’t have
57:55 any interaction so it’s completely
57:56 disordered

slide 14

57:59 of course you can do the same procedure
58:01 for um
58:04 for the 2d system yeah and there is of
58:06 course
58:07 again much more complicated then this 2d
58:10 system
58:12 you get a flow diagram that looks like
58:15 this here
58:17 on the bottom so here you suddenly
58:21 have another fixed point an unstable
58:23 fixed point
58:25 in between these two extremes now this
58:28 unstable fixed point
58:30 here if you start to the left of this
58:33 unstable fixed point you were driven to
58:36 a state
58:37 without that that we had previously
58:40 where the coupling is very low
58:42 or that corresponds to the system at
58:43 very high temperature
58:45 if you start to the right of this you’ll
58:47 be driven to a state
58:49 where you have order you know where your
58:51 coupling is basically infinity or your
58:53 temperature effective temperature
58:55 is zero and because you have now this
58:58 fixed point here this new fixed point
59:00 right you get
59:02 this singularity or this discontinuity
59:05 of the free energy because if you go a
59:06 little bit to the left
59:08 you go to another a different
59:10 macroscopic state
59:11 then if you go a little bit to the right
59:14 and of course you can test that with
59:15 numerical simulations
59:17 now so that’s here from the book of
59:19 cardi which is a very very nice book um
59:22 scaling and renormalization and
59:23 statistical physics
59:26 and i have to say that
59:29 the class don’t look very good on this
59:32 ipad
59:34 so what you see here are just
59:36 simulations of the two the eisenmann
59:39 and what they did is they performed one
59:42 block spin removalization procedure
59:45 that’s that’s what we did right now
59:47 that’s the coarse grain so one step of
59:50 this coarse graining
59:52 and uh if you are right at the critical
59:55 point on the left hand side
59:57 you do the coarse graining step yeah
59:59 then
60:00 the system remains invariant if you
60:02 start in a fixed point you’ll stay there
60:05 if you start a little bit this course
60:07 grading procedure
60:08 and here that’s not there’s nothing
60:10 fancier they just took the simulations
60:13 and they did one course grading step
60:15 they averaged you have maybe over a
60:17 block of spins or so
60:19 and if you are above this
60:23 critical temperature that means you
60:24 start here on the left of this fixed
60:26 point
60:27 and if you do this course grading step
60:29 then your system looks more disordered
60:31 than before
60:33 yeah so this year it looks like a higher
60:35 temperature than this year because you
60:37 have a lot of these small domains
60:39 now this is just a reflection of this
60:41 stat like an intuitive
60:43 picture of how you go in this
60:45 reunionization procedure
60:47 to the left to a state that has no
60:49 coupling at all
60:50 k equals zero and if you’re coupling a
60:53 zero that’s the same as when your
60:54 temperature
60:55 is very large and if you want to read
60:58 more about this
60:59 have a look at the book of john carty
61:01 and there’s also of what i just showed
61:03 you
61:04 there is a there is a nice book by um
61:11 so there’s a nice book there’s a nice
61:13 article by karanov
61:15 um by teaching the immunization group
61:17 and he does these calculations also for
61:19 the 2d
61:20 model okay so now
61:23 we state in a real space yeah and in
61:26 real space
61:28 uh is very intuitive now
61:31 and it works for the one deising model
61:33 for the 2d eisen model gets already
61:34 complicated
61:36 and it’s basically impractical to do
61:38 that
61:39 for general um
61:43 for for general physical models now it
61:46 gets very complicated to do that
61:47 procedure
61:48 in real space and the reason is that
61:50 there’s no small parameter involved
61:53 that you can use for an expansion

slide 15

61:57 then there was another guy called wilson
61:59 who came up with another idea
62:01 now that was actually and that’s called
62:03 the wilson
62:05 momentum shell idea
62:09 that’s the world’s in momentum shell
62:11 idea so what does it mean so what was
62:14 wilson’s idea was that we caused grain
62:18 by integrating out fast degrees of
62:21 freedom
62:22 or degrees of freedom that have a very
62:24 short wavelength
62:26 in fourier space that’s the way you do
62:29 that
62:30 is uh you look at free space also this
62:33 is our
62:34 free space let’s say we have two
62:35 directions in free space
62:38 then we have here
62:44 a maximum wave vector that’s the maximum
62:52 wave vector and this wave vector let’s
62:55 call it
62:56 capital omega is the same
63:00 just given by that’s the smallest
63:02 structure we can have in the system
63:04 that’s a microscopic length scale
63:06 yeah and that’s in these lattice systems
63:08 this typical
63:09 uh one over the the lattice the
63:12 microscopic level the lattice spacing
63:15 yeah so we cannot go any smaller than
63:17 that
63:19 now starting from the smallest length
63:21 scale now so a description of our system
63:24 on the smallest length scale
63:27 we now integrate out the blue stuff here
63:32 that’s this one here that’s the momentum
63:35 shell
63:42 and we integrate out this momentum shell
63:45 until we reach
63:48 a new wavelength a new y vector of a new
63:51 length scale
63:54 omega prime is equal to the original
63:57 omega
63:58 divided sum by some number lambda
64:01 yeah so
64:05 we integrate out one bit in momentum
64:08 space
64:10 at a time and that means that we perform
64:13 an integration
64:15 uh on a momentum shell on a tiny shell
64:18 in momentum space and also our new field
64:26 in the momentum shell
64:32 is then called typically something like
64:35 feel
64:36 find larger of q
64:40 and this is just undefined by
64:43 phi of q
64:46 with q in this interval
64:51 omega over lambda to omega
64:56 yeah so we integrate
65:00 out one step and now wilson’s scheme is
65:03 actually very similar
65:08 to what we’ve done already yes the first
65:10 step will be
65:12 recall screen
65:18 now by rescaling oops sorry
65:29 no by rescaling
65:34 and that will give rise to some
65:38 change in the coefficients
65:46 in the action
65:49 and the second step is that we perform
65:56 this integration in the momentum shell
66:00 integrate out
66:04 short range
66:08 fluctuations
66:12 or momentum
66:19 a second step and we’ll in the next
66:22 lecture
66:22 we do exactly this yeah we performed
66:24 this wilson’s ruralization group
66:26 procedure that is much more practical
66:29 than the
66:29 block spin immunization group that we
66:32 had in the
66:32 beginning of this lecture and the good
66:36 thing about this wilson’s
66:38 of wilson’s idea is that it actually has
66:40 a small parameter
66:41 this momentum shell is very small
66:44 and uh yes so this
66:48 has a small parameter that means that we
66:50 can actually then
66:51 hope to get some approximative um
66:55 uh scheme out of this approximative
67:00 so that we can approximate our integrals
67:02 that we get from the course grade
67:04 yeah so we’ll do that uh next week
67:07 for our little epidemic model and we’ll
67:10 derive
67:11 the renalization group flow from our
67:13 equity epidemic model
67:15 and from this flow we’ll then get the
67:17 exponents that derive that describe
67:20 the action of the the the behavior
67:23 of this epidemic model near the critical
67:26 point
67:27 yeah and um
67:30 exactly yeah so so that’s what we’ll do
67:33 i’ll
67:33 just leave it for here today because and
67:35 then next week we do the calculation and
67:37 if you’re
67:38 not interested in calculating that
67:41 because it’s so
67:42 uh so uh
67:46 so uh short before christmas if you’re
67:48 free to skip the next lecture
67:50 yeah and uh officially i think it’s not
67:53 a lecture data
67:56 but i wanted to get that done before
67:57 christmas that we can after christmas at
68:00 january what is that fifth or so i can’t
68:03 remember
68:03 um we’ll start actually done with data
68:06 science and
68:07 to look at some real data yeah okay
68:10 great
68:11 so that was today only the intuitive
68:13 part about romanization so next week
68:15 we’ll do
68:15 we’ll see that in action and see how it
68:18 actually works in the non-equilibrium
68:19 system
68:22 bye

answering a question

68:28 excuse me yes
68:31 um could you please explain again why
68:33 were we trying to reach the fixed point
68:35 from the critical point
68:37 so so we’re not trying to
68:41 it’s just that you assume for this to
68:44 work that there is such a fixed point
68:47 that determines the flow of the
68:50 renormalization group yeah in this
68:53 generality you have to assume that that
68:55 there is
68:56 such a fixed point and from nonlinear
68:58 dynamics dynamical system
69:00 lecture that we have we know that once
69:02 we have such a fixed point
69:04 we basically know already how the system
69:07 behaves also in other parts
69:09 of uh of the faith yeah so that’s why
69:12 the system these fixed points are so
69:14 important now we have to assume that
69:16 there exists some verb
69:17 uh but we have to basically for every
69:19 individual model that we look at we have
69:21 to show that they actually
69:23 that they that we have actually
69:25 meaningful
69:27 or moral more than one thing of course
69:31 yeah so but once you have the flow it’s
69:33 a problem in normal dynamics
69:35 that’s that once you have the flow you
69:36 do what you do in non-aerodynamic
69:38 dynamic dynamics so typically in these
69:40 books of linearization they use a
69:42 different language that’s a little bit
69:43 disconnected from this dynamical systems
69:47 field
69:48 no but what you do is you have
69:50 non-linear
69:51 differential equations and then you just
69:53 want to see what happens to these
69:55 nonlinear differential equations
69:57 and then if you ask this question then
70:00 and non-in your system you need to ask
70:01 about the fixed points
70:03 and about the stability of yeah and
70:05 that’s why this fixed point
70:07 in the randomization group flow is so
70:10 important
70:11 now if there wasn’t any fixed point at
70:13 all yeah then
70:14 it would be a very good system so we
70:16 have to have such a big
70:19 point on the critical manifold for this
70:22 uh for this procedure to work
70:25 now we have to assume at this stage here
70:27 we have to assume
70:29 that it exists and if it exists then it
70:31 will determine
70:32 our flow yeah
70:36 but we don’t want to we don’t want to do
70:38 so once we have to fix the the system
70:40 the renewalization group will
70:41 automatically
70:42 carry us to the fixed point now we don’t
70:44 we don’t want to go the system to
70:47 the system to go there if the fixed
70:49 point exists
70:50 we’ll have to uh we know that the flow
70:54 will determine we will be determined by
70:56 this
70:58 yeah so that’s the that’s the that’s the
71:01 idea
71:02 but there’s non-linear systems that in
71:04 principle has not much to do with
71:06 renormalization
71:07 yeah it is a general property of the
71:10 dynamic
71:14 okay thank you okay great
71:24 any other questions
71:28 um i have a question um so
71:32 does this does this um
71:35 require you to have a very
71:38 good understanding of the microscopic
71:42 um dynamics i guess
71:45 you have to have a model to start with i
71:48 mean
71:48 um but that’s kind of what i mean is
71:51 if you have some model that’s leaving
71:55 out certain things that are
71:57 that maybe
72:01 you know you thought wasn’t so important
72:03 or something like that
72:04 but then as you core scream more and
72:08 more
72:08 like does it like um
72:11 will it give rise to like a different rg
72:14 flow than
72:15 if you had included you know other other
72:19 uh okay i think i think that that
72:21 depends on
72:22 what you leave away and what you leave
72:24 out and what you can use so in principle
72:28 for the usual model including this si
72:30 model that we have here
72:32 that we’re looking at and also but also
72:34 for the icing model and
72:36 they they fall a little bit out of the
72:37 blue when somebody presents them to you
72:40 and for example in the lecture now they
72:43 completely make sense that they
72:45 destroyed these systems
72:47 but usually people have used
72:50 renormalization studies
72:52 to show that additional terms don’t
72:55 don’t matter for these yeah also for
72:57 this active
72:58 uh for these active systems that we
73:00 talked about
73:02 with these aligning uh with these
73:04 aligning directions
73:05 and so on now for these systems i showed
73:08 you briefly a larger way equation and
73:09 people did a
73:10 lot of work to show that this is
73:12 actually the simplest
73:14 uh description that you can have that
73:15 describes this system now because the
73:18 rigid rimmelization they found
73:20 that all other terms so this is the
73:22 minimal set of terms that i need
73:24 to describe still the same
73:26 renewalization
73:27 or still renumerization yeah
73:31 and uh whether you get new terms i
73:34 so i i would i wouldn’t expect that if
73:35 you
73:37 i wouldn’t expect to get new relevant
73:39 terms
73:40 out of nothing yeah in general you know
73:43 otherwise you could start with just
73:45 nothing at all and see what happens and
73:47 then you get like a theory of everything
73:49 i wouldn’t expect these things to pop up
73:52 yeah but of course you get all kind of
73:53 messy things
73:55 now you can have if you start with the
73:57 icing more then in reality then you get
73:59 higher order
74:00 interactions and so on and
74:04 and you have to basically have to get
74:06 rid of them
74:08 okay i think let’s uh okay great
74:18 okay so if there are no more questions
74:20 then uh
74:21 see you all next week so as as
74:25 as usual next week i’ll uh
74:28 start with a repetition of this and
74:31 explain explain it again before we
74:33 actually do the
74:34 real calculation okay bye see you next
74:40 week
74:45 you


About Joyk


Aggregate valuable and interesting links.
Joyk means Joy of geeK