Hubbry Logo
RenormalizationRenormalizationMain
Open search
Renormalization
Community hub
Renormalization
logo
8 pages, 0 posts
0 subscribers
Be the first to start a discussion here.
Be the first to start a discussion here.
Renormalization
Renormalization
from Wikipedia

Renormalization is a collection of techniques in quantum field theory, statistical field theory, and the theory of self-similar geometric structures, that is used to treat infinities arising in calculated quantities by altering values of these quantities to compensate for effects of their self-interactions. But even if no infinities arose in loop diagrams in quantum field theory, it could be shown that it would be necessary to renormalize the mass and fields appearing in the original Lagrangian.[1]

For example, an electron theory may begin by postulating an electron with an initial mass and charge. In quantum field theory a cloud of virtual particles, such as photons, positrons, and others surrounds and interacts with the initial electron. Accounting for the interactions of the surrounding particles (e.g. collisions at different energies) shows that the electron-system behaves as if it had a different mass and charge than initially postulated. Renormalization, in this example, mathematically replaces the initially postulated mass and charge of an electron with the experimentally observed mass and charge. Mathematics and experiments prove that positrons and more massive particles such as protons exhibit precisely the same observed charge as the electron – even in the presence of much stronger interactions and more intense clouds of virtual particles.

Renormalization specifies relationships between parameters in the theory when parameters describing large distance scales differ from parameters describing small distance scales. Physically, the pileup of contributions from an infinity of scales involved in a problem may then result in further infinities. When describing spacetime as a continuum, certain statistical and quantum mechanical constructions are not well-defined. To define them, or make them unambiguous, a continuum limit must carefully remove "construction scaffolding" of lattices at various scales. Renormalization procedures are based on the requirement that certain physical quantities (such as the mass and charge of an electron) equal observed (experimental) values. That is, the experimental value of the physical quantity yields practical applications, but due to their empirical nature the observed measurement represents areas of quantum field theory that require deeper derivation from theoretical bases.

Renormalization was first developed in quantum electrodynamics (QED) to make sense of infinite integrals in perturbation theory. Initially viewed as a suspect provisional procedure even by some of its originators, renormalization eventually was embraced as an important and self-consistent actual mechanism of scale physics in several fields of physics and mathematics. Despite his later skepticism, it was Paul Dirac who pioneered renormalization.[2][3]

Today, the point of view has shifted: on the basis of the breakthrough renormalization group insights of Nikolay Bogolyubov and Kenneth Wilson, the focus is on variation of physical quantities across contiguous scales, while distant scales are related to each other through "effective" descriptions. All scales are linked in a broadly systematic way, and the actual physics pertinent to each is extracted with the suitable specific computational techniques appropriate for each. Wilson clarified which variables of a system are crucial and which are redundant.

Renormalization is distinct from regularization, another technique to control infinities by assuming the existence of new unknown physics at new scales.

Self-interactions in classical physics

[edit]
Figure 1. Renormalization in quantum electrodynamics: The simple electron/photon interaction that determines the electron's charge at one renormalization point is revealed to consist of more complicated interactions at another.

The problem of infinities first arose in the classical electrodynamics of point particles in the 19th and early 20th century.

The mass of a charged particle should include the mass–energy in its electrostatic field (electromagnetic mass). Assume that the particle is a charged spherical shell of radius re. The mass–energy in the field is which becomes infinite as re → 0. This implies that the point particle would have infinite inertia and thus cannot be accelerated. Incidentally, the value of re that makes equal to the electron mass is called the classical electron radius, which (setting and restoring factors of c and ) turns out to be where is the fine-structure constant, and is the reduced Compton wavelength of the electron.

Renormalization: The total effective mass of a spherical charged particle includes the actual bare mass of the spherical shell (in addition to the mass mentioned above associated with its electric field). If the shell's bare mass is allowed to be negative, it might be possible to take a consistent point limit.[citation needed] This was called renormalization, and Lorentz and Abraham attempted to develop a classical theory of the electron this way. This early work was the inspiration for later attempts at regularization and renormalization in quantum field theory.

(See also regularization (physics) for an alternative way to remove infinities from this classical problem, assuming new physics exists at small scales.)

When calculating the electromagnetic interactions of charged particles, it is tempting to ignore the back-reaction of a particle's own field on itself. (Analogous to the back-EMF of circuit analysis.) But this back-reaction is necessary to explain the friction on charged particles when they emit radiation. If the electron is assumed to be a point, the value of the back-reaction diverges, for the same reason that the mass diverges, because the field is inverse-square.

The Abraham–Lorentz theory had a noncausal "pre-acceleration". Sometimes an electron would start moving before the force is applied. These problems remain in the relativistic version of the Abraham-Lorentz equation. This is a sign that the point limit is inconsistent, or/and that a quantum mechanical treatment is required.[4]

The trouble was worse in classical field theory than in quantum field theory, because in quantum field theory a charged particle experiences Zitterbewegung[citation needed] due to interference with virtual particle–antiparticle pairs, thus effectively smearing out the charge over a region comparable to the Compton wavelength. In quantum electrodynamics at small coupling, the electromagnetic mass only diverges as the logarithm of the radius of the particle.

Divergences in quantum electrodynamics

[edit]

(a) Vacuum polarization, a.k.a. charge screening. This loop has a logarithmic ultraviolet divergence.
(b) Self-energy diagram in QED
(c) Example of a "penguin" diagram

When developing quantum electrodynamics in the 1930s, Max Born, Werner Heisenberg, Pascual Jordan, and Paul Dirac discovered that in perturbative corrections many integrals were divergent (see The problem of infinities).

One way of describing the perturbation theory corrections' divergences was discovered in 1947–49 by Hans Kramers,[5] Hans Bethe,[6] Julian Schwinger,[7][8][9][10] Richard Feynman,[11][12][13] and Shin'ichiro Tomonaga,[14][15][16][17][18][19][20] and systematized by Freeman Dyson in 1949.[21] The divergences appear in radiative corrections involving Feynman diagrams with closed loops of virtual particles in them.

While virtual particles obey conservation of energy and momentum, they can have any energy and momentum, even one that is not allowed by the relativistic energy–momentum relation for the observed mass of that particle (that is, is not necessarily the squared mass of the particle in that process, e.g. for a photon it could be nonzero). Such a particle is called off-shell. When there is a loop, the momentum of the particles involved in the loop is not uniquely determined by the energies and momenta of incoming and outgoing particles. A variation in the energy of one particle in the loop can be balanced by an equal and opposite change in the energy of another particle in the loop, without affecting the incoming and outgoing particles. Thus many variations are possible. So to find the amplitude for the loop process, one must integrate over all possible combinations of energy and momentum that could travel around the loop.

These integrals are often divergent, that is, they give infinite answers. The divergences that are significant are the "ultraviolet" (UV) ones. An ultraviolet divergence can be described as one that comes from

  • the region in the integral where all particles in the loop have large energies and momenta,
  • very short wavelengths and high-frequencies fluctuations of the fields, in the path integral for the field,
  • very short proper-time between particle emission and absorption, if the loop is thought of as a sum over particle paths.

So these divergences are short-distance, short-time phenomena.

Shown in the pictures at the right margin, there are exactly three one-loop divergent loop diagrams in quantum electrodynamics:[22]

  1. A photon creates a virtual electron–positron pair, which then annihilates. This is a vacuum polarization diagram.
  2. An electron quickly emits and reabsorbs a virtual photon, called a self-energy.
  3. An electron emits a photon, emits a second photon, and reabsorbs the first. This process is shown in the section below in figure 2, and it is called a vertex renormalization. The Feynman diagram for this is also called a "penguin diagram" due to its shape resembling a penguin.

The three divergences correspond to the three parameters in the theory under consideration:

  1. The field normalization Z.
  2. The mass of the electron.
  3. The charge of the electron.

The second class of divergence called an infrared divergence, is due to massless particles, like the photon. Every process involving charged particles emits infinitely many coherent photons of infinite wavelength, and the amplitude for emitting any finite number of photons is zero. For photons, these divergences are well understood. For example, at the 1-loop order, the vertex function has both ultraviolet and infrared divergences. In contrast to the ultraviolet divergence, the infrared divergence does not require the renormalization of a parameter in the theory involved. The infrared divergence of the vertex diagram is removed by including a diagram similar to the vertex diagram with the following important difference: the photon connecting the two legs of the electron is cut and replaced by two on-shell (i.e. real) photons whose wavelengths tend to infinity; this diagram is equivalent to the bremsstrahlung process. This additional diagram must be included because there is no physical way to distinguish a zero-energy photon flowing through a loop as in the vertex diagram and zero-energy photons emitted through bremsstrahlung. From a mathematical point of view, the IR divergences can be regularized by assuming fractional differentiation w.r.t. a parameter, for example: is well defined at p = a but is UV divergent; if we take the 32-th fractional derivative with respect to a2, we obtain the IR divergence so we can cure IR divergences by turning them into UV divergences.[clarification needed]

A loop divergence

[edit]
Figure 2. A diagram contributing to electron–electron scattering in QED. The loop has an ultraviolet divergence.

The diagram in Figure 2 shows one of the several one-loop contributions to electron–electron scattering in QED. The electron on the left side of the diagram, represented by the solid line, starts out with 4-momentum pμ and ends up with 4-momentum rμ. It emits a virtual photon carrying rμpμ to transfer energy and momentum to the other electron. But in this diagram, before that happens, it emits another virtual photon carrying 4-momentum qμ, and it reabsorbs this one after emitting the other virtual photon. Energy and momentum conservation do not determine the 4-momentum qμ uniquely, so all possibilities contribute equally and we must integrate.

This diagram's amplitude ends up with, among other things, a factor from the loop of

The various γμ factors in this expression are gamma matrices as in the covariant formulation of the Dirac equation; they have to do with the spin of the electron. The factors of e are the electric coupling constant, while the provide a heuristic definition of the contour of integration around the poles in the space of momenta. The important part for our purposes is the dependency on qμ of the three big factors in the integrand, which are from the propagators of the two electron lines and the photon line in the loop.

This has a piece with two powers of qμ on top that dominates at large values of qμ (Pokorski 1987, p. 122):

This integral is divergent and infinite, unless we cut it off at finite energy and momentum in some way.

Similar loop divergences occur in other quantum field theories.

Renormalized and bare quantities

[edit]

The solution was to realize that the quantities initially appearing in the theory's formulae (such as the formula for the Lagrangian), representing such things as the electron's electric charge and mass, as well as the normalizations of the quantum fields themselves, did not actually correspond to the physical constants measured in the laboratory. As written, they were bare quantities that did not take into account the contribution of virtual-particle loop effects to the physical constants themselves. Among other things, these effects would include the quantum counterpart of the electromagnetic back-reaction that so vexed classical theorists of electromagnetism. In general, these effects would be just as divergent as the amplitudes under consideration in the first place; so finite measured quantities would, in general, imply divergent bare quantities.

To make contact with reality, then, the formulae would have to be rewritten in terms of measurable, renormalized quantities. The charge of the electron, say, would be defined in terms of a quantity measured at a specific kinematic renormalization point or subtraction point (which will generally have a characteristic energy, called the renormalization scale or simply the energy scale). The parts of the Lagrangian left over, involving the remaining portions of the bare quantities, could then be reinterpreted as counterterms, involved in divergent diagrams exactly canceling out the troublesome divergences for other diagrams.

Renormalization in QED

[edit]
Figure 3. The vertex corresponding to the Z1 counterterm cancels the divergence in Figure 2.

For example, in the Lagrangian of QED the fields and coupling constant are really bare quantities, hence the subscript B above. Conventionally the bare quantities are written so that the corresponding Lagrangian terms are multiples of the renormalized ones:

Gauge invariance, via a Ward–Takahashi identity, turns out to imply that we can renormalize the two terms of the covariant derivative piece together (Pokorski 1987, p. 115), which is what happened to Z2; it is the same as Z1.

A term in this Lagrangian, for example, the electron–photon interaction pictured in Figure 1, can then be written

The physical constant e, the electron's charge, can then be defined in terms of some specific experiment: we set the renormalization scale equal to the energy characteristic of this experiment, and the first term gives the interaction we see in the laboratory (up to small, finite corrections from loop diagrams, providing such exotica as the high-order corrections to the magnetic moment). The rest is the counterterm. If the theory is renormalizable (see below for more on this), as it is in QED, the divergent parts of loop diagrams can all be decomposed into pieces with three or fewer legs, with an algebraic form that can be canceled out by the second term (or by the similar counterterms that come from Z0 and Z3).

The diagram with the Z1 counterterm's interaction vertex placed as in Figure 3 cancels out the divergence from the loop in Figure 2.

Historically, the splitting of the "bare terms" into the original terms and counterterms came before the renormalization group insight due to Kenneth Wilson.[23] According to such renormalization group insights, detailed in the next section, this splitting is unnatural and actually unphysical, as all scales of the problem enter in continuous systematic ways.

Running couplings

[edit]

To minimize the contribution of loop diagrams to a given calculation (and therefore make it easier to extract results), one chooses a renormalization point close to the energies and momenta exchanged in the interaction. However, the renormalization point is not itself a physical quantity: the physical predictions of the theory, calculated to all orders, should in principle be independent of the choice of renormalization point, as long as it is within the domain of application of the theory. Changes in renormalization scale will simply affect how much of a result comes from Feynman diagrams without loops, and how much comes from the remaining finite parts of loop diagrams. One can exploit this fact to calculate the effective variation of physical constants with changes in scale. This variation is encoded by beta-functions, and the general theory of this kind of scale-dependence is known as the renormalization group.

Colloquially, particle physicists often speak of certain physical "constants" as varying with the energy of interaction, though in fact, it is the renormalization scale that is the independent quantity. This running does, however, provide a convenient means of describing changes in the behavior of a field theory under changes in the energies involved in an interaction. For example, since the coupling in quantum chromodynamics becomes small at large energy scales, the theory behaves more like a free theory as the energy exchanged in an interaction becomes large – a phenomenon known as asymptotic freedom. Choosing an increasing energy scale and using the renormalization group makes this clear from simple Feynman diagrams; were this not done, the prediction would be the same, but would arise from complicated high-order cancellations.

For example, is ill-defined.

To eliminate the divergence, simply change lower limit of integral into εa and εb:

Making sure εb/εa → 1, then I = ln a/b.

Regularization

[edit]

Since the quantity ∞ − ∞ is ill-defined, in order to make this notion of canceling divergences precise, the divergences first have to be tamed mathematically using the theory of limits, in a process known as regularization (Weinberg, 1995).

An essentially arbitrary modification to the loop integrands, or regulator, can make them drop off faster at high energies and momenta, in such a manner that the integrals converge. A regulator has a characteristic energy scale known as the cutoff; taking this cutoff to infinity (or, equivalently, the corresponding length/time scale to zero) recovers the original integrals.

With the regulator in place, and a finite value for the cutoff, divergent terms in the integrals then turn into finite but cutoff-dependent terms. After canceling out these terms with the contributions from cutoff-dependent counterterms, the cutoff is taken to infinity and finite physical results are recovered. If physics on scales we can measure is independent of what happens at the very shortest distance and time scales, then it should be possible to get cutoff-independent results for calculations.

Many different types of regulator are used in quantum field theory calculations, each with its advantages and disadvantages. One of the most popular in modern use is dimensional regularization, invented by Gerardus 't Hooft and Martinus J. G. Veltman,[24] which tames the integrals by carrying them into a space with a fictitious fractional number of dimensions. Another is Pauli–Villars regularization, which adds fictitious particles to the theory with very large masses, such that loop integrands involving the massive particles cancel out the existing loops at large momenta.

Yet another regularization scheme is the lattice regularization, introduced by Kenneth Wilson, which pretends that hyper-cubical lattice constructs our spacetime with fixed grid size. This size is a natural cutoff for the maximal momentum that a particle could possess when propagating on the lattice. And after doing a calculation on several lattices with different grid size, the physical result is extrapolated to grid size 0, or our natural universe. This presupposes the existence of a scaling limit.

A rigorous mathematical approach to renormalization theory is the so-called causal perturbation theory, where ultraviolet divergences are avoided from the start in calculations by performing well-defined mathematical operations only within the framework of distribution theory. In this approach, divergences are replaced by ambiguity: corresponding to a divergent diagram is a term which now has a finite, but undetermined, coefficient. Other principles, such as gauge symmetry, must then be used to reduce or eliminate the ambiguity.

Attitudes and interpretation

[edit]

The early formulators of QED and other quantum field theories were, as a rule, dissatisfied with this state of affairs. It seemed illegitimate to do something tantamount to subtracting infinities from infinities to get finite answers.

Dyson argued that these infinities are of a basic nature and cannot be eliminated by any formal mathematical procedures, such as the renormalization method.[25][26]

Dirac's criticism was the most persistent.[27] As late as 1975, he was saying:[28]

Most physicists are very satisfied with the situation. They say: 'Quantum electrodynamics is a good theory and we do not have to worry about it any more.' I must say that I am very dissatisfied with the situation because this so-called 'good theory' does involve neglecting infinities which appear in its equations, ignoring them in an arbitrary way. This is just not sensible mathematics. Sensible mathematics involves disregarding a quantity when it is small – not neglecting it just because it is infinitely great and you do not want it!

Another important critic was Feynman. Despite his crucial role in the development of quantum electrodynamics, he wrote the following in 1985:[29]

The shell game that we play to find n and j is technically called 'renormalization'. But no matter how clever the word, it is still what I would call a dippy process! Having to resort to such hocus-pocus has prevented us from proving that the theory of quantum electrodynamics is mathematically self-consistent. It's surprising that the theory still hasn't been proved self-consistent one way or the other by now; I suspect that renormalization is not mathematically legitimate.

Feynman was concerned that all field theories known in the 1960s had the property that the interactions become infinitely strong at short enough distance scales. This property called a Landau pole, made it plausible that quantum field theories were all inconsistent. In 1974, David Gross, Hugh David Politzer and Frank Wilczek showed that another quantum field theory, quantum chromodynamics, does not have a Landau pole. Feynman, along with most others, accepted that quantum chromodynamics was a fully consistent theory.[citation needed]

The general unease was almost universal in texts up to the 1970s and 1980s. Beginning in the 1970s, however, inspired by work on the renormalization group and effective field theory, and despite the fact that Dirac and various others—all of whom belonged to the older generation—never withdrew their criticisms, attitudes began to change, especially among younger theorists. Kenneth G. Wilson and others demonstrated that the renormalization group is useful in statistical field theory applied to condensed matter physics, where it provides important insights into the behavior of phase transitions. In condensed matter physics, a physical short-distance regulator exists: matter ceases to be continuous on the scale of atoms. Short-distance divergences in condensed matter physics do not present a philosophical problem since the field theory is only an effective, smoothed-out representation of the behavior of matter anyway; there are no infinities since the cutoff is always finite, and it makes perfect sense that the bare quantities are cutoff-dependent.

If quantum field theory holds all the way down past the Planck length (where it might yield to string theory, causal set theory or something different), then there may be no real problem with short-distance divergences in particle physics either; all field theories could simply be effective field theories. In a sense, this approach echoes the older attitude that the divergences in quantum field theory speak of human ignorance about the workings of nature, but also acknowledges that this ignorance can be quantified and that the resulting effective theories remain useful.

Be that as it may, Abdus Salam's remark[30] in 1972 seems still relevant

Field-theoretic infinities – first encountered in Lorentz's computation of electron self-mass – have persisted in classical electrodynamics for seventy and in quantum electrodynamics for some thirty-five years. These long years of frustration have left in the subject a curious affection for the infinities and a passionate belief that they are an inevitable part of nature; so much so that even the suggestion of a hope that they may, after all, be circumvented — and finite values for the renormalization constants computed – is considered irrational. Compare Russell's postscript to the third volume of his autobiography The Final Years, 1944–1969 (George Allen and Unwin, Ltd., London 1969),[31] p. 221:

In the modern world, if communities are unhappy, it is often because they have ignorances, habits, beliefs, and passions, which are dearer to them than happiness or even life. I find many men in our dangerous age who seem to be in love with misery and death, and who grow angry when hopes are suggested to them. They think hope is irrational and that, in sitting down to lazy despair, they are merely facing facts.

In quantum field theory, the value of a physical constant, in general, depends on the scale that one chooses as the renormalization point, and it becomes very interesting to examine the renormalization group running of physical constants under changes in the energy scale. The coupling constants in the Standard Model of particle physics vary in different ways with increasing energy scale: the coupling of quantum chromodynamics and the weak isospin coupling of the electroweak force tend to decrease, and the weak hypercharge coupling of the electroweak force tends to increase. At the colossal energy scale of 1015 GeV (far beyond the reach of our current particle accelerators), they all become approximately the same size (Grotz and Klapdor 1990, p. 254), a major motivation for speculations about grand unified theory. Instead of being only a worrisome problem, renormalization has become an important theoretical tool for studying the behavior of field theories in different regimes.

If a theory featuring renormalization (e.g. QED) can only be sensibly interpreted as an effective field theory, i.e. as an approximation reflecting human ignorance about the workings of nature, then the problem remains of discovering a more accurate theory that does not have these renormalization problems. As Lewis Ryder has put it, "In the Quantum Theory, these [classical] divergences do not disappear; on the contrary, they appear to get worse. And despite the comparative success of renormalisation theory, the feeling remains that there ought to be a more satisfactory way of doing things."[32]

Renormalizability

[edit]

From this philosophical reassessment, a new concept follows naturally: the notion of renormalizability. Not all theories lend themselves to renormalization in the manner described above, with a finite supply of counterterms and all quantities becoming cutoff-independent at the end of the calculation. If the Lagrangian contains combinations of field operators of high enough dimension in energy units, the counterterms required to cancel all divergences proliferate to infinite number, and, at first glance, the theory would seem to gain an infinite number of free parameters and therefore lose all predictive power, becoming scientifically worthless. Such theories are called nonrenormalizable.

The Standard Model of particle physics contains only renormalizable operators, but the interactions of general relativity become nonrenormalizable operators if one attempts to construct a field theory of quantum gravity in the most straightforward manner (treating the metric in the Einstein–Hilbert Lagrangian as a perturbation about the Minkowski metric), suggesting that perturbation theory is not satisfactory in application to quantum gravity.

However, in an effective field theory, "renormalizability" is, strictly speaking, a misnomer. In nonrenormalizable effective field theory, terms in the Lagrangian do multiply to infinity, but have coefficients suppressed by ever-more-extreme inverse powers of the energy cutoff. If the cutoff is a real, physical quantity—that is, if the theory is only an effective description of physics up to some maximum energy or minimum distance scale—then these additional terms could represent real physical interactions. Assuming that the dimensionless constants in the theory do not get too large, one can group calculations by inverse powers of the cutoff, and extract approximate predictions to finite order in the cutoff that still have a finite number of free parameters. It can even be useful to renormalize these "nonrenormalizable" interactions.

Nonrenormalizable interactions in effective field theories rapidly become weaker as the energy scale becomes much smaller than the cutoff. The classic example is the Fermi theory of the weak nuclear force, a nonrenormalizable effective theory whose cutoff is comparable to the mass of the W particle. This fact may also provide a possible explanation for why almost all of the particle interactions we see are describable by renormalizable theories. It may be that any others that may exist at the GUT or Planck scale simply become too weak to detect in the realm we can observe, with one exception: gravity, whose exceedingly weak interaction is magnified by the presence of the enormous masses of stars and planets.[citation needed]

Renormalization schemes

[edit]

In actual calculations, the counterterms introduced to cancel the divergences in Feynman diagram calculations beyond tree level must be fixed using a set of renormalisation conditions. The common renormalization schemes in use include:

Besides, there exists a "natural" definition of the renormalized coupling (combined with the photon propagator) as a propagator of dual free bosons, which does not explicitly require introducing the counterterms.[33]

In statistical physics

[edit]

History

[edit]

A deeper understanding of the physical meaning and generalization of the renormalization process, which goes beyond the dilatation group of conventional renormalizable theories, came from condensed matter physics. Leo P. Kadanoff's paper in 1966 proposed the "block-spin" renormalization group.[34] The blocking idea is a way to define the components of the theory at large distances as aggregates of components at shorter distances.

This approach covered the conceptual point and was given full computational substance[23] in the extensive important contributions of Kenneth Wilson. The power of Wilson's ideas was demonstrated by a constructive iterative renormalization solution of a long-standing problem, the Kondo problem, in 1974, as well as the preceding seminal developments of his new method in the theory of second-order phase transitions and critical phenomena in 1971. He was awarded the Nobel Prize in Physics for these decisive contributions in 1982.

Principles

[edit]

In more technical terms, let us assume that we have a theory described by a certain function of the state variables and a certain set of coupling constants . This function may be a partition function, an action, a Hamiltonian, etc. It must contain the whole description of the physics of the system.

Now we consider a certain blocking transformation of the state variables , the number of must be lower than the number of . Now let us try to rewrite the function only in terms of the . If this is achievable by a certain change in the parameters, , then the theory is said to be renormalizable.

The possible macroscopic states of the system, at a large scale, are given by this set of fixed points.

Renormalization group fixed points

[edit]

The most important information in the RG flow is its fixed points. A fixed point is defined by the vanishing of the beta function associated to the flow. Then, fixed points of the renormalization group are by definition scale invariant. In many cases of physical interest scale invariance enlarges to conformal invariance. One then has a conformal field theory at the fixed point.

The ability of several theories to flow to the same fixed point leads to universality.

If these fixed points correspond to free field theory, the theory is said to exhibit quantum triviality. Numerous fixed points appear in the study of lattice Higgs theories, but the nature of the quantum field theories associated with these remains an open question.

See also

[edit]

References

[edit]

Further reading

[edit]
[edit]
Revisions and contributorsEdit on WikipediaRead on Wikipedia
from Grokipedia
Renormalization is a systematic procedure in (QFT) that addresses ultraviolet divergences—infinities arising in perturbative expansions—by redefining bare parameters like masses, charges, and fields in terms of finite, observable quantities through the introduction of counterterms, thereby rendering calculations predictive and consistent with experimental data. An earlier historical example of resolving infinities arising from incompatible physical theories is the ultraviolet catastrophe of the late 19th century. Classical electromagnetism combined with the equipartition theorem of statistical mechanics predicted infinite energy emission from blackbodies at short (ultraviolet) wavelengths according to the Rayleigh-Jeans law. Max Planck resolved this in 1900 by introducing the hypothesis that energy is emitted and absorbed in discrete quanta, leading to Planck's law and the origin of quantum theory. This resolution required a fundamental modification to the underlying theory through quantization. In contrast, divergences in quantum electrodynamics (QED) were handled through renormalization by redefining parameters rather than altering the theory's core framework. The origins of renormalization trace back to the late 1940s amid challenges in (QED), where combining quantum mechanics with special relativity and classical electromagnetism produced infinite divergences (e.g., electron self-energy and vacuum polarization) in perturbative calculations during the 1930s–1940s. Calculations of processes like the yielded infinite results due to interactions with the vacuum. In 1947, pioneered its application by computing the electromagnetic shift in energy levels, effectively absorbing the divergence into the electron's mass renormalization to match the observed of approximately 1057 MHz. This insight was rapidly generalized by , , and Shin'ichiro Tomonaga, who developed covariant formulations of QED, with providing a rigorous perturbative framework in 1949 that demonstrated the renormalizability of the theory to all orders in the . These efforts transformed QED from a plagued theory into one capable of predictions accurate to parts per billion, such as the electron's anomalous . Beyond QED, renormalization proved essential for non-Abelian gauge theories, including (QCD) and electroweak theory, forming the backbone of the . In 1971, established the renormalizability of these theories, confirming that infinities could be absorbed into a finite number of parameters while preserving gauge invariance. 't Hooft and Martinus Veltman later introduced in 1972 as a key tool for handling these infinities. The (RG), first conceptualized by Ernst Stueckelberg and André Petermann in 1951 as a transformation group acting on coupling constants, later revealed how parameters "run" with energy scale via beta functions, explaining phenomena like the unification of forces at high energies. Wilson's 1971 formulation of the RG for lattice models bridged QFT with , enabling the study of and phase transitions, for which he received the 1982 . Today, renormalization remains indispensable for beyond-Standard-Model physics, effective field theories, and lattice simulations, with techniques like the Bogoliubov-Parasiuk-Hepp-Zimmermann (BPHZ) theorem ensuring its mathematical rigor in handling all-order divergences. Its success underscores QFT's power in describing nature across scales, from subatomic particles to condensed matter systems.

Motivations from

Self-interactions in electrodynamics

In classical electrodynamics, a interacts with its own , leading to self-interactions that manifest as both radiation reaction and contributions. These effects become problematic for point-like charges, as the particle's own field exerts a back-reaction force during , known as the radiation reaction or self-force. The Abraham-Lorentz formula captures this self-force in the non-relativistic limit, expressing it as Frad=2e23c3a˙\mathbf{F}_{\mathrm{rad}} = \frac{2 e^2}{3 c^3} \dot{\mathbf{a}}, where ee is the charge, cc is the , and a˙\dot{\mathbf{a}} is the time derivative of the (jerk). This formula arises from the Larmor radiation power and momentum conservation, but its derivation assumes a finite-sized charge distribution to avoid immediate infinities in the field; for a true point particle, the accelerating charge would experience an ill-defined, divergent self-force due to the singular nature of its own Coulomb field. A related issue emerges in the computation of the electron's electromagnetic , which represents the infinite energy stored in the particle's own . For a point charge, the is obtained by integrating the electrostatic u=E28πu = \frac{E^2}{8\pi} (in ) over all space: U=e28πr44πr2dr=e22rmindrr2U = \int \frac{e^2}{8\pi r^4} \, 4\pi r^2 \, dr = \frac{e^2}{2} \int_{r_{\min}}^\infty \frac{dr}{r^2}, which diverges linearly as 1/rmin1/r_{\min} when the lower rmin0r_{\min} \to 0. To render this finite, early models introduced a RR as a , modeling the electron as a charged ; the resulting is then U=23e2RU = \frac{2}{3} \frac{e^2}{R}, which diverges as R0R \to 0 and implies an infinite electromagnetic mass contribution mem=23e2Rc2m_{\mathrm{em}} = \frac{2}{3} \frac{e^2}{R c^2}. This divergence highlighted the instability of point-particle models, as the would dominate and render the electron's total mass infinite without an arbitrary . In 1938, addressed these infinities in his analysis of the classical theory of radiating , proposing a model where the possesses a finite radius to ensure a well-defined and avoid divergent self-energies and forces. 's approach treated the as a rigid charged , deriving that incorporated radiation reaction while maintaining finite quantities, though it required additional stabilizing mechanisms that remained unresolved. These classical divergences in self-interactions foreshadowed similar infinities in , necessitating renormalization techniques.

The Ultraviolet Catastrophe

An earlier historical example of incompatible infinities in classical physics is the ultraviolet catastrophe. In the late 19th century, classical electromagnetism combined with the equipartition theorem of statistical mechanics predicted that a blackbody at thermal equilibrium would emit infinite energy at short (ultraviolet) wavelengths, according to the Rayleigh–Jeans law. This law implied that spectral energy density increases with the square of frequency, leading to unbounded energy radiation at high frequencies. This prediction conflicted with experimental observations of blackbody spectra, which exhibit finite energy emission peaking at finite frequencies. The issue was resolved in 1900 by Max Planck, who introduced the concept of energy quanta—discrete packets of energy proportional to frequency (E=hνE = h\nu). This hypothesis yielded Planck's law for blackbody radiation, accurately matching experiments and eliminating the divergence at high frequencies, laying the foundation for quantum theory.

Early historical attempts

In the early , classical theory faced challenges from the infinite of point charges, prompting initial efforts to manage these divergences through adjustments. developed a model of the as a charged sphere with a finite radius, introducing a natural cutoff to the electromagnetic integral; this allowed the divergent electromagnetic mass to be absorbed into the observed inertial mass, marking an early form of mass renormalization. Henri Poincaré extended this framework in his 1905 and 1906 papers, addressing inconsistencies such as the "4/3 problem" where the electromagnetic momentum did not match the expected inertial mass. He proposed balancing the electromagnetic mass with non-electromagnetic "Poincaré stresses" or cohesive forces within the , effectively renormalizing the total mass by compensating for the electromagnetic contribution without altering the underlying theory.) By the 1930s, as advanced toward field theory formulations, and encountered similar issues in their attempts to quantize electrodynamics. In their work on quantum dynamics of wave fields, they identified divergent integrals in electron self-energy and , suggesting the inclusion of infinite counterterms to cancel these infinities and restore finite observables, though without a systematic procedure. Heisenberg further elaborated in 1934 on positron theory, highlighting logarithmic divergences that required such counterterms for consistency in perturbative expansions. A pivotal semi-empirical advance came in 1947 with Hans Bethe's calculation of the , observed experimentally that year. Bethe treated the effect as a finite energy shift by imposing a cutoff at the electron's , yielding a correction of approximately 1040 MHz to the 2S state energy without invoking full renormalization; this approach reconciled theory with measurement and foreshadowed modern techniques.

Divergences in Quantum Field Theory

Loop divergences in QED

In (QED), perturbative calculations reveal ultraviolet divergences arising from loop diagrams in Feynman , where virtual particles propagate in closed loops with arbitrarily high momenta. These infinities first became apparent in the late 1940s through detailed computations of higher-order corrections to basic processes. and independently identified these divergences in 1949, demonstrating that they appear systematically in QED's loop expansions and necessitate a reappraisal of the theory's foundational parameters. A prominent example is the one-loop correction to the electron self-energy, depicted in the Feynman diagram where an electron line emits a that loops back to rejoin the same line. This diagram contributes to the electron's mass renormalization, yielding a divergent shift δmmln(Λ/m)\delta m \propto m \ln(\Lambda / m). The is logarithmic, arising from the high-momentum region of the involving the photon and propagators. This echoes classical electrodynamics' infinite for a point charge, but in QED, it emerges quantum mechanically from the photon's massless . Another key loop diagram is , where a propagator is corrected by a closed loop of electron-positron pairs. This insertion modifies the photon's effective charge screening and introduces a logarithmic in the photon self-energy function Π(q2)\Pi(q^2), proportional to ln(Λ2/m2)\ln(\Lambda^2 / m^2) at large momenta, where mm is the . Dyson showed that such loop contributions pervade QED amplitudes, with divergences isolated to a few primitive graphs like and vertex corrections.

General perturbative expansions

In perturbative quantum field theories beyond (QED), such as scalar and gauge theories, the expansion in powers of the reveals divergences arising from loop integrals in Feynman diagrams, similar to the loop corrections observed in QED. These divergences manifest in higher-order terms of the elements or functions, necessitating regularization and renormalization to extract finite physical predictions. The structure of these expansions depends on the field's spin and the form of interactions, with power-counting providing an initial assessment of potential divergences. Infrared divergences can also appear in theories with massless particles, requiring additional resummation techniques like the Bloch-Nordsieck theorem in QED. A prototypical example is scalar ϕ4\phi^4 theory in four dimensions, described by the Lagrangian L=12μϕμϕ12m2ϕ2λ4!ϕ4\mathcal{L} = \frac{1}{2} \partial_\mu \phi \partial^\mu \phi - \frac{1}{2} m^2 \phi^2 - \frac{\lambda}{4!} \phi^4. At one loop, the tadpole diagram—a propagator closing into a loop attached to an external leg—generates a quadratic in the Σ(p2)\Sigma(p^2), which shifts the bare mass parameter and requires mass renormalization to absorb the . At two loops, the sunset diagram, consisting of two propagators forming a loop connected by a third propagator to the external legs, introduces further divergences, including logarithmic terms that contribute to both mass and field-strength (wave-function) renormalizations, as well as influencing the coupling constant through higher-order vertex corrections. These diagrams illustrate how subdivergences and overall divergences in ϕ4\phi^4 theory can be systematically absorbed into redefinitions of the bare parameters, rendering the theory renormalizable. In non-Abelian gauge theories, such as (QCD), perturbative expansions encounter more complex divergences due to the self-interacting nature of gauge bosons. To handle and preserve the structure of the theory, Faddeev-Popov fields are introduced, which are anticommuting scalar fields that cancel unphysical in the path integral. The Becchi-Rouet-Stora-Tyutin (BRST) , a nilpotent global transformation mixing gauge fields, s, and antighosts, ensures that renormalization respects gauge invariance, allowing divergent contributions from gluon self-interactions and quark loops to be absorbed into renormalized parameters without violating the Ward-Slavnov-Taylor identities. A key distinction in non-Abelian theories like QCD is the phenomenon of , where the strong αs\alpha_s decreases as the energy scale increases, in contrast to the running in QED that grows logarithmically at high energies. This behavior, arising from the negative contribution of non-Abelian vertex terms in the , implies that perturbative expansions become more reliable at short distances or high momenta. The discovery of resolved the issue of confinement in strong interactions and validated QCD as the theory of the strong force. To classify the severity of divergences in general perturbative diagrams, power-counting yields the superficial degree of divergence D=4LpropagatorsδpIp+verticesδvVvD = 4L - \sum_{\rm propagators} \delta_p I_p + \sum_{\rm vertices} \delta_v V_v, where LL is the number of loops, δp\delta_p is the dimension of the propagator (e.g., 2 for scalars/photons, 1 for fermions), IpI_p the number of internal propagators of that type, δv\delta_v the engineering dimension of the vertex coupling (0 for marginal interactions like ϕ4\phi^4 or gauge couplings), and VvV_v the number of such vertices. If D0D \geq 0, the diagram is potentially divergent, guiding the identification of counterterms needed for renormalization. For QED specifically, D=432EeEγD = 4 - \frac{3}{2} E_e - E_\gamma, where EeE_e and EγE_\gamma are the numbers of external electron and photon legs, respectively.

Regularization Techniques

Momentum cutoff methods

Momentum cutoff methods regularize ultraviolet divergences in by imposing an upper limit Λ\Lambda on the momenta entering loop integrals, thereby rendering them finite while preserving the four-dimensional structure of the theory. This artificial scale Λ\Lambda represents a high-energy , beyond which contributions from virtual particles are neglected, allowing perturbative calculations to proceed systematically. The method introduces no fundamental new physics but serves as a temporary tool, with physical predictions obtained in the limit Λ\Lambda \to \infty after renormalization absorbs cutoff-dependent terms into redefined parameters. A prominent implementation is the hard cutoff, which sharply restricts integration to regions where the p<Λ|p| < \Lambda. This is often applied directly to propagators, modifying the free scalar propagator to 1p2m2+iϵθ(Λp),\frac{1}{p^2 - m^2 + i\epsilon} \theta(\Lambda - |p|), where θ\theta denotes the Heaviside step function. Such a truncation makes divergent integrals converge but can violate symmetries like gauge invariance unless carefully adjusted. To address these issues while maintaining covariance, the Pauli-Villars regulator introduces auxiliary "ghost" fields with large fictitious masses MiΛM_i \gg \Lambda, which contribute oppositely to the original fields and cancel divergences in combinations like i(1)i/(p2Mi2)\sum_i (-1)^i / (p^2 - M_i^2). This technique, originally developed for invariant regularization in relativistic quantum field theories, effectively mimics a cutoff through the rapid falloff of the regulator propagators at high energies. In contrast, soft cutoff schemes apply gradual suppression to high momenta, avoiding the discontinuities of hard cutoffs that may complicate analytic continuation or numerical stability. A typical form involves multiplying integrands by factors like eδk2/Λ2e^{-\delta k^2 / \Lambda^2}, where δ>0\delta > 0 controls the decay rate, providing an exponentially damped tail that preserves more symmetries and eases the treatment of Lorentz-invariant theories. These smooth regulators are particularly advantageous in momentum-space formulations where sharp boundaries might introduce artifacts. Momentum cutoff regularization was instrumental in Kenneth G. Wilson's pioneering work on the , where it naturally arises in lattice discretizations of field theories, enforcing a finite range through the inverse lattice spacing a1Λa^{-1} \sim \Lambda.

Dimensional regularization

Dimensional regularization is a technique in that addresses ultraviolet divergences by analytically continuing Feynman integrals from four dimensions to a general complex dimension d=4ϵd = 4 - \epsilon, where ϵ\epsilon is a small positive parameter, before expanding around ϵ=0\epsilon = 0 to isolate and handle the resulting poles. This method was introduced in 1972 by several groups, including Gerard 't Hooft and Martinus Veltman in their paper on gauge field renormalization, as well as independently by L. J. C. Biedenharn et al. and F. J. Yndurain, providing a framework particularly suited for theories with symmetries, such as gauge invariance, which might be violated by other regularization approaches. In practice, loop momentum integrals are evaluated in [d](/page/D)[d](/page/D*) dimensions, yielding expressions that are finite for non-integer [d](/page/D)[d](/page/D*) but develop simple poles 1/[ϵ](/page/Epsilon)1/[\epsilon](/page/Epsilon) as [ϵ](/page/Epsilon)0[\epsilon](/page/Epsilon) \to 0, corresponding to the logarithmic divergences of the original four-dimensional theory. These poles arise from the analytic structure of the integrals and are systematically subtracted in the renormalization procedure, while finite parts contribute to physical predictions. The approach introduces an arbitrary mass scale μ\mu to maintain dimensional consistency, as the coupling constants and fields acquire anomalous dimensions under this continuation. A key tool in evaluating these integrals is the representation in terms of s, which facilitate the . For instance, the basic scalar takes the form ddk(2π)d1k2α=Γ(αd2)(4π)d/2Γ(α)μd2α,\int \frac{d^d k}{(2\pi)^d} \frac{1}{k^{2\alpha}} = \frac{\Gamma\left(\alpha - \frac{d}{2}\right)}{(4\pi)^{d/2} \Gamma(\alpha)} \mu^{d - 2\alpha}, where μ\mu is the dimensional regulator scale, and the Γ(z)\Gamma(z) encodes the poles when αd/2\alpha - d/2 is a non-positive . This formula, derived from hyperspherical coordinates and the properties of the , allows explicit computation of one-loop and higher-order integrals by reducing them to combinations of such propagators. Dimensional regularization proves especially advantageous for massless theories, as it automatically preserves gauge invariance without introducing spurious mass scales that could break Lorentz or gauge symmetries, unlike explicit cutoff methods which impose hard limits. Additionally, scaleless s—those without a characteristic scale, such as ddk(k2)α\int d^d k \, (k^2)^{-\alpha}—vanish identically in this scheme for α>0\alpha > 0, because the integral over all scales uniformly and the analytic continuation yields zero due to the identity Γ(β)Γ(1β)=π/sin(πβ)\Gamma(\beta) \Gamma(1 - \beta) = \pi / \sin(\pi \beta) balancing UV and IR divergences. This property simplifies calculations in conformal or massless limits but requires careful handling of potential infrared issues separately.

Core Concepts of Renormalization

Bare versus renormalized parameters

In , perturbative calculations often yield divergent expressions due to divergences, necessitating a distinction between bare parameters, which are unphysical quantities in the original Lagrangian, and renormalized parameters, which correspond to observable physical quantities. The bare parameters, such as the bare mass m0m_0 and bare charge e0e_0, are infinite in the limit where the regulator is removed, but they arise as limits of finite, regulator-dependent values that absorb these divergences. The relationship between bare and renormalized parameters is established through renormalization constants, for example, m0=Zmmm_0 = Z_m m and e0=Zeee_0 = Z_e e, where mm and ee are the renormalized and charge, respectively, and the constants take the form Zm=1+δmZ_m = 1 + \delta m with δm\delta m containing divergent contributions. These renormalization constants ZZ are determined order by order in by requiring that physical observables remain finite and match experimental values. The bare Lagrangian density is expressed in terms of the renormalized fields and parameters through field rescalings ϕ0=Zϕ1/2ϕR\phi_0 = Z_\phi^{1/2} \phi_R and renormalization constants that vary by term, ensuring each operator in LR\mathcal{L}_R is multiplied by its appropriate ZZ factor (e.g., kinetic term by ZϕZ_\phi, mass term by ZmZ_m). This framework ensures that the theory's predictions are regulator-independent and physically meaningful, with bare parameters serving as auxiliary constructs rather than directly measurable entities.

Counterterms and renormalization conditions

In , counterterms are additional terms introduced into the Lagrangian to systematically cancel the ultraviolet divergences arising from loop integrals in perturbative expansions. These counterterms are constructed such that their divergent contributions precisely offset the infinities in the bare Green's functions, rendering the finite after renormalization. The counterterm Lagrangian typically takes the form δL=δmψˉψδZψψˉi\slashψ+\delta \mathcal{L} = -\delta m \bar{\psi} \psi - \delta Z_\psi \bar{\psi} i \slash{\partial} \psi + \cdots , where δm\delta m and δZψ\delta Z_\psi are the mass and field renormalization counterterms, respectively, for a fermionic field ψ\psi. This structure ensures that the divergent parts are absorbed into redefinitions of the bare parameters. The procedure for implementing counterterms involves computing the perturbative expansion of Green's functions, isolating their divergent portions using a regularization scheme (such as dimensional regularization), and then subtracting these divergences through the counterterms. For instance, the one-particle irreducible self-energy function Σ(p)\Sigma(p) of a field receives divergent corrections from loops, and the counterterms are chosen to cancel the poles (e.g., 1/ϵ1/\epsilon terms in dimensional regularization) in the Laurent expansion around the physical point. This subtraction leaves finite remainders that correspond to physical observables. In renormalizable theories, this process applies multiplicatively to all orders, meaning that renormalized Green's functions are related to bare ones by finite renormalization factors ZiZ_i, ensuring consistent finite predictions without introducing new divergences at higher loops. Renormalization conditions are imposed to fix the finite portions of the counterterms, thereby defining the renormalized parameters in terms of measurable quantities. In the on-shell scheme, for example, the self-energy is required to vanish at the physical mass shell, Σ(=m)=0\Sigma(\not{p} = m) = 0, and the wave function renormalization ensures a unit residue at the pole, ddΣ()=m=0\frac{d}{d\not{p}} \Sigma(\not{p}) \big|_{\not{p}=m} = 0
Add your contribution
Related Hubs
User Avatar
No comments yet.