<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0"
	xmlns:atom="http://www.w3.org/2005/Atom"
	xmlns:content="http://purl.org/rss/1.0/modules/content/">
	<channel>
		<title>ields</title>
		<description>
			
				Fields - blog about mathematics, physics and philosophy.
			
		</description>
		<link>https://fields007.github.io</link>
		<atom:link href="https://fields007.github.io/articles/feed.xml" rel="self" type="application/rss+xml" />
		
			<item>
				<title>Emergent gravity</title>
				
					<category>general relativity></category>
				
					<category>holography></category>
				
					<category>black holes></category>
				
					<category>thermodynamics></category>
				
					<category>academic work></category>
				
				<description>This article is based on my research project1 in mathematical physics. I outline the central ideas in broader terms. The project examines the possibility that gravity may be an emergent phenomenon rather than a fundamental interaction. It begins with black hole thermodynamics and entropy bounds, including Bousso’s bound, and discusses the holographic principle and its realization in the AdS/CFT correspondence. Finally, it presents selected approaches to emergent gravity, most notably Jacobson’s thermodynamic derivation of Einstein’s equations and Verlinde’s entropic gravity.
</description>
				<image>
					<url>http://www.w3schools.com/images/logo.gif</url>
					<title>Illustration of the AdS/CFT correspondence</title>
					<link>/data/img/thumbnails/anicky_kresba.jpeg</link>
				</image>
				<content type="html">
					
						<b>Warning:</b> this post contains math blocks rendered with the
						<a href="https://www.mathjax.org/">MathJax</a> JavaScript library. If you're using an RSS reader or
						otherwise have Javascript disabled, none of them will display, so you should instead read this article at
						its <a href="https://fields007.github.io/articles/emergent_gravity">source</a>.
					
					<p>This article is based on my research project<sup id="link1"><a href="#note1">1</a></sup> in mathematical physics. I outline the central ideas in broader terms. The project examines the possibility that gravity may be an emergent phenomenon rather than a fundamental interaction. It begins with black hole thermodynamics and entropy bounds, including Bousso’s bound, and discusses the holographic principle and its realization in the AdS/CFT correspondence. Finally, it presents selected approaches to emergent gravity, most notably Jacobson’s thermodynamic derivation of Einstein’s equations and Verlinde’s entropic gravity.</p>

<h2 id="why-consider-gravity-as-emergent">Why consider gravity as emergent?</h2>

<p>One of the central open problems in modern theoretical physics is the formulation of a consistent theory of quantum gravity. General relativity describes gravity as curvature of spacetime, governed by the Einstein equations</p>

\[R_{\mu\nu} - \frac{1}{2} R g_{\mu\nu}
= 8\pi G \, T_{\mu\nu}.\]

<p>At low energies this classical theory works remarkably well. However, attempts to quantize gravity directly face serious difficulties. From the point of view of quantum field theory, general relativity is non-renormalizable, even though it can be treated as an effective field theory at low energies.</p>

<p>These obstacles motivate a different question. Instead of quantizing gravity in the same way as other interactions, could gravity itself be an emergent phenomenon, analogous to hydrodynamics or elasticity?</p>

<p>An important hint in this direction appeared already in the 1970s through an unexpected connection between black holes and thermodynamics.</p>

<h2 id="black-hole-thermodynamics">Black hole thermodynamics</h2>

<p>In 1973, Bardeen, Carter and Hawking formulated the <strong>four laws of black hole dynamics</strong><sup id="link2"><a href="#note2">2</a></sup>.<br />
These are theorems of classical general relativity, yet they closely resemble the laws of thermodynamics.</p>

<p>For a stationary black hole:</p>

<ol>
  <li>
    <p>The surface gravity \(\kappa\) is constant over the event horizon.</p>
  </li>
  <li>
    <p>For stationary perturbations one has</p>

\[\delta M =
\frac{\kappa}{8\pi} \, \delta A
+ \omega \, \delta J
+ \varphi \, \delta Q ,\]

    <p>where \(M\) is the mass, \(A\) the horizon area, \(J\) the angular momentum and \(Q\) the charge.</p>
  </li>
  <li>
    <p>The area of the event horizon never decreases.</p>
  </li>
  <li>
    <p>It is impossible to reduce \(\kappa\) to zero by a finite sequence of physical operations.</p>
  </li>
</ol>

<p>If one formally identifies surface gravity with temperature and area with entropy, the analogy with thermodynamics becomes apparent. Initially, this was widely regarded as just that, an analogy. In classical general relativity, black holes are perfect absorbers and do not radiate. It was therefore unclear in what sense they could have a temperature.</p>

<p>Jacob Bekenstein argued that black holes should carry entropy proportional to their horizon area. Otherwise, when matter with entropy falls into a black hole, there would be no way to verify that the second law of thermodynamics continues to hold. His proposal was controversial at the time.</p>

<p>Stephen Hawking set out to show that Bekenstein was wrong. Instead, in his semiclassical calculation of particle creation in a black hole background<sup id="link3"><a href="#note3">3</a></sup>, he discovered that black holes radiate as black bodies with temperature</p>

\[T_H = \frac{\kappa}{2\pi}.\]

<p>The analogy became physical. Combining this result with the first law fixes the entropy of a black hole to be</p>

\[S_{BH} = \frac{A}{4}\]

<p>in Planck units.</p>

<p>This so called <strong>Bekenstein–Hawking formula</strong> shows that black hole entropy is proportional to horizon area rather than volume. This fact will play a central role in what follows.</p>

<p>It suggests that gravity may impose a fundamental limitation on the number of independent degrees of freedom that can be associated with a spatial region. We now turn to entropy bounds and the holographic principle, where this idea is developed in a more general and covariant form.</p>

<h2 id="entropy-bounds-and-holography">Entropy bounds and holography</h2>

<p>The proportionality \(S_{BH} \propto A\) raises a natural question. Is the area scaling a special feature of black holes, or does it reflect a more general principle about gravity?</p>

<p>In ordinary local quantum field theory without gravity, entropy typically scales with volume. One expects the number of degrees of freedom in a region to be proportional to its volume. The Bekenstein–Hawking formula suggests something radically different, the maximal entropy contained in a region may scale with the area of its boundary.</p>

<p>This idea is sharpened by entropy bounds.</p>

<p>Among them, the most robust and covariant formulation is the <strong>Bousso’s covariant entropy bound</strong><sup id="link4"><a href="#note4">4</a></sup>:</p>

\[S[L(B)] \le \frac{A(B)}{4}.\]

<p>Here \(B\) is a \(d-2\) dimensional spacelike surface and \(L(B)\) is a <strong>light-sheet</strong> constructed from it.</p>

<p>A light-sheet is defined as a null hypersurface generated by light rays orthogonal to \(B\) whose expansion is initially non-positive. The null geodesics are followed until they reach a caustic or their expansion becomes positive. Intuitively, one selects those light rays that are locally contracting.</p>

<p><img src="/data/img/emergent_gravity/light-sheets.png" alt="Illustration of null congruences and light-sheets" style="width: 70%; max-width: 500px; display: block; margin: 0 auto;" /></p>

<p>To illustrate this notion, consider for example \(B\) to be a circle in 3D Minkowski space as in the image above. There are four null congruences orthogonal to it: future-outgoing \(\mathcal{C}_1\), past-ingoing \(\mathcal{C}_2\), future-ingoing \(\mathcal{C}_3\) and past-outgoing \(\mathcal{C}_4\). In the figure, the congruences that qualify as light-sheets are the contracting cones \(\mathcal{C}_2\) and \(\mathcal{C}_3\).</p>

<p>The focusing of null geodesics by matter, encoded in the Raychaudhuri equation, ensures that entropy flux through such contracting congruences is constrained by the initial area.</p>

<p>The von Neumann entropy of a quantum system is bounded above by the logarithm of the dimension of its Hilbert space. The covariant bound therefore suggests a limitation on the dimension of the Hilbert space describing a region.</p>

<p>We may formulate this idea precisely as follows.</p>

<blockquote>
  <p><strong>Conjecture (Holographic principle):</strong><br />
In some possibly emergent spacetime, let \(B\) be a \(d-2\) dimensional spacelike surface and let \(L(B)\) denote one of its light-sheets. If \(\mathcal{H}[L(B)]\) is the Hilbert space describing the degrees of freedom on \(L(B)\) in the final theory, then</p>

\[\dim \mathcal{H}[L(B)] \le e^{A(B)/4}.\]
</blockquote>

<p>If such a bound holds universally, the fundamental number of degrees of freedom associated with a region is controlled by boundary area rather than bulk volume. This statement is known as the holographic principle.</p>

<p>Its most concrete realization to date is the AdS/CFT correspondence.</p>

<h2 id="adscft-correspondence">AdS/CFT correspondence</h2>

<p>A particularly explicit realization of the holographic principle is provided by the AdS/CFT correspondence, originally conjectured by Maldacena<sup id="link5"><a href="#note5">5</a></sup>. It is widely regarded as one of the most concrete and far-reaching insights in the search for a consistent theory of quantum gravity.</p>

<p>In its most studied example, the duality states that a theory of quantum gravity, namely type IIB superstring theory in an asymptotically \(\mathrm{AdS}_5 \times S^5\) spacetime with \(N\) units of five-form flux on \(S^5\), is exactly equivalent to a quantum field theory without gravity — the 3+1 dimensional \(\mathcal{N}=4\) supersymmetric Yang–Mills theory with gauge group \(U(N)\).</p>

<p>The bulk theory is ten-dimensional and contains dynamical gravity. The boundary theory is four-dimensional and contains no graviton at all. Nevertheless, the two descriptions are conjectured to be fully equivalent. Every physical observable in the gravitational theory has a precise counterpart in the field theory. In this sense, the correspondence realizes the idea that spacetime and gravity may emerge from more fundamental, non-gravitational degrees of freedom.</p>

<p>The duality is encoded in the GKPW prescription (Gubser–Klebanov–Polyakov–Witten), which states that the generating functional of the conformal field theory equals the partition function of the bulk theory with specified boundary conditions:</p>

\[Z_{\mathrm{CFT}}[J]
=
Z_{\mathrm{string}}\big[\phi \,\big|\, \phi|_{\partial \mathrm{AdS}} = J\big].\]

<p>Here \(J(x)\) is a source coupled to some local operator \(\mathcal{O}(x)\) in the boundary theory, and \(\phi\) is the corresponding bulk field whose boundary value acts as that source. In this way, bulk fields and boundary operators are put into one-to-one correspondence.</p>

<p>In general this relation involves the full string theory path integral. However, in the limit of large \(N\) and large ‘t Hooft coupling, the bulk theory simplifies and is well approximated by classical supergravity. In appropriate low-energy truncations, supergravity reduces to Einstein’s gravity (with a negative cosmological constant and additional matter fields), so the dynamics is governed by classical general relativity in AdS spacetime. In this regime the partition function is dominated by its classical saddle:</p>

\[Z_{\mathrm{string}}[J]
=
e^{i N^2 S_{\mathrm{AdS}}[\phi_{\mathrm{cl}}]}.\]

<p>Here \(S_{\mathrm{AdS}}\) is the classical supergravity action evaluated on-shell in the AdS background, and \(\phi_{\mathrm{cl}}\) denotes the classical solution of the bulk equations of motion subject to the boundary condition</p>

\[\phi|_{\partial \mathrm{AdS}} = J.\]

<p>The factor of \(N^2\) reflects the scaling of the number of degrees of freedom in the gauge theory. In the large-\(N\) limit the bulk path integral reduces to a stationary phase approximation, so that quantum gravity in the bulk becomes classical gravity, while the boundary theory remains fully quantum.</p>

<p>This equivalence provides a concrete realization of holography: a gravitational theory in five bulk dimensions is completely encoded in a non-gravitational theory living on its four-dimensional boundary. The scaling of degrees of freedom as \(N^2\) matches the area scaling suggested by entropy bounds, giving a precise framework in which the holographic principle is realized.</p>

<p>Black holes in AdS correspond to thermal states in the boundary CFT. In the classical limit, black hole thermodynamics maps directly to ordinary thermodynamics of the gauge theory. In this way, horizon entropy and other gravitational phenomena acquire a microscopic interpretation in terms of quantum field theory degrees of freedom.</p>

<p>Below is an artistic illustration of the correspondence (drawing by my sister):</p>

<p><img src="/data/img/emergent_gravity/anicky_kresba.png" alt="Artistic illustration of AdS/CFT correspondence" style="width: 70%; max-width: 500px; display: block; margin: 0 auto;" /></p>

<p>The cylindrical geometry represents anti–de Sitter spacetime. The circular section at the top can be viewed as it’s spatial slice depicted here as a hyperbolic Poincaré disc. The vertical direction suggests time evolution, while the radial direction into the bulk encodes the energy scale of the boundary theory. The structure inside the cylinder is meant to evoke bulk string dynamics whose effects are fully captured at the boundary, reflecting the central idea that quantum gravity in the bulk is encoded in a lower-dimensional quantum field theory.</p>

<h2 id="jacobsons-emergent-gravity">Jacobson’s emergent gravity</h2>

<p>One of the most conceptually precise emergent-gravity proposals is due to Ted Jacobson<sup id="link6"><a href="#note6">6</a></sup>.</p>

<p>Instead of assuming Einstein’s equations, Jacobson assumes:</p>

<ul>
  <li>proportionality between entropy and horizon area,</li>
  <li>
    <p>validity of the Clausius relation</p>

\[\delta Q = T \, \delta S,\]
  </li>
  <li>Unruh temperature for local Rindler horizons.</li>
</ul>

<p>The key idea is local. Around any spacetime point one introduces a local Rindler horizon and associates to it an entropy proportional to its area and a temperature given by the Unruh temperature of an accelerated observer. Requiring that the Clausius relation holds for all such local horizons leads directly to the Einstein equations.</p>

<p>In this approach, the Einstein equations resemble equations of state, similar to how the Navier–Stokes equations emerge from statistical mechanics. Gravity is not introduced as a fundamental interaction but appears as a thermodynamic consistency condition imposed on local causal horizons.</p>

<p>There is also a natural connection to entropy bounds discussed earlier. The assumption \(\delta S \propto \delta A\) can be viewed as a local saturation of an area-entropy relation. In particular, Jacobson’s reasoning is closely aligned with the covariant spirit of Bousso’s bound, since the entropy is associated with local light-like horizons rather than global, static screens. In this sense, the framework is conceptually closer to the covariant entropy bound than to the original Bekenstein bound, which applies to isolated systems in more restricted settings.</p>

<h2 id="verlindes-entropic-gravity">Verlinde’s entropic gravity</h2>

<p>Erik Verlinde, influenced by Jacobson, proposed understanding gravity as an entropic force<sup id="link7"><a href="#note7">7</a></sup>. The basic idea comes from statistical mechanics. Consider, for example, a polymer immersed in a thermal bath. When stretched, the number of available microscopic configurations decreases, lowering the entropy. The system responds with an effective macroscopic force that tends to restore the polymer to a configuration of higher entropy. In such situations the force is not fundamental but emergent, and is given by</p>

\[F = T \nabla S.\]

<p>Here the entropy plays a role analogous to a potential, while the temperature sets the proportionality scale.</p>

<p>Verlinde applies this intuition to gravity. He considers holographic screens, surfaces that encode information about the bulk degrees of freedom. When a particle moves relative to such a screen, the entropy associated with the microscopic degrees of freedom changes. This change in entropy induces an effective force on the particle.</p>

<p>Using Bekenstein’s entropy bound together with holographic arguments and identifying the temperature with the Unruh temperature experienced by an accelerated observer, one can recover Newton’s law of gravitation in this framework.</p>

<p>While this approach provides appealing intuition and connects naturally to holography and AdS/CFT, it is structurally more restrictive. It relies directly on Bekenstein’s bound, which is less covariant than Bousso’s entropy bound and applies most cleanly in quasi-static situations. Extending the argument beyond the Newtonian regime requires working in special settings of static spacetimes and has not been fully carried out in a general relativistic formulation.</p>

<p>In the project, I compare this framework with Jacobson’s and analyse conceptual limitations that arise when attempting a modern relativistic formulation.</p>

<h2 id="what-this-project-aimed-to-do">What this project aimed to do</h2>

<p>This research project is primarily conceptual and structural rather than model-building. The goal was to:</p>

<ul>
  <li>review black hole thermodynamics rigorously,</li>
  <li>analyse entropy bounds and their geometric content,</li>
  <li>explain the holographic principle in modern language,</li>
  <li>clarify the relation between Jacobson’s and Verlinde’s approaches,</li>
  <li>discuss how generalized entropy functionals may modify cosmological equations.</li>
</ul>

<p>The aim of this work was not to claim that gravity is definitively emergent. Rather, the recurring appearance of thermodynamic, holographic and entanglement-based structures across different approaches suggests that they play a central role in any attempt to understand gravity at a more microscopic level. Whether these structures are fundamental or effective remains open, but they consistently point toward a deep relation between spacetime dynamics and quantum degrees of freedom.</p>

<hr />

<p><span id="note1"><a href="#link1">1</a></span>:
J. Masák,
<em>Emergent gravity</em>,
Research Project,
Faculty of Nuclear Sciences and Physical Engineering, Czech Technical University in Prague (2025),
<a href="/data/vyzkumny_ukol.pdf" target="_blank">PDF</a>.</p>

<p><span id="note2"><a href="#link2">2</a></span>:
J. M. Bardeen, B. Carter, S. W. Hawking,
<em>The four laws of black hole mechanics</em>,
<em>Commun. Math. Phys.</em> <strong>31</strong> (1973) 161–170,
<a href="https://doi.org/10.1007/BF01645742" target="_blank">doi:10.1007/BF01645742</a>.</p>

<p><span id="note3"><a href="#link3">3</a></span>:
S. W. Hawking,
<em>Particle creation by black holes</em>,
<em>Commun. Math. Phys.</em> <strong>43</strong> (1975) 199–220,
<a href="https://doi.org/10.1007/BF02345020" target="_blank">doi:10.1007/BF02345020</a>.</p>

<p><span id="note4"><a href="#link4">4</a></span>:
R. Bousso,
<em>The holographic principle</em>,
<em>Rev. Mod. Phys.</em> <strong>74</strong> (2002) 825–874,
<a href="https://doi.org/10.1103/RevModPhys.74.825" target="_blank">doi:10.1103/RevModPhys.74.825</a>.</p>

<p><span id="note5"><a href="#link5">5</a></span>:
J. M. Maldacena,
<em>The Large N Limit of Superconformal Field Theories and Supergravity</em>,
<em>Int. J. Theor. Phys.</em> <strong>38</strong> (1999) 1113–1133,
<a href="https://doi.org/10.1023/A:1026654312961" target="_blank">doi:10.1023/A:1026654312961</a>.</p>

<p><span id="note6"><a href="#link6">6</a></span>:
T. Jacobson,
<em>Thermodynamics of Spacetime: The Einstein Equation of State</em>,
<em>Phys. Rev. Lett.</em> <strong>75</strong> (1995) 1260–1263,
<a href="https://doi.org/10.1103/PhysRevLett.75.1260" target="_blank">doi:10.1103/PhysRevLett.75.1260</a>.</p>

<p><span id="note7"><a href="#link7">7</a></span>:
E. Verlinde,
<em>On the origin of gravity and the laws of Newton</em>,
<em>JHEP</em> <strong>2011</strong> (2011) 029,
<a href="https://doi.org/10.1007/JHEP04(2011)029" target="_blank">doi:10.1007/JHEP04(2011)029</a>.</p>
>
				</content>
				<pubDate>Fri, 15 Aug 2025 00:00:00 +0000</pubDate>
				<link>https://fields007.github.io/articles/emergent_gravity</link>
				<guid isPermaLink="true">https://fields007.github.io/articles/emergent_gravity</guid>
			</item>
		
			<item>
				<title>Topological insulators</title>
				
					<category>condensed matter></category>
				
					<category>topology></category>
				
					<category>graphene></category>
				
					<category>phase transitions></category>
				
					<category>academic work></category>
				
				<description>This article is based on my seminar report1 on topological insulators, prepared at Leiden University. The work gives a pedagogical overview of the topological description of band insulators, with a primary focus on the Kane–Mele model of graphene and the emergence of the ℤ₂ topological phase. Along the way, I introduce the necessary mathematical tools like vector bundles and their topological invariants and illustrate them on the Haldane and Kane–Mele models.
</description>
				<image>
					<url>http://www.w3schools.com/images/logo.gif</url>
					<title>Band structure of the Kane-Mele model</title>
					<link>/data/img/thumbnails/kane-mele_bands.png</link>
				</image>
				<content type="html">
					
						<b>Warning:</b> this post contains math blocks rendered with the
						<a href="https://www.mathjax.org/">MathJax</a> JavaScript library. If you're using an RSS reader or
						otherwise have Javascript disabled, none of them will display, so you should instead read this article at
						its <a href="https://fields007.github.io/articles/topological_insulators">source</a>.
					
					<p>This article is based on my seminar report<sup id="link1"><a href="#note1">1</a></sup> on topological insulators, prepared at Leiden University. The work gives a pedagogical overview of the topological description of band insulators, with a primary focus on the Kane–Mele model of graphene and the emergence of the ℤ₂ topological phase. Along the way, I introduce the necessary mathematical tools like vector bundles and their topological invariants and illustrate them on the Haldane and Kane–Mele models.</p>

<h2 id="introduction">Introduction</h2>

<p>Topological insulators are phases of matter which cannot be distinguished from ordinary insulators by the usual Landau picture based on spontaneous symmetry breaking and local order parameters. Instead, what matters is the global structure of the electronic bands. Roughly speaking, the valence bands of an insulator can carry non-trivial topology, and this topology cannot be changed continuously unless the energy gap closes.</p>

<p>One of the main reasons this topic became so important is that non-trivial bulk topology is accompanied by robust conducting states at the boundary. These edge or surface states are not just a mathematical curiosity. They are probably the most interesting feature of topological phases from the point of view of applications, because they can be protected against certain kinds of disorder and backscattering. The subject is therefore interesting not only conceptually, but also technologically. In the broader family of topological phases, related ideas also appear in the search for Majorana quasiparticles and in proposals for topological quantum computing, which promises error protection already at the hardware level. Microsoft is one of the companies actively pursuing this direction, as illustrated for example by the recent Nature paper<sup id="link6"><a href="#note6">6</a></sup>.￼ At the same time, I think it is important to keep some scepticism here. There is a lot of excitement around these applications, but it is still not clear how close we are to fully practical devices built on these principles.</p>

<p>The modern study of topological insulators was shaped in large part by the two seminal papers of Kane and Mele<sup id="link2"><a href="#note2">2</a></sup><sup id="link3"><a href="#note3">3</a></sup>, which showed that graphene with spin–orbit coupling provides a natural setting for a new kind of time-reversal invariant topological phase. Kane also later wrote a very nice pedagogical introduction to the topic, which I found especially useful for building intuition.<sup id="link4"><a href="#note4">4</a></sup> Although the original Kane–Mele model is not realized in actual graphene because the spin–orbit coupling there is too weak, closely related physics was later realized experimentally in HgTe quantum wells.<sup id="link5"><a href="#note5">5</a></sup></p>

<h2 id="from-bands-to-topology">From bands to topology</h2>

<p>To understand where topology enters band theory, it is useful to think of the eigenstates of the Bloch Hamiltonian as varying over momentum space. In a crystal, Bloch’s theorem tells us that single-particle states can be labelled by crystal momentum \(\mathbf{k}\), which lives in the Brillouin zone. In two dimensions, the Brillouin zone is topologically a torus \(\mathbb{T}^2\).</p>

<p>At each momentum \(\mathbf{k}\), the Bloch Hamiltonian has a finite-dimensional Hilbert space of eigenstates. If we focus on the states below the Fermi energy, that is the valence bands, these states form what is called the <strong>valence Bloch bundle</strong> over the Brillouin zone. This is the object whose topology we study.</p>

<p>One can think of a vector bundle very roughly as assigning a vector space to each point of some base space. Here the base space is the Brillouin zone and the fibres are the spaces of valence states. A trivial bundle is just a simple product, where one can choose the states continuously everywhere in momentum space without running into any obstruction. A non-trivial bundle is twisted in a global way.</p>

<p>An intuitive picture is the difference between a cylinder and a Möbius strip. Locally they both look simple, but globally the Möbius strip contains a twist that cannot be removed continuously.</p>

<p><img src="/data/img/topological_insulators/mobius_bundle.jpg" alt="Illustration of a trivial and non-trivial bundle over the circle." style="width: 70%; max-width: 500px; display: block; margin: 0 auto;" /></p>

<p>This picture should not be taken too literally, but it captures the essential idea. In band theory, non-trivial topology means that the valence states cannot be globally chosen in a completely smooth and symmetry-compatible way over the whole Brillouin zone. As long as the gap between valence and conduction bands stays open, this topology cannot change under adiabatic deformations of the Hamiltonian.</p>

<h2 id="graphene">Graphene</h2>

<p>The basic arena for both the Haldane and Kane–Mele models is graphene. Graphene has a honeycomb lattice, which is not itself a Bravais lattice but can be viewed as two interpenetrating triangular sublattices, usually denoted \(A\) and \(B\).</p>

<p><img src="/data/img/topological_insulators/lattice.png" alt="Graphene lattice with two sublattices." style="width: 70%; max-width: 500px; display: block; margin: 0 auto;" /></p>

<p>Because of this two-sublattice structure, the low-energy electronic states naturally come with a two-component structure. Near the special points \(\pm\mathbf{K}\) in the Brillouin zone one may write the momentum as \(\mathbf{k} = \pm \mathbf{K} + \mathbf{q}\), where \(\mathbf{q}\) measures the deviation from the Dirac points. In this regime the effective Hamiltonian takes the form of a two-dimensional Dirac Hamiltonian</p>

\[H_{\mathrm{eff}}(\mathbf q) = v_F \,\mathbf q \cdot \boldsymbol{\sigma} + \Delta \,\sigma_z ,\]

<table>
  <tbody>
    <tr>
      <td>where \(\mathbf q \cdot \boldsymbol{\sigma} =\mp q_x \sigma_x + q_y \sigma_y\), \(v_F\) is the Fermi velocity and the Pauli matrices act on the sublattice degree of freedom. The parameter \(\Delta\) controls the gap in the spectrum (the full band gap is $$2</td>
      <td>\Delta</td>
      <td>$$).</td>
    </tr>
  </tbody>
</table>

<p>In this low-energy description the topological properties are controlled precisely by how this gap parameter behaves at the two valleys \(\pm\mathbf{K}\). This is one of the reasons the Dirac picture is so useful: it makes the origin of the topology very transparent. What matters is not every microscopic detail of the lattice model, but the structure of the gaps in the effective theory near the points where the bands would otherwise touch.</p>

<h2 id="chern-insulators-and-the-haldane-model">Chern insulators and the Haldane model</h2>

<p>Before turning to the Kane–Mele model, it is useful to look at the simpler case of a Chern insulator. Here the relevant topological invariant is the <strong>Chern number</strong>.</p>

<p>To see where it comes from, recall that the valence states of an insulator form the valence Bloch bundle over the Brillouin zone \(\mathbb{T}^2\). A natural gauge freedom appears because the overall phase of a single-particle wave function is not observable. This means that the bundle carries a \(U(1)\) gauge symmetry, very much like in ordinary gauge theory.</p>

<p>Once such a gauge freedom is present, it is natural to introduce a connection on the bundle. In condensed matter physics one usually works with the <strong>Berry connection</strong>, which is the connection associated with the adiabatic variation of the Bloch Hamiltonian with momentum \(k\). But since we are computing a topological invariant one may choose completely arbitrary connection.</p>

<p>Every connection \(A\) has an associated curvature two-form</p>

\[F = dA .\]

<p>The topology of the bundle can then be characterized using Chern–Weil theory, which associates topological invariants to curvature forms. For a single valence band the resulting invariant is the first Chern number</p>

\[c_1 = \frac{1}{2\pi}\int_{T^2} F ,\]

<p>and the remarkable fact is that this quantity is always an integer, which ultimately follows from the underlying algebraic topology of the bundle.</p>

<p>Physically, the Chern number is important because of the so called TKNN relation, which connects it directly to Hall conductivity:</p>

\[\sigma_{xy} = \frac{e^2}{h}\, c_1 .\]

<p>So in this case topology is not just abstract mathematics. It becomes directly observable as a quantized transport coefficient.</p>

<p>A famous theoretical model realizing this physics on the graphene lattice is <strong>Haldane’s model</strong>. It describes a system with no net magnetic flux through the unit cell, but with a pattern of complex next-nearest-neighbour hopping terms that still breaks time-reversal symmetry. In this way one obtains a quantized Hall effect without an overall external magnetic field, the so-called anomalous quantum Hall effect.</p>

<p>In the low-energy description discussed in the previous section, the Hamiltonian still has the Dirac form (see the equation above). The effect of the Haldane terms is simply to generate a valley-dependent gap parameter \(\Delta = \lambda_v \pm 3\sqrt{3}\,t_2 \sin\phi\), where the sign again depends on around which Dirac point one expands.</p>

<p>As also already mentioned the topology is determined by the behaviour of \(\Delta\) at the Dirac points, specifically the Chern number can be expressed as</p>

\[c_1 = \frac{1}{2}\left[\operatorname{sgn}(\Delta(\mathbf{K}))-\operatorname{sgn}(\Delta(-\mathbf{K}))\right].\]

<p>One sees that, if the gaps at the two valleys have opposite signs, the system is in a topological phase with \(c_1=\pm1\). If the signs are the same, the contributions cancel and the phase is topologically trivial.</p>

<p>This model is conceptually very important even though it is not naturally realized in graphene itself. It provides the cleanest prototype of a Chern insulator and, more importantly for us, it serves as the building block for understanding the Kane–Mele model<sup id="link2"><a href="#note2">2</a></sup><sup id="link3"><a href="#note3">3</a></sup>.</p>

<h2 id="kanemele-model-and-the-quantum-spin-hall-effect">Kane–Mele model and the quantum spin Hall effect</h2>

<p>The Kane–Mele model extends the graphene Dirac picture by including the electron spin. The Hamiltonian can be written as two blocks corresponding to the two spin sectors plus a spin mixing term</p>

\[H_{KM}(\mathbf{k}) =
\begin{pmatrix}
H_H^{\uparrow}(\mathbf{k}) &amp; 0 \\
0 &amp; H_H^{\downarrow}(\mathbf{k})
\end{pmatrix} + H_R(\mathbf{k}) \,,\]

<p>where \(H_H^{\uparrow}\) and \(H_H^{\downarrow}\) are Haldane Hamiltonians with opposite effective fluxes. In particular they correspond to the Haldane model with parameters \(t_2 = \lambda_{so}\) and \(\phi = \pm \frac{\pi}{2}\).</p>

<p>For sufficiently strong intrinsic spin–orbit coupling \(\lambda_{so}\) these two Haldane models therefore lie in topological phases with Chern numbers \(c_1^{\uparrow}=+1\) and \(c_1^{\downarrow}=-1\).</p>

<p>This structure originates from <strong>spin–orbit coupling</strong>, which couples the electron spin to its motion in the lattice potential. In graphene the intrinsic spin–orbit interaction effectively generates opposite Haldane-type terms for the two spin sectors.</p>

<p>In the low-energy theory near the Dirac points the intrinsic spin–orbit interaction takes the simple form</p>

\[H_{so}^{\mathrm{eff}} = \pm \Delta_{so}\, \sigma^z \otimes \sigma^z \, , \qquad \Delta_{so} := 3\sqrt{3}\,\lambda_{so}.\]

<p>The Kane–Mele model can also contain a Rashba spin–orbit term, which mixes the spin sectors. In the effective theory it reads</p>

\[H_R^{\mathrm{eff}}(\mathbf{k}) = \Delta_R(\pm \sigma^y \otimes \sigma^x - \sigma^x \otimes \sigma^y)\,, \qquad \Delta_R := \frac{3}{2}\lambda_R .\]

<p>Unlike the intrinsic spin–orbit term, this interaction does not preserve the spin \(z\)-component and therefore prevents us from viewing the model simply as two independent Haldane systems. In the regime where the Rashba coupling is small, however, the two spin sectors remain approximately decoupled.</p>

<p>In that simplified regime the model effectively splits into two copies of Haldane’s model: one for spin up and one for spin down. These two copies are related by time-reversal symmetry, so their Chern numbers are opposite \(c_1^\uparrow = -c_1^\downarrow\). As a result the total charge Hall conductivity vanishes, which is exactly what one expects in a time-reversal invariant system. However, one can still define a spin Hall conductivity. If we denote it by</p>

\[\sigma_{xy}^s := \frac{\hbar}{e}\, \frac{\sigma_{xy}^{\uparrow}-\sigma_{xy}^{\downarrow}}{2},\]

<p>then using the TKNN relation one finds</p>

\[\sigma_{xy}^s = \frac{e}{2\pi}\, n_s ,\]

<p>where \(n_s = \frac{c_1^{\uparrow}-c_1^{\downarrow}}{2}\).</p>

<p>Thus the spin Hall response is quantized. This is the <strong>quantum spin Hall (QSH) phase</strong> discussed in the original Kane–Mele paper.<sup id="link2"><a href="#note2">2</a></sup></p>

<p>From this point of view the \(\mathbb{Z}_2\) invariant admits a very intuitive interpretation. One may define \(z_2 \equiv n_s \pmod{2}\).</p>

<p>If the two spin sectors carry opposite Chern numbers, then \(z_2=1\) and the phase is topologically non-trivial. If they are both trivial, then \(z_2=0\). In this regime one may therefore think of the quantum spin Hall phase simply as two time-reversed Haldane models glued together.</p>

<p>This is not yet the full story. In the more general Kane–Mele model the Rashba interaction mixes the spin sectors and the separate Chern numbers are no longer well defined. Nevertheless the non-trivial \(\mathbb{Z}_2\) phase survives beyond this simplified regime. For the purposes of this post I kept the discussion at the level of the quantum spin Hall regime, but more about the treatment of the general case can be found in the work itself.<sup id="link1"><a href="#note1">1</a></sup></p>

<h2 id="edge-states-and-why-they-matter">Edge states and why they matter</h2>

<p>The most striking consequence of non-trivial bulk topology is the existence of boundary modes. In a Chern insulator, these are <strong>chiral edge states</strong>: they move only in one direction along the boundary. In the quantum spin Hall regime of the Kane–Mele model, they become <strong>helical edge states</strong>: opposite directions of motion are tied to opposite spin sectors.</p>

<p>This edge physics is in many ways the most important part of the story. At the interface between a topological and a trivial phase, the bulk gap must close, and this produces gapless states localized near the boundary. In the topological regime these edge states cross the bulk gap and connect valence and conduction bands. In the trivial regime they do not.</p>

<p><img src="/data/img/topological_insulators/band_structure_plot.jpg" alt="Energy spectrum for a graphene strip." style="width: 70%; max-width: 500px; display: block; margin: 0 auto;" /></p>

<p>In the helical case, a right-moving edge mode and a left-moving edge mode form a time-reversal pair. Because of this, elastic backscattering from non-magnetic disorder is strongly suppressed. To scatter backwards, the electron would have to flip into its time-reversed partner, and time-reversal symmetry prevents generic perturbations of this kind from opening a gap. This robustness is the main reason edge states are so interesting for transport and for applications.</p>

<p>Of course, one should not oversell this. Real materials are never ideal, interactions can matter, disorder is not always harmless, and practical device design is much harder than the clean theoretical picture suggests. Still, the general mechanism is extremely robust conceptually: bulk topology enforces protected boundary states, and those boundary states can carry current in an unusually stable way.</p>

<p>This is one of the central lessons of topological condensed matter. The important physics is not only in the local form of the Hamiltonian, but in its global topological structure. And once that topology is non-trivial, the boundary has no choice but to respond.</p>

<h2 id="why-i-find-this-topic-interesting">Why I find this topic interesting</h2>

<p>What I find especially attractive about this subject is how naturally it connects abstract mathematics with concrete physics. Objects like vector bundles, Berry curvature and topological invariants are not added artificially. They arise quite directly once one starts taking the global structure of Bloch states seriously.</p>

<p>This topic is a nice example of a broader theme that appears again and again in theoretical physics: ideas that originally look abstract or formal often end up describing very real phenomena. In this case, concepts from geometry and topology became part of the language used to understand actual quantum materials, and possibly future technologies built from them.</p>

<hr />

<p><span id="note1"><a href="#link1">1</a></span>:
J. Masák,
<em>Topological insulators</em>,
Seminar Report,
Science faculty, Leiden university (2025),
<a href="/data/topological_insulators-report.pdf" target="_blank">PDF</a>.</p>

<p><span id="note2"><a href="#link2">2</a></span>:
C. L. Kane, E. J. Mele,
<em>Quantum Spin Hall Effect in Graphene</em>,
<em>Phys. Rev. Lett.</em> <strong>95</strong> (2005) 226801,
<a href="https://doi.org/10.1103/PhysRevLett.95.226801" target="_blank">doi:10.1103/PhysRevLett.95.226801</a>.</p>

<p><span id="note3"><a href="#link3">3</a></span>:
C. L. Kane, E. J. Mele,
<em>$\mathbb{Z}_2$ Topological Order and the Quantum Spin Hall Effect</em>,
<em>Phys. Rev. Lett.</em> <strong>95</strong> (2005) 146802,
<a href="https://doi.org/10.1103/PhysRevLett.95.146802" target="_blank">doi:10.1103/PhysRevLett.95.146802</a>.</p>

<p><span id="note4"><a href="#link4">4</a></span>:
C. L. Kane,
<em>Topological Band Theory and the $\mathbb{Z}_2$ Invariant</em>,
in <em>Topological Insulators</em>, Contemporary Concepts of Condensed Matter Science <strong>6</strong>,
Elsevier (2013) 3–34,
<a href="https://doi.org/10.1016/B978-0-444-63314-9.00001-9" target="_blank">doi:10.1016/B978-0-444-63314-9.00001-9</a>.</p>

<p><span id="note5"><a href="#link5">5</a></span>:
M. König, S. Wiedmann, C. Brüne, A. Roth, H. Buhmann, L. W. Molenkamp, X.-L. Qi, S.-C. Zhang,
<em>Quantum Spin Hall Insulator State in HgTe Quantum Wells</em>,
<em>Science</em> <strong>318</strong> (2007) 766–770,
<a href="https://doi.org/10.1126/science.1148047" target="_blank">doi:10.1126/science.1148047</a>.</p>

<p><span id="note6"><a href="#link6">6</a></span>:
M. Aghaee, A. Alcaraz Ramirez, Z. Alam, et al.,
<em>Interferometric single-shot parity measurement in InAs–Al hybrid devices</em>,
<em>Nature</em> <strong>638</strong> (2025),
<a href="https://doi.org/10.1038/s41586-024-08445-2" target="_blank">doi:10.1038/s41586-024-08445-2</a>.</p>
>
				</content>
				<pubDate>Wed, 25 Jun 2025 00:00:00 +0000</pubDate>
				<link>https://fields007.github.io/articles/topological_insulators</link>
				<guid isPermaLink="true">https://fields007.github.io/articles/topological_insulators</guid>
			</item>
		
			<item>
				<title>XY model simulation</title>
				
					<category>statistical physics></category>
				
					<category>phase transitions></category>
				
					<category>Monte Carlo></category>
				
					<category>computational physics></category>
				
					<category>academic work></category>
				
				<description>This article is based on a computational physics project1 carried out at Leiden University. In the project, the two-dimensional XY model was simulated on a square lattice using the Metropolis Monte Carlo algorithm with periodic boundary conditions. The goal was to explore the thermodynamic behaviour of the model and to reproduce the Kosterlitz–Thouless phase transition. Here I give a short conceptual overview of the work. The simulation code used in the project is available on GitHub.
</description>
				<image>
					<url>http://www.w3schools.com/images/logo.gif</url>
					<title>High temperature XY model plot.</title>
					<link>/data/img/thumbnails/XY_model.png</link>
				</image>
				<content type="html">
					
						<b>Warning:</b> this post contains math blocks rendered with the
						<a href="https://www.mathjax.org/">MathJax</a> JavaScript library. If you're using an RSS reader or
						otherwise have Javascript disabled, none of them will display, so you should instead read this article at
						its <a href="https://fields007.github.io/articles/XY_model">source</a>.
					
					<p>This article is based on a computational physics project<sup id="link1"><a href="#note1">1</a></sup> carried out at Leiden University. In the project, the two-dimensional XY model was simulated on a square lattice using the Metropolis Monte Carlo algorithm with periodic boundary conditions. The goal was to explore the thermodynamic behaviour of the model and to reproduce the Kosterlitz–Thouless phase transition. Here I give a short conceptual overview of the work. The simulation code used in the project is available on <a href="https://github.com/fields007/XY-model" target="_blank">GitHub</a>.</p>

<h2 id="introduction">Introduction</h2>

<p>The XY model imagines a two-dimensional lattice \(\Lambda\) of sites \(i \in \Lambda\), at each of which there is a vector \(\mathbf{s}_i \in \mathbb{S}^1\) of unit length. This can be viewed as a generalisation of the Ising model because \(\mathbb{S}^0 = \{-1,1\}\). Each configuration is then represented by a set of spins \(\{\mathbf{s}_i\}_{i \in \Lambda}\) or, equivalently, through the parametrisation \(\mathbf{s}_i = (\cos \theta_i, \sin \theta_i)\), with \(\theta_i \in [-\pi,\pi)\), by a set of angles \(\Theta = \{\theta_i\}_{i \in \Lambda}\), which is the representation I will use most often.</p>

<p>The Hamiltonian used in the simulations can be written as a sum of two parts,</p>

\[H[\Theta] = H_0[\Theta] + H_h[\Theta].\]

<p>The first term</p>

\[H_0[\Theta] = -J \sum_{\langle i,j \rangle} \mathbf{s}_i \cdot \mathbf{s}_j = -J \sum_{\langle i,j \rangle} \cos(\theta_i-\theta_j)\]

<p>describes the interaction between nearest neighbours and favours alignment of neighbouring spins. The second term</p>

\[H_h[\Theta] = -\mathbf{h}\cdot \sum_{i \in \Lambda} \mathbf{s}_i = -h \sum_{i \in \Lambda} \cos \theta_i\]

<p>represents the coupling to an external magnetic field of magnitude \(h = \lVert \mathbf{h} \rVert\) pointing in the \(x\) direction.</p>

<p>The key property of the model lies in the symmetry of the interaction term \(H_0\). Because it depends only on angle differences, it has the symmetry</p>

\[H_0[\Theta] = H_0[\Theta+\Delta\theta], \qquad \Delta\theta \in [-\pi,\pi),\]

<p>which is a global \(U(1)\) symmetry. This is probably the most important difference from the Ising model, which also has a global symmetry, but only a discrete \(\mathbb Z_2\) symmetry. Because of this continuous symmetry, we have the Mermin–Wagner theorem<sup id="link2"><a href="#note2">2</a></sup>, which states that in a system with such local interactions, continuous symmetries cannot be spontaneously broken at finite temperature when the spatial dimension is \(\leq 2\). This means that in the thermodynamic limit the system cannot develop a non-zero mean magnetisation serving as an ordinary order parameter.</p>

<p>The physical mechanism behind this is the existence of long-wavelength spin waves. Deviations from order can then be quantified by the momentum-space integral</p>

\[\int \frac{d^2k}{k^2}.\]

<p>If the lattice is \(N\times N\), this integral scales as \(\ln N\) and therefore diverges in the thermodynamic limit. Since in the simulation the lattice is finite, the divergence is only logarithmic, and as we will see, the system can still look magnetised at low temperatures. This is one of the reasons finite-size effects are important in the XY model.</p>

<p>The absence of spontaneous magnetisation also means that the system does not fit the standard Landau picture of phase transitions based on a local order parameter. Nevertheless, the two-dimensional XY model still exhibits a phase transition, first described by Kosterlitz and Thouless<sup id="link3"><a href="#note3">3</a></sup>. It is a topological phase transition, associated with the behaviour of vortices. In our work, we also studied the effect of an external magnetic field. Since the term \(H_h\) explicitly breaks the global \(U(1)\) symmetry, one expects the Kosterlitz–Thouless transition to disappear once the field is switched on. This is indeed what the simulation suggests.</p>

<h2 id="monte-carlo-integration-and-the-metropolis-algorithm">Monte Carlo integration and the Metropolis algorithm</h2>

<p>The Monte Carlo method is a powerful approach for studying the properties of a system at equilibrium when we are not interested in its dynamics and need to calculate high-dimensional integrals. In statistical physics, we typically have to evaluate integrals of the form</p>

\[\langle A \rangle = \int_M \mathcal{D}\Theta \, \frac{e^{-\beta H[\Theta]}}{Z} \, A[\Theta], \qquad Z = \int_M \mathcal{D}\Theta \, e^{-\beta H[\Theta]},\]

<p>with \(\mathcal{D}\Theta = \prod_{i\in\Lambda} d\theta_i\) and \(M = [-\pi,\pi)^{N^2}\), which can be thought of as the configuration space of the system. We therefore have a very high-dimensional integral, and additional complexity arises from the fact that the function \(A[\Theta]e^{-\beta H[\Theta]}\) is usually very sharply peaked. This requires that we do not sample over the configuration space with the uniform distribution, as the simplest Monte Carlo method does, but instead use importance sampling.</p>

<p>Importance sampling can be implemented by means of the Metropolis algorithm<sup id="link4"><a href="#note4">4</a></sup>. This algorithm is also convenient because it assumes that we know the target distribution only up to a normalisation factor. This is exactly the case with the Boltzmann distribution</p>

\[\pi[\Theta(t)] = \frac{e^{-\beta H[\Theta(t)]}}{Z},\]

<p>because we do not know the partition function \(Z\).</p>

<p>The Metropolis algorithm can be understood using the concept of a Markov chain: a set of random states \(\Theta(t)\), where the probability of \(\Theta(t+1)\) depends only on \(\Theta(t)\). A Markov chain is fully described by transition probabilities \(T(\Theta\rightarrow\Theta')\) such that</p>

\[\int_M \mathcal{D}\Theta' \; T(\Theta\rightarrow\Theta') = 1.\]

<p>In the Metropolis algorithm, one divides these probabilities as</p>

\[T(\Theta\rightarrow\Theta') = w(\Theta' \mid \Theta)\times A(\Theta' \mid \Theta).\]

<p>The first factor \(w(\Theta' \mid \Theta)\) denotes the probability to propose a new state \(\Theta'\) given a state \(\Theta\), and the second factor \(A(\Theta' \mid \Theta)\) denotes the probability to accept the proposed new state \(\Theta'\) if the system was before in the state \(\Theta\).</p>

<p>To ensure that a Markov process has a unique stationary distribution, two conditions must be met. First, there has to exist a stationary distribution. A sufficient, though not necessary, condition is the detailed balance condition</p>

\[\pi[\Theta]\,T(\Theta\rightarrow\Theta') = \pi[\Theta']\,T(\Theta'\rightarrow\Theta).\]

<p>Second, the stationary distribution should be unique. This is guaranteed by ergodicity of the Markov chain, a property commonly assumed in the Metropolis algorithm. A key component of ergodicity is irreducibility, which ensures that the Markov chain can explore the entire configuration space over time. As will be important later, in our system this assumption may break down at low temperatures because of finite-size effects.</p>

<p>If we assume that \(w\) is symmetric, that is,</p>

\[w(\Theta' \mid \Theta) = w(\Theta \mid \Theta'),\]

<p>then the detailed balance condition can be written as</p>

\[\frac{A(\Theta' \mid \Theta)}{A(\Theta \mid \Theta')} = \frac{\pi[\Theta']}{\pi[\Theta]},\]

<p>which can be satisfied with</p>

\[A(\Theta' \mid \Theta)=
\begin{cases}
1 &amp; \text{if } \pi[\Theta']&gt;\pi[\Theta],\\
\pi[\Theta']/\pi[\Theta] &amp; \text{if } \pi[\Theta']&lt;\pi[\Theta].
\end{cases}\]

<p>The Metropolis algorithm thus consists of the following steps:</p>

<ol>
  <li>Start with a random initial state \(\Theta(0)\).</li>
  <li>Generate a state \(\Theta'\) from \(\Theta(t)\).</li>
  <li>If \(\pi[\Theta']&gt;\pi[\Theta(t)]\), then \(A(\Theta'\mid\Theta)=1\) and set \(\Theta(t+1)=\Theta'\). Otherwise, accept the move with probability \(p=\pi[\Theta']/\pi[\Theta(t)]\).</li>
  <li>Continue with step 2.</li>
</ol>

<p>In our case, \(\pi[\Theta(t)]\) is the Boltzmann distribution. Thus, a trial move that lowers the energy is always accepted, while a move that increases the energy by an amount</p>

\[\Delta E = H[\Theta'] - H[\Theta]\]

<p>is accepted with probability \(e^{-\beta \Delta E}\). This means that the system tries to move towards lower total energy, but the higher the temperature, the higher is the probability to go to a state with more energy.</p>

<p>For practical implementation, we chose the trial-step probabilities to be</p>

\[w(\Theta' \mid \Theta)
=
\begin{cases}
1/N^2 &amp; \text{if } \Theta \text{ and } \Theta' \text{ differ by one spin},\\
0 &amp; \text{otherwise}.
\end{cases}\]

<p>This means that when the spins are in state \(\Theta\), we create a trial state \(\Theta'\) by picking one spin at random and changing its angle by a random increment from an interval \([-\Delta,\Delta]\subset[-\pi,\pi]\). In the simulation, this increment was chosen so that the average acceptance probability stayed close to \(0.5\), which allows the configuration space to be explored efficiently. After that, one computes the energy difference \(\Delta E\) by calculating the change in the interaction energy of the chosen spin with its four neighbours.</p>

<h2 id="measured-quantities">Measured quantities</h2>

<p>The following is an overview of the quantities measured in the simulation. We denote the total magnetisation by \(\mathbf M = \sum_{i\in\Lambda}\mathbf s_i\). The total magnetisation per spin is then \(\mathbf m = \mathbf M / N^2\). Vector norms are denoted simply by \(M = \lVert \mathbf{M} \rVert\) and \(m = \lVert \mathbf{m} \rVert\). The time averages that should converge to ensemble averages are denoted as</p>

\[\langle A \rangle = \frac{1}{T}\sum_{t=1}^{T} A(t),\qquad \langle A(t) \rangle = \frac{1}{T}\sum_{t'=1}^{T} A(t'+t).\]

<p>One of the outputs of the simulation is the average magnitude of total magnetisation per spin, \(\langle m \rangle = \langle M \rangle / N^2\). Another possibility would be to measure \(\langle \mathbf m \rangle\), which should be zero if \(h=0\), but in practice it is unstable at low temperatures. Below the critical temperature, the system acquires a magnetisation and it would take a very long time for it to average back to zero. Because of this, \(\langle \mathbf m \rangle\) is not a stable quantity and it is more useful to measure \(\langle m \rangle\). Analogously, the energy per spin is measured as \(e = E/N^2\), where \(E\) is the total energy of the system.</p>

<p>The standard deviation of magnetisation, and analogously of energy, is calculated as</p>

\[\sigma = \sqrt{\frac{2\tau}{t_{\max}}\left(\langle m^2\rangle - \langle m\rangle^2\right)}\,,\]

<p>where \(\tau\) is the autocorrelation time, that is the time it takes for the system to forget its previous state. The autocorrelation time is calculated from the autocorrelation function</p>

\[\chi(t) = \langle \mathbf m \cdot \mathbf m(t) \rangle - \langle \mathbf m \rangle \cdot \langle \mathbf m(t) \rangle.\]

<p>The autocorrelation function should decay as \(\chi(t) \sim e^{-t/\tau}\). In the simulation, \(\tau\) is approximated as</p>

\[\sum_{t=0}^{t_{\max}} \frac{\chi(t)}{\chi(0)} \simeq \sum_{t=0}^{t_{\max}} e^{-t/\tau} \sim \int_0^\infty e^{-t/\tau}\,\mathrm{d}t = \tau \,,\]

<p>where the last approximation holds for large \(t_{\max}\). So in practice the autocorrelation time is estimated using this sum, with \(t_{\max}\) chosen so that the summation stops once \(\chi(t)&lt;0,\) which signals the onset of poor statistics.</p>

<p>Additional quantities measured in the simulation are the magnetic susceptibility per spin,</p>

\[\chi_M = \frac{1}{N^2 k_B T} \left(\langle \mathbf M \cdot \mathbf M \rangle - \langle \mathbf M \rangle \cdot \langle \mathbf M \rangle \right),\]

<p>and the specific heat per spin,</p>

\[C = \frac{1}{N^2 k_B T^2} \left(\langle E^2\rangle - \langle E\rangle^2\right).\]

<p>An interesting feature that appears in the XY model at low temperatures is the presence of vortex–antivortex pairs. To quantify this phenomenon, the simulation measures the average vortex density and plots the distribution of vortex-pair member distances. The presence of bound vortex pairs in the low-temperature phase, and their unbinding in the high-temperature phase, is one of the central aspects of the Kosterlitz–Thouless picture<sup id="link5"><a href="#note5">5</a></sup>.</p>

<h2 id="vortices-and-the-kosterlitzthouless-transition">Vortices and the Kosterlitz–Thouless transition</h2>

<p>The Kosterlitz–Thouless transition is not best understood through an ordinary local order parameter. Its central objects are vortices, that is topological defects of the spin field. At low temperatures vortices and antivortices tend to appear as tightly bound pairs. At higher temperatures these pairs unbind, and this unbinding is precisely the mechanism underlying the topological phase transition.</p>

<div style="display: flex; justify-content: center; gap: 20px; flex-wrap: wrap;">

  <img src="/data/img/XY_model/last_state/low-T.png" alt="Final spin configuration at low temperature showing the effective symmetry breaking." style="width: 45%; max-width: 400px;" />

  <img src="/data/img/XY_model/last_state/high-T.png" alt="Final spin configuration at high temperature where the spins are disordered." style="width: 45%; max-width: 400px;" />

</div>

<p>At low temperature one still sees large correlated regions, even though in the thermodynamic limit true long-range order is absent. Because the lattice in the simulation is finite, the system can effectively select one global orientation and fluctuate only weakly around it. This behaviour is related to the finite-size effects discussed earlier: although long-wavelength spin waves prevent true spontaneous symmetry breaking in the thermodynamic limit, on a finite lattice the system may still appear magnetised for long periods of time.</p>

<p>The same finite-size effects also affect the Monte Carlo sampling. As mentioned earlier in the discussion of the Metropolis algorithm, the assumption that the Markov chain efficiently explores the whole configuration space can become problematic at low temperatures. In practice the simulation may remain trapped near one orientation of the spins, which is why the low-temperature configuration above shows relatively small angular fluctuations.</p>

<p>At high temperature the configuration is visibly more disordered.</p>

<figure style="text-align: center;">
  <video controls="" preload="metadata" width="450">
    <source src="/data/img/XY_model/video.mp4" type="video/mp4" />
  </video>
  <figcaption style="display:none;">
    Vortex–antivortex dynamics at low temperature.
  </figcaption>
</figure>

<p>The video illustrates the same physics more directly: vortices and antivortices appear as localized defects in the spin field, and at low temperatures they tend to remain pairwise bound.</p>

<h2 id="results-and-discussion">Results and discussion</h2>

<p>A particularly clear quantity is the autocorrelation time. The two plots below show the autocorrelation function at \(T=0.7\) and \(T=1.1\) for \(h=0\).</p>

<p><img src="/data/img/XY_model/autocorrelation/low-T.png" alt="Spin autocorrelation function at low temperature showing slow decay." style="width: 60%; max-width: 450px; display: block; margin: 0 auto;" /></p>

<p><img src="/data/img/XY_model/autocorrelation/high-T.png" alt="Spin autocorrelation function at high temperature with rapid decay." style="width: 60%; max-width: 450px; display: block; margin: 0 auto;" /></p>

<p>If the temperature is high, the autocorrelation function follows the exponential approximation quite accurately, at least in the initial regime where \(\chi(t)&gt;0\). At low temperatures the decay is slower and the difference between the exponential fit and the actual values becomes significant. This supports the concern already mentioned above: at low temperatures the finite lattice may fail to explore configuration space properly and can get stuck around a non-zero magnetisation.</p>

<p>The corresponding autocorrelation times are shown below.</p>

<p><img src="/data/img/XY_model/autocorrelation_time.png" alt="Autocorrelation time as a function of temperature showing critical slowing down near the Kosterlitz–Thouless transition." style="width: 60%; max-width: 450px; display: block; margin: 0 auto;" /></p>

<p>Without an external magnetic field, one can clearly see the existence of a critical point around \(T_C \approx 0.9\). The autocorrelation time grows strongly there, as expected from the Kosterlitz–Thouless transition theory. At low temperatures the autocorrelation time is not properly defined and its value differs between runs. This again arises from the finite-size effects already discussed.</p>

<p>The same figure also displays the effect of external magnetic fields. Even in the case of a weak field \(h=0.05\), the autocorrelation time no longer diverges at the critical point. Although the values are still somewhat larger at low temperatures, they remain in a reasonable range. A stronger magnetic field \(h=0.4\) decreases the autocorrelation time dramatically. This supports the main message of the simulation: the external magnetic field effectively removes the topological phase transition, although some critical behaviour remains.</p>

<p>The vortex-pair distance distributions tell a similar story.</p>

<div style="display: flex; justify-content: center; gap: 20px; flex-wrap: wrap;">

  <img src="/data/img/XY_model/vortex_historgram/low-T.png" alt="Histogram of vortex–antivortex pair separations at low temperature showing tightly bound pairs." style="width: 45%; max-width: 400px;" />

  <img src="/data/img/XY_model/vortex_historgram/high-T.png" alt="Histogram of vortex–antivortex pair separations at high temperature showing many unbound vortices." style="width: 45%; max-width: 400px;" />
</div>

<p>At low temperature there are relatively few vortices and they are mostly pairwise bound, with distances typically of one lattice spacing. At high temperature the distribution becomes much broader, signalling the unbinding of vortex pairs. This is one of the clearest geometric manifestations of the Kosterlitz–Thouless transition.</p>

<p>The simulation therefore reproduces the basic theoretical picture rather well. Without external magnetic field the model undergoes a topological phase transition associated with vortex unbinding and critical slowing down. When a magnetic field is applied, the transition disappears because the global \(U(1)\) symmetry is broken explicitly, but the system still exhibits some residual critical-looking behaviour.</p>

<h2 id="conclusion">Conclusion</h2>

<p>What I find especially attractive about this project is that it makes a rather subtle phase transition visible in a very concrete way. The two-dimensional XY model does not fit the usual symmetry-breaking paradigm, yet it still exhibits non-trivial thermodynamic behaviour governed by topological defects. The simulation captures this picture quite well: one sees critical slowing down, vortex proliferation above the critical region, and the qualitative disappearance of the transition once an external magnetic field is applied.</p>

<hr />

<p><span id="note1"><a href="#link1">1</a></span>:
J. Masák, E. Tõkke,
<em>Computational Physics Project 2: Monte Carlo simulation of the two-dimensional XY model</em>,
Course Project Report,
Science faculty, Leiden university (2025),
<a href="/data/XY_model-report.pdf" target="_blank">PDF</a>.</p>

<p><span id="note2"><a href="#link2">2</a></span>:
N. D. Mermin, H. Wagner,
<em>Absence of ferromagnetism or antiferromagnetism in one- or two-dimensional isotropic Heisenberg models</em>,
<em>Phys. Rev. Lett.</em> <strong>17</strong> (1966) 1133,
<a href="https://doi.org/10.1103/PhysRevLett.17.1133" target="_blank">doi:10.1103/PhysRevLett.17.1133</a>.</p>

<p><span id="note3"><a href="#link3">3</a></span>:
J. M. Kosterlitz, D. J. Thouless,
<em>Ordering, Metastability and Phase Transitions in Two-Dimensional Systems</em>,
<em>J. Phys. C: Solid State Phys.</em> <strong>6</strong> (1973) 1181,
<a href="https://doi.org/10.1088/0022-3719/6/7/010" target="_blank">doi:10.1088/0022-3719/6/7/010</a>.</p>

<p><span id="note4"><a href="#link4">4</a></span>:
N. Metropolis, A. W. Rosenbluth, M. N. Rosenbluth, A. H. Teller, E. Teller,
<em>Equation of State Calculations by Fast Computing Machines</em>,
<em>J. Chem. Phys.</em> <strong>21</strong> (1953) 1087,
<a href="https://doi.org/10.1063/1.1699114" target="_blank">doi:10.1063/1.1699114</a>.</p>

<p><span id="note5"><a href="#link5">5</a></span>:
J. Tobochnik, G. V. Chester,
<em>Monte Carlo Study of the Planar Spin Model</em>,
<em>Phys. Rev. B</em> <strong>20</strong> (1979) 3761,
<a href="https://doi.org/10.1103/PhysRevB.20.3761" target="_blank">doi:10.1103/PhysRevB.20.3761</a>.</p>
>
				</content>
				<pubDate>Mon, 05 May 2025 00:00:00 +0000</pubDate>
				<link>https://fields007.github.io/articles/XY_model</link>
				<guid isPermaLink="true">https://fields007.github.io/articles/XY_model</guid>
			</item>
		
			<item>
				<title>Argon gas simulation</title>
				
					<category>phase transitions></category>
				
					<category>statistical physics></category>
				
					<category>computational physics></category>
				
					<category>academic work></category>
				
				<description>This article is based on a computational physics project1 carried out at Leiden University. In the project, the motion of a relatively small number of Argon atoms was simulated using simple Newtonian dynamics with the Lennard–Jones potential, integrated using the Verlet algorithm with periodic boundary conditions. The aim was to see how this microscopic model reproduces the three phases of matter and to study their basic properties. Here I give a short conceptual exposition of the work. The simulation code used in the project is available on GitHub.
</description>
				<image>
					<url>http://www.w3schools.com/images/logo.gif</url>
					<title>Argon gas-discharge lamp.</title>
					<link>/data/img/thumbnails/argon_gas.png</link>
				</image>
				<content type="html">
					
						<b>Warning:</b> this post contains math blocks rendered with the
						<a href="https://www.mathjax.org/">MathJax</a> JavaScript library. If you're using an RSS reader or
						otherwise have Javascript disabled, none of them will display, so you should instead read this article at
						its <a href="https://fields007.github.io/articles/argon_gas">source</a>.
					
					<p>This article is based on a computational physics project<sup id="link1"><a href="#note1">1</a></sup> carried out at Leiden University. In the project, the motion of a relatively small number of Argon atoms was simulated using simple Newtonian dynamics with the Lennard–Jones potential, integrated using the Verlet algorithm with periodic boundary conditions. The aim was to see how this microscopic model reproduces the three phases of matter and to study their basic properties. Here I give a short conceptual exposition of the work. The simulation code used in the project is available on <a href="https://github.com/fields007/XY-model" target="_blank">GitHub</a>.</p>

<h2 id="molecular-dynamics-and-argon">Molecular dynamics and Argon</h2>

<p>Molecular dynamics aims to follow the motion of many interacting particles by numerically solving their equations of motion. Early simulations of this kind already demonstrated phase behaviour in simple interacting systems<sup id="link2"><a href="#note2">2</a></sup>. In this approach each atom is treated as a classical particle and its trajectory is obtained by integrating Newton’s equations. Even a relatively simple interaction model can already reproduce many properties of real materials.</p>

<p>In the simulation each particle obeys Newton’s second law</p>

\[m \frac{d^2 \mathbf{x}}{dt^2} = \mathbf{F}(\mathbf{x}) = - \nabla U(\mathbf{x}) .\]

<p>If the interaction potential is a sum of pair interactions, the force on particle \(\alpha\) can be written as</p>

\[\mathbf{F}_\alpha = - \sum_{\beta \neq \alpha} \nabla U(\mathbf{x}_\alpha - \mathbf{x}_\beta).\]

<p>Once the potential is specified, the task of the simulation is therefore to repeatedly compute the forces and update the particle positions.</p>

<h2 id="lennardjones-potential">Lennard–Jones potential</h2>

<p>For Argon atoms the interaction can be modeled by the Lennard–Jones two-particle potential</p>

\[U(r) = 4\varepsilon \left[\left(\frac{\sigma}{r}\right)^{12} - \left(\frac{\sigma}{r}\right)^6 \right].\]

<p>Here \(r\) is the distance between the two particles, while \(\varepsilon\) and \(\sigma\) set the energy and length scales of the interaction.</p>

<p>The \(r^{-12}\) term produces a strong short-range repulsion, preventing atoms from occupying the same position. The \(r^{-6}\) term represents an attractive van der Waals interaction. Together they create a preferred separation between atoms, corresponding to the minimum of the potential at \(r_{\min} = 2^{1/6}\sigma\).</p>

<p>This preferred distance later appears as the first peak in the pair-correlation function.</p>

<h2 id="numerical-integration">Numerical integration</h2>

<p>To evolve the system in time we integrate the equations of motion numerically. In this project the velocity Verlet algorithm<sup id="link4"><a href="#note4">4</a></sup> was used because it is simple and stable for long simulations.</p>

<p>The positions are updated according to</p>

\[\mathbf{x}(t+h)=\mathbf{x}(t)+h\mathbf{v}(t)+\frac{h^2}{2m}\mathbf{F}(\mathbf{x}(t))+\mathcal{O}(h^3)\,.\]

<p>The velocity update uses the forces at the beginning and end of the step</p>

\[\mathbf{v}(t+h)=\mathbf{v}(t)+\frac{h}{2m}(\mathbf{F}(\mathbf{x}(t+h))+\mathbf{F}(\mathbf{x}(t)))+\mathcal{O}(h^3)\,.\]

<p>In the continuous equations of motion the total energy is conserved. This follows from the invariance of the physical laws under time translations, as expressed by Noether’s theorem. When the equations are discretised this symmetry is generally broken. For example, a simple Euler integration scheme treats the forward and backward directions in time differently and typically leads to a gradual drift of the energy.</p>

<p>The Verlet algorithm improves this behaviour by making the update rule explicitly symmetric under the transformation \(t \to -t\). Although the energy is not exactly conserved, it typically oscillates around the correct value instead of drifting systematically. This makes the method particularly well suited for long molecular dynamics simulations.</p>

<h2 id="periodic-boundary-conditions-and-the-minimum-image-convention">Periodic boundary conditions and the minimum image convention</h2>

<p>The number of particles in a simulation is necessarily limited. To mimic a much larger system we use <strong>periodic boundary conditions</strong>.</p>

<p>Particles move inside a cubic box of side length \(L\). Whenever a particle leaves the box through one side it re-enters through the opposite side. One can think of the simulation box as being surrounded by infinitely many copies of itself. This removes artificial boundary effects and makes the system behave more like a small piece of bulk matter.</p>

<p>Under periodic boundary conditions each particle has infinitely many periodic images. When computing forces we therefore need a rule that determines which image interacts with a given particle. The <strong>minimum image convention</strong> states that particles interact only with the nearest periodic image as illustrated in the image below.</p>

<p><img src="/data/img/argon_gas/minimum_image.png" alt="Illustration of the minimum image convention." style="width: 50%; max-width: 500px; display: block; margin: 0 auto;" /></p>

<p>Formally the relative position between two particles can be written as</p>

\[r_{\alpha\beta} = \min_{k,l,m \in \{-1,0,1\}} \left| \mathbf{x}_\alpha - \mathbf{x}_\beta + (k,l,m)\,L \right|.\]

<p>Because the simulation box is orthogonal, the shortest displacement can be obtained by shifting each coordinate difference into the interval \((-L/2, L/2)\). In practice this can be implemented using the modulo operation</p>

\[\Delta x = (x_\alpha - x_\beta + L/2)\ \mathrm{mod}\ L - L/2 ,\]

<p>and similarly for \(\Delta y\) and \(\Delta z\). The displacement vector is then \(\mathbf{r}_{\alpha\beta} = (\Delta x, \Delta y, \Delta z)\).</p>

<h2 id="initial-configuration-and-equilibration">Initial configuration and equilibration</h2>

<p>To avoid extremely large forces the atoms are initially placed on a face-centered cubic lattice, which is the crystal structure into which Argon solidifies. The initial separation between atoms is chosen to correspond to the observed lattice spacing of solid Argon. In practice such structural parameters can be obtained from experiment or from electronic structure calculations. The unit cell of the face-centered cubic lattice is illustrated in the image below.</p>

<p><img src="/data/img/argon_gas/lattice.png" alt="A unit cell of the face-centered cubic lattice." style="width: 30%; max-width: 500px; display: block; margin: 0 auto;" /></p>

<p>Initial velocities are chosen from a Maxwell distribution corresponding to the desired temperature. The system is then evolved for a number of steps while rescaling the velocities so that the kinetic energy approaches the expected value</p>

\[E_{\mathrm{kin}} = (N-1)\frac{3}{2}k_B T .\]

<p>This equilibration stage allows the system to settle into a representative state before measurements are taken.</p>

<h2 id="pair-correlation-function">Pair-correlation function</h2>

<p>One of the most informative observables is the <strong>pair-correlation function</strong>, which was famously analysed in Rahman’s early molecular dynamics simulation of liquid Argon<sup id="link3"><a href="#note3">3</a></sup>. In the project some of the results reported in that work were also reproduced numerically, allowing for a direct comparison with the classical simulation of liquid Argon.</p>

<p>In the simulation the pair-correlation function is defined as</p>

\[g(r) = \frac{2V}{N(N-1)} \frac{\langle n(r)\rangle}{4\pi r^2 \Delta r}.\]

<p>Here \(n(r)\) denotes the number of particles found at a distance between \(r\) and \(r+\Delta r\) from a chosen reference particle.</p>

<p>Roughly speaking, \(g(r)\) measures how likely it is to find another atom at distance \(r\) compared to a completely uniform distribution. The structure of this function therefore reveals the spatial organization of the system.</p>

<h2 id="gas-phase">Gas phase</h2>

<p>In the gas phase the particles move freely and there is little spatial structure. The trajectories in the animation below show that the atoms travel through the simulation box with frequent collisions but without forming any persistent structure.</p>

<p><img src="/data/img/argon_gas/gas/trajectories.gif" alt="Molecular dynamics trajectories of Argon atoms in the gas phase." style="width: 70%; max-width: 500px; display: block; margin: 0 auto;" /></p>

<p>The corresponding pair-correlation function is shown in the plot below.</p>

<p><img src="/data/img/argon_gas/gas/correlation.png" alt="Pair-correlation function for Argon in the gas phase." style="width: 70%; max-width: 500px; display: block; margin: 0 auto;" /></p>

<p>It is nearly flat with \(g(r) \approx 1\) except at very small distances where the strong repulsive core of the Lennard–Jones potential suppresses close approaches between atoms.</p>

<h2 id="liquid-phase">Liquid phase</h2>

<p>In the liquid phase the atoms still move throughout the system but remain correlated with their neighbours. This can already be seen in the trajectories below, where particles continuously change their neighbours while remaining part of a relatively dense local environment.</p>

<p><img src="/data/img/argon_gas/liquid/trajectories.gif" alt="Molecular dynamics trajectories of Argon atoms in the liquid phase." style="width: 70%; max-width: 500px; display: block; margin: 0 auto;" /></p>

<p>The pair-correlation function shown below reflects this short-range order.</p>

<p><img src="/data/img/argon_gas/liquid/correlation.png" alt="Pair-correlation function for Argon in the liquid phase." style="width: 70%; max-width: 500px; display: block; margin: 0 auto;" /></p>

<p>A pronounced first peak appears near the preferred interatomic separation, together with weaker oscillations at larger distances. These peaks indicate that atoms tend to arrange themselves in shells around each other, even though the long-range order of a crystal is absent.</p>

<h2 id="solid-phase">Solid phase</h2>

<p>In the solid phase atoms remain localized near lattice sites and only oscillate around their equilibrium positions. This behaviour can be seen clearly in the trajectories below.</p>

<p><img src="/data/img/argon_gas/solid/trajectories.gif" alt="Molecular dynamics trajectories of Argon atoms in the solid phase." style="width: 70%; max-width: 500px; display: block; margin: 0 auto;" /></p>

<p>The corresponding pair-correlation function shows a series of sharp peaks.</p>

<p><img src="/data/img/argon_gas/solid/correlation.png" alt="Pair-correlation function for Argon in the solid phase." style="width: 70%; max-width: 500px; display: block; margin: 0 auto;" /></p>

<p>These peaks correspond to the well-defined distances between atoms in the underlying crystal lattice. In contrast to the liquid case, the peaks remain pronounced even at larger distances, reflecting the long-range order characteristic of the solid phase.</p>

<h2 id="conclusion">Conclusion</h2>

<p>Although the model is very simple, it already reproduces the qualitative behaviour of gases, liquids and solids. Even with a relatively small number of particles the characteristic differences between these phases become visible in the particle trajectories and in the pair-correlation function.</p>

<p>This illustrates a central idea of statistical physics: macroscopic properties of matter can often be understood as emerging from simple microscopic dynamics.</p>

<hr />

<p><span id="note1"><a href="#link1">1</a></span>:
J. Masák, E. Tõkke,
<em>Computational Physics Project 1: Molecular Dynamics Simulation of Argon Atoms</em>,
Course Project Report,
Science faculty, Leiden university (2025),
<a href="/data/argon_gas-report.pdf" target="_blank">PDF</a>.</p>

<p><span id="note2"><a href="#link2">2</a></span>:
B. J. Alder, T. E. Wainwright,
<em>Phase Transition for a Hard Sphere System</em>,
<em>J. Chem. Phys.</em> <strong>27</strong> (1957) 1208,
<a href="https://doi.org/10.1063/1.1743957" target="_blank">doi:10.1063/1.1743957</a>.</p>

<p><span id="note3"><a href="#link3">3</a></span>:
A. Rahman,
<em>Correlations in the Motion of Atoms in Liquid Argon</em>,
<em>Phys. Rev.</em> <strong>136</strong> (1964) A405,
<a href="https://doi.org/10.1103/PhysRev.136.A405" target="_blank">doi:10.1103/PhysRev.136.A405</a>.</p>

<p><span id="note4"><a href="#link4">4</a></span>:
L. Verlet,
<em>Computer “Experiments” on Classical Fluids. I. Thermodynamical Properties of Lennard-Jones Molecules</em>,
<em>Phys. Rev.</em> <strong>159</strong> (1967) 98,
<a href="https://doi.org/10.1103/PhysRev.159.98" target="_blank">doi:10.1103/PhysRev.159.98</a>.</p>
>
				</content>
				<pubDate>Thu, 20 Mar 2025 00:00:00 +0000</pubDate>
				<link>https://fields007.github.io/articles/argon_gas</link>
				<guid isPermaLink="true">https://fields007.github.io/articles/argon_gas</guid>
			</item>
		
			<item>
				<title>Generalized uncertainty relation and its use in cosmology</title>
				
					<category>thesis></category>
				
					<category>general relativity></category>
				
					<category>black holes></category>
				
					<category>academic work></category>
				
				<description>This article is based on my Bachelor’s thesis1 in mathematical physics. I outline the main ideas and motivations of the work in broader terms. In the thesis, I investigate how deformations of the standard uncertainty relations can be interpreted, at the classical level, as deformations of the Poisson algebra of observables. I then explore how this modified structure affects Hamiltonian dynamics, with applications to geodesic motion in Schwarzschild and Kerr spacetimes and to black hole thermodynamics.
</description>
				<image>
					<url>http://www.w3schools.com/images/logo.gif</url>
					<title>Illustration of a fuzzy black hole horizon</title>
					<link>/data/img/thumbnails/GUP.png</link>
				</image>
				<content type="html">
					
						<b>Warning:</b> this post contains math blocks rendered with the
						<a href="https://www.mathjax.org/">MathJax</a> JavaScript library. If you're using an RSS reader or
						otherwise have Javascript disabled, none of them will display, so you should instead read this article at
						its <a href="https://fields007.github.io/articles/gup">source</a>.
					
					<p>This article is based on my Bachelor’s thesis<sup id="link1"><a href="#note1">1</a></sup> in mathematical physics. I outline the main ideas and motivations of the work in broader terms. In the thesis, I investigate how deformations of the standard uncertainty relations can be interpreted, at the classical level, as deformations of the Poisson algebra of observables. I then explore how this modified structure affects Hamiltonian dynamics, with applications to geodesic motion in Schwarzschild and Kerr spacetimes and to black hole thermodynamics.</p>

<h2 id="uncertainty-relations">Uncertainty relations</h2>

<p>In quantum mechanics, physical quantities such as position or momentum are represented by operators. Given two observables \(a\) and \(b\), their commutator is defined as \([a,b] = ab - ba\). If \([a,b]=0\), the observables and said to commute. Physically, this can be understood as measuring \(a\) and then \(b\) giving the same result as measuring them in the opposite order. When the commutator does not vanish, the order of measurements matters.</p>

<p>The uncertainty relation follows directly from this non-commutativity. For two observables \(a\) and \(b\),</p>

\[\Delta a\, \Delta b \ge \frac{1}{2}|\langle [a,b]\rangle| .\]

<p>Here \(\Delta a\) denotes the standard deviation of the measurements of observable \(a\) and \(\langle \cdots \rangle\) is an expectation value (average of measurement values in this case of the commutator operator). In this work we assume vanishing expectation values of \(a\) and \(b\), so for instance \(\Delta a\) directly characterizes the typical size of fluctuations.</p>

<p>For position \(x\) and momentum \(p\), the canonical commutation relation \([x,p] = i\hbar\) leads to the familiar Heisenberg uncertainty relation</p>

\[\Delta x\, \Delta p \ge \frac{\hbar}{2} .\]

<p>A basic physical intuition already follows from this formula: if one tries to localize a particle more and more precisely, the momentum uncertainty tends to grow.</p>

<h2 id="motivation-for-generalized-uncertainty-relations">Motivation for generalized uncertainty relations</h2>

<p>Many theoretical frameworks suggest that the standard uncertainty relation may be modified at very short distances. These include for example numerous thought experiments, string theory, curved momentum space, noncommutative geometry, … Basis of many of these ideas is in one way or another existence of an effective minimal length scale.</p>

<p>A commonly studied deformation is a <strong>generalized uncertainty relation</strong> (GUP) of the form,</p>

\[\Delta x\, \Delta p \ge \frac{\hbar}{2}\left(1 + \beta (\Delta p)^2\right) ,\]

<p>\(\beta\) being some constant of appropriate dimensions. The plot below illustrates the qualitative consequence of this relation for the case \(\beta = \ell_p^2/\hbar^2\), where \(\ell_p\) is the Planck length.</p>

<p><img src="/data/img/GUP/minimal_length.svg" alt="Illustration of a minimal length" style="width: 70%; max-width: 500px; display: block; margin: 0 auto;" /></p>

<p>In this plot the coloured region represents the allowed values of \(\Delta x\) and \(\Delta p\). One can see that there is a minimum value of \(\Delta x\) (in this case precisely \(\ell_p\)) which can be interpreted as existence of a <strong>minimal length</strong>, below which localization is no longer possible. While the precise interpretation varies between frameworks, this qualitative feature appears very robust. For instance, similar behaviour arises in string theory due to the extended nature of strings.</p>

<h2 id="reverse-quantisation-and-deformed-algebra">Reverse quantisation and deformed algebra</h2>

<p>In classical mechanics, observables are functions on phase space, and their algebraic structure is encoded in the Poisson bracket \(\{a,b\}\). The equations of motion then take the form</p>

\[\dot{x} = \{x,H\}, \qquad \dot{p} = \{p,H\} ,\]

<p>where dots denote time derivatives and \(H\) is the Hamiltonian. These equations determine the trajectories of classical particles.</p>

<p>When one wants to promote a classical theory to a quantum one follows procedure known as quantisation. The canonical quantisation procedure is to replace Poisson brackets as</p>

\[\{a,b\} \longrightarrow \frac{1}{i\hbar}[a,b] .\]

<p>In my thesis, I take the reverse perspective. Because commutators give us uncertainty relations, GUP motivates a deformed quantum algebra. In my work, I ask what classical structure corresponds to one such deformed algebra.</p>

<p>Particularly I look at the <strong>Snyder algebra</strong><sup id="link2"><a href="#note2">2</a></sup>, given in the quantum theory by</p>

\[[x^\mu,x^\nu] = -i\hbar \beta (x^\mu p^\nu - p^\mu x^\nu), \qquad
  [x^\mu,p_\nu] = i\hbar \left(\delta^\mu_\nu + \beta \, p_i p_j\right), \qquad
  [p_i,p_j] = 0 .\]

<p>Where \(x^\mu\) denotes the spacetime coordinate operators and \(p^\mu\) four-momentum operators (four-momentum combines momentum and energy). Taking the classical limit leads to a deformed Poisson algebra with the same structure. This then leads to modified Hamiltonian equations, and therefore to modified classical trajectories.</p>

<h2 id="deformed-solutions-of-general-relativity">Deformed solutions of general relativity</h2>

<p>In general relativity, the geometry of spacetime is encoded in the metric tensor \(g_{\mu\nu}\). Free test particles then follow geodesics, which can be derived using the Hamiltonian formalism.</p>

<p>The Schwarzschild solution of general relativity for example describes the spacetime outside a spherically symmetric mass. It applies to systems like the Solar System as well as, for example non-rotating black holes. A common visualization of this spacetime provides the Flamm’s paraboloid, shown below, which represents a spatial slice in the radial direction.</p>

<p><img src="/data/img/GUP/schwarzshild.png" alt="Flamm's paraboloid for the Schwarzschild solution." style="width: 70%; max-width: 500px; display: block; margin: 0 auto;" /></p>

<p>The Kerr solution generalizes this to rotating black holes (though basically just to black holes) and features additional structures such as an ergosphere and an inner horizon.</p>

<p><img src="/data/img/GUP/kerr.png" alt="Schematic structure of the Kerr spacetime." style="width: 70%; max-width: 500px; display: block; margin: 0 auto;" /></p>

<p>In my work, I study particle motion in both Schwarzschild and Kerr spacetimes using the <strong>pseudo-Newtonian limit</strong>. In this approximation one assumes weak gravitational field and spatial components of canonical four-momenta much smaller than the time component. Implementing the Snyder-deformed Poisson algebra in this setting then leads to modified geodesic equations and hence to deformed trajectories.</p>

<h2 id="what-was-done-and-what-was-left-out">What was done and what was left out</h2>

<p>The main contribution of the thesis is a logically consistent implementation of a deformation motivated by GUP, together with its application to geodesic motion in curved spacetimes within a pseudo-Newtonian approximation. While similar deformations of Schwarzschild geodesics have been studied before<sup id="link3"><a href="#note3">3</a></sup>, the approach taken here differs both conceptually and technically.</p>

<p>Many aspects were left out of this short overview. The thesis<sup id="link1"><a href="#note1">1</a></sup> contains for example extended discussion on the motivations of such deformation and explicit calculations of the modified equations of motion. At the end applications of GUP to black hole thermodynamics are also discussed. Interested readers can find all details in the full text.</p>

<hr />
<p><span id="note1"><a href="#link1">1</a></span>:
J. Masák,
<em>Generalized uncertainty relation and its use in cosmology</em>,
Bachelor’s thesis,
Faculty of Nuclear Sciences and Physical Engineering, Czech Technical University in Prague (2023),
<a href="https://dspace.cvut.cz/handle/10467/111416" target="_blank">dspace.cvut.cz/handle/10467/111416</a>.</p>

<p><span id="note2"><a href="#link2">2</a></span>:
H. S. Snyder,
<em>Quantized Space-Time</em>,
Physical Review 71 (1947), 38–41,
<a href="https://doi.org/10.1103/PhysRev.71.38" target="_blank">doi:10.1103/PhysRev.71.38</a>.</p>

<p><span id="note3"><a href="#link3">3</a></span>:
S. Mignemi and R. Strajn,
<em>Geodesics in the Schwarzschild metric with Snyder dynamics</em>,
Phys. Rev. D <strong>90</strong>, 044019 (2014),
<a href="https://doi.org/10.1103/PhysRevD.90.044019" target="_blank">doi:10.1103/PhysRevD.90.044019</a>.</p>
>
				</content>
				<pubDate>Fri, 01 Sep 2023 00:00:00 +0000</pubDate>
				<link>https://fields007.github.io/articles/gup</link>
				<guid isPermaLink="true">https://fields007.github.io/articles/gup</guid>
			</item>
		
	</channel>
</rss>
