Introduction To Loop Quantum Gravity

In summary, Loop Quantum Gravity (LQG) is the attempt to unify General Theory of Relativity and Quantum Mechanics. This is a challenging task as these two theories have different foundations - position is uncertain in Quantum Mechanics due to the Heisenberg principle, while it is not the case in General Theory of Relativity. In order to quantize the GTR, gauge fields on a manifold are needed and must be quantized. This requires obeying two laws - diffeomorphism invariance and gauge invariance. Mathematicians like Gauss and Riemann have taught us that a manifold is described by connections, with the most familiar example being the metrical connection. In LQG, all possible metrics were initially used
  • #36
Quantizing Gen Rel gets rid of singularities

Gen Rel is an amazingly accurate theory of spacetime geometry whever it is applicable, where it doesn't break down and fail to compute.
Where it is applicable it predicts very fine differences in angles and times out to many decimal places. People have tried for decades to improve on it, or to test it and find it wrong out at the 6th decimal place. But they haven't succeeded yet.

But Gen Rel famously has places where it blows up and predicts infinities, in other words it is flawed. It has singularities.

this has been the case with other classical theories and it has been found that if you can QUANTIZE a classical theory it will often extend the applicability and get rid of places where it breaks down.

So a big aim of quantizing Gen Rel is to get rid of the classical singularities. mainly the "bigbang" and "blackhole" singularities.

The main reason why LQG is so active these days is that it appears to have removed Gen Rel singularities. the main reason Martin Bojowald is a key LQG figure is that he has been in the forefront in this and has gathered a considerable group of people who are working on this.

the first break came in 2001 when MB removed the bigbang classical singularity in a certain case.

To get the history since then, just go to arxiv.org and get the list of all Bojowald papers since 2001. the people active in this field are the people who have co-authored papers with Bojowald, and you can click on their names and find all the papers they have published independently.

the black hole singularity is being removed just now, starting at end 2004 and very much at the present. a bojo paper on that came out this month (March 2005)

When a classical singularity is removed then you can run the model THROUGH where it used to be. the machine no longer blows up or stalls at that point. So you can explore BEYOND the classical singularity and that is interesting. It is expected that one way to check LQG is to look for traces in the cosmic microwave background of what LQG predicts about the bigbang that is different from classical Gen Rel, different because of it having removed the singularity.

So that is a very important feature of LQG, the fact that it doesn't encounter these irritating singularities in Gen Rel that have bothered people so long.

If you want a non-math way to think about it, focus on the uncertainty of a quantum theory. For the universe or a black hole to collapse all the way to a point would just be too certain, wouldn't it? Too definite for real nature to allow :smile:. So it doesn't happen. At a certain indeterminate very high density there is a "bounce" according to the math (a time-evolution difference equation model) and contraction turns into expansion. And conditions for inflation are automatically generated.

recent papers
arxiv.org/gr-qc/0503020
Bojo
the early universe in Loop Quantum Cosmology

arxiv.org/gr-qc/0503041
Bojo, Goswami, Maartens, Singh
a black hole mass threshold from non-singular quantum gravitational collapse

if you glance at these papers you will not see anything about thinking of space as divided up into little bits, or grains :smile:
because that is not what real LQG is about,
but you will get a taste of what is going on with the overcoming of the Gen Rel singularities at bigbang and black hole.
 
Physics news on Phys.org
  • #37
What we have to do, where we have to go

We have to give an introduction to LQG that is reasonably faithful to the textbook version that researchers actually use----not just some possibly misleading verbal imagery. but it has to be understandable as an introduction.

that's hard. it will take several tries and the first ones will fail

It would be so great if people could just go and read a paper like
smolin "An Invitation to LQG" and have that suffice, but it somehow does not work. "Invitation" is too condensed for many people, or not explanatory enough.

Well I will try to get moving on this. I also want to keep those background points handy, from the previous 3 or 4 posts. So here, as a reminder, are the headings from post #31 onwards:

Introduction to LQG Part II

LQG can mean several things

There are textbook-level LQG sources

Textbook LQG is based on a differentiable manifold

QUANTUM" is a way of handling uncertainty and incomplete information realistically

Quantizing Gen Rel gets rid of singularities
 
  • #38
Can we get word definitions of some of the technical concepts? Nightcleaner's interest in BF theory suggests that we could show what the Ashtekar variables are, at some honest, non-confusing level. Then Wilson action, and why it takes values in the Lie Algebra of the group, for that matter how the group comes in (ation on the manifold, forget oll the bundle staff), and Circle functions and so on. I go up and down on this; I think it would be boring if it wasn't impossible, and then I think it's a duty to get this across to the bright, self-selected audiance we have here.
 
  • #39
A differentiable manifold is a shapeless smooth set

Differentiable manifold has 8 or 9 syllables and it is easier to say smooth set, which only has 2 syllables. And that is what one is. It is a set with a bunch of coordinate charts that work smoothly together.
Typically you can't get the whole set on one coordinate chart so you have several overlapping charts

that is like you can't get the whole Earth on one square map, but you can plaster maps all over the Earth so you have overlapping coverage.
On every patch of surface there is some map that is good at least on that local region.

the typical set used to represent space in LQG is the "3-sphere" where the surface of a balloon is a 2-sphere and you have to imagine going up one dimension. a local chart looks like regular 3-D graph paper or familiar euclidean 3-space

Only thing is we ignore the geometry you might have thought we had when I said 3-sphere. If we were thinking of the 2-sphere balloon as an analogy, the air is out of the balloon and it is crumpled up and thrown into your sock drawer. it has no shape. In the same way, by analogy, the 3-sphere has no shape. It is just a set of points, without a boundary, that has been equipped with an adequate bunch of coordinate maps

the "smooth" part is that wherever the charts overlap if you want to start on one map and find the corresponding point on the other map, and do a whole transference thing that remaps you from one to the other, well that
remapping (from one patch of 3-D graph paper to another) is smooth. that is to say differentiable, as in calculus, you can take the derivative as many times as you want. In other words the coordinate charts are COMPATIBLE with each other because whenever you remap between two that overlap you find you can DO CALCULUS at will on the function taking you from one to the other. this is an example of a technical condition that basically doesn't say very much except that we won't have nasty surprises when we get around to using the charts. The charts are smoothly compatible with each other.

The idea of a differentiable manifold was given us by George Riemann in 1854 when he was trying to get a job as lecturer at Göttingen and had to give a sample lecture, and it is actually SIMPLER than euclidean space because it does not have any geometry! Euclidean space has all kinds of rich structure immediately availabe, like you can say what a straight line is and you can measure the angle between two intersecting straight lines!

what we have here is a SHAPELESS SMOOTH SET and you can't do any of that. It has the absolute mininum of structure for something that can serve as a useful model of a CONTINUUM.

this is why it was a good idea of Riemann, because it is simpler and less structured than Euclidean space and so it is more able to adapt to the wonders of the universe. mathematics was changed very much in 1854.
George Riemann lived 1826 to 1866.

Here is his 1854 talk, in full:
http://www.ru.nl/w-en-s/gmfw/bronnen/riemann1.html

I think what he called a "stetige Mannigfaltigkeit" here in this talk we would call a smooth manifold. But thereafter the name "differenzierbare Mannigfaltigkeit" became prevalent and is what we call differentiable manifold.
 
Last edited by a moderator:
  • #40
stetig Maniigfaltigkeit translates literally to continuous manyfoldedness. The manyfoldedness is the number of independent variables, which if you think of them geometrically, become dimensions. Our word manifold is the result of a long history of trying to express the idea of multiplicity of variables in English. The French of course say variete, with a grave accent I am too lazy to supply. Personally I like continuum, which did for Einstein, and which he compared, as to its smoothness, to a marble table top (Soooo nineteenth century!).
 
  • #41
selfAdjoint said:
Can we get word definitions of some of the technical concepts? ...

I would consider it a favor if you stepped in and supplied some. I will not be going very fast so it will be possible to step in at pretty much any point and add or improve definitions. right now I think of this as "Introduction Mark II" and the aim is to get a preliminary description out there which is at least not too misleading. that means there will be topics to expand later in a "Introduction Mark III"
At some point some student of Ashtekar or Rovelli will probably write something that makes all this unnecessary. a real beginner textbook for Loop Gravity. but we can't afford to wait around for that because we don't know when it will happen
 
  • #42
Ashtekar varaibles

marcus said:
I would consider it a favor if you stepped in and supplied some. I will not be going very fast so it will be possible to step in at pretty much any point and add or improve definitions. right now I think of this as "Introduction Mark II" and the aim is to get a preliminary description out there which is at least not too misleading. that means there will be topics to expand later in a "Introduction Mark III"
At some point some student of Ashtekar or Rovelli will probably write something that makes all this unnecessary. a real beginner textbook for Loop Gravity. but we can't afford to wait around for that because we don't know when it will happen

I am up for F, but the "densitized dual 2-form" B (or E) has me buffaloed. A dual 2-form maps a pair of vectors multilinearly into the ground field, either reals or complex numbers. Dentsitizing makes it integrable, so far so good, but B or E has values in the Lie Algebra just like F, rather than the ground field. Why?
 
  • #43


It is suggestive that you mention Ashtekar variables, and also mention the variables of BF theory (which Freidel tries to reform us so that we write EF thinking that it makes better sense than BF). Let me tell you what my sense of direction tells me. I listened to a (January, Toronto?) recorded talk by Vafa and I heard something ring in his voice when he said "form theories of gravity"----and I went back and looked at the current paper Dijkgraaf, Gukov, Neitzke, Vafa just to make sure. there was a sense of relief. it represents a hopeful general idea for him.

from my perspective, Ashtekar variables and BF are foremostly examples of "form theories" and there could be modifications and other "form theories" we don't know about yet. there is a mental compass needle pointing in this general direction.

it we want to play the game of making verbal (non-math) definitions for an intelligent reader, then it is important the ORDER we define the concepts and also the GOAL or where we are going. I think the direction is that we want to get to where we can say what a "form defined on a manifold" is, or to be more official we should always say "differential form" defined on a "differentiable manifold". So we need to say what the "tangent vectors" are at a point in a manifold.

the obstacle here is that these concepts are unmotivated, have too many syllables if you try to speak correctly, and seem kind of arbitrary and technical.

So I am thinking like this. the thing about a tangent vectors and forms is that they are BACKGROUND INDEPENDENT. All that means, basically, is that you don't have to have a metric. A background independent approach to any kind of physics simply means in practice that you start with a manifold as usual (a "continuum" you say Einstein liked to say) and you refrain from giving yourself a metric.

Well, how can you do physics on a manifold that (at least for now at the beginning) has no metric? What kind of useful objects can you define without a metric? Well, you do have infinitesimal directions because you have coordinates and you can take the derivative at any point, so at a microscopic level you do have a vectorspace of directions-----call them TANGENTS. and on any vectorspace one can readily define the dual space of linear functionals of the vectors-----things that eat the vectors up and give a number. The dual space of the tangents is called the FORMS.

and also the forms don't have to be number-valued, they can be "matrix" valued, one form can eat a tangent vector and produce therefrom not simply one number but 3 numbers or 4 numbers, or a matrix of numbers, but that is not quite right let's say it eats the tangent vector and produces not a number but an element of some Lie algebra. then it is a ALGEBRA-VALUED form.

now this already seems disgustingly complicated so let's see why it might appeal to Cumrun Vafa arguably the world's top string theorist still functioning as such.
I think it appeals to Cumrun Vafa because it is a background independent way to do physics. that is essentially what "form theory of gravity" means.

And string theorists have been held up for two decades by not having a background independent approach. And it JUST HAPPENS that the Ashtekar variables are forms, and the B and F of BF theory are forms, and (no matter what detractors say) Loop has been making a lot of progress lately, and Vafa says "hey, this might be the way to get background independence" and he creates a new fashion called "topological Mtheory" which is a way of focussing on forms and linking up with "form theories of gravity".

So maybe the point is not that this or that particular approach is good or not, but simply that one should work with a manifold sans metric, and do physics with the restricted set of tools that can be defined without a metric. And that means that, painfully abstract as it sounds, nightcleaner has to understand 3 things:

1. the tangent space at a point of a manifold is a vectorspace
2. any vectorspace has a dual space (the things that eat the vectors) and that dual space IS ITSELF a vectorspace.
3. the dual of the tangentspace is the forms and you can do stuff with forms.

Like, you can multiply two forms together (the cute "wedge" symbol), and you can construct more complicate forms that eat two vectors at once or that produce something more jazzy, in place of a number.

The hardest thing in the world to accept is that this is not merely something that mathematicians have invented to do for fun, a genteel and slightly exasperating amusement. The hardest thing to accept is that nature wants us to consider these things because it is practically the only thing you can do with a manifold that doesn't require a metric!

So instead of talking about BF theory or Ashtekar variables in particular, my compass is telling me to wait for a while and see if anyone is interested in "forms on a manifold" that is to say in the clunky polysyllabic language "differential forms defined on the tangent space of a differentiable manifold" UGH.

Also, selfAdjoint, you mentioned the word "bundle". Bundles may be going too far but they are in this general area of discussion, and there is also "connection"
A "connection" is a type of form. So if you understand "form" then you can maybe understand connection.

there is also this extremely disastrous thing that "form" is a misleading term. In real English it means "shape" but a differential form is not a shape at all. Richard being a serious fan of words will insist that it means shape. But no. Some frenchman happened accidentally to call a machine that eats tangent vectors and spits out numbers by the name "form" and so that is what it is called, even tho it is in nowise a shape. It is more like an incometax form, than it is a shape-form. And it is not like an incometax form either.

And as a final ace in the hole we can always say that Gen Rel is an example of a physical theory defined on a manifold without a metric. The metric is a variable that you eventually solve the equation to get. you start without a metric and you do physics and you eventually get a metric.
If there is any useful sense to Kuhntalk then this is a "paradigm". and when Vafa has a good word to say about "form theories of gravity" then this might be the kind of softening that accompanies a shift in perspective.
 
Last edited:
  • #44
Could you please try to give an example of a calculation? I would like to get some feeling of how to handle all up to the mapping from a Lie algebra element to a Lie group element via parallel transport along a loop. Additionally, I'd like to see what a projection from the manifold into the tangent space looks like in practice.

Just to make things more clear: what exactly is the nature of this manifold you are talking of? Is there any physical interpretation?
 
  • #45
Cinquero said:
Could you please try to give an example of a calculation? I would like to get some feeling of how to handle all up to the mapping from a Lie algebra element to a Lie group element via parallel transport along a loop. Additionally, I'd like to see what a projection from the manifold into the tangent space looks like in practice.

Just to make things more clear: what exactly is the nature of this manifold you are talking of? Is there any physical interpretation?

hello Cinquero, have you by any chance looked at the beginning treatment of LQG in Rovelli and Upadhya's paper? This was my introduction to the subject back in 2003. Several of us at Physics Forum were reading that paper back then.

It is short (on the order of 10 pages) and shows how a number of things are calculated. If you are interested in learning LQG, then I could review the paper myself, and read some of it with you.

If you do not already have Rovelli/Upadhya and would like the link, please let me know. the date at arxiv is about 1998.
 
  • #46
meanwhile, a manifold is a topological space locally homeomorphic to Rd by mappings phi, psi,...which have the following differentiability property: where the domains of two maps overlap,
going from Rd to Rd by the composition of one with the inverse of the other is (either continuously differentiable a certain number of times or) infinitely differentiable.

LQG is usually developed in the d=3 case and the manifold that physically represents space is taken to be "smooth"----which means that the mappings from Rd to Rd which I just mentioned are infinitely differentiable.

LQG can be defined in any dimension d. It is not limited to the d = 3 case, and indeed has been studied in some other cases besides d = 3. But typically the manifold representing space is a compact smooth 3-manifold, a "continuum", denoted by the letter M.

You can get all this from any beginning treatment of LQG like, e.g. Rovelli/Upadhya, or Rovelli/Gaul. Again, if you need links, let me know.

All I have done to supplement the standard treatment that you find there is to define a differentiable manifold. I assume this is very familiar to you Cinquero but some other reader might conceivably want it defined.
 
Last edited:
  • #47
As a background note, the classical "ADM" treatment of General Relativity has been posted on arxiv!

http://arxiv.org/gr-qc/0405109

this is a reprint of something published in 1962! You might say this is where the manifold M, that Cinquero is asking about, comes from:
the 1962 Arnowitt, Deser, Misner treatment of classical Gen Rel. Instead of a purely spacetime development, ADM looked at the metric restricted to an embedded 3-d spatial hypersurface.

Around 1986, two other people, Sen and Ashtekar, adapted the ADM approach by shifting attention to connections defined on the 3-d manifold. The connection then, rather than the metric, represented the variable geometry on space.

In the 1990s, when LQG started to develop, much of the context (concepts and notation) was already in place because of this prior work in classical Gen Rel. It was a matter of quantizing the ADM/Ashtekar version of Gen Rel, which had already been established for some time and was familiar to relativists. Here is some of that context (notation has not been fully standardized)

M smooth compact manifold represent space
A connections on M, representing the set of all geometries
K complex-valued functions on A, quantum states of geometry

K is too big and needs to be collapsed down (by applying constraints and equivalences) to a separable Hilbert space----the physical state space---of quantum states of geometry.

But already, with this bare minimum of concepts, one can begin to get oriented. K is a linear space of functions defined on A, the set of connections. One can think of A as the "configurations" and K as "wave functions" familiar from common QM. It is interesting to look for a BASIS of K----a minimal spanning set of complex-valued functions defined on connections.

Already, even with this abbreviated roadmap of the subject, I am touching on concepts that would be a lot of work to define and are better to read about. So if there is further interest I will get some links.

this little sketch is typed from memory, I haven't reviewed the definitions and history for quite a while. And I'm not omniscient either! So suggestions and improvements, including links to articles that develop LQG formalism, are welcome.
 
Last edited by a moderator:
  • #48
hi Cinquero, I got the link for Rov/Upad in case you want it

http://arxiv.org/gr-qc/9806079

I checked and they are using notation L where I wrote K, but otherwise no change

BTW, I see you are a new member, welcome!
 
Last edited by a moderator:
  • #49
Thx! That article is very helpful.

Question:

in II.B, the very first sentence: A is defined on M, right? But then, why does AV make sense? V maps from M to SU(2), but A is defined on M! What am I missing?
 
  • #50
Cinquero said:
Thx! That article is very helpful.

Question:

in II.B, the very first sentence: A is defined on M, right? But then, why does AV make sense? V maps from M to SU(2), but A is defined on M! What am I missing?

Hi Cinquero, I just saw your post. Sorry for not replying earlier!

you remember on page 1, section II A, they say

"Let A be an SU(2) connection on M; that is, A is a smooth 1-form with values in su(2), the Lie algebra of SU(2)."

that means if you specify a point and a direction you get a matrix

(lets imagine that a basis has been chosen so that things are less abstract and all the SU(2) things and su(2) things are actually just 2x2 matrices :smile:)

but at every point of M, the function V also gives a matrix! so we can conjugate A by V and have
the new matrix V-1A V
what is meant by writing them together this way is just matrix multiplication

this is how to interpret the first sentence of II B, where the notation
AV is defined

==============
do I need to be more rigorous and formal, and spell this out in more detail?
or is this OK?
 
Last edited:
  • #51
marcus said:
do I need to be more rigorous and formal, and spell this out in more detail?
or is this OK?

No, it's ok. Thx!
 
  • #52
Someone might be curious as to what "Introduction to LQG(Mark II)" is all about, what is the main direction, if one were to look ahead. Someone teaching a course for juniors/seniors in LQG might, by the end of the semester, want to be in sight of this body of work:
http://arxiv.org/find/grp_physics/1/au:+bojowald/0/1/0/all/0/1

and in particular within striking distance of this 2001 trailblazer
http://arxiv.org/abs/gr-qc/0102069
Absence of Singularity in Loop Quantum Cosmology

it is just 4 pages. the classical BB singularity is replaced by a bounce (from a prior gravitational collapse)
later it was discovered that conditions at the bounce automatically trigger a brief episode of inflation (without fine-tuning or elaborate "extras")
see for example http://arxiv.org/abs/gr-qc/0407069, "Genericness of Inflation in LQC" (also just 4 pages) and references thererin.

here are the papers which have cited this key paper:
http://arxiv.org/cits/gr-qc/0102069
there are currently about 75 papers which have cited it, and about half of these appeared after January 2004
 
Last edited:
  • #53
One's approach to LQG, in an introduction, depends a lot on where one wants to be at the end of it and their are several equally valid goals that one could have. For me what stands out is that quantizing General Relativity gets rid of some important singularities where classical GR broke down and allows one to study in more detail what goes on there.

For example it might be good to aim for making contact not only with "Absence of Singularity in LQC" but also
http://arxiv.org/abs/gr-qc/0503041
which treats what emerges when the classical black hole singularity is removed by LQG.
 
  • #54
marlon, so are you saying that there is some sort of ''ether'', a actual physical property to space that changes with the interaction of particles on particles? The space itself warps, changes value, and interacts with the particles themself - this being the gravity ?
 
  • #55
It's common knowledge that the three people most responsible for initiating the LQG approach to quantizing General Relativity are Abhay Ashtekar, Carlo Rovelli, and Lee Smolin.

A good way to get a sense of what LQG is about is to keep an eye on major books and survey articles by these people, since they are like the "founding fathers" of the field.

From Smolin we have an excellent recent survey article "An Invitation to LQG" which gives useful information for the trained physicist considering getting into LQG research----main results, experimental tests, and a list of unsolved problems to work on.

From Rovelli we have his book Quantum Gravity which came out November 2004 published by Cambridge Press. the December 2003 draft is still online. He also has some earlier surveys and popular articles.

From Ashtekar there are several valuable surveys. Here are links to a couple of the more recent ones that might be useful.
http://arxiv.org/abs/gr-qc/0410054
http://arxiv.org/abs/gr-qc/0404018
But what is especially interesting right now is a book Ashtekar is preparing, to be published by World Scientific, called
A Hundred Years of Relativity.

this book has a broad scope including all of General Relativity, and it will show how Ashtekar sees LQG and other allied approaches to quantizing Gen Rel in their wider context.

Interestingly, several chapters of this book "100Y.of R." are already online as preprints!

I will get links for some preprint chapters.

Martin Bojowald
[he has contributed an article called "Loop Quantum Cosmology"
which I have not yet seen online]

Larry Ford
http://arxiv.org/abs/gr-qc/0504096

Rodolfo Gambini and Jorge Pullin
http://arxiv.org/abs/gr-qc/0505023

Hermann Nicolai
["Gravitational Billiards, Dualities and Hidden Symmetries" not yet online]

Thanu Padmanabhan
http://arxiv.org/abs/gr-qc/0503107

Alan Rendall
http://arxiv.org/abs/gr-qc/0503112

Clifford Will
http://arxiv.org/abs/gr-qc/0504086

Although Bojowald's article may not be available yet, see
http://edoc.mpg.de/display.epl?mode=people&fname=Martin&svir=0&name=Bojowald

and also
http://edoc.mpg.de/display.epl?mode=doc&id=213885&col=6&grp=84
 
Last edited:
  • #56
More information has come in, so i will revise parts of the preceding post.
From Ashtekar there are several valuable surveys. Here are links to a couple of recent ones that might be useful.
http://arxiv.org/abs/gr-qc/0410054
Gravity and the Quantum
http://arxiv.org/abs/gr-qc/0404018
Background Independent Quantum Gravity: A Status Report

What is especially interesting right now is a book Ashtekar is preparing, to be published by World Scientific, called
A Hundred Years of Relativity.

This book has a broad scope covering all of General Relativity, including numerical GR and testing. It will show how Ashtekar sees LQG and allied approaches to quantizing Gen Rel in the wider context. Several chapters of this book are already online as preprints:

Martin Bojowald
http://arxiv.org/abs/gr-qc/0505057
Elements of Loop Quantum Cosmology

Larry Ford
http://arxiv.org/abs/gr-qc/0504096

Rodolfo Gambini and Jorge Pullin
http://arxiv.org/abs/gr-qc/0505023
Discrete space-time

Hermann Nicolai
["Gravitational Billiards, Dualities and Hidden Symmetries" not yet online]

Thanu Padmanabhan
http://arxiv.org/abs/gr-qc/0503107
Understanding Our Universe: Current Status and Open Issues

Alan Rendall
http://arxiv.org/abs/gr-qc/0503112

Clifford Will
http://arxiv.org/abs/gr-qc/0504086
Was Einstein Right? Testing Relativity at the Centenary
 
Last edited:
  • #57
LQG means what comes to the Loops 05 Conference in October

Because LQG contains some leading-edge lines of research it cannot be given a fixed definition. The practical definition is that it is what Loop people do-----and in practice that means the research lines that are featured in this year's major Loop conference(s).

So for an operational definition of Loop-and-allied approaches to Quantum Gravity, watch the Programme of the October 10-14 conference at Potsdam AEI. Here's the link and some available details:

http://loops05.aei.mpg.de/index_files/Home.html

http://loops05.aei.mpg.de/index_files/Programme.html

The topics of this conference will include:

Background Independent Algebraic QFT
Causal Sets
Dynamical Triangulations
Loop Quantum Gravity
Non-perturbative Path Integrals
String Theory

A detailed programme will be available in July.

Invited Speakers will include:
Abhay Ashtekar (USA)
John Baez (USA)
John Barrett (UK)
Alejandro Corichi (MEX)
Robbert Dijkgraaf (NL)
Fay Dowker (UK)
Laurent Freidel (FR and CA)
Karel Kuchar (USA)
Jurek Lewandowski (POL)
Renate Loll (NL)
Roy Martens (UK)
Hugo Morales Tecotl (MEX)
Alejandro Perez (FR)
Jorge Pullin (USA)
Martin Reuter (GER)
Carlo Rovelli (FR)
Lee Smolin (CA)
Rafael Sorkin (USA)
Stefan Theisen (GER)
Rainer Verch (GER)
-------------------------------
My comment: because these are fast developing areas of research, it makes sense not to nail down the TITLES of the invited speaker's talks until shortly before the conference (July is 3 months before, plenty of time) but it's nice to know WHO will be giving the plenary talks.

I would say that String and old-style LQG are no longer leading edge, and I don't have a big interest in Causal Sets. So I would narrow the exciting topics down to these:

Dynamical Triangulations
Background Independent Algebraic QFT
Non-perturbative Path Integrals

1. Notice that Renate Loll is on the invited list. She will talk about CDT, causal dynam. triang.
This is currently the deepest part of Loop-and-allied research. Anyone interested in LQG, or quantum gravity in general for that matter, should know about it.

2. What they mean by "Background Independent" QFT is basically that it is done on a (metric-less) differentiable manifold. the way you work on a shapeless continuum without first introducing a prior geometry is you use
DIFFERENTIAL FORMS and stuff like bundles and connections. Cumrun Vafa's term for one case of this is "form theories of gravity". A key invited speaker in this line would be Laurent Freidel.

I am not sure what Background Independent "Algebraic" QFT means. I think the papers of Rainer Verch (which I don't know) could touch on this.

the moment you posit a manifold you have already specified a dimension like D = 4 and you already have patches of coordinates but notice that the CDT of Renate Loll does not have a prior commitment to a dimension and it uses NO COORDINATES AT ALL. the brilliant Tullio Regge figured in 1950 how to do Einstein Gen Rel without coordinates. and, in CDT which is basically a child of Regge, the dimension emerges rather than being specified in advance and the dimension can vary with scale----it can be 4D at macro and run smoothly down to around 2D at micro-scale.

This is why I cannot escape concluding that CDT is deeper-probing. It may be WRONG we don't know about right or wrong. However it seems to have Gen Rel as its classical limit, and integrate out to a simple quantum cosmology associated with Hawking as a kind of semiclassical limit.

3. Non-perturbative Path Integrals might be an improved and more general term for what used to be called Spin Foams, but it also includes CDT because in CDT you get a path integral. Which, however, is evaluated barbarically using Monte Carlo runs on the computer.

This is going to be an interesting Loops 05 Conference and I guess it is the conference that defines the field (more than the other way round).
So we will see in Potsdam in October what LQG is.
 
  • #58
short reading list for CDT

in case anyone is interested in getting a tast of causal dynamical triangulations (CDT) here is a short reading list.

A new monograph "Reconstructing the Universe" is due to come out this month. It will replace the 2001 paper which I link to here. the 3 short papers from 2004 and 2005 give the highlights of recent research results.
It is better to first read the 3 short recent papers before getting into the details in the 2001 paper IMHO.

1.
http://arxiv.org/hep-th/0105267
Dynamically Triangulating Lorentzian Quantum Gravity
J. Ambjorn (NBI, Copenhagen), J. Jurkiewicz (U. Krakow), R. Loll (AEI, Golm)
41 pages, 14 figures
Nucl.Phys. B610 (2001) 347-382
"Fruitful ideas on how to quantize gravity are few and far between. In this paper, we give a complete description of a recently introduced non-perturbative gravitational path integral whose continuum limit has already been investigated extensively in d less than 4, with promising results. It is based on a simplicial regularization of Lorentzian space-times and, most importantly, possesses a well-defined, non-perturbative Wick rotation. We present a detailed analysis of the geometric and mathematical properties of the discretized model in d=3,4..."

2.
http://arxiv.org/abs/hep-th/0404156
Emergence of a 4D World from Causal Quantum Gravity
J. Ambjorn (1 and 3), J. Jurkiewicz (2), R. Loll (3) ((1) Niels Bohr Institute, Copenhagen, (2) Jagellonian University, Krakow, (3) Spinoza Institute, Utrecht)
11 pages, 3 figures; final version to appear in Phys. Rev. Lett
Phys.Rev.Lett. 93 (2004) 131301
"Causal Dynamical Triangulations in four dimensions provide a background-independent definition of the sum over geometries in nonperturbative quantum gravity, with a positive cosmological constant. We present evidence that a macroscopic four-dimensional world emerges from this theory dynamically."

3.
http://arxiv.org/abs/hep-th/0411152
Semiclassical Universe from First Principles
J. Ambjorn, J. Jurkiewicz, R. Loll
15 pages, 4 figures
Phys.Lett. B607 (2005) 205-213
"Causal Dynamical Triangulations in four dimensions provide a background-independent definition of the sum over space-time geometries in nonperturbative quantum gravity. We show that the macroscopic four-dimensional world which emerges in the Euclidean sector of this theory is a bounce which satisfies a semiclassical equation. After integrating out all degrees of freedom except for a global scale factor, we obtain the ground state wave function of the universe as a function of this scale factor."

4.
http://arxiv.org/abs/hep-th/0505113
Spectral Dimension of the Universe
J. Ambjorn (NBI Copenhagen and U. Utrecht), J. Jurkiewicz (U. Krakow), R. Loll (U. Utrecht)
10 pages, 1 figure
SPIN-05/05, ITP-UU-05/07

"We measure the spectral dimension of universes emerging from nonperturbative quantum gravity, defined through state sums of causal triangulated geometries. While four-dimensional on large scales, the quantum universe appears two-dimensional at short distances. We conclude that quantum gravity may be "self-renormalizing" at the Planck scale, by virtue of a mechanism of dynamical dimensional reduction."
 
Last edited by a moderator:
  • #59
AFAICS a breakthrough form of simplex path-integral gravity called causal dynamical triangulations (CDT) is the most important current development in Quantum Gravity going on. In case anyone is interested in getting a taste of CDT here is a short reading list.

this is an update of what I listed earlier:


1.
http://arxiv.org/hep-th/0105267
Dynamically Triangulating Lorentzian Quantum Gravity
J. Ambjorn (NBI, Copenhagen), J. Jurkiewicz (U. Krakow), R. Loll (AEI, Golm)
41 pages, 14 figures
Nucl.Phys. B610 (2001) 347-382
"Fruitful ideas on how to quantize gravity are few and far between. In this paper, we give a complete description of a recently introduced non-perturbative gravitational path integral whose continuum limit has already been investigated extensively in d less than 4, with promising results. It is based on a simplicial regularization of Lorentzian space-times and, most importantly, possesses a well-defined, non-perturbative Wick rotation. We present a detailed analysis of the geometric and mathematical properties of the discretized model in d=3,4..."

2.
http://arxiv.org/abs/hep-th/0404156
Emergence of a 4D World from Causal Quantum Gravity
J. Ambjorn (1 and 3), J. Jurkiewicz (2), R. Loll (3) ((1) Niels Bohr Institute, Copenhagen, (2) Jagellonian University, Krakow, (3) Spinoza Institute, Utrecht)
11 pages, 3 figures; final version to appear in Phys. Rev. Lett
Phys.Rev.Lett. 93 (2004) 131301
"Causal Dynamical Triangulations in four dimensions provide a background-independent definition of the sum over geometries in nonperturbative quantum gravity, with a positive cosmological constant. We present evidence that a macroscopic four-dimensional world emerges from this theory dynamically."

3.
http://arxiv.org/abs/hep-th/0411152
Semiclassical Universe from First Principles
J. Ambjorn, J. Jurkiewicz, R. Loll
15 pages, 4 figures
Phys.Lett. B607 (2005) 205-213
"Causal Dynamical Triangulations in four dimensions provide a background-independent definition of the sum over space-time geometries in nonperturbative quantum gravity. We show that the macroscopic four-dimensional world which emerges in the Euclidean sector of this theory is a bounce which satisfies a semiclassical equation. After integrating out all degrees of freedom except for a global scale factor, we obtain the ground state wave function of the universe as a function of this scale factor."

4.
http://arxiv.org/abs/hep-th/0505113
Spectral Dimension of the Universe
J. Ambjorn (NBI Copenhagen and U. Utrecht), J. Jurkiewicz (U. Krakow), R. Loll (U. Utrecht)
10 pages, 1 figure
SPIN-05/05, ITP-UU-05/07

"We measure the spectral dimension of universes emerging from nonperturbative quantum gravity, defined through state sums of causal triangulated geometries. While four-dimensional on large scales, the quantum universe appears two-dimensional at short distances. We conclude that quantum gravity may be "self-renormalizing" at the Planck scale, by virtue of a mechanism of dynamical dimensional reduction."

5.
http://arxiv.org/hep-th/0505154
Reconstructing the Universe
J. Ambjorn (NBI Copenhagen and U. Utrecht), J. Jurkiewicz (U. Krakow), R. Loll (U. Utrecht)
52 pages, 20 figures
Report-no: SPIN-05/14, ITP-UU-05/18

"We provide detailed evidence for the claim that nonperturbative quantum gravity, defined through state sums of causal triangulated geometries, possesses a large-scale limit in which the dimension of spacetime is four and the dynamics of the volume of the universe behaves semiclassically. This is a first step in reconstructing the universe from a dynamical principle at the Planck scale, and at the same time provides a nontrivial consistency check of the method of causal dynamical triangulations. A closer look at the quantum geometry reveals a number of highly nonclassical aspects, including a dynamical reduction of spacetime to two dimensions on short scales and a fractal structure of slices of constant time."

this is a landmark paper.
I have been looking also for a reader-friendly introductor paper. there is one that is lecture notes aimed at the graduate student level

6.
http://arxiv.org/hep-th/0212340
A discrete history of the Lorentzian path integral
R. Loll (U. Utrecht)
38 pages, 16 figures
SPIN-2002/40
Lect.Notes Phys. 631 (2003) 137-171
"In these lecture notes, I describe the motivation behind a recent formulation of a non-perturbative gravitational path integral for Lorentzian (instead of the usual Euclidean) space-times, and give a pedagogical introduction to its main features. At the regularized, discrete level this approach solves the problems of (i) having a well-defined Wick rotation, (ii) possessing a coordinate-invariant cutoff, and (iii) leading to_convergent_ sums over geometries. Although little is known as yet about the existence and nature of an underlying continuum theory of quantum gravity in four dimensions, there are already a number of beautiful results in d=2 and d=3 where continuum limits have been found. They include an explicit example of the inequivalence of the Euclidean and Lorentzian path integrals, a non-perturbative mechanism for the cancellation of the conformal factor, and the discovery that causality can act as an effective regulator of quantum geometry."

Loll wrote this as an introduction to CDT for Utrecht graduate students who might want to get into her line of research. It is a good beginning. It is already 2 years out of date so it does not have the latest headline results but that is OK.
 
Last edited by a moderator:
  • #60
Here is a reminder of the importance of background independence (no prior metric) and diffeomorphism invariance (general covariance) in the case where the spacetime model is a differential manifold.
these are principles are basic to LQG, and to all allied approaches to quantum gravity.

in fact these two features are basic to classical 1915 General Relativity! So any approach that really tries to quantize Gen Rel is going to exhibit these features or the equivalent

Anyway this is sometimes pointed out as one of the troubles with string theory---that it doesn't have background independence etc. And people debate this. I will not take a stand but simply point out that these principles are really important---and implementing them has shaped LQG and some related approaches---and that one can get into trouble if one does not.

this was illustrated by something posted a few minutes ago in "Third Road" sticky-thread,

http://arxiv.org/gr-qc/0505138
Fibered Manifolds, Natural Bundles, Structured Sets, G-Sets and all that: The Hole Story from Space Time to Elementary Particles
J. Stachel, M. Iftime
40 pages

The article had this in the conclusions (partly already quoted in "third road" thread, but we can use them too as emphasizing how crucial background independence is) :

<<...Perturbative string theory fails this test, since the background spacetime (of no matter how many dimensions) is only invariant under a finite parameter Lie subgroup of the group of all possible diffeomorphisms of its elements. This point now seems to be widely acknowledged in the string community. I quote from two recent review articles. Speaking of the original string theory Michael Green[19] notes: “This description of string theory is wedded to a semiclassical perturbative formulation in which the string is viewed as a particle moving through a fixed background geometry ... Although the series of superstring diagrams has an elegant description in terms of two-dimensional surfaces embedded in spacetime, this is only the perturbative approximation to some underlying structure that must include a description of the quantum geometry of the target space as well as the strings propagating through it ( p. A78). ... A conceptually complete theory of quantum gravity cannot be based on a background dependent perturbation theory ..."

"In ... a complete formulation the notion of string-like particles would arise only as an approximation, as would the whole notion of classical spacetime (p. A 86) ” Speaking of the more recent development of M-theory, Green says: “An even worse problem with the present formulation of the matrix model is that the formalism is manifestly background dependent. This may be adequate for understanding M theory in specific backgrounds but is obviously not the fundamental way of describing quantum gravity (p. A 96).”

And in a review of matrix theory, Thomas Banks comments: String theorists have long fantasized about a beautiful new physical principle which will replace Einsteins marriage of Riemannian geometry and gravitation. Matrix theory most emphatically does not provide us with such a principle. Gravity and geometry emerge in a rather awkward fashion, if at all. Surely this is the major defect of the current formulation, and we need to make a further conceptual step in order to overcome it (pp. 181-182). It is my hope that emphasis on the importance of the principle of dynamic individuation of the fundamental entities, with its corollary requirement of invariance of the theory under the entire permutation group acting on these entities, constitutes a small contribution to the taking of that further conceptual step. >>

[19] Green(1999) Superstrings, M-theory and quantum gravity, Classical and Quantum Gravity, 16, A77-A100

Michael Green and Thomas Banks are major figures in string/M research---originators----and speak with authority. They may be wrong (they are the experts on string, not me, so I cannot judge if they are right or not) but in any case these strong words help give adequate emphasis to the issues of background independence and invariance under diffeomorphic mappings.
 
Last edited by a moderator:
  • #61
somebody might wish to ask how Loll-style "Triangulations" gravity implements background independence and diffeo invariance (or reasonable substitute, since it doesn't have any diffeos)
 
  • #62
Introduction to "Triangulations" quantum gravity

the triangulations QG approach of Loll and coworkers looks like the most interesting, and perhaps promising, development being pursued by the people participating in this years "Loops 05" conference.

It is one of the broadly defined "Loop-and-allied" approaches that Loop people do----not narrowly defined core LQG. there are a bunch of approaches that deal with similar stuff but differ in details.

this thread can serve a useful purpose as an INTRODUCTION to more than just one of the Loop-and-allied approaches. Probably the most timely to consider at the moment is CDT-style Triangulations.

As a point of departure here is how the abstract of a recent landmark CDT paper starts off:

"We provide detailed evidence for the claim that nonperturbative quantum gravity, defined through state sums of causal triangulated geometries, possesses a large-scale limit in which the dimension of spacetime is four and the dynamics of the volume of the universe behaves semiclassically. This is a first step in reconstructing the universe from a dynamical principle at the Planck scale,..."

this is from
http://arxiv.org/hep-th/0505154
Reconstructing the Universe
J. Ambjorn (NBI Copenhagen and U. Utrecht), J. Jurkiewicz (U. Krakow), R. Loll (U. Utrecht)
52 pages, 20 figures

here is a short reading list
https://www.physicsforums.com/showpost.php?p=585294&postcount=59

Now what I want to do is describe the CDT "Triangulations" method as simply as I can.
 
Last edited by a moderator:
  • #63
First there is a kind of birdseye view illustrated by a talk that Renate Loll gave in 2002 called
"Quantum gravity IS counting geometries"

It is a possible approach---sometimes called "state sum" or "path integral"
Its roots go back to the Feynman path integral for a particle where you add up all the possible (approximate piecewise straight) paths the particle might take to get from here to there----with complex weights to make a kind of weighted average. It is a way to get probability amplitudes and calculate things about the quantum path the particle takes. This turns out to be 100 percent of the time very ROUGH, nowhere is it even differentiable, but nevertheless intuitively it kind of blurs or fuzzes out to resemble a smooth continuously differentiable classical path that you might expect from freshman calculus.

In the "state sum" approach to a particle's path the calculation is in effect counting lots of different (piecewise straight) paths. And adding them all together, with a system of weighting that embodies the microscopic dynamics, to get answers. It has been a very successful method. the CDT authors found out how to apply it to spacetimes.

A spacetime is like a path, from space being this way to space having evolved to be that way, or more grandly from the beginning of a universe to its end. In QUANTUM gravity, that is in quantum spacetime dynamics, one is not certain exactly which path it took. One only has amplitudes of various ways of evolving from this shape to that shape. It is very much analogous to the particle path. You can even think of the universe wandering around in the space of all geometries and its evolution an actual path, but I can see no compelling reason to think so abstractly as that about it.

to put it simply, the CDT authors found a way to approximate (by piecewise flat geometries, made of flat Minkowski building blocks) all the possible spacetimes that get you from here to there, or from the beginning to the end. And they found a way to compute experimental answers from the STATE SUM of all these geometries.

After that it almost seems obvious and really straightforward. They can generate random spacetimes, random histories of the universe, as 4D worlds living in the computer memory, and they can HAVE LITTLE IMAGINARY MEN RUN AROUND IN THEM TO EXPLORE THEM, by taking random walks----a so called diffusion process---which is a way of finding out about the geometry, like what dimension it really is in there.

then after studying each random example they can add everything up with the usual weights (there are actually two sets of weights connected by the Wick trick, one set is simply real numbers like ordinary probabilities and the other set of weights is complex amplitude-type numbers but this doesn't matter to the overall picture) So they add everything up in a weighted average and get the state sum report (from the little imaginary men) on how it is in there.

Now having done this, Loll and co-workers are catching results like the fish are running. They are just pulling them in hand over hand. throw in the line and hook one every time.
This is a big change from the Nineties when many people worked the state sum triangulations approach but didnt catch anything edible. everything they got was the wrong dimension.

so this is part of an overview.

what I have to EXPLAIN is how they set up one of these layered triangulated geometries----and how they then shuffle the cards so as to get a series of random geometries. this is the nutsandbolts part.

a 4-simplex is the 4D analog of a triangle and they build these appoximate piecewise flat geometries out of two TYPES of 4-simples, the
"level"-kind and the "tilt"-kind

they call them the (4,1) kind and the (3,2) kind. it is how the vertices are destributed between two causal layers

I have to balance giving an overview with giving some introductory nutsandbolts.
 
  • #64
Here is some more overview. It elucidates the "state sum" idea of adding up all possible geometries. and the essential business of reducing the calculation to COUNTING.


In this case what we have is something from an American Physical Society publiscation

http://focus.aps.org/story/v14/st13

The American Physical Society sponsors the major peer-review journals series Phys. Rev. and Physical Review Letters. And they pick out articles for highlighting journalistically in the accompanying publication Physical Review Focus.

This is from Adrian Cho's Focus article on a paper by Loll and co-workers.

<<The researchers added up all the possible spacetimes to see if something like a large-scale four-dimensional spacetime would emerge from the sum. That was not guaranteed, even though the tiny bits of spacetime were four-dimensional. On larger scales the spacetime could curve in ways that would effectively change its dimension, just as a two-dimensional sheet of paper can be wadded into a three-dimensional ball or rolled into a nearly one-dimensional tube. This time the researchers found that they could achieve something that appeared to have one time dimension and three space dimensions--like the universe we know and love.

"It's exceedingly important" work, says Lee Smolin of the Perimeter Institute for Theoretical Physics in Waterloo, Canada. "Now at least we know one way to do this." Des Johnston of Heriot-Watt University in Edinburgh, Scotland, agrees the work is "very exciting" and says it underlines the importance of causality. "The other neat thing about this work is that you're essentially reducing general relativity to a counting problem," Johnston says. "It's a very minimalist approach to looking at gravity.">>
 
  • #65
...
...this is part of an overview.

what I have to explain is how they set up one of these layered triangulated geometries----and how they then shuffle the cards so as to get a series of random geometries. this is the nutsandbolts part.

a 4-simplex is the 4D analog of a triangle and they build these appoximate piecewise flat geometries out of two TYPES of 4-simples, the
"level"-kind and the "tilt"-kind

they call them the (4,1) kind and the (3,2) kind. it is how the vertices are destributed between two causal layers

I have to balance giving an overview with giving some introductory nutsandbolts.

the best source on the basics is http://arxiv.org/hep-th/0105267

we have to COUNT THE CAUSAL GEOMETRIES of spacetime, it sounds terribly hard but it isn't and they managed to program it, and it's the basic job we can't get around

causal means LAYERED, each model of spacetime gets laid down in sheets or slices, like a book with pages or a tree-trunk with rings, an event in one layer can only be caused by something from a deeper layer----or think of it like a many-storied building.

so we have to BUILD ALL POSSIBLE LAYERED SPACETIME GEOMETRIES in such a way that we can COUNT THEM or anyway explore to find what are the most numerous kind or the most likely kind, or somehow average them.

Maybe in the end we won't be able to count them exactly but we will have statistics and averages and random samples about them just as if we could actually count them. We will take the census of these layered spacetime geometries.

The technique will be to learn how to build layered geometries using "triangular" building blocks cut out of the txyz space of special relativity

these blocks will all be the same size----their spatial edges will be a fixed length 'a' that we will successively make smaller and smaller---and they will be of two kinds. the LEVEL kind and the TILT kind. The level kind is like a pyramid which has a 3D spatial tetrahedron as its base, on one floor of the building, and its 5-th vertex on the floor above or, the upsidedown version, the floor below.

the authors write the level kind as either (4,1) or (1,4), because it has 4 vertexes (the 4 vertices of the tetrahedron) on this floor and 1 vertex on the floor above, or viceversa one vertex on this floor and 4 on the floor above

intuitively one layer is all of 3D space, and the spacetime history of the universe is being built 3D layer by 3D layer, so it is like a book except the pages are 3D.

a LEVEL kind of building block has 4 timelike edges going from each of the four corners of its spatial tetrahedron base up to the vertex on the floor above, or else going down to the solitary vertex on the floor below. the other kind of buildingblock is like the LEVEL kind but tilted over so that now one of those timelike edges becomes a ridge and is entirely in the floor above, and instead of sitting on a full tetrahedron base it is now only sitting on a triangle side of it.

the authors write the TILT kind of buildingblock as either (3,2) or (2,3)
because it has 3 vertices on one floor, that make its spatial triangle base, and it has 2 vertices on the floor above or below, that make this ridge I mentioned. Like the ridge of a roof or the keel of a boat, depending it is up or down.

the TILT kind has 4 spacelike edges (three for the triangle and one for the base) and it has 6 timelike edges, whereas the LEVEL kind had 6 spacelike (that you need to make a tetrahedron) and 4 timelike.

the quickest way to understand this business is to follow through the analogous 3D case which is spelled out in
http://arxiv.org/hep-th/0105267

there, the building blocks are tetrahedrons---spatial layers are intuitively 2D, like the pages of a book---everything is easy to imagine, and they have a lot of drawings

but I am trying to discuss modeling 4D spacetime geometry without first going thru the 3D case.
 
Last edited by a moderator:
  • #66
after that it is not too hard to say, in general terms, how the method works

you want to get the EFFECT of building a layered geometry with, say, a halfmillion identical LEVEL kind buildingblocks
and howevermany you need (which will also be about halfmillion) TILT kind blocks to fill in.

because when you try to build layers with the LEVEL kind it always turns out that you get gaps which are just right to fit the other kind into, so it turns out that to build up layer by layer you need approximately the same number of the other kind.

NOW YOU DONT ACTUALLY BUILD EVERY POSSIBLE LAYERED SPACETIME GEOMETRY with these million virtually identical blocks

it is like taking an opinion poll where you don't talk to everybody, you take a random sample.

you want the EFFECT of having built all of them, and studied each one, and counted and made statistics about how they all are. you don't want to actually do it. you want the effect as if you did it.

this is where "shuffling the deck of cards" comes in. the CDT authors call it "thermalizing" the geometry. you set up a very simple plain geometry to start with, in computer memory, and then you do RANDOMIZING PASSES thru it, until it gradually becomes totally unrecognizable.


like, have a look at Figures 4,5 and 6 in "Reconstructing the Universe"
http://arxiv.org/hep-th/0505154
they are all three quite different-looking but they all come from starting with a simple initial geometry and doing randomizing passes.

the authors call each pass "making a sweep", and each sweep involves doing a million or so "Monte Carlo moves" which are individual shuffles that change some of the building blocks around.

they use a lot of computer time. thermalizing (thoroughly randomizing) a geometry can take a week on a workstation. then you study it and measure things

when you have a random geometry you can run random walks in it, or diffusion process, and you can measure distances and volumes and see how they relate...
 
Last edited by a moderator:
  • #67
here is an example.
say you have built a layered spacetime geometry in your computer and you pick just one spatial layer and you want to explore that by a random walk.

well a spacelike slice is just made of the tetrahedrons which were the bases of the LEVEL buildingblocks!
So you have a set of things in your computer which are little 4-face pyramids (equilateral triangle bottom, so 3 side faces and a bottom face, except no way to distinguish the bottom from the other faces). And this bunch of tetrahedrons are fitted together some way so that every face of one is up against the face of some other.

So you can pick a random block to start in, and then TOSS A FOURSIDED COIN to select which face to go out of

and when you go out one face you are now in a new tetrahedron and you can toss the foursided coin to choose which face to pass thru, and again and again.

it will be a clue to the actual dimensionality of the spatial slice to see if you get completely lost by doing this random walk, or if you now and then get back home to where you started. the authors determine the probabilities EMPIRICALLY by actually running the random walks in the computer, and this tells them about the dimensionality of the spatial slices

the nice thing is the answers gotten this way are weird and quite Alice-in-Wonderland. at microscopic level the continuum (as pictured by CDT) is a non-classical, unexpected world which Lewis Carroll would have loved.
 
Last edited:
  • #68
the Einstein Hilbert action

in the path integral or state sum approach you make a WEIGHTED sum using a badness or handicap function S(path) which is large for really kooky unphysical paths

this is how you introduce the classical dynamics into the quantum picture.

Feynmann has an essay about the "least action principle" in his Lectures, it is one of the core things in the Feynman physics textbook. It is terribly important and he suddenly takes a serious tone of voice when he gets to it.

the way you do classical physics is you consider all the possible paths and you DIFFERENTIATE S(path) and set to zero so as pick the one and only one path that MINIMIZES S(path)-------you pick the unique path that minimizes the "action", which is a word for badness or silliness of paths.

we have inherited our "action" functions from great old classical guys like Lagrange and Einstein and Hamilton, directly or indirectly, whom we revere, and they have the feature that minimizing them gets you the expected classical equations of motion

the way you do quantum physics is when you consider all possible paths you don't try to pick the unique winner, you ADD THEM ALL TOGETHER, but you don't do this in a completely indiscriminate way! You handicap each one by putting a little real or complex number "weight" on it. this is incidentally how they used to handicap racehorses, with a little weight, but you can do it at a different level with the betting odds too.

one weight you might consider putting on is exp( - S(path))
that is:
"eee to the minus badness"

e-badness

so if the badness is large it make the weight exponentially very small and then the path tagged with that weight will not count for very much in the sum or weighted average of all paths

YOU DO NOT JUST PICK ONE HORSE THAT IS YOUR FAVORITE, you add together all the horses, but you weight each one so your favorites count for more and the bad ones count for less-------you get a "composite" horse.

another kind of weight you might consider putting on is exp(iS(path))

"eee to the eye badness"

ei badness

As you may know from elementary complex numbers "eee to the eye theta" is complex numbers going around and around the unit radius circle.
and this is very clever because if you go around the circle, around zero, very fast it will average out to ZERO ITSELF by simple vector addition.

taking a step N and S and E and W adds up to going nowhere

so if you are averaging things with rapidly increasing badness and tagging them with "eee to the eye badness" numbers then these things with lots of badness will CANCEL EACH OTHER OUT in the sum and not have much influence on the sum

this is the two kinds of handicaps, the real number weights and the complex number weights from around the unit radius circle.

YOU CAN GET FROM ONE SET OF WEIGHTS TO THE OTHER SET OF WEIGHTS by the simple expedient of changing the eye into a minus sign, or the minus into an eye. this is called the WICK ROTATION, in honor of Joe Wick born in Torino around 1906.
 
  • #69
the Einstein Hilbert action and the Wick rotation

the Einstein action measures roughly speaking how much some spacetime is off from being a well behaved classical solution of the classical Einstein equation of General Relativity. So if you minimize the einstein action it you get the classical equation back.
so it measures how much the "path" is screwing up and getting distracted from its studies and cutting classes and taking dope and all what it isn't supposed to be doing---how "busy" it is with messing up---that "busy-ness" is the action. believe me you want to cut down on it.

a spacetime is just a path from the beginning of the world to the end, a path in "geometry space" if you can picture the space of all geometries.

the quantum idea is the universe doesn't just follow one path, it is a fuzzy mixture, well that is a rather distracting idea so let's not get into that.

the extremely beautiful thing is that with simplex geometries, with geometries built of triangles, YOU CAN IMPLEMENT THE ACTION FUNCTION JUST BY COUNTING DIFFERENT KINDS OF TRIANGLES

so even a computer, merely able to count up things in its memory, can do it

so we get back to our story where Loll and friends are running a computer model of spacetime, and the model is doing "sweeps" consisting of a million or so "Monte Carlo moves" which are localized elementary rearrangements of the simplex building blocks

each time they roll the dice and pick a monty move at random, they calculated some "badness" or "action" numbers to see whether to ACCEPT OR REJECT the proposed move!

this is how the localized microscopic "DYNAMICAL PRINCIPLE" that Loll talks about enters into the picture

it is this action principle operating at a microscopic Planckian or even maybe sub-Planckian level that the overall spacetime grows from. However it looks, whether it has 4 dimensions or 3 dimensions or some fraction etc, whatever its geometry, it grows out of many many local applications of the action principle at micro-scale.

I will try to find a quote.
 
  • #70
Yeah, this is going to seem very dry and overdetailed but it shows how the quantum spacetime dynamics, the path integral action principle, was implemented:

<<3. Numerical implementation

We have investigated the infinite-volume limit of the ensemble of causal triangulated four-dimensional geometries with the help of Monte Carlo simulations at finite four-volumes N4 = N(4,1) + N(3,2) of up to 362,000 four-simplices. A simplicial geometry is stored in the computer as a set of lists, where the lists consist of dynamic sequences of labels for simplices of dimension n from zero to four, together with their position and orientation with respect to the time direction. Additional list data include information about nearest neighbours, i.e. how the triangulation “hangs together”, and other discrete data (for example, how many four-simplices meet at a given edge) which help improve the acceptance rate of Monte Carlo moves. The simulation is set up to generate a random walk in the ensemble of causal geometries of a fixed time extension t. The local updating algorithm consists of a set of moves that change the geometry of the simplicial manifold locally, without altering its topological properties. These can be understood as a Lorentzian variant of (a simplified version of) the so-called Alexander moves [21, 22, 23], in the sense that they are compatible with the discrete time slicing of our causal geometries. For example, the subdivision of a four-simplex into five four-simplices by placing a new vertex at its centre is not allowed, because vertices can only be located at integer times tau . Details of the local moves can be found in [8]. As usual, each suggested local change of triangulation is accepted or rejected according to certain probabilities depending on the change in the action and the local geometry. (Note that a move will always be rejected if the resulting triangulation violates the simplicial manifold property.) The moves are called in random order, with probabilities chosen in such a way as to ensure that the numbers of actually performed moves of each type are approximately equal. We attained a rather high average acceptance rate of about 12.5%, which was made possible by keeping ...>>

By the way Alexander wrote his book in 1930. that is how far these "Monty Carlo moves" go back. they are just modified Alexander moves. Pachner is also cited. So this his how they "shuffle the deck".

Now I can say what part the Wick rotation plays. The Wick rotation changes the complex weights into real weights which can be dealt with as PROBABILITIES in this process of choosing the next random move, in "shuffling the deck" or randomizing spacetime geometry by Monte Carlo moves.

the probabilities enter each time you do a local rearrangement of some building blocks, you check whether that local microscopic rearrangement would be favored or disfavored by the Einstein equation. you do that by comparing badness. and it is still random----there is still always a chance that you can do a move that increases the badness, that happens lots in fact---but the probabilities are weighted against it (the House of general relativity wins over the long run). well maybe that is too impressionistic an impression.

I promised in Quantum Graffiti thread to say something about Wick rotation and Einstein Hilbert action
 
Last edited:

Similar threads

Replies
2
Views
735
Replies
7
Views
2K
Replies
15
Views
2K
Replies
13
Views
3K
Replies
13
Views
1K
Replies
3
Views
2K
Back
Top