What’s up with arXiv?

First of all, I wanted to take a moment to say that this is the one-year anniversary of this blog. I’ve been posting every week, (almost always) on Friday, since I first was motivated to start blogging back in November 2012. It’s been a fun ride, through ups and downs, Ars Technica and Amplituhedra, and I hope it’s been fun for you, the reader, as well!

I’ve been giving links to arXiv since my very first post, but I haven’t gone into detail about what arXiv is. Since arXiv is a rather unique phenomenon, it could use a more full description.

arXivpic

The word arXiv is pronounced much like the normal word archive, just think of the capital X like a Greek letter Chi.

Much as the name would suggest, arXiv is an archive, specifically a preprint archive. A pre-print is in a sense a paper before it becomes a paper; more accurately, it is a scientific paper that has not yet been published in a journal. In the past, such preprints would be kept by individual universities, or passed between interested individuals. Now arXiv, for an increasing range of fields (first physics and mathematics, now also computer science, quantitative biology, quantitative finance, and statistics) puts all of the preprints in one easily accessible, free to access place.

Different fields have different conventions when it comes to using arXiv. As a theoretical physicist, I can only really speak to how we use the system.

When theoretical physicists write a paper, it is often not immediately clear which journal we should submit it to. Different journals have different standards, and a paper that will gather more interest can be published in a more prestigious journal. In order to gauge how much interest a paper will raise, most theoretical physicists will put their papers up on arXiv as preprints first, letting them sit there for a few months to drum up attention and get feedback before formally submitting the paper to a journal.

The arXiv isn’t just for preprints, though. Once a paper is published in a journal, a copy of the paper remains on arXiv. Often, the copy on arXiv will be updated when the paper is updated, changed to the journal’s preferred format and labeled with the correct journal reference. So arXiv, ultimately, contains almost all of the papers published in theoretical physics in the last decade or two, all free to read.

But it’s not just papers! The digital format of arXiv makes it much easier to post other files alongside a paper, so that many people upload not just their results, but the computer code they used to generate them, or their raw data in long files. You can also post papers too long or unwieldy to publish in a journal, making arXiv an excellent dropping-off point for information in whatever format you think is best.

We stand at the edge of a new age of freely accessible science. As more and more disciplines start to use arXiv and similar services, we’ll have more flexibility to get more information to more people, while still keeping the advantage of peer review for publication in actual journals. It’s going to be very interesting to see where things go from here.

Braaains…Boltzmann Braaaains…

In honor of Halloween yesterday, let me tell you a scary physics story:

Sarah was an ordinary college student, in an ordinary dorm room, ordinary bean bag chairs strewn around an ordinary bed with ordinary pink sheets. If she concentrated, she could imagine her ordinary parents back home in ordinary Minnesota. In her ordinary physics textbook on her ordinary desk, ordinary laws of physics were written, described as the result of centuries of experimentation.

Unbeknownst to Sarah, the universe was much more chaotic and random than she realized, and also much more vast. Arbitrary collections of matter formed and dissipated, and over the universe’s long history, any imaginable combination might come to be.

Combinations like Sarah.

You see, Sarah too was a random combination, a chance arrangement of particles formed only a bare few moments ago. In truth, she had no ordinary parents, nor was she surrounded by an ordinary college, and the laws of physics that her textbook asserted were discovered through centuries of experimentation were just a moment’s distribution of ink on a page.

And as she got up to open the door into the vast dark of the outside, her world dissipated, and she ceased to exist.

That’s the life of a Boltzmann Brain. If a universe is random and old enough, it is inevitable that such minds exist. They might have memories of an extended, orderly world, but these would just be illusions, chance arrangements of their momentary neurons. What’s more, they may think they know the laws of physics through careful experiment and reasoning, but such knowledge would be illusory as well. And most frightening of all, if the universe is truly ancient and unimaginably vast, there would be many orders of magnitude more Boltzmann Brains than real humans…so many, that it would almost certainly be the case that you are in fact a Boltzmann Brain right now!

This is legitimately worrying to some physicists. The situation gets a bit more interesting when you remember that, as a Boltzmann Brain, anything you know about physics may well be a lie, since the history of research you think exists might not have. The problem is, if you manage to prove that you are probably a Boltzmann Brain, you had to use physics to do it. But your physics is probably wrong!

This, as Sean Carroll argues is why the concept of a Boltzmann Brain is self-defeating. It is, in a way, a logical impossibility. And if a universe of Boltzmann Brains is logically impossible, then any physics that makes Boltzmann Brains more likely than normal humans must similarly be wrong. That’s Carroll’s argument, one that he uses to argue for specific physical conclusions about the real world, namely a proposal about the properties of the Higgs boson.

It might seem philosophically illegitimate to use such a paradox to argue about the real world. However, philosophers have a similar argument when it comes to such “reality is a lie” scenarios. In general, modern philosophers point out that any argument that proves that all of our knowledge is false or meaningless by necessity also proves itself false or meaningless. This is what allows analytical philosophy to carry forward and make progress, even if it can’t reject the idea that reality is an illusion by more objective means.

With that said, there seems to be a difference between simply rejecting arguments that “show” that the world is an illusion or that we are all Boltzmann Brains, and using those arguments to draw conclusions about other parts of the world. I would be curious if there are similar arguments to Carroll’s in philosophy, arguments that draw conclusions more specific than “we exist and can know things”. Any philosopher readers should feel welcome to chime in in the comments!

And for the rest of you, you probably aren’t a Boltzmann Brain. But if the outside world looks a little too dark tonight…

Visiting Brandeis

I gave a talk at Brandeis’s High Energy Theory Seminar this week. Brandeis is much easier to park at than Brown, but it’s proportionately easier to get lost in. While getting lost, I happened to run into this:

Many campuses have buildings that look like castles. Usen Castle is the only one I’ve seen that officially calls itself a castle. It’s a dorm, and the students there can honestly say that they live in a castle.

If I were a responsible, mature blogger, I’d tell you about the history of the place. I’d tell you who built it, and why they felt it was appropriate to make a literal castle the center of their college.

As I’m not a responsible, mature blogger, I’ll just leave you with this thought: they have a castle!

Blackboards, Again

Recently I had the opportunity to give a blackboard talk. I’ve talked before about the value of blackboards, how they facilitate collaboration and can even be used to get work done. What I didn’t feel the need to explain was their advantages when giving a talk.

No, the blackboard behind me isn't my talk.

No, the blackboard behind me isn’t my talk.

When I mentioned I was giving a blackboard talk, some of my friends in other fields were incredulous.

“Why aren’t you using PowerPoint? Do you people hate technology?”

So why do theorists (and mathematicians) do blackboard talks, when many other fields don’t?

Typically, a chemist can’t bring chemicals to a talk. A biologist can’t bring a tank of fruit flies or zebrafish, and a psychologist probably shouldn’t bring in a passel of college student test subjects. As a theorist though, our test subjects are equations, and we can absolutely bring them into the room.

In the most ideal case, a talk by a theorist walks you through their calculation, reproducing it on the blackboard in enough detail that you can not only follow along, but potentially do the calculation yourself. While it’s possible to set up a calculation step by step in PowerPoint, you don’t have the same flexibility to erase and add to your equations, which becomes especially important if you need to clarify a point in response to a question.

Blackboards also often give you more space than a single slide. While your audience still only pays attention to a slide-sized area of the board at one time, you can put equations up in one area, move away, and then come back to them later. If you leave important equations up, people can remind themselves of them on their own time, without having to hold everybody up while you scroll back through the slides to the one they want to see.

Using a blackboard well is a fine art, and one I’m only beginning to learn. You have to know what to erase and what to leave up, when to pause to allow time to write or ask questions, and what to say while you’re erasing the board. You need to use all the quirks of the medium to your advantage, to show people not just what you did, but how and why you did it.

That’s why we use blackboards. And if you ask why we can’t do the same things with whiteboards, it’s because whiteboards are terrible. Everybody knows that.

What are Vacua? (A Point about the String Landscape)

A couple weeks back, there was a bit of a scuffle between Matt Strassler and Peter Woit on the subject of predictions in string theory (or more properly, the question of whether any predictions can be made at all). As a result, Strassler has begun a series on the subject of quantum field theory, string theory, and predictions.

Strassler hasn’t gotten to the topic of string vacua yet, but he’s probably going to cover the subject in a future post. While his take on the subject is likely to be more expansive and precise than mine, I think my perspective on the problem might still be of interest.

Let’s start with the basics: one of the problems often cited with string theory is the landscape problem, the idea that string theory has a metaphorical landscape of around 10^500 vacua.

What are vacua?

Vacua is the plural of vacuum.

Ok, and?

A vacuum is empty space.

That’s what you thought, right? That’s the normal meaning of vacuum. But if a vacuum is empty, how can there be more than one of them, let alone 10^500?

“Empty” is subjective.

Now we’re getting somewhere. The problem with defining a concept like “empty space” in string theory or field theory is that it’s unclear what precisely it should be empty of. Naively, such a space should be empty of “stuff”, or “matter”, but our naive notions of “matter” don’t apply to field theory or string theory. In fact, there is plenty of “stuff” that can be present in “empty” space.

Think about two pieces of construction paper. One is white, the other is yellow. Which is empty? Neither has anything drawn on it, so while one has a color and the other does not, both are empty.

“Empty space” doesn’t come in multiple colors like construction paper, but there are equivalent parameters that can vary. In quantum field theory, one option is for scalar fields to take different values. In string theory, different dimensions can be curled up in different ways (as an aside, when string theory leads to a quantum field theory often these different curling-up shapes correspond to different values for scalar fields, so the two ideas are related).

So if space can have “stuff” in it and still count as empty, are there any limits on what can be in it?

As it turns out, there is a quite straightforward limit. But to explain it, I need to talk a bit about why physicists care about vacua in the first place.

Why do physicists care about vacua?

In physics, there is a standard modus operandi for solving problems. If you’ve taken even a high school physics course, you’ve probably encountered it in some form. It’s not the only way to solve problems, but it’s one of the easiest. The idea, broadly, is the following:

First get the initial conditions, and then use the laws of physics to see what happens next.

In high school physics, this is how almost every problem works: your teacher tells you what the situation is, and you use what you know to figure out what happens next.

In quantum field theory, things are a bit more subtle, but there is a strong resemblance. You start with a default state, and then find the perturbations, or small changes, around that state.

In high school, your teacher told you what the initial conditions were. In quantum field theory, you need another source for the “default state”. Sometimes, you get that from observations of the real world. Sometimes, though, you want to make a prediction that goes beyond what your observations tell you. In that case, one trick often proves useful:

To find the default state, find which state is stable.

If your system starts out in a state that is unstable, it will change. It will keep changing until eventually it changes into a stable state, where it will stop changing. So if you’re looking for a default state, that state should be one in which the system is stable, where it won’t change.

(I’m oversimplifying things a bit here to make them easier to understand. In particular, I’m making it sound like these things change over time, which is a bit of a tricky subject when talking about different “default” states for the whole of space and time. There’s also a cool story connected to this about why tachyons don’t exist, which I’d love to go into for another post.)

Since we know that the “default” state has to be stable, if there is only one stable state, we’ve found the default!

Because of this, we can lay down a somewhat better definition:

A vacuum is a stable state.

There’s more to the definition than this, but this should be enough to give you the feel for what’s going on. If we want to know the “default” state of the world, the state which everything else is just a small perturbation on top of, we need to find a vacuum. If there is only one plausible vacuum, then our work is done.

When there are many plausible vacua, though, we have a problem. When there are 10^500 vacua, we have a huge problem.

That, in essence, is why many people despair of string theory ever making any testable predictions. String theory has around 10^500 plausible vacua (for a given, technical, meaning of plausible).

It’s important to remember a few things here.

First, the reason we care about vacuum states is because we want a “default” to make predictions around. That is, in a sense, a technical problem, in that it is an artifact of our method. It’s a result of the fact that we are choosing a default state and perturbing around it, rather than proving things that don’t depend on our choice of default state. That said, this isn’t as useful an insight as it might appear, and as it turns out there is generally very little that can be predicted without choosing a vacuum.

Second, the reason that the large number of vacua is a problem is that if there was only one vacuum, we would know which state was the default state for our world. Instead, we need some other method to pick, out of the many possible vacua, which one to use to make predictions. That is, in a sense, a philosophical problem, in that it asks what seems ostensibly to be a philosophical question: what is the basic, default state of the universe?

This happens to be a slightly more useful insight than the first one, and it leads to a number of different approaches. The most intuitive solution is to just shrug and say that we will see which vacuum we’re in by observing the world around us. That’s a little glib, since many different vacua could lead to very similar observations. A better tactic might be to try to make predictions on general grounds by trying to see what the world we can already observe implies about which vacua are possible, but this is also quite controversial. And there are some people who try another approach, attempting to pick a vacuum not based on observations, but rather on statistics, choosing a vacuum that appears to be “typical” in some sense, or that satisfies anthropic constraints. All of these, again, are controversial, and I make no commentary here about which approaches are viable and which aren’t. It’s a complicated situation and there are a fair number of people working on it. Perhaps, in the end, string theory will be ruled un-testable. Perhaps the relevant solution is right under peoples’ noses. We just don’t know.

Brown, Blue, and Birds

I gave a talk at Brown this week, so this post may be shorter than usual. On the topic of Brown I don’t have much original to say: the people were friendly, the buildings were brownish-colored, and bringing a car there was definitely a bad idea. Don’t park at Brown. Not even then.

There’s a quote from Werner Heisenberg that has been making the rounds of the internet. It comes out of a 1976 article by Felix Bloch where he describes taking a walk with Heisenberg, when the discussion turned to the subject of space and time:

I had just read Weyl’s book Space, Time and Matter, and under its influence was proud to declare that space was simply the field of linear operations.

“Nonsense,” said Heisenberg, “space is blue and birds fly through it.”

Heisenberg’s point is that sometimes in physics you need to ask what your abstractions are really describing. You need to make sure that you haven’t stretched your definitions too badly away from their original inspiration.

When people first hear that string theory requires eleven dimensions, many wonder if this point applies. In mathematics, it’s well known that a problem can be described in many dimensions more than the physical dimensions of space. There’s a lovely example in the book Flatterland (a sequel to Flatland, a book which any math-y person should read at least once) of the dimensions of a bike. The bike’s motion through space gives three dimensions: up/down, backward/forward, and left/right. However, the bike can move in other ways: its gears can each be in a different position, as can its handlebars, as can the wheels…in the end, a bike can be envisioned as having many more “dimensions” than our normal three-dimensional space, each corresponding to some internal position.

Is string theory like this? No.

The first hint of the answer comes from something called F theory. String theory is part of something larger called M theory, and since M theory has eleven dimensions this is usually the number of dimensions given. But F theory contains string theory in a certain sense as well, only F theory contains twelve dimensions.

So why don’t string theorists say that the world has twelve dimensions?

As it turns out, the extra dimension added by F theory isn’t “really” a dimension. It’s much more like the mathematical dimensions of a bike’s gears and wheels than it is like the other eleven dimensions of M theory.

What’s the difference? What, according to a string theorist, is the definition of a dimension of space?

It’s simple: Space is “blue” (or colorless, I suppose). Birds (and particles, and strings, and membranes) fly in it.

We’re using the same age-old distinction that Heisenberg was, in a way. What is space? Space is just a place where things can move, in the same way they move in our usual three dimensions. Space is where you have momentum, where that momentum can change your position. Space is where forces act, the set of directions in which something can be pulled or pushed in a symmetric way. Space can’t be reduced, at least not without a lot of tricks: a bird flying isn’t just another description of a lizard crawling, not in the way a bicycle’s gears moving can be thought of as turning through our normal three dimensions without any extra ones. And while F theory doesn’t fit this criterion, M theory really does. The membranes of M theory fly around in eleven dimensional space-time, just like a bird moves through three space and one time dimensions.

Space for a string theorist isn’t any crazier or more abstract than it is for you. It’s just a place where things can move.

Planar vs. Non-Planar: A Colorful Story

Last week, I used two terms, planar theory and non-planar theory, without defining them. This week, I’m going to explain what they mean, and why they’re important.

Suppose you’re working with a Yang-Mills theory (not necessarily N=4 super Yang-Mills. To show you the difference between planar and non-planar, I’ll draw some two-loop Feynman diagrams for a process where two particles go in and two particles come out:

planarity1

The diagram on your left is planar, while the diagram on your right is non-planar. The diagram on the left can be written entirely on a flat page (or screen), with no tricks. By contrast, with the diagram on the right I have to cheat and make one of the particle lines jump over another one (that’s what the arrow is meant to show). Try as you might, you can’t twist that diagram so that it lies flat on a plane (at least not while keeping the same particles going in and out). That’s the difference between planar and non-planar.

Now, what does it mean for a theory to be planar or non-planar?

Let’s review some facts about Yang-Mills theories. (For a more detailed explanation, see here). In Yang-Mills there are a certain number of colors, where each one works a bit like a different kind of electric charge. The strong force, the force that holds protons and neutrons together, has three colors, usually referred to as red, blue, and green (this is of course just jargon, not the literal color of the particles).

Forces give rise to particles. In the case of the strong force, those particles are called gluons. Each gluon has a color and an anti-color, where you can think of the color like a positive charge and the anti-color like a negative charge. A given gluon might be red-antiblue, or green-antired, or even red-antired.

While the strong force has three colors, for this article it will be convenient to pretend that there are four: red, green, blue, and yellow.

An important principle of Yang-Mills theories is that color must be conserved. Since anti-colors are like negative colors, they can cancel normal colors out. So if you’ve got a red-antiblue gluon that collides with a blue-antigreen gluon, the blue and antiblue can cancel each other out, and you can end up with, for example, red-antiyellow and yellow-antigreen instead.

Let’s consider that process in particular. There are lots of Feynman diagrams you can draw for it, let’s draw one of the simplest ones first:

planarity2

The diagram on the left just shows the process in terms of the particles involved: two gluons go in, two come out.

The other diagram takes into account conservation of colors. The red from the red-antiblue gluon becomes the red in the red-antiyellow gluon on the other side. The antiblue instead goes down and meets the blue from the blue-antigreen gluon, and both vanish in the middle, cancelling each other out. It’s as if the blue color entered the diagram, then turned around backwards and left it again. (If you’ve ever heard someone make the crazy-sounding claim that antimatter is normal matter going backwards in time, this is roughly what they mean.)

From this diagram, we can start observing a general principle: to make sure that color is conserved, each line must have only one color.

Now let’s try to apply this principle to the two-loop diagrams from the beginning of the article. If you draw double lines like we did in the last example, fill in the colors, and work things out, this is what you get:

planarity3

What’s going on here?

In the diagram on the left, you see the same lines as the earlier diagram on the outside. On the inside, though, I’ve drawn two loops of color, purple and pink.

I drew the lines that way because, just based on the external lines, you don’t know what color they should be. They could be red, or yellow, or green, or blue. Nothing tells you which one is right, so all of them are possible.

Remember that for Feynman diagrams, we need to add up every diagram we can draw to get the final result. That means that there are actually four times four or sixteen copies of this diagram, each one with different colors in the loops.

Now let’s look at the other diagram. Like the first one, it’s a diagram with two loops. However, in this case, the inside of both loops is blue. If you like, you can try to trace out the lines in the loops. You’ll find that they’re all connected together. Because this diagram is non-planar, color conservation fixes the color in the loops.

So while there are sixteen copies of the first diagram, there is only one possible version of the second one. Since you add all the diagrams together, that means that the first diagram is sixteen times more important than the second diagram.

Now suppose we had more than four colors. Lots more.

More than that…

With ten colors, the planar diagrams are a hundred times more important. With a hundred colors, they are ten thousand times more important. Keep increasing the number of colors, and it gets to the point where you can honestly say that the non-planar diagrams don’t matter at all.

What, then, is a “planar theory”?

A planar theory is a theory with a very large (infinite) number of colors.

In a planar theory, you can ignore the non-planar diagrams and focus only on the planar ones.

Nima Arkani-Hamed’s Amplituhedron method applies to the planar version of N=4 super Yang-Mills. There is a lot of progress on the planar version of the theory, and it is because the restriction to planar diagrams makes things simpler.

However, sometimes you need to go beyond planar diagrams. There are relationships between planar and non-planar diagrams, based on the ways that you can pair different colors together in the theory. Fully understanding this relationship is powerful for understanding Yang-Mills theory, but, as it turns out, it’s also the key to relating Yang-Mills theory to gravity! But that’s a story for another post.

The Amplituhedron and Other Excellently Silly Words

Nima Arkani-Hamed recently gave a talk at the Simons Center on the topic of what he and Jaroslav Trnka are calling the Amplituhedron.

There’s an article on it in Quanta Magazine. The article starts out a bit hype-y for my taste (too much language of importance, essentially), but it has several very solid descriptions of the history of the situation. I particularly like how the author concisely describes the Feynman diagram picture in the space of a single paragraph, and I would recommend reading that part even if you don’t have time to read the whole article. In general it’s worth it to get a picture of what’s going on.

That said, I obviously think I can clear a few things up, otherwise I wouldn’t be writing about it, so here I go!

“The” Amplituhedron

Nima’s new construction, the Amplituhedron, encodes amplitudes (building blocks of probabilities in particle physics) in N=4 super Yang-Mills as the “area” of a multi-dimensional analog of a polyhedron (hence, Amplitu-hedron).

Now, I’m a big supporter of silly-sounding words with amplitu- at the beginning (amplitudeologist, anyone?), and this is no exception. Anyway, the word Amplitu-hedron isn’t what’s confusing people. What’s confusing people is the word the.

When the Quanta article says that Nima has found “the” Amplituhedron, it makes it sound like he has discovered one central formula that somehow contains the whole universe. If you read the comments, many readers went away with that impression.

In case you needed me to say it, that’s not what is going on. The problem is in the use of the word “the”.

Suppose it was 1886, and I told you that a fellow named Carl Benz had invented “the Automobile”, a marvelous machine that can get everyone to work on time (as well as become the dominant form of life on Long Island).

My use of “the” might make you imagine that Benz invented some single, giant machine that would roam across the country, picking people up and somehow transporting everyone to work. You’d be skeptical of this, of course, expecting that long queues to use this gigantic, wondrous machine would swiftly ruin any speed advantage it might possess…

The Automobile, here to take you to work.

Or, you could view “the” in another light, as indicating a type of thing.

Much like “the Automobile” is a concept, manifested in many different cars and trucks across the country, “the Amplituhedron” is a concept, manifested in many different amplituhedra, each corresponding to a particular calculation that we might attempt.

Advantages…

Each amplituhedron has to do with an amplitude involving a specific number of particles, with a particular number of internal loops. (The Quanta article has a pretty good explanation of loops, here’s mine if you’d rather read that). Based on the problem you’re trying to solve, there are a set of rules that you use to construct the particular amplituhedron you need. The “area” of this amplituhedron (in quotation marks because I mean the area in an abstract, mathematical sense) is the amplitude for the process, which lets you calculate the probability that whatever particle physics situation you’re describing will happen.

Now, we already have many methods to calculate these probabilities. The amplituhedron’s advantage is that it makes these calculations much simpler. What was once quite a laborious and complicated four-loop calculation, Nima claims can be done by hand using amplituhedra. I didn’t get a chance to ask whether the same efficiency improvement holds true at six loops, but Nima’s description made it sound like it would at least speed things up.

[Edit: Some of my fellow amplitudeologists have reminded me of two things. First, that paper I linked above paved the way to more modern methods for calculating these things, which also let you do the four-loop calculation by hand. (You need only six or so diagrams). Second, even back then the calculation wasn’t exactly “laborious”, there were some pretty slick tricks that sped things up. With that in mind, I’m not sure Nima’s method is faster per se. But it is a fast method that has the other advantages described below.]

The amplituhedron has another, more sociological advantage. By describing the amplitude in terms of a geometrical object rather than in terms of our usual terminology, we phrase things in a way that mathematicians are more likely to understand. By making things more accessible to mathematicians (and the more math-headed physicists), we invite them to help us solve our problems, so that together we can come up with more powerful methods of calculation.

Nima and the Quanta article both make a big deal about how the amplituhedron gets rid of the principles of locality and unitarity, two foundational principles of quantum field theory. I’m a bit more impressed by this than Woit is. The fine distinction that needs to be made here is that the amplituhedron isn’t simply “throwing out” locality and unitarity. Rather, it’s written in such a way that it doesn’t need locality and unitarity to function. In the end, the formulas it computes still obey both principles. Nima’s hope is that, now that we are able to write amplitudes without needing locality and unitarity, if we end up having to throw out either of those principles to make a new theory we will be able to do so. That’s legitimately quite a handy advantage to have, it just doesn’t mean that locality and unitarity must be thrown out right now.

…and Disadvantages

It’s important to remember that this whole story is limited to N=4 super Yang-Mills. Nima doesn’t know how to apply it to other theories, and nobody else seems to have any good ideas either. In addition, this only applies to the planar part of the theory. I’m not going to explain what that term means here; for now just be aware that while there are tricks that let you “square” a calculation in super Yang-Mills to get a similar calculation in quantum gravity, those tricks rely on having non-planar data, or information beyond the planar part of the theory. So at this point, this doesn’t give us any new hints about quantum gravity. It’s conceivable that physicists will find ways around both of these limits, but for now this result, though impressive, is quite limited.

Nima hasn’t found some sort of singular “jewel at the heart of physics”. Rather, he’s found a very slick, very elegant, quite efficient way to make calculations within one particular theory. This is profound, because it expresses things in terms that mathematicians can address, and because it shows that we can write down formulas without relying on what are traditionally some of the most fundamental principles of quantum field theory. Only time will tell whether Nima or others can generalize this picture, taking it beyond planar N=4 super Yang-Mills and into the tougher theories that still await this sort of understanding.

Hype versus Miscommunication, or the Language of Importance

A fellow amplitudes-person was complaining to me recently about the hype surrounding the debate regarding whether black holes have “firewalls”. New York Times coverage seems somewhat excessive for what is, in the end, a fairly technical debate, and its enthusiasm was (rightly?) mocked in several places.

There’s an attitude I often run into among other physicists. The idea is that when hype like this happens, it’s because senior physicists are, at worst, cynically manipulating the press to further their positions or, at best, so naïve that they really see what they’re working on as so important that it deserves hype-y coverage. Occasionally, the blame will instead be put on the journalists, with largely the same ascribed motivations: cynical need for more page views, or naïve acceptance of whatever story they’re handed.

In my opinion, what’s going on there is a bit deeper, and not so easily traceable to any particular person.

In the articles on the (2, 0) theory I put up in the last few weeks, I made some disparaging comments about the tone of this Scientific American blog post. After exchanging a few tweets with the author, I think I have a better idea of what went down.

The problem here is that when you ask a scientist about something they’re excited about, they’re going to tell you why they’re excited about it. That’s what happened here when Nima Arkani-Hamed was interviewed for the above article: he was asked about the (2, 0) theory, and he seems to have tried to convey his enthusiasm with a metaphor that explained how the situation felt to him.

The reason this went wrong and led to a title as off-base and hype-sounding as “the Ultimate Ultimate Theory of Physics” was that we (scientists and science journalists) are taught to express enthusiasm in the language of importance.

There has been an enormous resurgence in science communication in recent years, but it has come with a very us-vs.-them mentality. The prevailing attitude is that the public will only pay attention to a scientific development if they are told that it is important. As such, both scientists and journalists try to make whatever they’re trying to communicate sound central, either to daily life or to our understanding of the universe. When both sides of the conversation are operating under this attitude, it creates an echo chamber where a concept’s importance is blown up many times greater than it really deserves, without either side doing anything other than communicating science in the only way they know.

We all have to step back and realize that most of the time, science isn’t interesting because of its absolute “importance”. Rather, a puzzle is often interesting simply because it is a puzzle. That’s what’s going on with the (2, 0) theory, or with firewalls: they’re hard to figure out, and that’s why we care.

Being honest about this is not going to lose us public backing, or funding. It’s not just scientists who value interesting things because they are challenging. People choose the path of their lives not based on some absolute relevance to the universe at large, but because things make sense in context. You don’t fall in love because the target of your affections is the most perfect person in the universe, you fall in love because they’re someone who can constantly surprise you.

Scientists are in love with what they do. We need to make sure that that, and not some abstract sense of importance, is what we’re communicating. If we do that, if we calm down and make a bit more effort to be understood, maybe we can win back some of the trust that we’ve lost by appearing to promote Ultimate Ultimate Theories of Everything.

The (2, 0) Theory: What is it, though?

Part Three of a Series on the (2, 0) Theory

If you’ve been following this series, you know that the (2, 0) theory describes what it’s like to live on a five dimensional membrane in M theory. You know it’s got five scalar fields, and four chiral fermions (and hopefully you have a rough idea of what those things are). And if you’ve been reading for longer, you’ve probably heard me mention that a theory is essentially a list of quantum fields. So if I’m going to define the (2, 0) theory for you, I ought to, at the very least, list its quantum fields.

This is where things get tricky, and where unfortunately I will have to get a big vague. Some of the quantum fields in the (2, 0) theory are things I’ve talked about before: the five scalars and the four fermions. The remaining field, though, is different, and it’s the reason why the (2, 0) theory is so mysterious.

I’ll start by throwing around some terminology. Normally, I’d go back and explain it, but in this case there’s simply too much. My aim here is to give the specialists reading this enough to understand what I’m talking about. Then I’ll take a few paragraphs to talk about what the implications of all this jargon are for a general understanding.

The remaining field in the (2, 0) theory is a two-form, or an antisymmetric, two-index tensor, with a self-dual field strength. It comes from the gauge orientation zero modes of the M5-brane. It is not a Yang-Mills field. However, it is non-abelian, that is, it “interacts with itself” in a similar way to how a Yang-Mills field does.

While I can give examples of familiar Yang-Mills fields, fermions, and now with the Higgs even scalars, I can’t give you a similar example of a fundamental two-form field. That’s because in our four dimensional world, such a field doesn’t make sense. It only makes sense in six or more dimensions.

The problem with understanding this isn’t just a matter of not having examples in the real world, though. We can invent a wide variety of unobserved fields, and in general have no problem calculating their hypothetical properties. The problem is that, in the case of the two-form field of the (2, 0) theory, we don’t know how to properly do calculations about it.

There are a couple different ways to frame the issue. One is that, while we know roughly which fields should interact with which other fields, there isn’t a mathematically consistent way to write down how they do so. Any attempt results in a formula with some critical flaw that keeps it from being useful.

The other way to frame the problem is to point out that every Yang-Mills force has a number that determines how powerful it is, called the coupling constant. As I discuss here, it is the small size of the coupling constant that allows us to calculate only the simplest Feynman diagrams and still get somewhat accurate results.

The (2, 0) theory has no coupling constant. There is no parameter that, if it was small, would allow you to only look at some diagrams and not others. In the (2, 0) theory, every diagram is equally important.

When people say that the (2, 0) theory is “irreducibly quantum”, this is what they’re referring to: we can’t separate out the less-quantum (lower loops) bits from the more quantum (higher loops) bits. The theory simply is quantum, inherently and uniformly so.

This is what makes it so hard to understand, what makes it sexy and mysterious and Mara Jade-like. If we could understand it, the payoffs would be substantial: M theory has a similar problem, so a full understanding of the (2, 0) theory might pave the way to a full understanding of M theory, which, unlike the (2, 0) theory, really is supposed to be a theory of everything.

And there is progress…somewhat, anyway. Twisting one of the six dimensions of the (2, 0) theory around in a circle gives you N=4 super Yang-Mills in five dimensions, while another circle brings it down to four dimensions. Because super Yang-Mills is so well-understood, this gives us a tantalizing in-road to understanding the (2, 0) theory. I’ve worked a bit on this myself.

Perhaps a good way to summarize the situation would be to say that, while N=4 super Yang-Mills is interesting because of how much we know about it, the (2, 0) theory is interesting because, contrary to expectations, we can do something with it at all. Every time someone comes up with a novel method for understanding quantum field theories, you can rest assured that they will end up trying to apply it to the (2, 0) theory. One of them might even work.