Author Archives: 4gravitons

When to Look under the Bed

Last week, blogged about a rather interesting experiment, designed to test the quantum properties of gravity. Normally, quantum gravity is essentially unobservable: quantum effects are typically only relevant for very small systems, where gravity is extremely weak. However, there has been a lot of progress in putting larger and larger systems into interesting quantum states, and a team of experimentalists has recently proposed a setup. The experiment wouldn’t have enough detail to, for example, distinguish between rival models of quantum gravity, but it would provide evidence as to whether or not gravity is quantum at all.

Lubos Motl, meanwhile, argues that such an experiment is utterly pointless, because there is no possible way that gravity could not be quantum. I won’t blame you if you don’t read his argument since it’s written in his trademark…aggressive…style, but the gist is that it’s really hard to make sense of the idea that there are non-quantum things in an otherwise quantum world. It causes all sorts of issues with pretty much every interpretation of quantum mechanics, and throws the differences between those interpretations into particularly harsh and obvious light. From this perspective, checking to see if gravity might not actually be quantum (an idea called semi-classical gravity) is a bit like checking for a monster under the bed.

You might find semi-classical gravity!

In general, I share Motl’s reservations about semi-classical gravity. As I mentioned back when journalists were touting the BICEP2 results as evidence of quantum gravity, the idea that gravity could not be quantum doesn’t really make much sense. (Incidentally, Hossenfelder makes a similar point in her post.)

All that said, sometimes in science it’s absolutely worth looking under the bed.

Take another unlikely possibility, that of cell phone radiation causing cancer. Things that cause cancer do it by messing with the molecular bonds in DNA. In order to mess with molecular bonds, you need high-frequency light. That’s how UV light from the sun can cause skin cancer. Cell phones emit microwaves, which are very low-frequency light. It’s what allows them to be useful inside of buildings, where normal light wouldn’t reach. It also means it’s impossible for them to cause cancer.

Nevertheless, if nobody had ever studied whether cell phones cause cancer, it would probably be worth at least one study. If that study came back positive, it would say something interesting, either about the study’s design or about other possible causes of cancer. If negative, the topic could be put to bed more convincingly. As it happens, those studies have been done, and overall confirm the expectations we have from basic science.

Another important point here is that experimentalists and theorists have different priorities, due to their different specializations. Theorists are interested in confirmation for particular theories: they want not just an unknown particle, but a gluino, and not just a gluino, but the gluino predicted by their particular model of supersymmetry. By contrast, experimentalists typically aren’t very interested in proving or disproving one theory or another. Rather, they look for general signals that indicate broad classes of new physics. For example, experimentalists might use the LHC to look for a leptoquark, a particle that allows quarks and leptons to interact, without caring what theory might produce them. Experimentalists are also very interested in improving their techniques. Much like theorists, a lot of interesting work in the field involves pushing the current state-of-the-art as far as it will go.

So, when should we look under the bed?

Well, if nobody has ever looked under this particular bed before, and if seeing something strange under this bed would at least be informative, and if looking under the bed serves as a proving ground for the latest in bed-spelunking technology, then yes, we should absolutely look under this bed.

Just don’t expect to see any monsters.

Is Everything Really Astonishingly Simple?

Neil Turok gave a talk last week, entitled The Astonishing Simplicity of Everything. In it, he argued that our current understanding of physics is really quite astonishingly simple, and that recent discoveries seem to be confirming this simplicity.

For the right sort of person, this can be a very uplifting message. The audience was spellbound. But a few of my friends were pretty thoroughly annoyed, so I thought I’d dedicate a post to explaining why.

Neil’s talk built up to showing this graphic, one of the masterpieces of Perimeter’s publications department:

Looked at in this way, the laws of physics look astonishingly simple. One equation, a few terms, each handily labeled with a famous name of some (occasionally a little hazy) relevance to the symbol in question.

In a sense, the world really is that simple. There are only a few kinds of laws that govern the universe, and the concepts behind them are really, deep down, very simple concepts. Neil adroitly explained some of the concepts behind quantum mechanics in his talk (here represented by the Schrodinger, Feynman, and Planck parts of the equation), and I have a certain fondness for the Maxwell-Yang-Mills part. The other parts represent different kinds of particles, and different ways they can interact.

While there are only a few different kinds of laws, though, that doesn’t mean the existing laws are simple. That nice, elegant equation hides 25 arbitrary parameters, hidden in the Maxwell-Yang-Mills, Dirac, Kobayashi-Masakawa, and Higgs parts. It also omits the cosmological constant, which fuels the expansion of the universe. And there are problems if you try to claim that the gravity part, for example, is complete.

When Neil mentions recent discoveries, he’s referring to the LHC not seeing new supersymmetric particles, to telescopes not seeing any unusual features in the cosmic microwave background. The theories that were being tested, supersymmetry and inflation, are in many ways more complicated than the Standard Model, adding new parameters without getting rid of old ones. But I think it’s a mistake to say that if these theories are ruled out, the world is astonishingly simple. These theories are attempts to explain unlikely features of the old parameters, or unlikely features of the universe we observe. Without them, we’ve still got those unlikely, awkward, complicated bits.

Of course, Neil doesn’t think the Standard Model is all there is either, and while he’s not a fan of inflation, he does have proposals he’s worked on that explain the same observations, proposals that are also beyond the current picture. More broadly, he’s not suggesting here that the universe is just what we’ve figured out so far and no more. Rather, he’s suggesting that new proposals ought to build on the astonishing simplicity of the universe, instead of adding complexity, that we need to go back to the conceptual drawing board rather than correcting the universe with more gears and wheels.

On the one hand, that’s Perimeter’s mission statement in a nutshell. Perimeter’s independent nature means that folks here can focus on deeper conceptual modifications to the laws of physics, rather than playing with the sorts of gears and wheels that people already know how to work with.

On the other hand, a lack of new evidence doesn’t do anyone any favors. It doesn’t show the way for supersymmetry, but it doesn’t point to any of the “deep conceptual” approaches either. And so for some people, Neil’s glee at the lack of new evidence feels less like admiration for the simplicity of the cosmos and more like that one guy in a group project who sits back chuckling while everyone else fails. You can perhaps understand why some people felt resentful.

Hooray for Neutrinos!

Congratulations to Takaaki Kajita and Arthur McDonald, winners of this year’s Nobel Prize in Physics, as well as to the Super-Kamiokande and SNOLAB teams that made their work possible.

Congratulations!

Unlike last year’s Nobel, this is one I’ve been anticipating for quite some time. Kajita and McDonald discovered that neutrinos have mass, and that discovery remains our best hint that there is something out there beyond the Standard Model.

But I’m getting a bit ahead of myself.

Neutrinos are the lightest of the fundamental particles, and for a long time they were thought to be completely massless. Their name means “little neutral one”, and it’s probably the last time physicists used “-ino” to mean “little”. Neutrinos are “neutral” because they have no electrical charge. They also don’t interact with the strong nuclear force. Only the weak nuclear force has any effect on them. (Well, gravity does too, but very weakly.)

This makes it very difficult to detect neutrinos: you have to catch them interacting via the weak force, which is, well, weak. Originally, that meant they had to be inferred by their absence: missing energy in nuclear reactions carried away by “something”. Now, they can be detected, but it requires massive tanks of fluid, carefully watched for the telltale light of the rare interactions between neutrinos and ordinary matter. You wouldn’t notice if billions of neutrinos passed through you every second, like an unstoppable army of ghosts. And in fact, that’s exactly what happens!

Visualization of neutrinos from a popular documentary

In the 60’s, scientists began to use these giant tanks of fluid to detect neutrinos coming from the sun. An enormous amount of effort goes in to understanding the sun, and these days our models of it are pretty accurate, so it came as quite a shock when researchers observed only half the neutrinos they expected. It wasn’t until the work of Super-Kamiokande in 1998, and SNOLAB in 2001, that we knew the reason why.

As it turns out, neutrinos oscillate. Neutrinos are produced in what are called flavor states, which match up with the different types of leptons. There are electron-neutrinos, muon-neutrinos, and tau-neutrinos.

Radioactive processes usually produce electron-neutrinos, so those are the type that the sun produces. But on their way from the sun to the earth, these neutrinos “oscillate”: they switch between electron neutrinos and the other types! The older detectors, focused only on electron-neutrinos, couldn’t see this. SNOLAB’s big advantage was that it could detect the other types of neutrinos as well, and tell the difference between them, which allowed it to see that the “missing” neutrinos were really just turning into other flavors! Meanwhile, Super-Kamiokande measured neutrinos coming not from the sun, but from cosmic rays reacting with the upper atmosphere. Some of these neutrinos came from the sky above the detector, while others traveled all the way through the earth below it, from the atmosphere on the other side. By observing “missing” neutrinos coming from below but not from above, Super-Kamiokande confirmed that it wasn’t the sun’s fault that we were missing solar neutrinos, neutrinos just oscillate!

What does this oscillation have to do with neutrinos having mass, though?

Here things get a bit trickier. I’ve laid some of the groundwork in older posts. I’ve told you to think about mass as “energy we haven’t met yet”, as the energy something has when we leave it alone to itself. I’ve also mentioned that conservation laws come from symmetries of nature, that energy conservation is a result of symmetry in time.

This should make it a little more plausible when I say that when something has a specific mass, it doesn’t change. It can decay into other particles, or interact with other forces, but left alone, by itself, it won’t turn into something else. To be more specific, it doesn’t oscillate. A state with a fixed mass is symmetric in time.

The only way neutrinos can oscillate between flavor states, then, is if one flavor state is actually a combination (in quantum terms, a superposition) of different masses. The components with different masses move at different speeds, so at any point along their path you can be more or less likely to see certain masses of neutrinos. As the mix of masses changes, the flavor state changes, so neutrinos end up oscillating from electron-neutrino, to muon-neutrino, to tau-neutrino.

So because of neutrino oscillation, neutrinos have to have mass. But this presented a problem. Most fundamental particles get their mass from interacting with the Higgs field. But, as it turns out, neutrinos can’t interact with the Higgs field. This has to do with the fact that neutrinos are “chiral”, and only come in a “left-handed” orientation. Only if they had both types of “handedness” could they get their mass from the Higgs.

As-is, they have to get their mass another way, and that way has yet to be definitively shown. Whatever it ends up being, it will be beyond the current Standard Model. Maybe there actually are right-handed neutrinos, but they’re too massive, or interact too weakly, for them to have been discovered. Maybe neutrinos are Majorana particles, getting mass in a novel way that hasn’t been seen yet in the Standard Model.

Whatever we discover, neutrinos are currently our best evidence that something lies beyond the Standard Model. Naturalness may have philosophical problems, dark matter may be explained away by modified gravity…but if neutrinos have mass, there’s something we still have yet to discover. And that definitely seems worthy of a Nobel to me!

Hexagon Functions III: Now with More Symmetry

I’ve got a new paper up this week.

It’s a continuation of my previous work, understanding collisions involving six particles in my favorite theory, N=4 super Yang-Mills.

This time, we’re pushing up the complexity, going from three “loops” to four. In the past, I could have impressed you with the number of pages the formulas I’m calculating take up (eight hundred pages for the three-loop formula from that first Hexagon Functions paper). Now, though, I don’t have that number: putting my four-loop formula into a pdf-making program just crashes the program. Instead, I’ll have to impress you with file sizes: 2.6 MB for the three-loop formula, 96 MB for the four-loop one.

Calculating such a formula sounds like a pretty big task, and it was, the first time. But things got a lot simpler after a chat I had at Amplitudes.

We calculate these things using an ansatz, a guess for what the final answer should look like. The more vague our guess, the more parameters we need to fix, and the more work we have in general. If we can guess more precisely, we can start with fewer parameters and things are a lot easier.

Often, more precise guesses come from understanding the symmetries of the problem. If we can know that the final answer must be the same after making some change, we can rule out a lot of possibilities.

Sometimes, these symmetries are known features of the answer, things that someone proved had to be correct. Other times, though, they’re just observations, things that have been true in the past and might be true again.

We started out using an observation from three loops. That got us pretty far, but we still had a lot of work to do: 808 parameters, to be fixed by other means. Fixing them took months of work, and throughout we hoped that there was some deeper reason behind the symmetries we observed.

Finally, at Amplitudes, I ran into fellow amplitudeologist Simon Caron-Huot and asked him if he knew the source of our observed symmetry. In just a few days he was able to link it to supersymmetry, giving us justification for our jury rigged trick. However, we figured out that his explanation went further than any of us expected. In the end, rather than 808 parameters we only really needed to consider 34.

Thirty-four options to consider. Thirty-four possible contributions to a ~100 MB file. That might not sound like a big deal, but compared to eight hundred and eight it’s a huge deal. More symmetry means easier calculations, meaning we can go further. At this point going to the next step in complexity, to five loops rather than four, might be well within reach.

Bras and Kets, Trading off Instincts

Some physics notation is a joke, but that doesn’t mean it shouldn’t be taken seriously.

Take bras and kets. On the surface, as silly a physics name as any. If you want to find the probability that a state in quantum mechanics turns into another state, you write down a “bracket” between the two states:

\langle a | b\rangle

This leads, with typical physics logic, to the notation for the individual states: separate out the two parts, into a “bra” and a “ket”:

\langle a||b\rangle

It’s kind of a dumb joke, and it annoys the heck out of mathematicians. Not for the joke, of course, mathematicians probably have worse.

Mathematicians are annoyed when we use complicated, weird notation for something that looks like a simple, universal concept. Here, we’re essentially just taking inner products of vectors, something mathematicians have been doing in one form or another for centuries. Yet rather than use their time-tested notation we use our own silly setup.

There’s a method to the madness, though. Bras and kets are handy for our purposes because they allow us to leverage one of the most powerful instincts of programmers: the need to close parentheses.

In programming, various forms of parentheses and brackets allow you to isolate parts of code for different purposes. One set of lines might only activate under certain circumstances, another set of brackets might make text bold. But in essentially every language, you never want to leave an open parenthesis. Doing so is almost always a mistake, one that leaves the rest of your code open to whatever isolated region you were trying to create.

Open parentheses make programmers nervous, and that’s exactly what “bras” and “kets” are for. As it turns out, the states represented by “bras” and “kets” are in a certain sense un-measurable: the only things we can measure are the brackets between them. When people say that in quantum mechanics we can only predict probabilities, that’s a big part of what they mean: the states themselves mean nothing without being assembled into probability-calculating brackets.

This ends up making “bras” and “kets” very useful. If you’re calculating something in the real world and your formula ends up with a free “bra” or a “ket”, you know you’ve done something wrong. Only when all of your bras and kets are assembled into brackets will you have something physically meaningful. Since most physicists have done some programming, the programmer’s instinct to always close parentheses comes to the rescue, nagging until you turn your formula into something that can be measured.

So while our notation may be weird, it does serve a purpose: it makes our instincts fit the counter-intuitive world of quantum mechanics.

Scooped Is a Spectrum

I kind of got scooped recently.

I say kind of, because as I’ve been realizing being scooped isn’t quite the all-or-nothing thing you’d think it would be. Rather, being scooped is a spectrum.

Go ahead and scoop up a spectrum as you’re reading this.

By the way, I’m going to be a bit cagey about what exactly I got scooped on. As you’ll see, there are still a few things my collaborator and I need to figure out, and in the meantime I don’t want to put my foot in my mouth. Those of you who follow what’s going on in amplitudes might have some guesses. In case you’re worried, it has nothing to do with my work on Hexagon Functions.

When I heard about the paper that scooped us, my first reaction was to assume the project I’d been working on for a few weeks was now a dead end. When another group publishes the same thing you’ve been working on, and does it first, there doesn’t seem to be much you can do besides shake hands and move on.

As it turns out, though, things are a bit more complicated. The risk of publishing fast, after all, is making mistakes. In this case, it’s starting to look like a few of the obstructions that were holding us back weren’t solved by the other group, and in fact that they may have ignored those obstructions altogether in their rush to get something publishable.

This creates an interesting situation. It’s pretty clear the other group is beyond us in certain respects, they published first for a (good) reason. On the other hand, precisely because we’ve been slower, we’ve caught problems that it looks like the other group didn’t notice. Rather than rendering our work useless, this makes it that much more useful: complementing the other group’s work rather than competing with it.

Being scooped is a spectrum. If two groups are working on very similar things, then whoever publishes first usually wins. But if the work is different enough, then a whole range of roles opens up, from corrections and objections to extensions and completions. Being scooped doesn’t have to be the end of the world, in fact, it can be the beginning.

A Tale of Two CMB Measurements

While trying to decide what to blog about this week, I happened to run across this article by Matthew Francis on Ars Technica.

Apparently, researchers have managed to use Planck‘s measurement of the Cosmic Microwave Background to indirectly measure a more obscure phenomenon, the Cosmic Neutrino Background.

The Cosmic Microwave Background, or CMB is often described as the light of the Big Bang, dimmed and spread to the present day. More precisely, it’s the light released from the first time the universe became transparent. When electrons and protons joined to form the first atoms, light no longer spent all its time being absorbed and released by electrical charges, and was free to travel in a mostly-neutral universe.

This means that the CMB is less like a view of the Big Bang, and more like a screen separating us from it. Light and charged particles from before the CMB was formed will never be observable to us, because they would have been absorbed by the early universe. If we want to see beyond this screen, we need something with no electric charge.

That’s where the Cosmic Neutrino Background comes in. Much as the CMB consists of light from the first time the universe became transparent, the CNB consists of neutrinos from the first time the universe was cool enough for them to travel freely. Since this happened a bit before the universe was transparent to light, the CNB gives information about an earlier stage in the universe’s history.

Unfortunately, neutrinos are very difficult to detect, the low-energy ones left over from the CNB even more so. Rather than detecting the CNB directly, it has to be observed through its indirect effects on the CMB, and that’s exactly what these researchers did.

Now does all of this sound just a little bit familiar?

Gravitational waves are also hard to detect, hard enough that we haven’t directly detected any yet. They’re also electrically neutral, so they can also give us information from behind the screen of the CMB, letting us learn about the very early universe. And when the team at BICEP2 purported to measure these primordial gravitational waves indirectly, by measuring the CMB, the press went crazy about it.

This time, though? That Ars Technica article is the most prominent I could find. There’s nothing in major news outlets at all.

I don’t think that this is just a case of people learning from past mistakes. I also don’t think that BICEP2’s results were just that much more interesting: they were making a claim about cosmic inflation rather than just buttressing the standard Big Bang model, but (outside of certain contrarians here at Perimeter) inflation is not actually all that controversial. It really looks like hype is the main difference here, and that’s kind of sad. The difference between a big (premature) announcement that got me to write four distinct posts and an article I almost didn’t notice is just one of how the authors chose to make their work known.

Don’t Watch the Star, Watch the Crowd

I didn’t comment last week on Hawking’s proposed solution of the black hole firewall problem. The media buzz around it was a bit less rabid than the last time he weighed in on this topic, but there was still a lot more heat than light.

The impression I get from the experts is that Hawking’s proposal (this time made in collaboration with Andrew Strominger and Malcom Perry, the former of whom is famous for, among other things, figuring out how string theory can explain the entropy of black holes) resembles some earlier suggestions, with enough new elements to make it potentially interesting but potentially just confusing. It’s a development worth paying attention to for specialists, but it’s probably not the sort of long-awaited answer the media seems to be presenting it as.

This raises a question: how, as a non-specialist, are you supposed to tell the difference? Sure, you can just read blogs like mine, but I can’t report on everything.

I may have a pretty solid grounding in physics, but I know almost nothing about music. I definitely can’t tell what makes a song good. About the best I can do is see if I can dance to it, but that doesn’t seem to be a reliable indicator of quality music. Instead, my best bet is usually to watch the crowd.

Lasers may make this difficult.

Ask the star of a show if they’re doing good work, and they’re unlikely to be modest. Ask the average music fan, though, and you get a better idea. Watch music fans as a group, and you get even more information.

When a song starts playing everywhere you go, when people start pulling it out at parties and making their own imitations of it, then maybe it’s important. That might not mean it’s good, but it does mean it’s worth knowing about.

When Hawking or Strominger or Witten or anyone whose name you’ve heard of says they’ve solved the puzzle of the century, be cautious. If it really is worth your attention, chances are it won’t be the last you’ll hear about it. Other physicists will build off of it, discuss it, even spin off a new sub-field around it. If it’s worth it, you won’t have to trust what the stars of the physics world say: you’ll be able to listen to the crowd.

Romeo and Juliet, through a Wormhole

Perimeter is hosting this year’s Mathematica Summer School on Theoretical Physics. The school is a mix of lectures on a topic in physics (this year, the phenomenon of quantum entanglement) and tips and tricks for using the symbolic calculation program Mathematica.

Juan Maldacena is one of the lecturers, which gave me a chance to hear his Romeo and Juliet-based explanation of the properties of wormholes. While I’ve criticized some of Maldacena’s science popularization work in the past, this one is pretty solid, so I thought I’d share it with you guys.

You probably think of wormholes as “shortcuts” to travel between two widely separated places. As it turns out, this isn’t really accurate: while “normal” wormholes do connect distant locations, they don’t do it in a way that allows astronauts to travel between them, Interstellar-style. This can be illustrated with something called a Penrose diagram:

Static

Static “Greyish Black” Diagram

In the traditional Penrose diagram, time goes upward, while space goes from side to side. In order to measure both in the same units, we use the speed of light, so one year on the time axis corresponds to one light-year on the space axis. This means that if you’re traveling at a 45 degree line on the diagram, you’re going at the speed of light. Any lower angle is impossible, while any higher angle means you’re going slower.

If we start in “our universe” in the diagram, can we get to the “other universe”?

Pretty clearly, the answer is no. As long as we go slower than the speed of light, when we pass the event horizon of the wormhole we will end up, not in the “other universe”, but at the part of the diagram labeled Future Singularity, the singularity at the center of the black hole. Even going at the speed of light only keeps us orbiting the event horizon for all eternity, at best.

What use could such a wormhole be? Well, imagine you’re Romeo or Juliet.

Romeo has been banished from Verona, but he took one end of a wormhole with him, while the other end was left with Juliet. He can’t go through and visit her, she can’t go through and visit him. But if they’re already considering taking poison, there’s an easier way. If they both jump in to the wormhole, they’ll fall in to the singularity. Crucially, though, it’s the same singularity, so once they’re past the event horizon they can meet inside the black hole, spending some time together before the end.

Depicted here for more typical quantum protagonists, Alice and Bob.

This explains what wormholes really are: two black holes that share a center.

Why was Maldacena talking about this at a school on entanglement? Maldacena has recently conjectured that quantum entanglement and wormholes are two sides of the same phenomenon, that pairs of entangled particles are actually connected by wormholes. Crucially, these wormholes need to have the properties described above: you can’t use a pair of entangled particles to communicate information faster than light, and you can’t use a wormhole to travel faster than light. However, it is the “shared” singularity that ends up particularly useful, as it suggests a solution to the problem of black hole firewalls.

Firewalls were originally proposed as a way of getting around a particular paradox relating three states connected by quantum entanglement: a particle inside a black hole, radiation just outside the black hole, and radiation far away from the black hole. The way the paradox is set up, it appears that these three states must all be connected. As it turns out, though, this is prohibited by quantum mechanics, which only allows two states to be entangled at a time. The original solution proposed for this was a “firewall”, a situation in which anyone trying to observe all three states would “burn up” when crossing the event horizon, thus avoiding any observed contradiction. Maldacena’s conjecture suggests another way: if someone interacts with the far-away radiation, they have an effect on the black hole’s interior, because the two are connected by a wormhole! This ends up getting rid of the contradiction, allowing the observer to view the black hole and distant radiation as two different descriptions of the same state, and it depends crucially on the fact that a wormhole involves a shared singularity.

There’s still a lot of detail to be worked out, part of the reason why Maldacena presented this research here was to inspire more investigation from students. But it does seem encouraging that Romeo and Juliet might not have to face a wall of fire before being reunited.

The Theorist Exclusion Principle

There are a lot of people who think theoretical physics has gone off-track, though very few of them agree on exactly how. Some think that string theory as a whole is a waste of time, others that the field just needs to pay more attention to their preferred idea. Some think we aren’t paying enough attention to the big questions, or that we’re too focused on “safe” ideas like supersymmetry, even when they aren’t working out. Some think the field needs less focus on mathematics, while others think it needs even more.

Usually, people act on these opinions by writing strongly worded articles and blog posts. Sometimes, they have more power, and act with money, creating grants and prizes that only go to their preferred areas of research.

Let’s put the question of whether the field actually needs to change aside for the moment. Even if it does, I’m skeptical that this sort of thing will have any real effect. While grants and blogs may be very good at swaying experimentalists, theorists are likely to be harder to shift, due to what I’m going to call the Theorist Exclusion Principle.

The Pauli Exclusion Principle is a rule from quantum mechanics that states that two fermions (particles with half-integer spin) can’t occupy the same state. Fermions include electrons, quarks, protons…essentially, all the particles that make up matter. Many people learn about the Pauli Exclusion Principle first in a chemistry class, where it explains why electrons fall into different energy levels in atoms: once one energy level “fills up”, no more electrons can occupy the same state, and any additional electrons are “excluded” and must occupy a different energy level.

Those 1s electrons are such a clique!

In contrast, bosons (like photons, or the Higgs) can all occupy the same state. It’s what allows for things like lasers, and it’s why all the matter we’re used to is made out of fermions: because fermions can’t occupy the same state as each other, as you add more fermions the structures they form have to become more and more complicated.

Experimentalists are a little like bosons. While you can’t stuff two experimentalists into the same quantum state, you can get them working on very similar projects. They can form large collaborations, with each additional researcher making the experiment that much easier. They can replicate eachother’s work, making sure it was accurate. They can take some physical phenomenon and subject it to a battery of tests, so that someone is bound to learn something.

Theorists, on the other hand, are much more like fermions. In theory, there’s very little reason to work on something that someone else is already doing. Replication doesn’t mean very much: the purest theory involves mathematical proofs, where replication is essentially pointless. Theorists do form collaborations, but they don’t have the same need for armies of technicians and grad students that experimentalists do. With no physical objects to work on, there’s a limit to how much can be done pursuing one particular problem, and if there really are a lot of options they can be pursued by one person with a cluster.

Like fermions, then, theorists expand to fill the projects available. If an idea is viable, someone will probably work on it, and once they do, there isn’t much reason for someone else to do the same thing.

This makes theory a lot harder to influence than experiment. You can write the most beautiful thinkpiece possible to persuade theorists to study the deep questions of the universe, but if there aren’t any real calculations available nothing will change. Contrary to public perception, theoretical physicists aren’t paid to just sit around thinking all day: we calculate, compute, and publish, and if a topic doesn’t lend itself to that then we won’t get much mileage out of it. And no matter what you try to preferentially fund with grants, mostly you’ll just get people re-branding what they’re already doing, shifting a few superficial details to qualify.

Theorists won’t occupy the same states, so if you want to influence theorists you need to make sure there are open states where you’re trying to get them to go. Historically, theorists have shifted when new states have opened up: new data from experiment that needed a novel explanation, new mathematical concepts that opened up new types of calculations. You want there to be fewer string theorists, or more focus on the deep questions? Give us something concrete to do, and I guarantee you’ll get theorists flooding in.