Tag Archives: theoretical physics

Experimentalist Says What?

I’m a theoretical physicist. That means I work with pencil and paper, or with my laptop, or at most with a computer cluster. I don’t have a lab, and even if I did I wouldn’t have any equipment to store there.

By contrast, most physicists (and most scientists in general) are experimentalists, the people who actually do experiments, actually work in labs, and actually use piles and piles of expensive equipment. Naturally, these two groups have very different ways of doing things, spawned by different requirements for their jobs. This leads to very different ways of talking. We theorists sometimes get confused by the quaint turns of phrase used by experimentalists, so I’ve put together this handy translation guide:

 

Lab: Kind of like an office, but has a bunch of big machines in it for some reason. Also, in some of them they don’t even drink coffee, some nonsense about toxic contaminants. I don’t know how they get any work done with all those test tubes all over the place.

PI: Not Private Investigator, but close! The Primary Investigator is the big cheese among the experimentalists, the one who owns all the big machines. All of the others must bow before him or her, even fellow professors must grovel if they want to use the PI’s expensive equipment. Naturally, this makes experimentalists very hierarchical, a sharp contrast to theorists who are obviously totally egalitarian.

Poster: Let me tell you a secret about experimentalists: there are a lot of them. Way more than there are theorists. So many, that if they all go to a conference it’s impossible for them all to give talks! That’s where posters come in: some of the experimentalists all stand in a room in front of rectangles of cardboard covered in charts, while the others walk around and ask questions. Traditionally, these posters are printed an hour before the conference, obviously for maximum freshness and not at all because of procrastination.

Group: Like our Institutes, but (because there are a lot of experimentalists) there isn’t just one per university and (because of the shared lab) they actually have something to talk about. This leads to regular group meetings, because when you’re using expensive equipment you actually need to show you’re doing something worthwhile with it.

IRB: For the medical and psychological folks, the Internal Review Board is there to tell you that, no, you can’t infect monkeys with flesh-eating bacteria just to see what happens. They’re also the people who ask you whether a grammatical change in your online survey will pose risks to pregnant women, which is clearly exactly as important. Theorists don’t have these, because numbers are an oppressed underclass with no rights to speak of. EHS (Environmental Health and Safety) fills a similar role for those who only oppress yeast and their own grad students.

Annual Meeting: Experimentalists tend to be part of big organizations like the American Physical Society. And that’s all well and good, occupies a space on the CV and so forth. What’s somewhat more baffling is their tendency to trust those organizations to run conferences. Generally these are massive affairs, with people from all sorts of sub-fields participating. This only works because experimentalists have the mysterious ability to walk into each other’s talks and actually understand what’s going on, even if the subject matter is very different from what they’re used to. Experts suggest this has something to do with actually studying real things in the real world, but this is a hypothesis at best.

How do I get where you are?

I’ve mentioned before that this blog will be undergoing a redesign this summer, transitioning from 4gravitons.wordpress.com to just 4gravitons.wordpress.com. One part of that redesign will be the introduction of new categories to help people search for content, as well as new guides like the ones for N=4 super Yang-Mills and the (2,0) theory for some of those categories. Of those, one planned category/guide will discuss careers in physics, with an eye towards explaining some of the often-unstated assumptions behind the process.

I’ve already posted on being a graduate research assistant and on what a postdoc is. I haven’t said much yet about the process leading up to becoming a graduate student. In this post, I’m going to give an overview of a career in theoretical physics, with a focus on what happens before you find an advisor. This is going to be inherently biased, based as it will be on my experiences. In particular, each country’s education system is different, so much of this will only be relevant for students in the US.

Let’s start at the beginning.

A very good place to start.

If you want to become a theoretical physicist, you’d better start by taking physics and math courses in high school. Unfortunately, this is where socioeconomic status has a big effect. Some schools have Advanced Placement or International Baccalaureate courses that let you get a head-start on college, many don’t. Some schools don’t even have physics courses at all anymore. My only advice here is to get what you can, when you can. If you can take a physics course, do it. If you can take calculus, do it. If you can take classes that will give you university credit, take them.

After high school, you go to college for a Bachelor’s degree in physics. Getting into college these days is some sort of ridiculous popularity contest, and I don’t pretend to be able to give advice on that. What I can say is that once you’re in college, coursework is important, but research is more important. Graduate schools will look at how well you did in your courses and how advanced those courses were, but they will pay special attention to who you get recommendations from, and whether you did research with them. Whether or not your college has anyone who you can research with, you should consider doing summer research somewhere interesting. With programs like the NSF’s Research Experience for Undergraduates (or REU) you can apply to get hooked up with interesting projects and mentors. In addition to looking good on an application to grad school, doing research helps boost your self-confidence: knowing that you can do something real really helps you start feeling like a scientist. Research also teaches you specialized skills much faster than coursework can: I’ve learned much more about programming from having to use it on projects than from any actual programming course.

That said, coursework is also useful. You want courses that will familiarize you with basic tools of your field, physics courses on classical mechanics and quantum mechanics and electromagnetism and math courses on linear algebra and differential equations. You want to take a math course on group theory, but only if it’s taught by a physicist, as mathematicians focus on different aspects. More than any of that, though, you want to try to take at least a few graduate-level courses in while you’re still in college.

That’s important, because grad school in theoretical physics is kind of a mess. You’ll be there for around five years in total (I was in at the low end with four, some people take six or seven). However, you take most if not all of your courses in the first two years. In general, during that time you are paid as a Teaching Assistant. The school pays your tuition and a livable (if barely) wage, and in return you lead lab sections or grade papers. Teaching experience can be a positive thing, but you don’t want to keep doing it for too long, because the point of grad school isn’t teaching or courses, it’s research. Your goal is to find an advisor who is willing to pay you out of one of their (usually government) grants, so that you can transition from Teaching Assistant to Research Assistant. This is hard to do while you’re still taking courses: you won’t have time, and worse, you won’t know everything you need. Theoretical physics requires a lot of background, and much of it gets taught in grad school. Here at Stony Brook, you’d be taking graduate-level quantum mechanics, quantum field theory, and string theory. Until recently, each one of those was a one-year course, and the most logical way to take them was one after the other. Add that up, and that’s three years…kind of a problem when you want to start research after two. That’s why getting ahead in courses, however and whenever you can, is so important: not so much for the courses themselves, but so you can get past them and do research.

Research is what you do for the rest of your time in grad school. It’s what you do after you graduate, when you become a postdoc. It (and teaching) are what you do as a professor, what you are judged on when they decide whether or not you get tenure. Working through research is going to teach you more than any other experience you will have, so get as much of it as you can. And good luck!

The Four Ways Physicists Name Things

If you’re a biologist and you discover a new animal, you’ve always got Latin to fall back on. If you’re an astronomer, you can describe what you see. But if you’re a physicist, your only option appears to involve falling back on one of a few terrible habits.

The most reasonable option is just to name it after a person. Yang-Mills and the Higgs Boson may sound silly at first, but once you know the stories of C. N. Yang, Robert Mills, Peter Higgs and Satyendra Nath Bose you start appreciating what the names mean. While this is usually the most elegant option, the increasingly collaborative nature of physics means that many things have to be named with a series of initials, like ABJM, BCJ and KKLT.

A bit worse is the tendency to just give it the laziest name possible. What do you call the particles that “glue” protons and neutrons together? Why gluons, of course, yuk yuk yuk!

This is particularly common when it comes to supersymmetry, where putting the word “super” in front of something almost always works. If that fails, it’s time to go for more specific conventions: to find the partner of an existing particle, if the new particle is a boson, just add “s-” for “super”“scalar” apparently to the name. This creates perfectly respectable names like stau, sneutrino, and selectron. If the new particle is a fermion, instead you add “-ino” to the end, getting something like a gluino if you start with a gluon. If you’ve heard of neutrinos, you may know that neutrino means “little neutral one”. You might perfectly rationally expect that gluino means “little gluon”, if you had any belief that physicists name things logically. We don’t. A gluino is called a gluino because it’s a fermion, and neutrinos are fermions, and the physicists who named it were too lazy to check what “neutrino” actually means.

Pictured: the superpartner of Nidoran?

Worse still are names that are obscure references and bad jokes. These are mercifully rare, and at least memorable when they occur. In quantum mechanics, you write down probabilities using brackets of two quantum states, \langle a | b\rangle. What if you need to separate the two states, \langle a| and |b\rangle? Then you’ve got a “bra” and a “ket”!

Or have you heard the story of how quarks were named? Quarks, for those of you unfamiliar with them, are found in protons and neutrons in groups of three. Murray Gell-Mann, one of the two people who first proposed the existence of quarks, got their name from Finnegan’s Wake, a novel by James Joyce, which at one point calls for “Three quarks for Muster Mark!” While this may at first sound like a heartwarming tale of respect for the literary classics, it should be kept in mind that a) Finnegan’s Wake is a novel composed almost entirely of gibberish, read almost exclusively by people who pretend to understand it to seem intelligent and b) this isn’t exactly the most important or memorable line in the book. So Gell-Mann wasn’t so much paying homage to a timeless work of literature as he was referencing the most mind-numbingly obscure piece of nerd trivia before the invention of Mara Jade. Luckily these days we have better ways to remember the name.

Albeit wrinklier ways.

The final, worst category, though, don’t even have good stories going for them. They are the names that tell you absolutely nothing about the thing they are naming.

Probably the worst examples of this from my experience are the a-theorem and the c-theorem. In both cases, a theory happened to have a parameter in it labeled by a letter. When a theorem was proven about that parameter, rather than giving it a name that told you anything at all about what it was, people just called it by the name of the parameter. Mathematics is full of names like this too. Without checking Wikipedia, what’s the difference between a set, a group, and a category? What the heck is a scheme?

If you ever have to name something, be safe and name it after a person. If you don’t, just try to avoid falling into these bad habits of physics naming.

A Wild Infinity Appears! Or, Renormalization

Back when Numberphile’s silly video about the zeta function came up, I wrote a post explaining the process of regularization, where physicists take an incorrect infinite result and patch it over to get something finite. At the end of that post I mentioned a particular variant of regularization, called renormalization, which was especially important in quantum field theory.

Renormalization has to do with how we do calculations and make predictions in particle physics. If you haven’t read my post “What’s so hard about Quantum Field Theory anyway?” you should read it before trying to tackle this one. The important concepts there are that probabilities in particle physics are calculated using Feynman Diagrams, that those diagrams consist of lines representing particles and points representing the ways they interact, that each line and point in the diagram gives a number that must be plugged in to the calculation, and that to do the full calculation you have to add up all the possible diagrams you can draw.

Let’s say you’re interested in finding out the mass of a particle. How about the Higgs?

You can’t weigh it, or otherwise see how gravity affects it: it’s much too light, and decays into other particles much too fast. Luckily, there is another way. As I mentioned in this post, a particle’s mass and its kinetic energy (energy of motion) both contribute to its total energy, which in turn affects what particles it can turn into if it decays. So if you want to find a particle’s mass, you need the relationship between its motion and its energy.

Suppose we’ve got a Higgs particle moving along. We know it was created out of some collision, and we know what it decays into at the end. With that, we can figure out its mass.

higgstree

There’s a problem here, though: we only know what happens at the beginning and the end of this diagram. We can’t be certain what happens in the middle. That means we need to add in all of the other diagrams, every possible diagram with that beginning and that end.

Just to look at one example, suppose the Higgs particle splits into a quark and an anti-quark (the antimatter version of the quark). If they come back together later into a Higgs, the process would look the same from the outside. Here’s the diagram for it:

higgsloop

When we’re “measuring the Higgs mass”, what we’re actually measuring is the sum of every single diagram that begins with the creation of a Higgs and ends with it decaying.

Surprisingly, that’s not the problem!

The problem comes when you try to calculate the number that comes out of that diagram, when the Higgs splits into a quark-antiquark pair. According to the rules of quantum field theory, those quarks don’t have to obey the normal relationship between total energy, kinetic energy, and mass. They can have any kinetic energy at all, from zero all the way up to infinity. And because it’s quantum field theory, you have to add up all of those possible kinetic energies, all the way up. In this case, the diagram actually gives you infinity.

(Note that not every diagram with unlimited kinetic energy is going to be infinite. The first time theorists calculated infinite diagrams, they were surprised.

For those of you who know calculus, the problem here comes after you integrate over momentum. The two quarks each give a factor of one over the momentum, and then you integrate the result four times (for three dimensions of space plus time), which gives an infinite result. If you had different particles arranged in a different way you might divide by more factors of momentum and get a finite value.)

The modern understanding of infinite results like this is that they arise from our ignorance. The mass of the Higgs isn’t actually infinity, because we can’t just add up every kinetic energy up to infinity. Instead, at some point before we get to infinity “something else” happens.

We don’t know what that “something else” is. It might be supersymmetry, it might be something else altogether. Whatever it is, we don’t know enough about it now to include it in the calculations as anything more than a cutoff, a point beyond which “something” happens. A theory with a cutoff like this, one that is only “effective” below a certain energy, is called an Effective Field Theory.

While we don’t know what happens at higher energies, we still need a way to complete our calculations if we want to use them in the real world. That’s where renormalization comes in.

When we use renormalization, we bring in experimental observations. We know that, no matter what is contributing to the Higgs particle’s mass, what we observe in the real world is finite. “Something” must be canceling the divergence, so we simply assume that “something” does, and that the final result agrees with the experiment!

"Something"

“Something”

In order to do this, we accepted the experimental result for the mass of the Higgs. That means that we’ve lost any ability to predict the mass from our theory. This is a general rule for renormalization: we trade ignorance (of the “something” that happens at high energy) for a loss of predictability.

If we had to do this for every calculation, we couldn’t predict anything at all. Luckily, for many theories (called renormalizable theories) there are theorems proving that you only need to do this a few times to fix the entire theory. You give up the ability to predict the results of a few experiments, but you gain the ability to predict the rest.

Luckily for us, the Standard Model is a renormalizable theory. Unfortunately, some important theories are not. In particular, quantum gravity is non-renormalizable. In order to fix the infinities in quantum gravity, you need to do the renormalization trick an infinite number of times, losing an infinite amount of predictability. Thus, while making a theory of quantum gravity is not difficult in principle, in practice the most obvious way to create the theory results in a “theory” that can never make any predictions.

One of the biggest virtues of string theory (some would say its greatest virtue) is that these infinities never appear. You never need to renormalize string theory in this way, which is what lets it work as a theory of quantum gravity. N=8 supergravity, the gravity cousin of N=4 super Yang-Mills, might also have this handy property, which is why many people are so eager to study it.

Why we Physics

There are a lot of good reasons to study theories in theoretical physics, even the ones that aren’t true. They teach us how to do calculations in other theories, including those that do describe reality, which lets us find out fundamental facts about nature. They let us hone our techniques, developing novel methods that often find use later, in some cases even spinoff technology. (Mathematica came out of the theoretical physics community, while experimental high energy physics led to the birth of the modern internet.)

Of course, none of this is why physicists actually do physics. Sure, Nima Arkani-Hamed might need to tell himself that space-time is doomed to get up in the morning, but for a lot of us, it isn’t about proving any wide-ranging point about the universe. It’s not even all about the awesome, as some would have it: most of what we do on a day-to-day basis isn’t especially awesome. It goes a bit deeper than that.

Science, in the end, is about solving puzzles. And solving puzzles is immensely satisfying, on a deep, fundamental level.

There’s a unique feeling that you get when all the pieces come together, when you’re calculating something and everything cancels and you’re left with a simple answer, and for some people that’s the best thing in existence.

It’s especially true when you’re working with an ansatz or using some other method where you fix parameters and fill in uncertainties, one by one. You can see how close you are to the answer, which means each step gives you that little thrill of getting just that much closer. One of my colleagues describes the calculations he does in supergravity as not tedious but “delightful” for precisely this reason: a calculation where every step puts another piece in the right place just feels good.

Theoretical physicists are the kind of people who would get a Lego set for their birthday, build it up to completion, and then never play with it again (unless it was to take it apart and make something else). We do it for the pure joy of seeing something come together and become complete. Save what it’s “for” for the grant committees, we’ve got a different rush in mind.

The Royal We of Theoretical Physics

I’m about to show you an abstract from a theoretical physics paper. Don’t worry about what it says, just observe the grammar.

wittenabstract

Notice anything? Here, I’ll zoom in:

wittenwe

This paper has one author, Edward Witten. So who’s “we”?

As it turns out, it is actually quite common in theoretical physics for a paper to use the word “we”, even when it is written by a single author. While this tradition has been called stilted, pompous, and just plain bad writing, there is a legitimate reason behind it. “We” is convenient, because it represents several different important things.

While the paper I quoted was written by only one author, many papers are collaborative efforts. For a collaboration, depending on collaboration style, it is often hard to distinguish who did what in a consistent way. As such, “we” helps smooth over different collaboration styles in a consistent way.

What about single-authored papers, though? For a single author, and often even for multiple authors, “we” means the author plus the reader.

In principle, anyone reading a paper in theoretical physics should be able to follow along, doing the calculations on their own, and replicate the paper’s results. In practice this can often be difficult to impossible, but it’s still true that if you want to really retain what you read in theoretical physics, you need to follow along and do some of the calculation yourself. As a nod to this, it is conventional to write theoretical physics papers as if the reader was directly participating, leading them through the results point by point like exercises in a textbook. “We” do one calculation, then “we” use the result to derive the next point, and so on.

There are other meanings that “we” can occasionally serve, such as referring to everyone in a particular field, or a group in a hypothetical example.

While each of these meanings of “we” could potentially use a different word, that tends to make a paper feel cluttered, with jarring transitions between different subjects. Using “we” for everything gives the paper a consistent voice and feel, though it does come at the cost of obscuring some of the specific details of who did what. Especially for collaborations, the “we the collaborators” and “we the author plus reader” meanings can overlap and blur together. This usually isn’t a problem, but as I’ve been finding out recently it does make things tricky when writing for people who aren’t theoretical physicists, such as universities with guidelines that require a thesis to clearly specify who in a collaboration did what.

On an unrelated note, two papers went up this week pushing the hexagon function story to new and impressive heights. I wasn’t directly involved in either, I’ve been attacking a somewhat different part of the problem, and you can look forward to something on that in a few months.

What’s in a Thesis?

As I’ve mentioned before, I’m graduating this spring, which means I need to write that most foreboding of documents, the thesis. As I work on it, I’ve been thinking about how the nature of the thesis varies from field to field.

If you don’t have much experience with academics, you probably think of a thesis as a single, overarching achievement that structures a grad student’s career. A student enters grad school, designs an experiment, performs it, collects data, analyzes the data, draws some conclusion, then writes a thesis about it and graduates.

In some fields, the thesis really does work that way. In biology for example, the process of planning an experiment, setting it up, and analyzing and writing up the data can be just the right size so that, a reasonable percentage of the time, it really can all be done over the course of a PhD.

Other fields tend more towards smaller, faster-paced projects. In theoretical physics, mathematics, and computer science, most projects don’t have the same sort of large experimental overhead that psychologists or biologists have to deal with. The projects I’ve worked on are large-scale for theoretical physics, and I’ll still likely have worked on three distinct things before I graduate. Others, with smaller projects, will often have covered more.

In this situation, a thesis isn’t one overarching idea. Rather, it’s a compilation of work from past projects, sewed together with a pretense of an overall theme. It’s a bit messy, but because it’s the way things are expected to be done in these fields, no-one minds particularly much.

The other end of the spectrum is potentially much harder to deal with. For those who work on especially big experiments, the payoff might take longer to arrive than any reasonable degree. Big machines like colliders and particle detectors can take well over a decade before they start producing data, while longitudinal studies that follow a population as they grow and age take a long time no matter how fast you work.

In cases like this, the challenge is to chop off a small enough part of the project to make it feel like a thesis. A thesis could be written about designing one component for the eventual machine, or analyzing one part of the vast sea of data it produces. Preliminary data from a longitudinal study could be analyzed, even when the final results are many years down the line.

People in these fields have to be flexible and creative when it comes to creating a thesis, but usually the thesis committee is reasonable. In the end, a thesis is what you need to graduate, whatever that actually is for you.

Update on the Amplituhedron

Awhile back I wrote a post on the Amplituhedron, a type of mathematical object  found by Nima Arkani-Hamed and Jaroslav Trnka that can be used to do calculations of scattering amplitudes in planar N=4 super Yang-Mills theory. (Scattering amplitudes are formulas used to calculate probabilities in particle physics, from the probability that an unstable particle will decay to the probability that a new particle could be produced by a collider.) Since then, they published two papers on the topic, the most recent of which came out the day before New Year’s Eve. These papers laid out the amplituhedron concept in some detail, and answered a few lingering questions. The latest paper focused on one particular formula, the probability that two particles bounce off each other. In discussing this case, the paper serves two purposes:

1. Demonstrating that Arkani-Hamed and Trnka did their homework.

2. Showing some advantages of the amplituhedron setup.

Let’s talk about them one at a time.

Doing their homework

There’s already a lot known about N=4 super Yang-Mills theory. In order to propose a new framework like the amplituhedron, Arkani-Hamed and Trnka need to show that the new framework can reproduce the old knowledge. Most of the paper is dedicated to doing just that. In several sections Arkani-Hamed and Trnka show that the amplituhedron reproduces known properties of the amplitude, like the behavior of its logarithm, its collinear limit (the situation when two momenta in the calculation become parallel), and, of course, unitarity.

What, you heard the amplituhedron “removes” unitarity? How did unitarity get back in here?

This is something that has confused several commenters, both here and on Ars Technica, so it bears some explanation.

Unitarity is the principle that enforces the laws of probability. In its simplest form, unitarity requires that all probabilities for all possible events add up to one. If this seems like a pretty basic and essential principle, it is! However, it and locality (the idea that there is no true “action at a distance”, that particles must meet to interact) can be problematic, causing paradoxes for some approaches to quantum gravity. Paradoxes like these inspired Arkani-Hamed to look for ways to calculate scattering amplitudes that don’t rely on locality and unitarity, and with the amplituhedron he succeeded.

However, just because the amplituhedron doesn’t rely on unitarity and locality, doesn’t mean it violates them. The amplituhedron, for all its novelty, still calculates quantities in N=4 super Yang-Mills. N=4 super Yang-Mills is well understood, it’s well-behaved and cuddly, and it obeys locality and unitarity.

This is why the amplituhedron is not nearly as exciting as a non-physicist might think. The amplituhedron, unlike most older methods, isn’t based on unitarity and locality. However, the final product still has to obey unitarity and locality, because it’s the same final product that others calculate through other means. So it’s not as if we’ve completely given up on basic principles of physics.

Not relying on unitarity and locality is valuable. For those who research scattering amplitudes, it has often been useful to try to “eliminate” one principle or another from our calculations. 20 years ago, avoiding Feynman diagrams was the key to finding dramatic simplifications. Now, many different approaches try to sidestep different principles. (For example, while the amplituhedron calculates an integrand and leaves a final integral to be done, I’m working on approaches that never employ an integrand.)

If we can avoid relying on some “basic” principle, that’s usually good evidence that the principle might be a consequence of something even more basic. By showing how unitarity can arise from the amplituhedron, Arkani-Hamed and Trnka have shown that a seemingly basic principle can come out of a theory that doesn’t impose it.

Advantages of the Amplituhedron

Not all of the paper compares to old results and principles, though. A few sections instead investigate novel territory, and in doing so show some of the advantages and disadvantages of the amplituhedron.

Last time I wrote on this topic, I was unclear on whether the amplituhedron was more efficient than existing methods. At this point, it appears that it is not. While the formula that the amplituhedron computes has been found by other methods up to seven loops, the amplituhedron itself can only get up to three loops or so in practical cases. (Loops are a way that calculations are classified in particle physics. More loops means a more complex calculation, and a more precise final result.)

The amplituhedron’s primary advantage is not in efficiency, but rather in the fact that its mathematical setup makes it straightforward to derive interesting properties for any number of loops desired. As Trnka occasionally puts it, the central accomplishment of the amplituhedron is to find “the question to which the amplitude is the answer”. By being able to phrase this “question” mathematically, one can be very general, which allows them to discover several properties that should hold no matter how complex the rest of the calculation becomes. It also has another implication: if this mathematical question has a complete mathematical answer, that answer could calculate the amplitude for any number of loops. So while the amplituhedron is not more efficient than other methods now, it has the potential to be dramatically more efficient if it can be fully understood.

All that said, it’s important to remember that the amplituhedron is still limited in scope. Currently, it applies to a particular theory, one that doesn’t (and isn’t meant to) describe the real world. It’s still too early to tell whether similar concepts can be defined for more realistic theories. If they can, though, it won’t depend on supersymmetry or string theory. One of the most powerful techniques for making predictions for the Large Hadron Collider, the technique of generalized unitarity, was first applied to N=4 super Yang-Mills. While the amplituhedron is limited now, I would not be surprised if it (and its competitors) give rise to practical techniques ten or twenty years down the line. It’s happened before, after all.

Elegance, Not So Mysterious

You’ll often hear theoretical physicists in the media referring to one theory or another as “elegant”. String theory in particular seems to get this moniker fairly frequently.

It may often seem like mathematical elegance is some sort of mysterious sixth sense theorists possess, as inexplicable to the average person as color to a blind person. What’s “elegant” about string theory, after all?

Before explaining elegance, I should take a bit of time to say what it’s not. Elegance isn’t Occam’s razor. It isn’t naturalness, either. Both of those concepts have their own technical definitions.

Elegance, by contrast, is a much hazier, and yet much simpler, notion. It’s hazy enough that any definition could provoke arguments, but I can at least give you an approximate idea by telling you that an elegant theory is simple to describe, if you know the right terms. Often, it is simpler than the phenomenon that it explains.

How does this apply to something like string theory? String theory seems to be incredibly complicated: ten dimensions, curled up in a truly vast number of different ways, giving rise to whole spectrums of particles.

That said, the basic idea is quite simple. String theory asks the question: what if, in addition to fundamental point-particles (zero dimensional objects), there were fundamental objects of other dimensions? That idea leads to complicated consequences: if your theory is going to produce all the particles of the real world then you need the ten dimensions and the supersymmetry and yadda yadda. But the basic idea is simple to describe. An elegant theory can have very complicated consequences, but still be simple to describe.

This, broadly, is the sort of explanation theoretical physicists look for. Math is the kind of field where the same basic systems can describe very complex phenomena. Since theoretical physics is about describing the world in terms of math, the right explanation is usually the most elegant.

This can occasionally trip physicists up when they migrate to other careers. In biology, for example, the elegant solution is often not the right one, because evolution doesn’t care about elegance: evolution just grabs whatever is within reach. Financial systems and economics occasionally have similar problems. All this is to say that while elegance is an important thing for a physicist to strive for, sometimes we have to be careful about it.

Blackboards, Again

Recently I had the opportunity to give a blackboard talk. I’ve talked before about the value of blackboards, how they facilitate collaboration and can even be used to get work done. What I didn’t feel the need to explain was their advantages when giving a talk.

No, the blackboard behind me isn't my talk.

No, the blackboard behind me isn’t my talk.

When I mentioned I was giving a blackboard talk, some of my friends in other fields were incredulous.

“Why aren’t you using PowerPoint? Do you people hate technology?”

So why do theorists (and mathematicians) do blackboard talks, when many other fields don’t?

Typically, a chemist can’t bring chemicals to a talk. A biologist can’t bring a tank of fruit flies or zebrafish, and a psychologist probably shouldn’t bring in a passel of college student test subjects. As a theorist though, our test subjects are equations, and we can absolutely bring them into the room.

In the most ideal case, a talk by a theorist walks you through their calculation, reproducing it on the blackboard in enough detail that you can not only follow along, but potentially do the calculation yourself. While it’s possible to set up a calculation step by step in PowerPoint, you don’t have the same flexibility to erase and add to your equations, which becomes especially important if you need to clarify a point in response to a question.

Blackboards also often give you more space than a single slide. While your audience still only pays attention to a slide-sized area of the board at one time, you can put equations up in one area, move away, and then come back to them later. If you leave important equations up, people can remind themselves of them on their own time, without having to hold everybody up while you scroll back through the slides to the one they want to see.

Using a blackboard well is a fine art, and one I’m only beginning to learn. You have to know what to erase and what to leave up, when to pause to allow time to write or ask questions, and what to say while you’re erasing the board. You need to use all the quirks of the medium to your advantage, to show people not just what you did, but how and why you did it.

That’s why we use blackboards. And if you ask why we can’t do the same things with whiteboards, it’s because whiteboards are terrible. Everybody knows that.