Tag Archives: DoingScience

When and How Scientists Reach Out

You’ve probably heard of the myth of the solitary scientist. While Newton might have figured out calculus isolated on his farm, most scientists work better when they communicate. If we reach out to other scientists, we can make progress a lot faster.

Even if you understand that, you might not know what that reaching out actually looks like. I’ve seen far too many crackpots who approach scientific communication like a spammer: sending out emails to everyone in a department, commenting in every vaguely related comment section they can find. While commercial spammers hope for a few gullible people among the thousands they contact, that kind of thing doesn’t benefit crackpots. As far as I can tell, they communicate that way because they genuinely don’t know any better.

So in this post, I want to give a road map for how we scientists reach out to other scientists. Keep these steps in mind, and if you ever need to reach out to a scientist you’ll know what to do.

First, decide what you want to know. This may sound obvious, but sometimes people skip this step. We aren’t communicating just to communicate, but because we want to learn something from the other person. Maybe it’s a new method or idea, maybe we just want confirmation we’re on the right track. We don’t reach out just to “show our theory”, but because we hope to learn something from the response.

Then, figure out who might know it. To do this, we first need to decide how specialized our question is. We often have questions about specific papers: a statement we don’t understand, a formula that seems wrong, or a method that isn’t working. For those, we contact an author from that paper. Other times, the question hasn’t been addressed in a paper, but does fall under a specific well-defined topic: a particular type of calculation, for example. For those we seek out a specialist on that specific topic. Finally, sometimes the question is more general, something anyone in our field might in principle know but we happen not to. For that kind of question, we look for someone we trust, someone we have a prior friendship with and feel comfortable asking “dumb questions”. These days, we can supplement that with platforms like PhysicsOverflow that let us post technical questions and invite anyone to respond.

Note that, for all of these, there’s some work to do first. We need to read the relevant papers, bone up on a topic, even check Wikipedia sometimes. We need to put in enough work to at least try to answer our question, so that we know exactly what we need the other person for.

Finally, contact them appropriately. Papers will usually give contact information for one, or all, of the authors. University websites will give university emails. We’d reach out with something like that first, and switch to personal email (or something even more casual, like Skype or social media) only for people we already have a track record of communicating with in that way.

By posing and directing our questions well, scientists can reach out and get help when we struggle. Science is a team effort, we’re stronger when we work together.

To Elliptics and Beyond!

I’ve been busy running a conference this week, Elliptics and Beyond.

After Amplitudes was held online this year, a few of us at the Niels Bohr Institute were inspired. We thought this would be the perfect time to hold a small online conference, focused on the Calabi-Yaus that have been popping up lately in Feynman diagrams. Then we heard from the organizers of Elliptics 2020. They had been planning to hold a conference in Mainz about elliptic integrals in Feynman diagrams, but had to postpone it due to the pandemic. We decided to team up and hold a joint conference on both topics: the elliptic integrals that are just starting to be understood, and the mysterious integrals that lie beyond. Hence, Elliptics and Beyond.

I almost suggested Buzz Lightyear for the logo but I chickened out

The conference has been fun thus far. There’s been a mix of review material bringing people up to speed on elliptic integrals and exciting new developments. Some are taking methods that have been successful in other areas and generalizing them to elliptic integrals, others have been honing techniques for elliptics to make them “production-ready”. A few are looking ahead even further, to higher-genus amplitudes in string theory and Calabi-Yaus in Feynman diagrams.

We organized the conference along similar lines to Zoomplitudes, but with a few experiments of our own. Like Zoomplitudes, we made a Slack space for the conference, so people could chat physics outside the talks. Ours was less active, though. I suspect that kind of space needs a critical mass of people, and with a smaller conference we may just not have gotten there. Having fewer people did allow us a more relaxed schedule, which in turn meant we could mostly keep things on-time. We had discussion sessions in the morning (European time), with talks in the afternoon, so almost everyone could make the talks at least. We also had a “conference dinner”, which went much better than I would have expected. We put people randomly into Zoom Breakout Rooms of five or six, to emulate the tables of an in-person conference, and folks chatted while eating their (self-brought of course) dinner. People seemed to really enjoy the chance to just chat casually with the other folks at the conference. If you’re organizing an online conference soon, I’d recommend trying it!

Holding a conference online means that a lot of people can attend who otherwise couldn’t. We had over a hundred people register, and while not all of them showed up there were typically fifty or sixty people on the Zoom session. Some of these were specialists in elliptics or Calabi-Yaus who wouldn’t ordinarily make it to a conference like this. Others were people from the rest of the amplitudes field who joined for parts of the conference that caught their eye. But surprisingly many weren’t even amplitudeologists, but students and young researchers in a variety of topics from all over the world. Some seemed curious and eager to learn, others I suspect just needed to say they had been to a conference. Both are responding to a situation where suddenly conference after conference is available online, free to join. It will be interesting to see if, and how, the world adapts.

Grants at the Other End

I’m a baby academic. Two years ago I got my first real grant, a Marie Curie Individual Fellowship from the European Union. Applying for it was a complicated process, full of Word templates and mismatched expectations. Two years later the grant is over, and I get another new experience: grant reporting.

Writing a report after a grant is sort of like applying for a grant. Instead of summarizing and justifying what you intend to do, you summarize and justify what you actually did. There are also Word templates. Grant reports are probably easier than grant applications: you don’t have to “hook” your audience or show off. But they are harder in one aspect: they highlight the different ways different fields handle uncertainty.

If you do experiments, having a clear plan makes sense. You buy special equipment and hire postdocs and even technicians to do specific jobs. Your experiments may or may not find what you hope for, but at least you can try to do them on schedule, and describe the setbacks when you can’t.

As a theorist, you’re more nimble. Your equipment are computers, your postdocs have their own research. Overall, it’s easy to pick up new projects as new ideas come in. As a result, your plans change more. New papers might inspire you to try new things. They might also discourage you, if you learn the idea you had won’t actually work. The field can move fast, and you want to keep up with it.

Writing my first grant report will be interesting. I’ll need to thread the gap between expectations and reality, to look back on my progress and talk about why. And of course, I have to do it in Microsoft Word.

Kicking Students Out of Their Homes During a Pandemic: A Bad Idea

I avoid talking politics on this blog. There are a few issues, though, where I feel not just able, but duty-bound, to speak out. Those are issues affecting graduate students.

This week, US Immigration and Customs Enforcement (ICE) announced that, if a university switched to online courses as a response to COVID-19, international students would have to return to their home countries or transfer to a school that still teaches in-person.

This is already pretty unreasonable for many undergrads. But think about PhD students.

Suppose you’re a foreign PhD student at a US university. Maybe your school is already planning to have classes online this fall, like Harvard is. Maybe your school is planning to have classes in person, but will change its mind a few weeks in, when so many students and professors are infected that it’s clearly unreasonable to continue. Maybe your school never changes its mind, but your state does, and the school has to lock down anyway.

As a PhD student, you likely don’t live in the dorms. More likely you live in a shared house, or an apartment. You’re an independent adult. Your parents aren’t paying for you to go to school. Your school is itself a full-time job, one that pays (as little as the university thinks it can get away with).

What happens when your school goes online? If you need to leave the country?

You’d have to find some way out of your lease, or keep paying for it. You’d have to find a flight on short notice. You’d have to pack up all your belongings, ship or sell anything you can’t store, or find friends to hold on to it.

You’d have to find somewhere to stay in your “home country”. Some could move in with their parents temporarily, many can’t. Some of those who could in other circumstances, shouldn’t if they’re fleeing from an outbreak: their parents are likely older, and vulnerable to the virus. So you have to find a hotel, eventually perhaps a new apartment, far from what was until recently your home.

Reminder: you’re doing all of this on a shoestring budget, because the university pays you peanuts.

Can you transfer instead? In a word, no.

PhD students are specialists. They’re learning very specific things from very specific people. Academics aren’t the sort of omnidisciplinary scientists you see in movies. Bruce Banner or Tony Stark could pick up a new line of research on a whim, real people can’t. This is why, while international students may be good at the undergraduate level, they’re absolutely necessary for PhDs. When only three people in the world study the thing you want to study, you don’t have the luxury of staying in your birth country. And you can’t just transfer schools when yours goes online.

It feels like the people who made this decision didn’t think about any of this. That they don’t think grad students matter, or forgot they exist altogether. It seems frustratingly common for policy that affects grad students to be made by people who know nothing about grad students, and that baffles me. PhDs are a vital part of the academic career, without them universities in their current form wouldn’t even exist. Ignoring them is like if hospital policy ignored residencies.

I hope that this policy gets reversed, or halted, or schools find some way around it. At the moment, anyone starting school in the US this fall is in a very tricky position. And anyone already there is in a worse one.

As usual, I’m going to ask that the comments don’t get too directly political. As a partial measure to tone things down, I’d like to ask you to please avoid mentioning any specific politicians, political parties, or political ideologies. Feel free to talk instead about your own experiences: how this policy is likely to affect you, or your loved ones. Please also feel free to talk more technically on the policy/legal side. I’d like to know what universities can do to work around this, and whether there are plausible paths to change or halt the policy. Please be civil, and be kind to your fellow commenters.

The Citation Motivation Situation

Citations are the bread and butter of academia, or maybe its prison cigarettes. They link us together, somewhere between a map to show us the way and an informal currency. They’re part of how the world grades us, a measure more objective than letters from our peers but that’s not saying much. It’s clear why we we want to be cited, but why do we cite others?

For more reasons than you’d expect.

First, we cite to respect priority. Since the dawn of science, we’ve kept track not only of what we know, but of who figured it out first. If we use an idea in our paper, we cite its origin: the paper that discovered or invented it. We don’t do this for the oldest and most foundational ideas: nobody cites Einstein for relativity. But if the idea is at all unusual, we make sure to give credit where credit is due.

Second, we cite to substantiate our claims. Academic papers don’t stand on their own: they depend on older proofs and prior discoveries. If we make a claim that was demonstrated in older work, we don’t need to prove it again. By citing the older work, we let the reader know where to look. If they doubt our claim, they can look at the older paper and see what went wrong.

Those two are the most obvious uses of citations, but there are more. Another important use is to provide context. Academic work doesn’t stand alone: we choose what we work on in part based on how it relates to other work. As such, it’s important to cite that other work, to help readers understand our motivation. When we’re advancing the state of the art, we need to tell the reader what that state of the art is. When we’re answering a question or solving a problem, we can cite the paper that asked the question or posed the problem. When we’re introducing a new method or idea, we need to clearly say what’s new about it: how it improves on older, similar ideas.

Scientists are social creatures. While we often have a scientific purpose in mind, citations also follow social conventions. These vary from place to place, field to field, and sub-field to sub-field. Mention someone’s research program, and you might be expected to cite every paper in that program. Cite one of a pair of rivals, and you should probably cite the other one too. Some of these conventions are formalized in the form of “citeware“, software licenses that require citations, rather than payments, to use. Others come from unspoken cultural rules. Citations are a way to support each other, something that can slightly improve another’s job prospects at no real cost to your own. It’s not surprising that they ended up part of our culture, well beyond their pure academic use.

Zoomplitudes Retrospective

During Zoomplitudes (my field’s big yearly conference, this year on Zoom) I didn’t have time to write a long blog post. I said a bit about the format, but didn’t get a chance to talk about the science. I figured this week I’d go back and give a few more of my impressions. As always, conference posts are a bit more technical than my usual posts, so regulars be warned!

The conference opened with a talk by Gavin Salam, there as an ambassador for LHC physics. Salam pointed out that, while a decent proportion of speakers at Amplitudes mention the LHC in their papers, that fraction has fallen over the years. (Another speaker jokingly wondered which of those mentions were just in the paper’s introduction.) He argued that there is still useful work for us, LHC measurements that will require serious amplitudes calculations to understand. He also brought up what seems like the most credible argument for a new, higher-energy collider: that there are important properties of the Higgs, in particular its interactions, that we still have not observed.

The next few talks hopefully warmed Salam’s heart, as they featured calculations for real-world particle physics. Nathaniel Craig and Yael Shadmi in particular covered the link between amplitudes and Standard Model Effective Field Theory (SMEFT), a method to systematically characterize corrections beyond the Standard Model. Shadmi’s talk struck me because the kind of work she described (building the SMEFT “amplitudes-style”, directly from observable information rather than more complicated proxies) is something I’d seen people speculate about for a while, but which hadn’t been done until quite recently. Now, several groups have managed it, and look like they’ve gotten essentially “all the way there”, rather than just partial results that only manage to replicate part of the SMEFT. Overall it’s much faster progress than I would have expected.

After Shadmi’s talk was a brace of talks on N=4 super Yang-Mills, featuring cosmic Galois theory and an impressively groan-worthy “origin story” joke. The final talk of the day, by Hofie Hannesdottir, covered work with some of my colleagues at the NBI. Due to coronavirus I hadn’t gotten to hear about this in person, so it was good to hear a talk on it, a blend of old methods and new priorities to better understand some old discoveries.

The next day focused on a topic that has grown in importance in our community, calculations for gravitational wave telescopes like LIGO. Several speakers focused on new methods for collisions of spinning objects, where a few different approaches are making good progress (Radu Roiban’s proposal to use higher-spin field theory was particularly interesting) but things still aren’t quite “production-ready”. The older, post-Newtonian method is still very much production-ready, as evidenced by Michele Levi’s talk that covered, among other topics, our recent collaboration. Julio Parra-Martinez discussed some interesting behavior shared by both supersymmetric and non-supersymmetric gravity theories. Thibault Damour had previously expressed doubts about use of amplitudes methods to answer this kind of question, and part of Parra-Martinez’s aim was to confirm the calculation with methods Damour would consider more reliable. Damour (who was actually in the audience, which I suspect would not have happened at an in-person conference) had already recanted some related doubts, but it’s not clear to me whether that extended to the results Parra-Martinez discussed (or whether Damour has stated the problem with his old analysis).

There were a few talks that day that didn’t relate to gravitational waves, though this might have been an accident, since both speakers also work on that topic. Zvi Bern’s talk linked to the previous day’s SMEFT discussion, with a calculation using amplitudes methods of direct relevance to SMEFT researchers. Clifford Cheung’s talk proposed a rather strange/fun idea, conformal symmetry in negative dimensions!

Wednesday was “amplituhedron day”, with a variety of talks on positive geometries and cluster algebras. Featured in several talks was “tropicalization“, a mathematical procedure that can simplify complicated geometries while still preserving essential features. Here, it was used to trim down infinite “alphabets” conjectured for some calculations into a finite set, and in doing so understand the origin of “square root letters”. The day ended with a talk by Nima Arkani-Hamed, who despite offering to bet that he could finish his talk within the half-hour slot took almost twice that. The organizers seemed to have planned for this, since there was one fewer talk that day, and as such the day ended at roughly the usual time regardless.

We also took probably the most unique conference photo I will ever appear in.

For lack of a better name, I’ll call Thursday’s theme “celestial”. The day included talks by cosmologists (including approaches using amplitudes-ish methods from Daniel Baumann and Charlotte Sleight, and a curiously un-amplitudes-related talk from Daniel Green), talks on “celestial amplitudes” (amplitudes viewed from the surface of an infinitely distant sphere), and various talks with some link to string theory. I’m including in that last category intersection theory, which has really become its own thing. This included a talk by Simon Caron-Huot about using intersection theory more directly in understanding Feynman integrals, and a talk by Sebastian Mizera using intersection theory to investigate how gravity is Yang-Mills squared. Both gave me a much better idea of the speakers’ goals. In Mizera’s case he’s aiming for something very ambitious. He wants to use intersection theory to figure out when and how one can “double-copy” theories, and might figure out why the procedure “got stuck” at five loops. The day ended with a talk by Pedro Vieira, who gave an extremely lucid and well-presented “blackboard-style” talk on bootstrapping amplitudes.

Friday was a grab-bag of topics. Samuel Abreu discussed an interesting calculation using the numerical unitarity method. It was notable in part because renormalization played a bigger role than it does in most amplitudes work, and in part because they now have a cool logo for their group’s software, Caravel. Claude Duhr and Ruth Britto gave a two-part talk on their work on a Feynman integral coaction. I’d had doubts about the diagrammatic coaction they had worked on in the past because it felt a bit ad-hoc. Now, they’re using intersection theory, and have a clean story that seems to tie everything together. Andrew McLeod talked about our work on a Feynman diagram Calabi-Yau “bestiary”, while Cristian Vergu had a more rigorous understanding of our “traintrack” integrals.

There are two key elements of a conference that are tricky to do on Zoom. You can’t do a conference dinner, so you can’t do the traditional joke-filled conference dinner speech. The end of the conference is also tricky: traditionally, this is when everyone applauds the organizers and the secretaries are given flowers. As chair for the last session, Lance Dixon stepped up to fill both gaps, with a closing speech that was both a touching tribute to the hard work of organizing the conference and a hilarious pile of in-jokes, including a participation award to Arkani-Hamed for his (unprecedented, as far as I’m aware) perfect attendance.

The Sum of Our Efforts

I got a new paper out last week, with Andrew McLeod, Henrik Munch, and Georgios Papathanasiou.

A while back, some collaborators and I found an interesting set of Feynman diagrams that we called “Omega”. These Omega diagrams were fun because they let us avoid one of the biggest limitations of particle physics: that we usually have to compute approximations, diagram by diagram, rather than finding an exact answer. For these Omegas, we figured out how to add all the infinite set of Omega diagrams up together, with no approximation.

One implication of this was that, in principle, we now knew the answer for each individual Omega diagram, far past what had been computed before. However, writing down these answers was easier said than done. After some wrangling, we got the answer for each diagram in terms of an infinite sum. But despite tinkering with it for a while, even our resident infinite sum expert Georgios Papathanasiou couldn’t quite sum them up.

Naturally, this made me think the sums would make a great Master’s project.

When Henrik Munch showed up looking for a project, Andrew McLeod and I gave him several options, but he settled on the infinite sums. Impressively, he ended up solving the problem in two different ways!

First, he found an old paper none of us had seen before, that gave a general method for solving that kind of infinite sum. When he realized that method was really annoying to program, he took the principle behind it, called telescoping, and came up with his own, simpler method, for our particular case.

Picture an old-timey folding telescope. It might be long when fully extended, but when you fold it up each piece fits inside the previous one, resulting in a much smaller object. Telescoping a sum has the same spirit. If each pair of terms in a sum “fit together” (if their difference is simple), you can rearrange them so that most of the difficulty “cancels out” and you’re left with a much simpler sum.

Henrik’s telescoping idea worked even better than expected. We found that we could do, not just the Omega sums, but other sums in particle physics as well. Infinite sums are a very well-studied field, so it was interesting to find something genuinely new.

The rest of us worked to generalize the result, to check the examples and to put it in context. But the core of the work was Henrik’s. I’m really proud of what he accomplished. If you’re looking for a PhD student, he’s on the market!

The Point of a Model

I’ve been reading more lately, partially for the obvious reasons. Mostly, I’ve been catching up on books everyone else already read.

One such book is Daniel Kahneman’s “Thinking, Fast and Slow”. With all the talk lately about cognitive biases, Kahneman’s account of his research on decision-making was quite familiar ground. The book turned out to more interesting as window into the culture of psychology research. While I had a working picture from psychologist friends in grad school, “Thinking, Fast and Slow” covered the other side, the perspective of a successful professor promoting his field.

Most of this wasn’t too surprising, but one passage struck me:

Several economists and psychologists have proposed models of decision making that are based on the emotions of regret and disappointment. It is fair to say that these models have had less influence than prospect theory, and the reason is instructive. The emotions of regret and disappointment are real, and decision makers surely anticipate these emotions when making their choices. The problem is that regret theories make few striking predictions that would distinguish them from prospect theory, which has the advantage of being simpler. The complexity of prospect theory was more acceptable in the competition with expected utility theory because it did predict observations that expected utility theory could not explain.

Richer and more realistic assumptions do not suffice to make a theory successful. Scientists use theories as a bag of working tools, and they will not take on the burden of a heavier bag unless the new tools are very useful. Prospect theory was accepted by many scholars not because it is “true” but because the concepts that it added to utility theory, notably the reference point and loss aversion, were worth the trouble; they yielded new predictions that turned out to be true. We were lucky.

Thinking Fast and Slow, page 288

Kahneman is contrasting three theories of decision making here: the old proposal that people try to maximize their expected utility (roughly, the benefit they get in future), his more complicated “prospect theory” that takes into account not only what benefits people get but their attachment to what they already have, and other more complicated models based on regret. His theory ended up more popular, both than the older theory and than the newer regret-based models.

Why did his theory win out? Apparently, not because it was the true one: as he says, people almost certainly do feel regret, and make decisions based on it. No, his theory won because it was more useful. It made new, surprising predictions, while being simpler and easier to use than the regret-based models.

This, a theory defeating another without being “more true”, might bug you. By itself, it doesn’t bug me. That’s because, as a physicist, I’m used to the idea that models should not just be true, but useful. If we want to test our theories against reality, we have a large number of “levels” of description to choose from. We can “zoom in” to quarks and gluons, or “zoom out” to look at atoms, or molecules, or polymers. We have to decide how much detail to include, and we have real pragmatic reasons for doing so: some details are just too small to measure!

It’s not clear Kahneman’s community was doing this, though. That is, it doesn’t seem like he’s saying that regret and disappointment are just “too small to be measured”. Instead, he’s saying that they don’t seem to predict much differently from prospect theory, and prospect theory is simpler to use.

Ok, we do that in physics too. We like working with simpler theories, when we have a good excuse. We’re just careful about it. When we can, we derive our simpler theories from more complicated ones, carving out complexity and estimating how much of a difference it would have made. Do this carefully, and we can treat black holes as if they were subatomic particles. When we can’t, we have what we call “phenomenological” models, models built up from observation and not from an underlying theory. We never take such models as the last word, though: a phenomenological model is always viewed as temporary, something to bridge a gap while we try to derive it from more basic physics.

Kahneman doesn’t seem to view prospect theory as temporary. It doesn’t sound like anyone is trying to derive it from regret theory, or to make regret theory easier to use, or to prove it always agrees with regret theory. Maybe they are, and Kahneman simply doesn’t think much of their efforts. Either way, it doesn’t sound like a major goal of the field.

That’s the part that bothered me. In physics, we can’t always hope to derive things from a more fundamental theory, some theories are as fundamental as we know. Psychology isn’t like that: any behavior people display has to be caused by what’s going on in their heads. What Kahneman seems to be saying here is that regret theory may well be closer to what’s going on in people’s heads, but he doesn’t care: it isn’t as useful.

And at that point, I have to ask: useful for what?

As a psychologist, isn’t your goal ultimately to answer that question? To find out “what’s going on in people’s heads”? Isn’t every model you build, every theory you propose, dedicated to that question?

And if not, what exactly is it “useful” for?

For technology? It’s true, “Thinking Fast and Slow” describes several groups Kahneman advised, most memorably the IDF. Is the advantage of prospect theory, then, its “usefulness”, that it leads to better advice for the IDF?

I don’t think that’s what Kahneman means, though. When he says “useful”, he doesn’t mean “useful for advice”. He means it’s good for giving researchers ideas, good for getting people talking. He means “useful for designing experiments”. He means “useful for writing papers”.

And this is when things start to sound worryingly familiar. Because if I’m accusing Kahneman’s community of giving up on finding the fundamental truth, just doing whatever they can to write more papers…well, that’s not an uncommon accusation in physics as well. If the people who spend their lives describing cognitive biases are really getting distracted like that, what chance does, say, string theory have?

I don’t know how seriously to take any of this. But it’s lurking there, in the back of my mind, that nasty, vicious, essential question: what are all of our models for?

Bonus quote, for the commenters to have fun with:

I have yet to meet a successful scientist who lacks the ability to exaggerate the importance of what he or she is doing, and I believe that someone who lacks a delusional sense of significance will wilt in the face of repeated experiences of multiple small failures and rare successes, the fate of most researchers.

Thinking Fast and Slow, page 264

Thoughts on Doing Science Remotely

In these times, I’m unusually lucky.

I’m a theoretical physicist. I don’t handle goods, or see customers. Other scientists need labs, or telescopes: I just need a computer and a pad of paper. As a postdoc, I don’t even teach. In the past, commenters have asked me why I don’t just work remotely. Why go to conferences, why even go to the office?

With COVID-19, we’re finding out.

First, the good: my colleagues at the Niels Bohr Institute have been hard at work keeping everyone connected. Our seminars have moved online, where we hold weekly Zoom seminars jointly with Iceland, Uppsala and Nordita. We have a “virtual coffee room”, a Zoom room that’s continuously open with “virtual coffee breaks” at 10 and 3:30 to encourage people to show up. We’re planning virtual colloquia, and even a virtual social night with Jackbox games.

Is it working? Partially.

The seminars are the strongest part. Remote seminars let us bring in speakers from all over the world (time zones permitting). They let one seminar serve the needs of several different institutes. Most of the basic things a seminar needs (slides, blackboards, ability to ask questions, ability to clap) are present on online platforms, particularly Zoom. And our seminar organizers had the bright idea to keep the Zoom room open after the talk, which allows the traditional “after seminar conversation with the speaker” for those who want it.

Still, the setup isn’t as good as it could be. If the audience turns off their cameras and mics, the speaker can feel like they’re giving a talk to an empty room. This isn’t just awkward, it makes the talk worse: speakers improve when they can “feel the room” and see what catches their audience’s interest. If the audience keeps their cameras or mics on instead, it takes a lot of bandwidth, and the speaker still can’t really feel the room. I don’t know if there’s a good solution here, but it’s worth working on.

The “virtual coffee room” is weaker. It was quite popular at first, but as time went on fewer and fewer people (myself included) showed up. In contrast, my wife’s friends at Waterloo do a daily cryptic crossword, and that seems to do quite well. What’s the difference? They have real crosswords, we don’t have real coffee.

I kid, but only a little. Coffee rooms and tea breaks work because of a core activity, a physical requirement that brings people together. We value them for their social role, but that role on its own isn’t enough to get us in the door. We need the excuse: the coffee, the tea, the cookies, the crossword. Without that shared structure, people just don’t show up.

Getting this kind of thing right is more important than it might seem. Social activities help us feel better, they help us feel less isolated. But more than that, they help us do science better.

That’s because science works, at least in part, through serendipity.

You might think of scientific collaboration as something we plan, and it can be sometimes. Sometimes we know exactly what we’re looking for: a precise calculation someone else can do, a question someone else can answer. Sometimes, though, we’re helped by chance. We have random conversations, different people in different situations, coffee breaks and conference dinners, and eventually someone brings up an idea we wouldn’t have thought of on our own.

Other times, chance helps by providing an excuse. I have a few questions rattling around in my head that I’d like to ask some of my field’s big-shots, but that don’t feel worth an email. I’ve been waiting to meet them at a conference instead. The advantage of those casual meetings is that they give an excuse for conversation: we have to talk about something, it might as well be my dumb question. Without that kind of causal contact, it feels a lot harder to broach low-stakes topics.

None of this is impossible to do remotely. But I think we need new technology (social or digital) to make it work well. Serendipity is easy to find in person, but social networks can imitate it. Log in to facebook or tumblr looking for your favorite content, and you face a pile of ongoing conversations. Looking through them, you naturally “run into” whatever your friends are talking about. I could see something similar for academia. Take something like the list of new papers on arXiv, then run a list of ongoing conversations next to it. When we check the arXiv each morning, we could see what our colleagues were talking about, and join in if we see something interesting. It would be a way to stay connected that would keep us together more, giving more incentive and structure beyond simple loneliness, and lead to the kind of accidental meetings that science craves. You could even graft conferences on to that system, talks in the middle with conversation threads on the side.

None of us know how long the pandemic will last, or how long we’ll be asked to work from home. But even afterwards, it’s worth thinking about the kind of infrastructure science needs to work remotely. Some ideas may still be valuable after all this is over.

4gravitons, Spinning Up

I had a new paper out last week, with Michèle Levi and Andrew McLeod. But to explain it, I’ll need to clarify something about our last paper.

Two weeks ago, I told you that Andrew and Michèle and I had written a paper, predicting what gravitational wave telescopes like LIGO see when black holes collide. You may remember that LIGO doesn’t just see colliding black holes: it sees colliding neutron stars too. So why didn’t we predict what happens when neutron stars collide?

Actually, we did. Our calculation doesn’t just apply to black holes. It applies to neutron stars too. And not just neutron stars: it applies to anything of roughly the right size and shape. Black holes, neutron stars, very large grapefruits…

LIGO’s next big discovery

That’s the magic of Effective Field Theory, the “zoom lens” of particle physics. Zoom out far enough, and any big, round object starts looking like a particle. Black holes, neutron stars, grapefruits, we can describe them all using the same math.

Ok, so we can describe both black holes and neutron stars. Can we tell the difference between them?

In our last calculation, no. In this one, yes!

Effective Field Theory isn’t just a zoom lens, it’s a controlled approximation. That means that when we “zoom out” we don’t just throw out anything “too small to see”. Instead, we approximate it, estimating how big of an effect it can have. Depending on how precise we want to be, we can include more and more of these approximated effects. If our estimates are good, we’ll include everything that matters, and get a good approximation for what we’re trying to observe.

At the precision of our last calculation, a black hole and a neutron star still look exactly the same. Our new calculation aims for a bit higher precision though. (For the experts: we’re at a higher order in spin.) The higher precision means that we can actually see the difference: our result changes for two colliding black holes versus two colliding grapefruits.

So does that mean I can tell you what happens when two neutron stars collide, according to our calculation? Actually, no. That’s not because we screwed up the calculation: it’s because some of the properties of neutron stars are unknown.

The Effective Field Theory of neutron stars has what we call “free parameters”, unknown variables. People have tried to estimate some of these (called “Love numbers” after the mathematician A. E. H. Love), but they depend on the details of how neutron stars work: what stuff they contain, how that stuff is shaped, and how it can move. To find them out, we probably can’t just calculate: we’ll have to measure, observe an actual neutron star collision and see what the numbers actually are.

That’s one of the purposes of gravitational wave telescopes. It’s not (as far as I know) something LIGO can measure. But future telescopes, with more precision, should be able to. By watching two colliding neutron stars and comparing to a high-precision calculation, physicists will better understand what those neutron stars are made of. In order to do that, they will need someone to do that high-precision calculation. And that’s why people like me are involved.