Tag Archives: academia

On the Care and Feeding of International Employees

Science and scholarship are global. If you want to find out the truth about the universe, you’ll have to employ the people best at figuring out that truth, regardless of where they come from. Research shuffles people around, driving them together to collaborate and apart to share their expertise.

(If you don’t care about figuring out the truth, and just want to make money? You still may want international employees. For plenty of jobs, the difference between the best person in the world and the best person in your country can be quite substantial.)

How do you get these international employees? You could pay them a lot, I guess, but that’s by definition expensive, and probably will annoy the locals. Instead, most of what you need to do to attract international employees isn’t to give them extra rewards: instead, it’s more important to level the playing field, and cover for the extra disadvantages an international employee will have.

You might be surprised when I mention disadvantages, but while international employees may be talented people, that doesn’t make moving to another country easy. If you stay in the same country you were born, you get involved in that country’s institutions in a regular way. Your rights and responsibilities, everything from driving to healthcare to taxes, are set up gradually over the course of your life. For someone moving to a new country, that means all of this has to be set up all at once.

This means that countries that can process these things quickly are much better for international employees. If your country takes six months to register someone for national healthcare, then new employees are at risk during that time or will have to pay extra for private insurance. If a national ID number is required to get a bank account, then whatever processing time that ID number takes must pass before the new employee can get paid. It also matters if the rules are clearly and consistently communicated, as new international employees can waste a lot of time and money if they’re given incorrect advice, or if different bureaucrats enforce different rules at their own discretion.

It also means that employers have an advantage if they can smooth entry into these institutions. In some countries it can be quite hard to find a primary care physician, as most people have the same doctor as their parents, switching only when a doctor retires. When I worked with the Perimeter Institute, they had a relationship with a local clinic that would accept their new employees as clients. In a city where it was otherwise quite hard to find a doctor, that was a real boon. Employers can also offer consistent advice even when their government doesn’t. They can keep track of their employees experiences and make reliable guides for how to navigate the system. If they can afford it, they can even keep an immigration lawyer on staff to advise about these questions.

An extremely important institution is the language itself. Moving internationally will often involve moving somewhere where you don’t speak the language, or don’t speak it very well. This gives countries an advantage if their immigrant-facing institutions are proficient in a language that’s common internationally, which at the moment largely means English. It also means countries have a big advantage if their immigrant-facing institutions are digital. If you communicate with immigrants with text, they can find online translations and at least try to figure things out. If you communicate in person, or worse through a staticky phone line, then you will try the patience even of people who do passably speak the language.

In the long term, of course, one cannot get by in one’s native language alone. As such, it is also important for countries to have good ways for people to learn the language. While I lived there, Denmark went back and forth on providing free language lessons for recent immigrants, sometimes providing them and sometimes not.

All of these things become twice as important in the case of spouses. You might think the idea that a country or employer should help out a new employee’s spouse is archaic, a product of an era of housewives discouraged from supporting themselves. But it is precisely because we don’t live in such an era that countries and employers need to take spouses into account. For an employer, hiring someone from another country is already an unusual event. Two partners getting hired to move to the same country by different employers at the same time is, barring special arrangements, extremely unlikely. That means that spouses of international employees should not have to wait for an employer to give them the same rights as their spouse: they need the same right to healthcare and employment and the like as their spouse, on arrival, so that they can find jobs and integrate without an unfair disadvantage. An employer can level the playing field further. The University of Copenhagen’s support for international spouses included social events (important because it’s hard to make new friends in a new country without the benefit of work friends), resume help (because each country has different conventions and expectations for job seekers), and even legal advice. At minimum, every resource you provide your employees that could in principle also be of use to their spouses (language classes, help with bureaucracy) should be considered.

In all your planning, as a country or an employer, keep in mind that not everyone has the same advantages. You can’t assume that someone moving to a new country will be able to integrate on their own. You have to help them, if not for fairness’ sake, then because if you don’t you won’t keep getting international employees to come at all.

What RIBs Could Look Like

The journal Nature recently published an opinion piece about a new concept for science funding called Research Impact Bonds (or RIBs).

Normally, when a government funds something, they can’t be sure it will work. They pay in advance, and have to guess whether a program will do what they expect, or whether a project will finish on time. Impact bonds are a way for them to pay afterwards, so they only pay for projects that actually deliver. Instead, the projects are funded by private investors, who buy “impact bonds” that guarantee them a share of government funding if the project is successful. Here’s an example given in the Nature piece:

For instance, say the Swiss government promises to pay up to one million Swiss francs (US$1.1 million) to service providers that achieve a measurable outcome, such as reducing illiteracy in a certain population by 5%, within a specified number of years. A broker finds one or more service providers that think they can achieve this at a cost of, say, 900,000 francs, as well as investors who agree to pay these costs up front — thus taking on the risk of the project — for a potential 10% gain if successful. If the providers achieve their goals, the government pays 990,000 francs: 900,000 francs for the work and a 90,000-franc investment return. If the project does not succeed, the investors lose their money, but the government does not.

The author of the piece, Michael Hill, thinks that this could be a new way for governments to fund science. In his model, scientists would apply to the government to propose new RIBs. The projects would have to have specific goals and time-frames: “measure the power of this cancer treatment to this accuracy in five years”, for example. If the government thinks the goal is valuable, they commit to paying some amount of money if the goal is reached. Then investors can decide whether the investment is worthwhile. The projects they expect to work get investor money, and if they do end up working the investors get government money. The government only has to pay if the projects work, but the scientists get paid regardless.

Ok, what’s the catch?

One criticism I’ve seen is that this kind of model could only work for very predictable research, maybe even just for applied research. While the author admits RIBs would only be suitable for certain sorts of projects, I think the range is wider than you might think. The project just has to have a measurable goal by a specified end date. Many particle physics experiments work that way: a dark matter detector, for instance, is trying to either rule out or detect dark matter to a certain level of statistical power within a certain run time. Even “discovery” machines, that we build to try to discover the unexpected, usually have this kind of goal: a bigger version of the LHC, for instance, might try to measure the coupling of Higgs bosons to a certain accuracy.

There are a few bigger issues with this model, though. If you go through the math in Hill’s example, you’ll notice that if the project works, the government ends up paying one million Swiss francs for a service that only cost the provider 900,000 Swiss francs. Under a normal system, the government would only have had to pay 900,000. This gets compensated by the fact that not every project works, so the government only pays for some projects and not others. But investors will be aware of this, and that means the government can’t offer too many unrealistic RIBs: the greater the risk investors are going to take, the more return they’ll expect. On average then, the government would have to pay about as much as they would normally: the cost of the projects that succeed, plus enough money to cover the risk that some fail. (In fact, they’d probably pay a bit more, to give the investors a return on the investment.)

So the government typically won’t save money, at least not if they want to fund the same amount of research. Instead, the idea is that they will avoid risk. But it’s not at all clear to me that the type of risk they avoid is one they want to.

RIBs might appeal to voters: it might sound only fair that a government only funds the research that actually works. That’s not really a problem for the government itself, though: because governments usually pay for many small projects, they still get roughly as much success overall as they want, they just don’t get to pick where. Instead, RIBS put the government agency in a much bigger risk, the risk of unexpected success. As part of offering RIBs, the government would have to estimate how much money they would be able to pay when the projects end. They would want to fund enough projects so that, on average, they pay that amount of money. (Otherwise, they’d end up funding science much less than they do now!) But if the projects work out better than expected, then they’d have to pay much more than they planned. And government science agencies usually can’t do this. In many countries, they can’t plan far in advance at all: their budgets get decided by legislators year to year, and delays in decisions mean delays in funding. If an agency offered RIBs that were more successful than expected, they’d either have to cut funding somewhere else (probably firing a lot of people), or just default on their RIBs, weakening the concept for the next time they used them. These risks, unlike the risk of individual experiments not working, are risks that can really hurt government agencies.

Impact bonds typically have another advantage, in that they spread out decision-making. The Swiss government in Hill’s example doesn’t have to figure out which service providers can increase literacy, or how much it will cost them: it just puts up a budget, and lets investors and service providers figure out if they can make it work. This also serves as a hedge against corruption. If the government made the decisions, they might distribute funding for unrelated political reasons or even out of straight-up bribery. They’d also have to pay evaluators to figure things out. Investors won’t take bribes to lose money, so in theory would be better at choosing projects that will actually work, and would have a vested interest in paying for a good investigation.

This advantage doesn’t apply to Hill’s model of RIBs, though. In Hill’s model, scientists still need to apply to the government to decide which of their projects get offered as RIBs, so the government still needs to decide which projects are worth investing in. Then the scientists or the government need to take another step, and convince investors. The scientists in this equation effectively have to apply twice, which anyone who has applied for a government grant will realize is quite a lot of extra time and effort.

So overall, I don’t think Hills’ model of RIBs is useful, even for the purpose he imagines. It’s too risky for government science agencies to commit to payments like that, and it generates more, not less, work for scientists and the agency.

Hill’s model, though, isn’t the only way RIBs can work. And “avoiding risk” isn’t the only reason we might want them. There are two other reasons one might want RIBs, with very different-sounding motivations.

First, you might be pessimistic about mainstream science. Maybe you think scientists are making bad decisions, choosing ideas that either won’t pan out or won’t have sufficient impact, based more on fashion than on careful thought. You want to incentivize them to do better, to try to work out what impact they might have with some actual numbers and stand by their judgement. If that’s your perspective, you might be interested in RIBs for the same reason other people are interested in prediction markets: by getting investors involved, you have people willing to pay for an accurate estimate.

Second, you might instead be optimistic about mainstream science. You think scientists are doing great work, work that could have an enormous impact, but they don’t get to “capture that value”. Some projects might be essential to important, well-funded goals, but languish unrewarded. Others won’t see their value until long in the future, or will do so in unexpected ways. If scientists could fund projects based on their future impact, with RIBs, maybe they could fund more of this kind of work.

(I first started thinking about this perspective due to a talk by Sabrina Pasterski. The talk itself offended a lot of people, and had some pretty impractical ideas, like selling NFTs of important physics papers. But I think one part of the perspective, that scientists have more impact than we think, is worth holding on to.)

If you have either of those motivations, Hill’s model won’t help. But another kind of model perhaps could. Unlike Hill’s, it could fund much more speculative research, ideas where we don’t know the impact until decades down the line. To demonstrate, I’ll show how it could fund some very speculative research: the work of Peter van Nieuwenhuizen.

Peter van Nieuwenhuizen is one of the pioneers of the theory of supergravity, a theory that augments gravity with supersymmetric partner particles. From its beginnings in the 1970’s, the theory ended up having a major impact on string theory, and today they are largely thought of as part of the same picture of how the universe might work.

His work has, over time, had more practical consequences though. In the 2000’s, researchers working with supergravity noticed a calculational shortcut: they could do a complicated supergravity calculation as the “square” of a much simpler calculation in another theory, called Yang-Mills. Over time, they realized the shortcut worked not just for supergravity, but for ordinary gravity as well, and not just for particle physics calculations but for gravitational wave calculations. Now, their method may make an important contribution to calculations for future gravitational wave telescopes like the Einstein telescope, letting them measure properties of neutron stars.

With that in mind, imagine the following:

In 1967, Jocelyn Bell Burnell and Antony Hewish detected a pulsar, in one of the first direct pieces of evidence for the existence of neutron stars. Suppose that in the early 1970’s NASA decided to announce a future purchase of RIBs: in 2050, they would pay a certain amount to whoever was responsible for finding the equation of state of a neutron star, the formula that describes how its matter moves under pressure. They compute based on estimates of economic growth and inflation, and arrive at some suitably substantial number.

At the same time, but unrelatedly, van Nieuwenhuizen and collaborators sell RIBs. Maybe they use the proceeds to buy more computer time for their calculations, or to refund travel so they can more easily meet and discuss. They tell the buyers that, if some government later decides to reward their discoveries, the holders of the RIB would get a predetermined cut of the rewards.

The years roll by, and barring some unexpected medical advances the discoverers of supergravity die. In the meantime, researchers use their discovery to figure out how to make accurate predictions of gravitational waves from merging neutron stars. When the Einstein telescope turns out, it detects such a merger, and the accurate predictions let them compute the neutron star’s equation of state.

In 2050, then, NASA looks back. They make a list of everyone who contributed to the discovery of the neutron star’s equation of state, every result that was needed for the discovery, and try to estimate how important each contribution was. Then they spend the money they promised buying RIBs, up to the value for each contributor. This includes RIBs originally held by the investors in van Nieuwenhuizen and collaborators. Their current holders make some money, justifying whatever value they paid from their previous owners.

Imagine a world in which government agencies do this kind of thing all the time. Scientists could sell RIBs in their projects, without knowing exactly which agency would ultimately pay for them. Rather than long grant applications, they could write short summaries for investors, guessing at the range of their potential impact, and it would be up to the investors to decide whether the estimate made sense. Scientists could get some of the value of their discoveries, even when that value is quite unpredictable. And they would be incentivized to pick discoveries that could have high impact, and to put a bit of thought and math into what kind of impact that could be.

(Should I still be calling these things bonds, when the buyers don’t know how much they’ll be worth at the end? Probably not. These are more like research impact shares, on a research impact stock market.)

Are there problems with this model, then? Oh sure, loads!

I already mentioned that it’s hard for government agencies to commit to spending money five years down the line. A seventy-year commitment, from that perspective, sounds completely ridiculous.

But we don’t actually need that in the model. All we need is a good reason for investors to think that, eventually, NASA will buy some research impact shares. If government agencies do this regularly, then they would have that reason. They could buy a variety of theoretical developments, a diversified pool to make it more likely some government agency would reward them. This version of the model would be riskier, though, so they’d want more return in exchange.

Another problem is the decision-making aspect. Government agencies wouldn’t have to predict the future, but they would have to accurately assess the past, fairly estimating who contributed to a project, and they would have to do it predictably enough that it could give rise to worthwhile investments. This is itself both controversial and a lot of work. If we figure out the neutron star equation of state, I’m not sure I trust NASA to reward van Nieuwenhuizen’s contribution to it.

This leads to the last modification of the model, and the most speculative one. Over time, government agencies will get better and better at assigning credit. Maybe they’ll have better models of how scientific progress works, maybe they’ll even have advanced AI. A future government (or benevolent AI, if you’re into that) might decide to buy research impact shares in order to validate important past work.

If you believe that might happen, then you don’t need a track record of government agencies buying research impact shares. As a scientist, you can find a sufficiently futuristically inclined investor, and tell them this story. You can sell them some shares, and tell them that, when the AI comes, they will have the right to whatever benefit it bestows upon your research.

I could imagine some people doing this. If you have an image of your work saving humanity in the distant future, you should be able to use that image to sell something to investors. It would be insanely speculative, a giant pile of what-ifs with no guarantee of any of it cashing out. But at least it’s better than NFTs.

Learning for a Living

It’s a question I’ve now heard several times, in different forms. People hear that I’ll be hired as a researcher at an institute of theoretical physics, and they ask, “what, exactly, are they paying you to research?”

The answer, with some caveats: “Whatever I want.”

When a company hires a researcher, they want to accomplish specific things: to improve their products, to make new ones, to cut down on fraud or out-think the competition. Some government labs are the same: if you work for NIST, for example, your work should contribute in some way to achieving more precise measurements and better standards for technology.

Other government labs, and universities, are different. They pursue basic research, research not on any specific application but on the general principles that govern the world. Researchers doing basic research are given a lot of freedom, and that freedom increases as their careers go on.

As a PhD student, a researcher is a kind of apprentice, working for their advisor. Even then, they have some independence: an advisor may suggest projects, but PhD students usually need to decide how to execute them on their own. In some fields, there can be even more freedom: in theoretical physics, it’s not unusual for the more independent students to collaborate with other people than just their advisor.

Postdocs, in turn, have even more freedom. In some fields they get hired to work on a specific project, but they tend to have more freedom as to how to execute it than a PhD student would. Other fields give them more or less free rein: in theoretical physics, a postdoc will have some guidance, but often will be free to work on whatever they find interesting.

Professors, and other long-term researchers, have the most freedom of all. Over the climb from PhD to postdoc to professor, researchers build judgement, demonstrating a track record for tackling worthwhile scientific problems. Universities, and institutes of basic research, trust that judgement. They hire for that judgement. They give their long-term researchers free reign to investigate whatever questions they think are valuable.

In practice, there are some restrictions. Usually, you’re supposed to research in a particular field: at an institute for theoretical physics, I should probably research theoretical physics. (But that can mean many things: one of my future colleagues studies the science of cities.) Further pressure comes from grant funding, money you need to hire other researchers or buy equipment that can come with restrictions attached. When you apply for a grant, you have to describe what you plan to do. (In practice, grant agencies are more flexible about this than you might expect, allowing all sorts of changes if you have a good reason…but you still can’t completely reinvent yourself.) Your colleagues themselves also have an impact: it’s much easier to work on something when you can walk down the hall and ask an expert when you get stuck. It’s why we seek out colleagues who care about the same big questions as we do.

Overall, though, research is one of the free-est professions there is. If you can get a job learning for a living, and do it well enough, then people will trust your judgement. They’ll set you free to ask your own questions, and seek your own answers.

Enfin, Permanent

My blog began, almost eleven years ago, with the title “Four Gravitons and a Grad Student”. Since then, I finished my PhD. The “Grad Student” dropped from the title, and the mysterious word “postdoc” showed up on a few pages. For three years I worked as a postdoc at the Perimeter Institute in Canada, before hopping the pond and starting another three-year postdoc job in Denmark. With a grant from the EU, three years became four. More funding got me to five (with a fancier title), and now nearing on six. Each step, my contract has been temporary: at first three years at a time, then one-year extensions. Each year I applied, all over the world, looking for a permanent job: for a chance to settle down somewhere, to build my own research group without worrying about having to move the next year.

This year, things have finally worked out. In the Fall I will be moving to France, starting a junior permanent position with L’Institut de Physique Théorique (or IPhT) at CEA Paris-Saclay.

A photo of the entryway to the Institute, taken when I interviewed

It’s been a long journey to get here, with a lot of soul-searching. This year in particular has been a year of reassessment: of digging deep and figuring out what matters to me, what I hope to accomplish and what clues I have to guide the way. Sometimes I feel like I’ve matured more as a physicist in the last year than in the last three put together.

The CEA (originally Commissariat à l’énergie atomique, now Commissariat à l’énergie atomique et aux énergies alternatives, or Alternative Energies and Atomic Energy Commission, and yes that means they’re using the “A” for two things at the same time), is roughly a parallel organization to the USA’s Department of Energy. Both organizations began as a way to manage their nation’s nuclear program, but both branched out, both into other forms of energy and into scientific research. Both run a nationwide network of laboratories, lightly linked but independent from their nations’ universities, both with notable facilities for particle physics. The CEA’s flagship site is in Saclay, on the outskirts of Paris, and it’s their Institute for Theoretical Physics where I’ll be working.

My new position is genuinely permanent: unlike a tenure-track position in the US, I don’t go up for review after a fixed span of time, with the expectation that if I don’t get promoted I lose the job altogether. It’s also not a university, which in particular means I’m not required to teach. I’ll have the option of teaching, working with nearby universities. In the long run, I think I’ll pursue that option. I’ve found teaching helpful the past couple years: it’s helped me think about physics, and think about how to communicate physics. But it’s good not to have to rush into preparing a new course when I arrive, as new professors often do.

It’s also a really great group, with a lot of people who work on things I care about. IPhT has a long track record of research in scattering amplitudes, with many leading figures. They’ve played a key role in topics that frequent readers will have seen show up on this blog: on applying techniques from particle physics to gravitational waves, to the way Calabi-Yau manifolds show up in Feynman diagrams, and even recently to the relationship of machine learning to inference in particle physics.

Working temporary positions year after year, not knowing where I’ll be the next year, has been stressful. Others have had it worse, though. Some of you might have seen a recent post by Bret Deveraux, a military historian with a much more popular blog who has been in a series of adjunct positions. Deveraux describes the job market for the humanities in the US quite well. I’m in theoretical physics in Europe, so while my situation hasn’t been easy, it has been substantially better.

First, there’s the physics component. Physics has “adjunctified” much less than other fields. I don’t think I know a single physicist who has taken an adjunct teaching position, the kind of thing where you’re paid per course and only to teach. I know many who have left physics for other kinds of work, for Wall Street or Silicon Valley or to do data science for a bank or to teach high school. On the other side, I know people in other fields who do work as adjuncts, particularly in mathematics.

Deveraux blames the culture of his field, but I think funding also must have an important role. Physicists, and scientists in many other areas, rarely get professor positions right after their PhDs, but that doesn’t mean they leave the field entirely because most can find postdoc positions. Those postdocs are focused on research, and are often paid for by government grants: in my field in the US, that usually means the Department of Energy. People can go through two or sometimes even three such positions before finding something permanent, if they don’t leave the field before that. Without something like the Department of Energy or National Institutes of Health providing funding, I don’t know if the humanities could imitate that structure even if they wanted to.

Europe, in turn, has a different situation than the US. Most European countries don’t have a tenure-track: just permanent positions and fixed-term positions. Funding also works quite differently. Department of Energy funding in the US is spread widely and lightly: grants are shared by groups of theorists at a given university, each getting funding for a few postdocs and PhDs across the group. In Europe, a lot of the funding is much more concentrated: big grants from the European Research Council going to individual professors, with various national and private grants supplementing or mirroring that structure. That kind of funding, and the rarity of tenure, in turn leads to a different kind of temporary position: one not hired to teach a course but hired for research as long as the funding lasts. The Danish word for my current title is Adjunkt, but that’s as one says in France a faux ami: the official English translation is Assistant Professor, and it’s nothing like a US adjunct. I know people in a variety of forms of that kind of position in a variety of countries, people who landed a five-year grant where they could act like a professor, hire people and so on, but who in the end were expected to move when the grant was over. It’s a stressful situation, but at least it lets us further our research and make progress, unlike a US adjunct in the humanities or math who needs to spend much of their time on teaching.

I do hope Deveraux finds a permanent position, he’s got a great blog. And to return to the theme of the post, I am extremely grateful and happy that I have managed to find a permanent position. I’m looking forward to joining the group at Saclay: to learning more about physics from them, but also, to having a place where I can start to build something, and make a lasting impact on the world around me.

Extrapolated Knowledge

Scientists have famously bad work-life balance. You’ve probably heard stories of scientists working long into the night, taking work with them on weekends or vacation, or falling behind during maternity or paternity leave.

Some of this is culture. Certain fields have a very cutthroat attitude, with many groups competing to get ahead and careers on the line if they fail. Not every field is like that though: there are sub-fields that are more collaborative than competitive, that understand work-life balance and try to work together to a shared goal. I’m in a sub-field like that, so I know they exist.

Put aside the culture, and you’ve still got passion. Science is fun, it’s puzzle after puzzle, topics chosen because we find them fascinating. Even in the healthiest workplace you’d still have scientists pondering in the shower and scribbling notes on the plane, mixing business with pleasure because the work is genuinely both.

But there’s one more reason scientists are workaholics. I suspect, ultimately, it’s the most powerful reason. It’s that every scientist is, in some sense, irreplaceable.

In most jobs, if you go on vacation, someone can fill in when you’re gone. The replacement may not be perfect (think about how many times you watched movies in school with a substitute teacher), but they can cover for you, making some progress on your tasks until you get back. That works because you and they have a shared training, a common core that means they can step in and get what needs to be done done.

Scientists have shared training too, of course. Some of our tasks work the same way, the kind of thing that any appropriate expert can do, that just need someone to spend the time to do them.

But our training has a capstone, the PhD thesis. And the thing about a PhD thesis is that it is, always and without exception, original research. Each PhD thesis is an entirely new result, something no-one else had known before, discovered by the PhD candidate. Each PhD thesis is unique.

That, in turn, means that each scientist is unique. Each of us has our own knowledge, our own background, our own training, built up not just during the PhD but through our whole career. And sometimes, the work we do requires that unique background. It’s why we collaborate, why we reach out to different people around the world, looking for the unique few people who know how to do what we need.

Over time, we become a kind of embodiment of our accumulated knowledge. We build a perspective shaped by our experience, goals for the field and curiosity just a bit different from everyone else’s. We act as agents of that perspective, each the one person who can further our particular vision of where science is going. When we enter a collaboration, when we walk into the room at a conference, we are carrying with us all we picked up along the way, each a story just different enough to matter. We extrapolate from what we know, and try to do everything that knowledge can do.

So we can, and should, take vacations, yes, and we can, and should, try to maintain a work-life balance. We need to to survive, to stay sane. But we do have to accept that when we do, certain things won’t get done as fast. Our own personal vision, our extrapolated knowledge…will just have to wait.

Why Are Universities So International?

Worldwide, only about one in thirty people live in a different country from where they were born. Wander onto a university campus, though, and you may get a different impression. The bigger the university and the stronger its research, the more international its employees become. You’ll see international PhD students, international professors, and especially international temporary researchers like postdocs.

I’ve met quite a few people who are surprised by this. I hear the same question again and again, from curious Danes at outreach events to a tired border guard in the pre-clearance area of the Toronto airport: why are you, an American, working here?

It’s not, on the face of it, an unreasonable question. Moving internationally is hard and expensive. You may have to take your possessions across the ocean, learn new languages and customs, and navigate an unfamiliar bureaucracy. You begin as a temporary resident, not a citizen, with all the risks and uncertainty that involves. Given a choice, most people choose to stay close to home. Countries sometimes back up this choice with additional incentives. There are laws in many places that demand that, given a choice, companies hire a local instead of a foreigner. In some places these laws apply to universities as well. With all that weight, why do so many researchers move abroad?

Two different forces stir the pot, making universities international: specialization, and diversification.

Researchers may find it easier to live close to people who grew up with us, but we work better near people who share our research interests. Science, and scholarship more generally, are often collaborative: we need to discuss with and learn from others to make progress. That’s still very hard to do remotely: it requires serendipity, chance encounters in the corridor and chats at the lunch table. As researchers in general have become more specialized, we’ve gotten to the point where not just any university will do: the people who do our kind of work are few enough that we often have to go to other countries to find them.

Specialization alone would tend to lead to extreme clustering, with researchers in each area gathering in only a few places. Universities push back against this, though. A university wants to maximize the chance that one of their researchers makes a major breakthrough, so they don’t want to hire someone whose work will just be a copy of someone they already have. They want to encourage interdisciplinary collaboration, to try to get people in different areas to talk to each other. Finally, they want to offer a wide range of possible courses, to give the students (many of whom are still local), a chance to succeed at many different things. As a result, universities try to diversify their faculty, to hire people from areas that, while not too far for meaningful collaboration, are distinct from what their current employees are doing.

The result is a constant international churn. We search for jobs in a particular sweet spot: with people close enough to spur good discussion, but far enough to not overspecialize. That search takes us all over the world, and all but guarantees we won’t find a job where we were trained, let alone where we were born. It makes universities quite international places, with a core of local people augmented by opportune choices from around the world. It makes us, and the way we lead our lives, quite unusual on a global scale. But it keeps the science fresh, and the ideas moving.

Building the Railroad to Rigor

As a kid who watched far too much educational television, I dimly remember learning about the USA’s first transcontinental railroad. Somehow, parts of the story stuck with me. Two companies built the railroad from different directions, one from California and the other from the middle of the country, aiming for a mountain in between. Despite the US Civil War happening around this time, the two companies built through, in the end racing to where the final tracks were laid with a golden spike.

I’m a theoretical physicist, so of course I don’t build railroads. Instead, I build new mathematical methods, ways to check our theories of particle physics faster and more efficiently. Still, something of that picture resonates with me.

You might think someone who develops new mathematical methods would be a mathematician, not a physicist. But while there are mathematicians who work on the problems I work on, their goals are a bit different. They care about rigor, about stating only things they can carefully prove. As such, they often need to work with simplified examples, “toy models” well-suited to the kinds of theorems they can build.

Physicists can be a bit messier. We don’t always insist on the same rigor the mathematicians do. This makes our results less reliable, but it makes our “toy models” a fair amount less “toy”. Our goal is to try to tackle questions closer to the actual real world.

What happens when physicists and mathematicians work on the same problem?

If the physicists worked alone, they might build and build, and end up with an answer that isn’t actually true. The mathematicians, keeping rigor in mind, would be safe in the truth of what they built, but might not end up anywhere near the physicists’ real-world goals.

Together, though, physicists and mathematicians can build towards each other. The physicists can keep their eyes on the mathematicians, correcting when they notice something might go wrong and building more and more rigor into their approach. The mathematicians can keep their eyes on the physicists, building more and more complex applications of their rigorous approaches to get closer and closer to the real world. Eventually, like the transcontinental railroad, the two groups meet: the mathematicians prove a rigorous version of the physicists’ approach, or the physicists adopt the mathematicians’ ideas and apply them to their own theories.

A sort of conference photo

In practice, it isn’t just two teams, physicists and mathematicians, building towards each other. Different physicists themselves work with different levels of rigor, aiming to understand different problems in different theories, and the mathematicians do the same. Each of us is building our own track, watching the other tracks build towards us on the horizon. Eventually, we’ll meet, and science will chug along over what we’ve built.

Talking and Teaching

Someone recently shared with me an article written by David Mermin in 1992 about physics talks. Some aspects are dated (our slides are no longer sheets of plastic, and I don’t think anyone writing an article like that today would feel the need to put it in the mouth of a fictional professor (which is a shame honestly)), but most of it still holds true. I particularly recognized the self-doubt of being a young physicist sitting in a talk and thinking “I’m supposed to enjoy this?”

Mermin’s basic point is to keep things as light as possible. You want to convey motivation more than content, and background more than your own contributions. Slides should be sparse, both because people won’t be able to see everything but also because people can get frustrated “reading ahead” of what you say.

Mermin’s suggestion that people read from a prepared text was probably good advice for him, but maybe not for others. It can be good if you can write like he does, but I don’t think most people’s writing is that much better than what they say in talks (you can judge this by reading peoples’ papers!) Some are much clearer speaking impromptu. I agree with him that in practice people end up just reading from their slides, which indeed is bad, but reading from a normal physics paper isn’t any better.

I also don’t completely agree with him about the value of speech over text. Yes, putting text on your slides means people can read ahead (unless you hide some of the text, which is easier to do these days than in the days of overhead transparencies). But just saying things means that if someone’s attention lapses for just a moment, they’ll be lost. Unless you repeat yourself a lot (good practice in any case), you should avoid just saying anything you need your audience to remember, and make sure they can read it somewhere if they need it as well.

That said, “if they need it” is doing a lot of work here, and this is where I agree again with Mermin. Fundamentally, you don’t need to convey everything you think you do. (I don’t usually need to convey everything I think I do!) It’s a lesson I’ve been learning this year from pedagogy courses, a message they try to instill in everyone who teaches at the university. If you want to really convey something well, then you just can’t convey that much. You need to focus, pick a few things and try to get them across, and structure the rest of what you say to reinforce those things. When teaching, or when speaking, less is more.

All About the Collab

Sometimes, some scientists work alone. But mostly, scientists collaborate. We team up, getting more done together than we could alone.

Over the years, I’ve realized that theoretical physicists like me collaborate in a bit of a weird way, compared to other scientists. Most scientists do experiments, and those experiments require labs. Each lab typically has one principal investigator, or “PI”, who hires most of the other people in that lab. For any given project, scientists from the lab will be organized into particular roles. Some will be involved in the planning, some not. Some will do particular tests, gather data, manage lab animals, or do statistics. The whole experiment is at least roughly planned out from the beginning, and everyone has their own responsibility, to the extent that journals will sometimes ask scientists to list everyone’s roles when they publish papers. In this system, it’s rare for scientists from two different labs to collaborate. Usually it happens for a reason: a lab needs a statistician for a particularly subtle calculation, or one lab must process a sample so another lab can analyze it.

In contrast, theoretical physicists don’t have labs. Our collaborators sometimes come from the same university, but often they’re from a different one, frequently even in a different country. The way we collaborate is less like other scientists, and more like artists.

Sometimes, theoretical physicists have collaborations with dedicated roles and a detailed plan. This can happen when there is a specific calculation that needs to be done, that really needs to be done right. Some of the calculations that go into making predictions at the LHC are done in this way. I haven’t been in a collaboration like that (though in retrospect one collaborator may have had something like that in mind).

Instead, most of the collaborations I’ve been in have been more informal. They tend to start with a conversation. We chat by the coffee machine, or after a talk, anywhere there’s a blackboard nearby. It starts with “I’ve noticed something odd”, or “here’s something I don’t understand”. Then, we jam. We go back and forth, doing our thing and building on each other. Sometimes this happens in person, a barrage of questions and doubts until we hammer out something solid. Sometimes we go back to our offices, to calculate and look up references. Coming back the next day, we compare results: what did you manage to show? Did you get what I did? If not, why?

I make this sound spontaneous, but it isn’t completely. That starting conversation can be totally unplanned, but usually one of the scientists involved is trying to make it happen. There’s a different way you talk when you’re trying to start a collaboration, compared to when you just want to talk. If you’re looking for a collaboration, you go into more detail. If the other person is on the same wavelength, you start using “we” instead of “I”, or you start suggesting plans of action: “you could do X, while I do Y”. If you just want someone’s opinion, or just want to show off, then your conversation is less detailed, and less personal.

This is easiest to do with our co-workers, but we do it with people from other universities too. Sometimes this happens at conferences, more often during short visits for seminars. I’ve been on almost every end of this. As a visitor, I’ve arrived to find my hosts with a project in mind. As a host, I’ve invited a visitor with the goal of getting them involved in a collaboration, and I’ve received a visitor who came with their own collaboration idea.

After an initial flurry of work, we’ll have a rough idea of whether the project is viable. If it is, things get a bit more organized, and we sort out what needs to be done and a rough idea of who will do it. While the early stages really benefit from being done in person, this part is easier to do remotely. The calculations get longer but the concepts are clear, so each of us can work by ourselves, emailing when we make progress. If we get confused again, we can always schedule a Zoom to sort things out.

Once things are close (but often not quite done), it’s time to start writing the paper. In the past, I used Dropbox for this: my collaborators shared a folder with a draft, and we’d pass “control” back and forth as we wrote and edited. Now, I’m more likely to use something built for this purpose. Git is a tool used by programmers to collaborate on code. It lets you roll back edits you don’t like, and merge edits from two people to make sure they’re consistent. For other collaborations I use Overleaf, an online interface for the document-writing language LaTeX that lets multiple people edit in real-time. Either way, this part is also more or less organized, with a lot of “can you write this section?” that can shift around depending on how busy people end up being.

Finally, everything comes together. The edits stabilize, everyone agrees that the paper is good (or at least, that any dissatisfaction they have is too minor to be worth arguing over). We send it to a few trusted friends, then a few days later up on the arXiv it goes.

Then, the cycle begins again. If the ideas are still clear enough, the same collaboration might keep going, planning follow-up work and follow-up papers. We meet new people, or meet up with old ones, and establish new collaborations as we go. Our fortunes ebb and flow based on the conversations we have, the merits of our ideas and the strengths of our jams. Sometimes there’s more, sometimes less, but it keeps bubbling up if you let it.

The Many Varieties of Journal Club

Across disciplines, one tradition seems to unite all academics: the journal club. In a journal club, we gather together to discuss papers in academic journals. Typically, one person reads the paper in depth in advance, and comes prepared with a short presentation, then everyone else asks questions. Everywhere I’ve worked has either had, or aspired to have, a journal club, and every academic I’ve talked to recognizes the concept.

Beyond that universal skeleton, though, are a lot of variable details. Each place seems to interpret journal clubs just a bit differently. Sometimes a lot differently.

For example, who participates in journal clubs? In some places, journal clubs are a student thing, organized by PhD or Master’s students to get more experience with their new field. Some even have journal clubs as formal courses, for credit and everything. In other places, journal clubs are for everyone, from students up through the older professors.

What kind of papers? Some read old classic papers, knowing that without an excuse we’d never take the time to read them and would miss valuable insights. Some instead focus on the latest results, as a way to keep up with progress in the field.

Some variation is less intentional. Academics are busy, so it can be hard to find a volunteer to prepare a presentation on a paper every week. This leads journal clubs to cut corners, in once again a variety of ways. A journal club focused on the latest papers can sometimes only find volunteers interested in presenting their own work (which we usually already have a presentation prepared for). Sometimes this goes a step further, and the journal club becomes a kind of weekly seminar: a venue for younger visitors to talk about their work that’s less formal than a normal talk. Sometimes, instead of topic, the corner cut is preparation: people still discuss new papers, but instead of preparing a presentation they just come and discuss on the fly. This gets dangerous, because after a certain point people may stop reading the papers altogether, hoping that someone else will come having read it to explain it!

Journal clubs are tricky. Academics are curious, but we’re also busy and lazy. We know it would be good for us to discuss, to keep up with new papers or read the old classics… but actually getting organized, that’s another matter!