Joel Mokyr, Gita Chadha, Lawrence Krauss, Joseph Vijay

All that thinking about Joel Mokyr and his prescription to support society’s intellectual elite in order to ensure technological progress took me back to a talk Gita Chadha delivered in 2020, and to a dilemma I’d had at the time involving Lawrence Krauss. Chadha’s proposed resolution to it could in fact settle another matter I’ve been considering of late, involving Tamil actor-politician Joseph Vijay.

But first a recap. Gita Chadha is a sociologist and author, a professor at Azim Premji University, and an honorary senior fellow at the NCBS Archives. Her 2020 talk was titled ‘Exploring the idea of ‘Scientific Genius’ and its consequences’.

Lawrence M. Krauss is a cosmologist, sceptic, and author, former chair of the Board of Sponsors of the Bulletin of the Atomic Scientists, an alleged sexual predator, and a known associate (and defender) of convicted child-sex offender Jeffrey Epstein. In 2020 he was a year away from publishing a book entitled The Physics of Climate Change, combining two topics of great interest to me, but I wasn’t sure if I should read it. On the one hand there was the wisdom about separating the scholarship from the scholar but on the other I didn’t want to fill out Krauss’s wallet — or that of his publisher, who was trading on Krauss’s reputation — nor heighten the relevance of his book.

Finally, I’ve been thinking about Vijay more than I might’ve been if not for my friend, who’s a fan of Vijay the actor but not the politician. Whenever I express displeasure over her support for Vijay’s acting, she asks me to separate his films from Vijay the person, and politician. I haven’t been convinced. Is Vijay a good actor? I, like my friend, think so. My opinion of Vijay the politician declined however following the crowding disaster in Karur on September 27: while his party’s cadre whipped up a crowd whose size greatly exceeded what the location could safely hold, Vijay made a bad situation worse by first insisting on conducting a roadshow and then arriving late.

Today, I firmly believe separating the work from the person doesn’t make sense when the work itself produces the person’s power.

I first realised this when contemplating Krauss. When I asked during her talk about how we can separate the scholarship from the scholar, Chadha among other things said, “We need to critically start engaging with how the social location of a scholar impacts the kind of work that they do.” Her point was that, rather than consider whether knowledge remains usable once the person who originated it is revealed to have been unethical, we must remember prestige is never innocent: because it changes what institutions and audiences are prepared to excuse.

Broadly speaking, when society puts specific academics “on pedestals”, their eminence and the grant money they bring in become ways to excuse their harm. This is how people like Krauss were able to conduct themselves the way they did. Their work wasn’t just their contribution to the scientific knowledge of all humankind; it was also the reason for their universities to close ranks around them, in ways that the individuals also condoned, until the allegations became too inconvenient to ignore. The scholar benefited from what the scholarship was and the scholarship benefited from who the scholar was.

So a good response isn’t to pretend that it’s possible to cleanly separate the art from the artist but to pay attention to how the work builds social capital for the individual and to keep the individual — and the institutions within which they operate — from wielding that capital as a shield. Thus we must scrutinise Krauss, we must scrutinise his defenders, and we must ask ourselves why we uphold his scholarship above that of others.

(Note: We don’t have to read Krauss’s books, however. This is different from, say, the fact that we have to use Feynman diagrams in theoretical physics even as Richard Feynman was a misogynist and a creep. It doesn’t have to be one at the expense of the other; it can, and perhaps should, be both. I myself eventually decided to not read Krauss’s book: not because he defended Epstein but because I wanted to spend that time and attention on something completely new to me. I asked some friends for recommendations and that’s how I read When the Whales Leave by Yuri Rytkheu.)

The same rationale also clarified the problem I’d had with my friend’s suggestion that I separate Vijay’s work as an actor from Vijay himself. For starters: sure, an actor can play a role well and thus be deemed a good actor, but I think the sort of roles they pick to the exclusion of others ought to matter just as well to their reputation. And the parts he’s picked to play over the last decade or so have all been those of preachy alpha-males touting conservative views of women’s reproductive rights, male attitudes towards women, and retributive justice, among others. It’s also no coincidence that these morals genuflect smoothly to the pro-populist parts of his political messaging.

Similarly, Vijay’s alpha-male roles that I dislike aren’t just fictions: they’re part of the public persona that Vijay has deliberately converted into his newfound political authority. Once a ‘star’ enters electoral politics, “watching for entertainment” is hard to separate from participating in and enabling a machinery that’s generating legitimacy for the ‘star’. The tickets sold, the number of streams, the rallies attended, and the number of fans mobilised all help to manufacture the claim that Vijay has a mandate at all. As with Krauss, participation increased, and continues to increase, material power and relevance as well as paves the way to claim and probably receive immunity from the consequences of inflicting social harm.

But where the case of Vijay diverges from that of Krauss is that the former presents much less of a dilemma. When a person goes from cinema to electoral politics, separating their work from their personal identity is practically indefensible because the political leader himself is the vehicle of the power that he has cultivated through his film work. That is to say, the art and the artist are the same entity because the art fuels the artist’s social standing and the artist’s social standing fuels his particular kind of art.

Mokyr hearts Nobel Prizes

I don’t like Joel Mokyr’s history of progress and have written about that before. I also have a longer analysis and explanation of my issues coming soon in The Hindu. On December 8 I got more occasion to critique his thinking over his Nobel lecture in Stockholm, after receiving the prize that applied to his inchoate history of European Enlightenment a sheen of credibility I (and others) don’t think it deserves. In his lecture (transcribed in full here), Mokyr said:

I have argued at great length that these four conditions [incentives for elite innovators; a competitive “market for ideas”; talented people having the freedom to go where they like; a somewhat ‘activist state’] held increasingly in Europe between 1450 and 1750, the three centuries leading up to the industrial revolution. That is the kind of environment that led increasingly to incredibly creative innovative people: Baruch Spinoza, David Hume, James Watt, Adam Smith, Antoine Lavoisier, and Leonhard Euler, [something indecipherable], Ludwig van Beethoven. These are all people who came up with brand new ideas in an environment that supported them, if not perfectly at least far better than anything in the past.

The question is: do these conditions hold today? I would put it this way: the incentives in propositional knowledge in science are still there and they’re stronger and larger and more pervasive than ever. The market for ideas today provides unprecedented rewards and incentives to successful intellectual innovators, particularly in science. We have hundreds of thousands of people who work in intellectual endeavours, most of them (but not all) in universities. So what we do is we offer them what most of these people need more than anything else, which is financial security, which is tenure and of course in research institutions you get things like named chairs. Then we have rewards and of course there’s a whole pyramid of rewards at which the Nobel Prize and the Abel Prize presumably stand at the very peak, but there are many many other rewards, memberships in academies, and prizes for the best papers and the best books, and honorary degrees.

The funny thing is these incentives are cheap relative to the benefits that these people bestow on humankind, and that is I think a critical thing. Of course, in addition to all that there’s name recognition, fame through mass media, and, perhaps most important, these things lead to peer recognition: many academics really want to be respected by their peers, by other people like themselves, and of course in addition for a very few you know there’s lots of money to be made if they work in the right fields and get it right. Now, most of these incentives were already noticeable in about 1500, but in some ways the 20th century has done far better than anybody before.

In short Joel Mokyr treats rewards like the Nobel Prize to be a low-cost “pyramid” of incentives that helps the “upper tail” of society generate ideas. However the Nobel Prizes aren’t only incentives for innovation: they also exemplify how modern societies manage credit and legitimacy, which are examples of social relations, in ways that shape what innovation looks like and who benefits. The irony here is thus that the Nobel Prizes are part of the same system of social relations he underplays in his theory of progress — and a good example of his blindspot vis-à-vis the history of Europe’s progress.

According to Mokyr, ideas originate in the minds of an intellectual elite — his “upper tail” of society — and society’s job is to reward them. The diffusion of ideas is secondary in his framing. However, scientists, social science scholars, and historians of science have all critiqued the fact that the Nobel Prizes systematically individualise what’s almost always distributed work and sideline the science labour of laboratory managers, technicians, makers of instruments, graduate students, maintenance staff, supply chains, and of course state procurement. The Prizes advance a picture of “elite incentives” working to advance science when in fact it’s predicated equally, if not more, on questions of status and hierarchy, particularly on admission, patronage, language, funding, and geopolitics.

And in their turn the Prizes impose inefficiencies of their own. As we’ve seen before with the story of Brian Keating, they can reorganise research agendas — and more broadly they fetishise problems in science that can be solved and easily be verified to have been solved by a small number of people ‘first’.

Later in his lecture Mokyr further says:

… when technology changes, institutions have to adapt in various ways, and the problem is that they are usually slow to adapt. It takes decades until various parliamentary committees and political forces agree that some form of regulation or some form of control is necessary. And what evolutionary theory suggests is that adaptation to a changing environment is quite possible provided the changes in the technological environment are not too fast and not too abrupt. A sudden discontinuous shock will lead to mass extinctions and catastrophe; we know this is true from evolutionary history.

So there’s some concern that the acceleration in the rate of technological change in recent decades cannot be matched by institutional adaptation. What’s more, the acceleration implied by the last 10 years’ advances in AI and similar fields [suggests where] this is going to be a problem. Certain technological inventions have led to political polarisations, through social media for instance. This is something we haven’t fully solved and it’s not clear that we will be able to.

TL;DR: technology often advances faster than institutions adapt and politics, misinformation, nationalism, and xenophobia threaten the conditions for progress.

But then wouldn’t this mean then that incentives for the elites are the easy part? That is, Nobel Prizes or other incentives like them don’t fix these problems, in fact they may even distract from them by implying the main thing is to simply keep “geniuses” motivated.

Tracking down energetic cosmic rays

Once in a while, nature runs experiments that no human lab can match. Ultra-high-energy cosmic rays are a good example.

Cosmic rays are fast, high energy particles from space that strike Earth’s atmosphere with energies far beyond what even the most powerful particle accelerators can routinely create. Most particles are the nuclei of atoms such as hydrogen or helium and a smaller fraction are electrons.

When a cosmic-ray particle hits the atmosphere, it triggers a large ‘shower’ of particles spread over many square-kilometres. (One such shower may have caused a serious error in a computer onboard an Airbus A320 aircraft on October 30, causing it to suddenly drop 100 feet, injuring several people onboard.) By observing and measuring these showers, scientists hope to probe two big questions at once: what kinds of cosmic events accelerate matter to such extreme energies, and what happens to particle interactions at energies we can’t otherwise test.

The Pierre Auger Observatory in Argentina is one of the world’s main instruments for this work. Its latest results, published in Physical Review Letters on December 9, focus on a curiously simple idea: does the energy spectrum of these cosmic rays — i.e. how many particles arrive at each energy — look the same from every direction of the sky?

In the new study, the Pierre Auger Collaboration analysed data with three notable features. First, the Observatory recorded the spectrum above 2.5 EeV. One EeV is 1018 electron-volt (eV), a unit of energy applied to subatomic particles. Second, it recorded this spectrum across a wide declination range, from +44.8º to −90º. In this range, +90º is the celestial north pole and –90º is the celestial south pole. And third, the analysis included around 3.1 lakh events collected between 2004 and 2022.

The direction in which a cosmic ray is coming from carries information about its origins. If, say, a handful of galaxies or starburst regions are responsible for the highest energy cosmic rays, then the spectrum should show a bump in that part from the sky, and nowhere else.

Understanding the direction from which the most powerful cosmic rays come has become more important since the Collaboration found evidence before that these directions aren’t perfectly uniform. Above 8 EeV, for instance, the Observatory has reported a modest but clear imbalance across the sky. It has also sensed a similarly modest link between extremely energetic cosmic rays and specific parts of space.

Against this backdrop, the new study is part of a larger effort to elevate ultra-high-energy cosmic rays from a curiosity into a way to map the location of ‘cosmic accelerators’ in the nearby universe.

This is like the trajectory that neutrino astronomy has taken as well. For much of the 20th century, neutrinos frustrated physicists because they knew the particles were the universe’s second-most abundant, yet they remain extremely difficult to catch and study. Because neutrinos carry no electric charge and interact only weakly with matter, they pass through stars and magnetic fields almost untouched, making them unusually honest witnesses to violent places in the universe. However, physicists could take advantage of that only if they could build suitable detectors. And step by step that’s what happened. Today, experiments like IceCube in Antarctica realise neutrino astronomy: a way to study the universe using neutrinos. (Francis Halzen’s long push for this detector is why he’s been awarded the APS Medal for 2026.)

Cosmic rays stand on the cusp of a similar opportunity. To this end, the Pierre Auger Collaboration had to find whether the spectrum is dependent or independent of direction. A direction-independent spectrum would push the field towards models in which different types of cosmic sources produce high-energy cosmic rays; a direction-dependent spectrum would do the opposite.

The new result was firm: across declinations from −90° to +44.8°, the team didn’t find a meaningful change in the spectrum’s shape.

Cosmic ray researchers read the energy spectrum as a sort of forensic record. Over many decades, experiments have shown that the spectrum doesn’t increase smoothly. If you plotted the energy on the graph, in other words, you wouldn’t see a smooth curve. Instead you’d see the curve bending and changing how steep it is in places (see image below). These bumps reflect changes in the sources of cosmic rays, in the chemical makeup of the particles (protons v. nuclei), and how cosmic rays lose energy as they travel intergalactic space before reaching Earth.

The Collaboration’s new paper framed its findings in the context of two recent developments.

First, above 8 EeV, cosmic rays aren’t arriving in perfectly random directions around Earth. Instead roughly one half of the sky supplies 6% more such cosmic rays. Physicists have interpreted this to mean the distribution is being shaped by large-scale structure in the nearby universe, by magnetic fields or by both.

Second, the Collaboration previously identified evidence for an ‘instep’ feature near 10 EeV.

A plot of the cosmic ray flux (y-axis) versus the particles’ energy. The red dots show the rough ‘locations’ of the knee, ankle, and instep, from top to bottom in that order. Credit: Sven Lafebre (CC BY-SA)

If you look at the energy curve (shown above), you’ll notice a shift in slope at three points. Top to bottom, they’re called the ‘knee’, the ‘ankle’, and the ‘instep’. At each of these points, physicists believe, the set of physical effects producing the cosmic rays changes.

(LHAASO, a large detector array in China built to catch the particle showers made by very energetic gamma rays, recently reported signs that microquasar jets — where a stellar-mass black hole pulls in gas from a companion star and emits fast beams of radiation — could be accelerating cosmic rays to near the knee part of the spectrum.)

A spectrum whose shape changes depending on the direction is a way to connect these two aspects. If the instep is due to a small number of nearby, unusually strong sources, you might expect it to show up more strongly in the part of the sky where those sources are located. If the instep is a generic feature produced by many broadly similar sources, it should appear in essentially the same way across the sky (after accounting for the modest unevenness of the dipole). In the new study, the Collaboration tried to  make that distinction sharper.

The Pierre Auger Observatory detects showers of particles in the atmosphere with an array of water tanks spread over a large area. Showers arriving near-vertically and those arriving at an angle closer to the horizon behave differently because Earth’s magnetic field distorts their paths. So the analysis used different methods to reconstruct the cosmic rays based on their showers for angles up to 60º (vertical) and from 60º to 80º (inclined). Scientists inferred the energy in the rays based on the properties of the particles in the shower.

In the declination range −90º and +44.8º, the Observatory found the spectrum didn’t vary significantly with declination, with or without the dipole. In other words, once the Collaboration accounted for the disuniform intensity in the sky, the spectrum’s shape didn’t change depending on the direction.

The second major result was the ‘instep’, with the findings reinforcing previous evidence for this feature.

Now, if the instep was mainly caused by a small number of nearby sources, it would be reasonable to expect the spectrum would change depending on the declination. But the study found the spectrum to be indistinguishable across declinations. This in turn, per the Collaboration, disfavours the possibility of a small number of nearby sources contributing ultra-high-energy cosmic rays. Instead, the spectrum’s key features could be set by many sources and/or effects acting together.

The paper also suggested that the spectrum steepening near the instep could be due to a change in what the cosmic rays are made of: from lighter nuclei like helium to heavier nuclei like carbon and oxygen. If this bend in the curve is really due to a change in the cosmic rays’ composition, rather than in their sources, then cosmic rays coming from all directions should have this feature. And this is what the Pierre Auger Collaboration has reported: the spectrum’s shape doesn’t change by direction.

According to the paper’s authors, because the spectrum looks the same in different parts of the sky, the next clues to cracking the mystery of cosmic rays’ origins need to come from tests that measure their composition more clearly, helping to explain the instep.

Featured image: A surface detector tank of the Pierre Auger Observatory in 2007. Credit: Public domain.

Acting and speaking

Daily writing prompt
Have you ever performed on stage or given a speech?

I performed several times on stage in school, up to when I was 11 years old. I studies for around four years at a school in Tumkur where drama was part of the curriculum. Every year we’d have a playwright come over with a story, usually involving kings and conflict, and they’d cast us for various roles and have us rehearse for months. I was almost always one of the ‘extras’ but I didn’t mind: it was just good to be a part of it. We also got to do a lot of craft work in the process as we prepared the props and materials for the set, often involving bamboo, after school hours. Finally we’d put up a show for the whole town to watch. (Tumkur was a town then!) I still remember we had a particularly great time the year Velu Saravanan directed us. There was a lot of laughter, even if I don’t remember what the story was. It was a wonderful period.

I generally don’t like giving speeches and talks because I’m not convinced that that’s the most useful way to share what I know with others and vice versa. I’ve received quite a few opportunities and invitations in my professional years. On the few occasions I’ve accepted, I’ve insisted on being seated with the ‘audience’ and converting the chance to an AMA, where anyone can ask me anything. That way I can be sure what I say is useful for at least a few people.

So you use AI to write…

You’re probably using AI to write. Both ChatGPT and Google AI Studio prefer to construct their sentences in specific and characteristic ways and anyone who’s been a commissioning editor for at least a few years will find the signs of its influence hard to miss, even if you personally think it’s undetectable.

You’re probably going to continue using AI to write. There’s nothing I can do about that. In fact, as AI models improve, my ability to prove you used AI to write some text will only decline. And as a commissioning editor focusing articles about science, health, and environment research, I’m also not entirely at liberty to stop working with a writer if I notice AI-written text in their reports. I can only warn them. If I threaten to stop, I’m certain that one day I’ll have to impose the ultimate sanction, and after that I’ll have to look for new writers from a pool that’s already quite small. That’s a bad proposition for me.

To be sure, I realise I’m in a difficult position. People, especially those without a good grasp of English writing as well as those without the means to attain that grasp, should have the opportunity to be understood as precisely as they wish to be among English-speaking readers. If immaculate writing and grammar also allow their opinions to reach audiences that shunned them before for being too difficult to understand, all the more reason, no?

This has been something of an issue with the way The New Yorker wields its grammar rulebook, with its commas, hyphens, accents, and umlauts in just the right places. As the author Kyle Paoletta disputed in a 2017 critique in The Baffler:

For The New Yorker, a copy editor’s responsibility to avoid altering the substance of a writer’s prose is twisted into an utter disinterest in content writ large. … Content must be subordinated—thoughtfully, of course!—to the grammatical superstructure applied to it. Not only does this attitude treat the reader as somewhat dim, it allows the copy editor to establish a position of privilege over the writer. …

[Famed NYer copy-editor Mary] Norris frets over whether or not some of James Salter’s signature descriptive formulations (a “stunning, wide smile,” a “thin, burgundy dress”) rely on misused commas. When she solicits an explanation, he answers, “I sometimes ignore the rules about commas… Punctuation is for clarity and also emphasis, but I also feel that, if the writing warrants it, punctuation can contribute to the music and rhythm of the sentences.” Norris begrudgingly accepts this defence, but apparently only because a writer of no lesser stature than Salter is making it. Even in defeat, Norris, as the tribune of The New Yorker’s style, is cast as a grammatical arbiter that must be appealed to by even the most legendary writers.

I shouldn’t stand in judgment of when and how a writer wishes to wield the English language as they understand and have adopted it. But with AI in the picture, that could also mean trusting the writer to a degree that also overlooks whether they’ve used AI. Put another way: if it’s levelling the playing field for you, it’s getting rid of another sign of authenticity for me.

Perhaps the larger point is that as long as you make sure your own writing improves, we’re on the right track. To be clear: whose English should be improved? I’m saying that’s the contributor’s call. Because if they send a poorly written piece that makes good points, some of what I’m going to be doing to their piece is what AI will be as well (and tools like Grammarly have already been). Second, as AI use becomes more sophisticated, and I become less able to tell original and AI-composed copies apart, I’ll just have to accept at some point that there’s nothing I can do. In that context, all I can say is: “Use it to improve your English at least”.

Or perhaps, as in the words of Northern Illinois University professor David J. Gunkel, “LLMs may well signal the end of the author, but this isn’t a loss to be lamented. … these machines can be liberating: they free both writers and readers from the authoritarian control and influence of this thing we call the ‘author’. For better or for worse, however, I don’t see how a journalistic publication can adopt this line wholesale.

* * *

As a commissioning editor in science journalism specifically, I’ve come across many people who have useful and clever things to say but their English isn’t good enough. And from where I’m sitting, using AI to make their ideas clearer in English only seems like a good thing… That’s why I say it levels the playing field, and when you do that in general, the people who already have a skill because they trained to acquire it lose their advantages while the new entrants lose their disadvantages. The latter is of course unfair.

In the same vein, while I know AI tools have been enormously useful for data journalists, asking a programmer what they feel about that would likely elicit the same sort of reaction I and many of my peers have vis-à-vis people writing well without having studied and/or trained to do so. That would make the data journalists’ use problematic as well by virtue of depriving a trained programmer of their advantages in the job market, no?

Short of the world’s Big AI companies coming together and deciding they’re going to insert their products’ outputs with some kind of inalienable watermark, we have to deal with the inevitability of our being unable to tell AI-made versus human-made content apart.

Yes, AI is hurting what it means to be human, at least in terms of our creative expression. But another way to frame all this is that the AI tools we have represent the near-total of human knowledge collected on the internet. Isn’t it wonderful that we can dip into that pool and come away as more skilled, more knowledgeable, and perhaps better persons ourselves?

Science communication, as it focuses on communicating what scientists have found, might become redundant and low-value, at least among those that can afford these subscriptions (whose prices will also fall and accessibility will increase) because people can feed a model a paper and ask for the output. Hallucinations will always be there but it’s possible they could become less likely. (And it seems to me to be somewhat precarious to found our own futures as journalists on the persistence of hallucinations.) Instead where we can cut ahead, and perhaps stay that way, is by focusing on a journalism of new ideas, arguments, conversations, etc. And there the playing field will also be less about whether you can write better and more about what and how you’ve learnt to think about the world and its people.

We all agree democratisation is good. Earlier, and to a large degree still, education facilitated and facilitates that to a significant degree. Now the entry of AI models is short-circuiting that route, and it’s presenting some of us with significant disadvantages. Yes, there are many employers who will try to take advantage of this state to bleed their workforce and take the short way out. But it remains that — setting aside the AI models coming up with BS and equivocating on highly politicised issues — the models are also democratising knowledge. Our challenge is to square that fact (which can’t be un-facted in future) with preserving the advantages for those that ‘deserve’ it.

As if mirroring the question about who gets to decide who should improve their English, who gets to decide who gets to have advantages?

AI slop clears peer-review

Here’s an image from a paper that was published by Nature Scientific Reports on November 19 and retracted on December 5:

This paper made it through peer review at the journal. Let that sink in for a moment. Perhaps the reviewers wanted to stick it to the editors. Then again how the image made its way past the editors is also a mystery.

Nature Scientific Reports has had several problems before, enumerated on its Wikipedia page. It’s a ‘megajournal’ in the vein of PLOS One and follows the gold OA model, with an article processing charge of “£2190.00/$2690.00/€2390.00”.

Worlds between theory and experiment

Once Isaac Newton showed that a single gravitational law plus his rules of dynamics could reproduce the orbits of planets that Johannes Kepler had predicted, explain tides on Earth, and predict that a comet that had passed by once would return again, physicists considered Newtonian mechanics and gravitation to have been completely validated. After these successful tests, they didn’t wait to test every other possible prediction of Newton’s ideas before they considered them to be legitimate.

When Jean Perrin and others carefully measured Brownian motion and extracted Avogadro’s number in the early 20th century, they helped cement the science of the kinetic theory of gases and statistical mechanics that Ludwig Boltzmann and Josiah Willard Gibbs had developed. As with Newtonian mechanics, physicists didn’t also require every single consequence of kinetic theory to be rechecked from scratch. They considered them all to be fully and equally legitimate then on.

Similarly, in 1886-1889, Heinrich Hertz produced and detected electromagnetic waves in the laboratory, measured their speed and other physical properties, and showed that they behaved exactly as James Clerk Maxwell had predicted based on his (famous) equation. Hertz’s experiments didn’t test every possible configuration of charges and fields that Maxwell’s equations allowed, yet what they did test and confirm sufficed to convince all physicists that Maxwell’s theory could be treated as the correct classical theory of electromagnetism.

In all these cases, a theory won broad acceptance after scientists validated only a small (yet robust) subset of its predictions. They didn’t have to validate every single prediction in distinct experiments.

However, there are many ideas in high-energy particle physics that, even as they are derived from other theoretical constructs that have been tested to extreme precision, physicists insist on testing them anew as well. Why are they going to this trouble now?

“High-energy particle physics” is a four-word label for something you’ve likely already heard of: the physics of the search for the subatomic particles like the Higgs boson and the efforts to identify their properties.

In this enterprise, many scientific ideas follow from theories that have been validated by very large amounts of experimental data. Yet physicists want to test them at every single step because of the way such theories are built and the way unknown effects can hide inside their structures.

The overarching theory that governs particle physicists is called, simply, the Standard Model. It’s a quantum field theory, i.e. a theory that combines the precepts of quantum mechanics and special relativity*. Because the Standard Model is set up in this way, it makes predictions about the relations between different observable quantities, e.g. the mass of a subatomic particle called the W boson with a parameter that’s related to the decay of other particles called muons. Some of these relations connect measured quantities with others that have not yet been probed, e.g. the mass of the muon with the rate at which Higgs bosons decay to pairs of muons. (Yes, it’s all convoluted.) These ‘extra’ relations often depend on assumptions that go beyond the domains that experiments have already explored. New particles and new interactions between them can change particular parts of the structure while leaving other parts nearly unchanged.

(* Quantum field theory gives physicists a single, internally consistent framework in which they can impose both the rules of quantum theory and the requirements of special relativity, such as that information or matter can’t travel faster than light and that our spacetime conserves energy and momentum together, for example. However, quantum field theory does not unify quantum theory with general relativity; that’s the monumental and still unfinished purpose of the quantum gravity problem.)

For a more intricate example, consider the gauge sector of the Standard Model, i.e. the parts of the Model involving the gluons, W and Z bosons, and photons, their properties, and their interactions with other particles. The gauge sector has been thoroughly tested in experiments and is well-understood. Now, the gauge sector also interacts with the Higgs sector, and the Higgs sector interacts with other sectors. The result is new possibilities involving the properties of the Higgs boson, their implications for the gauge sector, and so on that — even if physicists have tested the gauge sector — need to be tested separately. The reason is that none of these possibilities follow directly from the basic principles of the gauge sector.

The search for ‘new physics’ also drives this attitude. ‘New physics’ refers to measurable entities and physical phenomena that lie beyond what the Standard Model can currently describe. For instance, most physicists believe a substance called dark matter exists (in order to explain some anomalous observations about the universe), but they haven’t been able to confirm what kind of particles dark matter is made of. One popular proposal is that dark matter is made of hitherto unknown entities called weakly interacting massive particles (WIMPs). The Standard Model in its contemporary form doesn’t have room for WIMPs, so the search for WIMPs is a search for new physics.

Physicists have also proposed many ways to ‘extend’ the Standard Model to accommodate new kinds of particles that ‘repair’ the cracks in reality left by the existing crop of particles. Some of these extensions predict changes to the Model that are most pronounced in sectors that are currently poorly pinned down by existing data. This means even a sizeable deviation from the Model’s structure in this sector would still be compatible with all current measurements. This is another important reason physicists want to collect more data and with ever-greater precision.

Earlier experience also plays an important role. Physicists may make some assumptions because they seem safe in some year but new data collected in the next two decades might reveal that they were mistaken. For instance, physicists believed neutrinos didn’t have mass, like photons, because that idea was consistent with many existing datasets. Yet dedicated experiments contradicted their belief (and won their performers the 2015 physics Nobel Prize).

(Aside: High-energy particle physics uses large machines called particle colliders to coerce subatomic particles into configurations where they interact with each other, then collect data of those interactions. Operating these instruments demands hundreds of people working together, using sophisticated technologies and substantial computing resources. Because the instruments are so expensive, these collaborations aim to collect as much data as possible, then maximise the amount of information they extract from each dataset.)

Thus, when a theory like the Standard Model predicts a specific process, that process becomes a thing to test. But even if the prediction seems simple or obvious, actually measuring it can still rule out whole families of rival theories offering to explain the same process. It also sharpens physicists’ estimates of the theory’s basic parameters, which then makes other predictions more precise and helps plan the next round of experiments. This is why, in high-energy physics, even predictions that follow from other, well-tested parts of a theory are expected to face experimental tests of their own. Each successful test can reduce the space for new physics to hide in — or in fact could reveal it.

A study published in Physical Review Letters on December 3 showcases a new and apt example of testing predictions made by a theory some of whose other parts have already survived testing. Tests at the Large Hadron Collider (LHC) — the world’s largest, most powerful particle collider — had until recently only weakly constrained the Higgs boson’s interaction with second-generation leptons (a particle type that includes muons). The new study provides strong, direct evidence for this coupling and significantly narrows that gap.

The LHC operates by accelerating two beams of protons in opposite directions to nearly the speed of light and smashing them head on. Its operation is divided into segments called ‘runs’. Between runs, the collaboration that manages the machine conducts maintenance and repair work and, sometimes, upgrades its detectors.

One of the LHC’s most prominent detectors is named ATLAS. To probe the interactions between Higgs bosons and leptons, the ATLAS collaboration collected and analysed data from the LHC’s run 2 and run 3. The motivation was to obtain direct evidence for Higgs bosons’ coupling to muons and to measure its strength. And in the December 3 paper, the collaboration reported that the coupling parameters were consistent with the Standard Model’s predictions.

So that’s one more patch of the Standard Model that has passed a test, and one more door to ‘new physics’ that has closed a little more.


Featured image: A view of the Large Hadron Collider inside its tunnel. Credit: CERN.

Robbing NISAR to pay ISRO

A.K. Anil Kumar, the director of ISRO’s Telemetry, Tracking, and Command Network (a.k.a. ISTRAC), has reportedly made some seriously misleading comments as part of his convocation address at a Maharishi University in Lucknow.* Kumar’s speech begins at the 1:38:10 mark in this video (hat-tip to Pradx):

A poorly written article in The Free Press Journal (which I couldn’t find online) has amplified Kumar’s claims without understanding that the two satellites Kumar was seemingly talking about are actually one: the NASA-ISRO Synthetic Aperture Radar (NISAR), developed jointly by the US and Indian space agencies. The article carries an image of NISAR but doesn’t caption it as such.

The article makes several dubious claims:

  • That the “satellite” can forecast earthquakes,
  • That NISAR can capture subsurface images of Earth, including of underground formations,
  • That India’s “satellite” didn’t require a 12-metre-long antenna the way NASA’s “satellite” did, and
  • That ISRO’s “satellite” was built at one-tenth of the cost of NASA’s “satellite”

To be clear, an ISRO satellite that can forecast earthquakes or image subsurface features and which the organisation built and launched for Rs 1,000 crore does not exist. What actually exists is NISAR, a part of which ISRO built. The claims are almost spiteful because they purport to come from a senior ISRO official whose work likely benefited from the ISRO-NASA collaboration and because he ought to have known better than to mislead.

NISAR is a dual-frequency (DF) synthetic aperture radar (SAR). The ‘DF’ bit means the satellite captures data on two radar frequencies, L-band and S-band. To quote from a piece I wrote for The Hindu on July 27:

At the time the two space organisations agreed to build NISAR, NASA and ISRO decided each body would contribute equivalent‑scale hardware, expertise, and funding. … [ISRO] supplied the I‑3K spacecraft bus, the platform that houses the controls to handle command and data, propulsion, and attitude, plus 4 kW of solar power. The same package also included the entire S‑band radar electronics, a high‑rate Ka‑band telecom subsystem, and a gimballed high‑gain antenna.

‘SAR’ refers to a remote-sensing technique in which a small antenna moves along a path while using a computer to combine the data it captures along the way, thus mimicking a much larger antenna. NISAR uses a 12-metre mesh antenna plus a reflector for this purpose. Both the S-band and L-band radars use it to perform their functions. As a result of using the SAR technique, the two radars onboard NISAR are able to produce high-resolution images of Earth’s surface irrespective of daylight or cloud cover and support studies of ground deformation, ice sheets, forests, and the oceans.

In this regard, for The Free Press Journal to claim NISAR “didn’t require India to install a separate 12-metre antenna, unlike NASA” gives the impression that ISRO’s S-band radar didn’t need the antenna. This is wrong: it does need the antenna. That NASA was the agency to build and deploy it on NISAR comes down to the terms of the collaboration agreement, which specified that ISRO would provide the spacecraft bus, the S-band radar (and its attendant components), and the launch vehicle while NASA would take care of everything else. This is the same reason why ISRO’s contributions to NISAR amounted to around Rs 980 crore — which Kumar rounded up to Rs 1,000 crore — whereas NASA’s cost was around Rs 10,000 crore.

The antenna is in fact an engineering marvel. Without it ISRO’s S-band radar wouldn’t be so performant and thus its data wouldn’t be so useful for decision-making for both research and disaster management. On the day ISRO launched NISAR, on July 30 this year, I got to interview Karen  St. Germain, the director of the Earth Science Division at the Science Mission Directorate at NASA. Here’s an excerpt from the interview about the antenna:

Both the L-band and the S-band radars use the same reflector. Since S-band has a shorter wavelength than the L-band, does this create any trade-offs in either L-band or S-band performance?

It doesn’t. And the reason for that is because this is a synthetic aperture radar. It creates its spatial resolution as it moves along. Each radar is taking snapshots as it moves along. You know, to get this kind of centimetre level fidelity and the kind of spatial resolution we’re achieving, if you were to use a solid antenna, it would have to be five miles long. Just like when you’re talking about a camera, if you want to be able to get high fidelity, you need a big lens. Same idea. But we can’t deploy an antenna that big. So what we do is we build up image after image after image to get that resolution. And because of this technique, it’s actually independent of wavelength. It works the same for S- and for L-bands. The only thing that’s a little different is because the antenna feeds for the L-band and the S-band can’t physically occupy the same space, they have to be next to each other and that means there’s a slight difference in the way their pulses reflect off the antenna. There’s that positioning difference, and that we can correct for.

Could you tell us a little bit more about that slight difference?

Karen St. Germain:
It’s the way a reflector works. You would ideally want to put the feed at the focal point of the reflector. But when you have two feeds, you can’t do that. So they’re slightly offset. That means they illuminate the reflector just slightly differently. The alignment is just a little bit different. The team optimised the design to minimise that difference and to make it so that they could correct it in post-processing.

And even for all these abilities, we (i.e. people everywhere) currently don’t know enough to be able to forecast earthquakes. What we can do today is make short-term predictions and we can prepare probabilistic forecasts over a longer period of time. That is, for instance, we can say “there’s a 20% chance of a quake of magnitude 8 or more occurring in the Himalaya in the next century” and we have the means to alert people in an area tens of seconds before an earthquake occurs. We can’t say “there will be an earthquake in Chennai at 3 pm tomorrow”.

The question for The Free Press Journal is thus what role a satellite can essay in this landscape. In a statement in 2021, ISRO had said “NISAR would provide a means of disentangling highly spatial and temporally complex processes ranging from ecosystem disturbances to ice sheet collapses and natural hazards including earthquakes, tsunamis, volcanoes and landslides.” This means NISAR will help scientists better piece together the intricate processes implicated in earthquakes — processes that are distributed over some area and happen over some time. Neither NISAR nor the S-band radar alone can forecast earthquakes.

On a related note, the L-band (1,000-2,000 MHz) and S-band (2,000-4,000 MHz) radar frequencies do overlap with the frequencies used in ground-penetrating radar (10-4,000 MHz). However, the lower the frequency, the further underground an electromagnetic wave can penetrate (while keeping the resolution fixed). Scientists have documented a ceiling of around 100 MHz for deep geological profiling, which is far from either of NISAR’s radars. Even the L-band radar, which has lower frequency than the S-band, can at best penetrate a few metres underground if the surface is extremely dry, like in a desert, or if the surroundings are made of water ice. What both radars can penetrate very well is cloud cover, heavy rain, and vegetation.

The ISRO + NASA collaboration that built NISAR was a wonderful thing that the agencies need to replicate in future even as it continued their less formalised engagements from before and whose benefits both host countries, India and the USA, continue to accrue in the satellite observation and remote-sensing domains. For Kumar to call the cost component into question in the way that he did, followed by the The Free Press Journal‘s shoddy coverage of his remarks, does no favours to the prospect of space literacy in the country.

* I updated this post at 7.45 pm on December 2, 2025, to make it clear that all but one of the objectionable claims were made by The Free Press Journal in its article; the exception was the cost comparison, which Kumar did make.

Part light, part matter

Consider a bunch of molecules that have been trapped between two mirrors facing each other very closely. In this ‘box’ the light can’t move around freely; it can only exist in certain fixed patterns, somewhat like sound inside a finely tuned musical instrument. Even when all light has been removed from this box, quantum physics causes small flickers of light to pop in and out of existence in fractions of a second. These immutable flickers of the electromagnetic field are sometimes collectively called the vacuum field.

Now, what would happen if the natural motion of matter inside molecules and these confined light patterns become strongly linked?

This isn’t an esoteric thought experiment. Over the last decade, physicists have shown in experiments that when this link is strong, new hybrid states appear that are part light and part matter — and these states can spread over many molecules at once. And in this condition, physicists have found that the materials in which these states occur suddenly have very different properties, all without changing the chemical formulae of their constituent molecules.

A new paper by PhD scholar Subha Biswas and assistant professor Anoop Thomas — both at the Department of Inorganic and Physical Chemistry of the Indian Institute of Science — brought the many findings in this area of research together and argued that the confined vacuum field reshapes the weak forces between molecules, which is the same force that determine how the molecules arrange themselves and how easily energy and charge can move between them. And the duo has shown that this explanation can account for a wide range of surprising findings. Their findings were published in ACS Applied Optical Materials on November 11.

The paper isn’t a single experiment; instead the authors read and compared many published papers about experiments in which other scientists placed molecules inside small mirror cavities and forced them to interact strongly with the trapped light.

Thus, the duo reported, energy could travel quite differently in these cavities. Normally, when energy passes from one molecule to the next, it does so through a series of short hops if the material is disordered. A simple example is an ordinary piece of glass. In a crystal, atoms or molecules sit on a tidy 3D grid, like seats in a stadium. In glass however they’re more like people milling around in a crowd: there’s some short-range order but no long-range one.

When motion inside the molecules is strongly coupled to the trapped light, however, the new light–matter states can extend over many molecules, creating a sort of highway that energy travel on faster, more efficiently, and across much larger regions than physicists generally expect in disordered materials.

For another, the duo reported that some organic materials, especially plastics that usually conduct electricity very poorly, are suddenly able to conduct a lot better when certain internal motions of their molecules are strongly coupled to the trapped light. In some experiments scientists had found that protons also moved faster through water, suggesting the cavity could in some way alter the network of hydrogen bonds in water.

Based on analysing these and other examples, the duo arrived at a central conclusion: that by modifying how light is confined around a material, scientists can influence how its molecules ‘feel’ each other and thus alter its bulk behaviour, even when it’s in its lowest-energy state. (Usually materials become capable of unusual things when they’re imparted more energy.) Put another way the cavity isn’t just a passive container but an active design tool that can reshapes the background electromagnetic field, and with that the landscape of intermolecular forces.

This is interesting because it’s a new way to control molecules in chemistry. Usually when scientists need to change the way a chemical reaction happens, they change the molecules that are in play. For instance they alter the molecules’ structure, add side groups, and/or ‘mix in’ new components. But the new study has found what seems to be an additional handle: scientists may be able to leave the molecules as they are but just place them in a carefully designed optical environment. And by carefully choosing the spacing between the mirrors, the materials that make up the apparatus, and how the cavity resonates with specific molecular motions, they can adjust the reaction rate, favour one crystal form over another, stabilise certain structures, shift the balance between different reaction products, and guide how large molecules assemble into various shapes, etc.

The implications are similarly broad for materials science, where physicists can use these cavities to improve energy transport in thin films, enhance conduction in soft or flexible materials, guide how polymers and other large molecules arrange themselves in devices, and tune how materials crystallise as they’re fabricated. Because the effect comes from light and matter coupling together over a longer range, it might be especially useful in systems that are otherwise disordered and where conventional design tools struggle.

For instance, the molecules of polystyrene are arranged in a messy, haphazard way. This disorder makes it very hard for scientists to ‘engineer’ clean paths within the material through which electrons can flow, so polystyrene is usually an excellent insulator. However, when scientists place thin films of such plastics in a carefully tuned optical cavity, experiments have shown that their ability to conduct electricity increases drastically even though the material’s chemical makeup hasn’t changed.

A simple salt solution, like of table salt dissolved in water, offers another good example. Here the water molecules and the salt ions move around in an incessantly shifting and disordered fashion, and chemists typically control them by changing how much salt or what additives there are. But when the solution is strongly coupled to the confined light, the way water molecules surround and move around the ions changes and the ions travel more easily through the liquid.

The authors have stressed that these effects are selective — which is to say not every material will change dramatically. Instead scientists will have to select details such as how molecules pack together and which internal motion is coupled to the light in the cavity all matter. This in turn means this new vacuum-field engineering won’t altogether replace chemical design but could in fact complement it. In fact the authors have sketched a future in which chemists and materials scientists routinely think together about what molecules to build and the kind of electromagnetic environment to place them in to coax new or improved properties out.

New labour codes

On November 21, the Union government said the four consolidated labour codes on wages, industrial relations, social security, and occupational safety and health had been brought into force, replacing 29 national labour laws, five years after Parliament had signed off on them and four years beyond the original 2021 deadline.

The gap was because labour is in the Concurrent List, so both the Union government and the states had to draft and finalise detailed rules under each code. The Centre had also said it wanted to bring all four codes into force together, which meant waiting until most of the states had completed their rule-making, of course many of them were slow to do this while others were reluctant (given trade-union opposition and criticism that the codes diluted protections and eased retrenchment). Then of course there was the COVID-19 pandemic, various Assembly elections, and the 2024 Lok Sabha election.

Anyway, in a statement accompanying the announcement, the Ministry of Labour and Employment presented the codes’ entering into force as an overhaul that would simplify compliance and extend social security and minimum wages to workers across sectors. In fact a substantial part of the Ministry’s claims rests on actual legal changes. They replace a patchwork of national and state laws with a more consolidated framework on wages, industrial relations, social security, and occupational safety and health such that employers already within the formal net now deal with a more coherent set of statutes and a single system of registrations and returns. The code on wages extends the minimum wage framework from “scheduled employments” to all employees, closing a formal gap that had left some occupations outside the statute even as some states expanded their schedules.

The codes also formally defined “gig workers”, “platform workers”, and “aggregators”, creating a legal basis on which the Union government could found schemes funded by “aggregator” contributions. The occupational safety code turns appointment letters into a statutory duty for all covered employers, including for existing workers, and introduces a general requirement of annual health check-ups for workers older than 40 years, extending a form of medical surveillance that was earlier confined largely to specified hazardous processes.

Unfortunately the announcement also has a pattern of overstatement that raises important questions about whether the Ministry is aware of the real constraints on what the new framework can deliver.

The foremost example is the Ministry’s language on social security coverage. According to the release, all workers “will get PF, ESIC, insurance and other social security benefits” — but the code does no such thing. It consolidates the statutes on PF, ESIC, maternity, gratuity, and employees’ compensation and empowers the Union government to frame schemes for new categories of workers and to extend the ESIC scheme to new sectors. Whether any given worker receives PF or ESI still depends on the notifications and the nature of the establishment.

Similarly, the ESI scheme has expanded gradually and its coverage still depends on notified areas and sectors; it’s also true that the codes give governments more flexibility to expand the scheme and makes it possible, in principle, to cover smaller establishments or individual workers involved in hazardous processes. However, the claim that “social protection coverage will be expanded to all workers” is really a policy aspiration tied to future notifications and resources rather than a statement of what the law already guarantees today.

Second, the shift from “inspectors” to “inspectors-cum-facilitators” means officials are now tasked not only with detecting violations and prosecuting violators but also with advising employers on how to comply and with using risk-based inspection plans that emphasise predictability and graded responses.

Risk-based inspections can in principle focus attention on more harmful sectors but in practice they can also narrow the number of surprise visits and turn inspection into a negotiated exercise. When the same official is expected to build cooperative relationships with employers while also being an enforcer, the incentive to treat violations as matters to be counselled away rather than penalised becomes stronger. And for workers in sectors where accidents are hidden and trade union presence is weak, the shift only weakens the prospect of government officials identifying and pursuing breaches.

The codes also preserve headcount thresholds for registration and for many duties on safety and social security. Many contractors and principal employers already respond to these thresholds — including in the last five years — by splitting operations across multiple small entities, by cycling workers through short-term engagements, and by labelling them as ‘partners’ or ‘volunteers’ so that no single establishment appears to cross the line that would trigger stricter obligations.

My ‘favourite’ example of such corruption in action is an incident in 2016 when “India’s 16 occupational safety laws couldn’t prevent two deaths” at the Anna University campus in Chennai. As Thomas Manuel wrote in The Wire at the time:

The needless tragedy of the two deaths is mirrored by the tragedy that is the state of the laws and public institutions set up to deal with incidents such as these. There are sixteen laws (and a dozen or more boards, directorates, inspectorates and other bodies) that deal broadly with the subject of working conditions and occupational health and safety (OHS) in India. But despite this seeming glut, the circumstances that led to the death of Shankar and Deepan seem to have contravened no law at all.

Like most such deaths, this one too was easily reduced to a question of compensation. U. Lakshmikanthan, Deputy Commissioner of Labour, stated that after reading the media reports on the deaths, he suo moto initiated a complaint, summoning representatives of the university and the employer to his office. But his intervention was only to adjudicate on the amount of compensation that would be paid to the family of the deceased as per the statutes of the Employees Compensation Act 1923. This Act discusses the terms on which compensation will be paid by the employer in the event of the death or disability of an employee. It doesn’t have anything to say on questions of occupational health in the strict sense. Lakshmikanthan did not disclose the compensation that had been decided but the maximum allowed under the law is approximately Rs 11 lakh.

The Directorate of Industrial Health and Safety (DISH) … reacted quite differently. P. Bose, the director of DISH, stated that after a preliminary investigation, it was decided that the case did not fall within his purview. It didn’t come under the Factories Act 1948 as Kavimeena Rubber Products employed less than ten people and the incident happened outside the manufacturing premises. It also wasn’t covered by the Building And Other Construction Workers (Regulation Of Employment And Conditions Of Service) Act, 1996, because – again – more than ten workers were required for the Act to apply as well as the fact that the cost of construction had to be more than ten lakh rupees.

An inspector-cum-facilitator abiding by risk-based protocols and time limits may never see the full picture of a fragmented workforce and may lack both the mandate and the institutional backing to piece it together. In the event of a serious injury or death, thus, principal employers and contractors can still argue over who — if anyone — had a direct employment relationship with the worker and whether the work even fell within the categories the codes explicitly regulate.

Finally, in the Ministry’s announcement, there are paragraphs on beedi rollers, plantation workers, media workers, mine workers, IT staff, dock workers, and export units — but no discussion of sanitation workers employed through layers of contractors and others who perform tasks that often amount to manual scavenging by another name, and where deaths are frequent.

For these workers, the concerns are still whether the principal employer can keep them off the books and whether inspectors and courts will push past the chain of contracts to fix liability on those who control the work and profit from it. The codes contain some tools that could help — e.g. treating workers supplied by an unlicensed contractor as if they were directly employed by the principal employer — but the communique overall only highlights more general welfare themes.

Indian labour law has long been more protective than practice on the ground would suggest. The codes plug some genuine gaps, update definitions, consolidate a mosaic of statutes, and extend certain obligations. But for informal workers or those working in small, unregistered units, reform continues to depend on whether employers and contractors are registered at all, whether inspectors have the mandate and capacity to look beyond paperwork, and whether state governments choose to use the discretion the codes give them to broaden coverage.