The sirens of overstatement

Visiting David Blandy’s installation Atomic Light at John Hansard Gallery, University of Southampton, for New Scientist, 22 March 2023

The Edge of Forever, one of four short films by Brighton-based video and installation artist David Blandy, opens with an elegaic pan of Cuckmere Haven in Sussex. A less apocalyptic landscape it would be hard to imagine. Cuckmere is one of the most ravishing spots in the Home Counties. Still, the voiceover insists that we contemplate “a ravaged Earth” and “forgotten peoples” as we watch two children exploring their post-human future. The only sign of former human habitation is a deserted observatory (the former Royal Observatory at Herstmonceux Castle in Sussex). The children enter and study the leavings of dead technologies and abandoned ambitions, steeped all the while in refracted sunlight: Claire Barrett’s elegiac camerawork is superb.

The films in Blandy’s installation “Atomic Light” connect three different kinds of fire: the fire of the sun; the wildfires that break out naturally all over the earth, but which are gathering force and frequency as the Earth’s climate warms; and the atomic blast that consumed the Japanese city of Hiroshima on 6 August 1945.

There’s a personal dimension to all this, beyond Blandy’s vaunted concern for the environment: his grandfather was a prisoner of the Japanese in Singapore during the second World War, and afterwards lived with the knowledge that, had upwards of 100,000 civilians not perished in Hiroshima blast, he almost certainly would not have survived.

Bringing this lot together is a job of work. In Empire of the Swamp
a man wanders through the mangrove swamps at the edge of Singapore, while Blandy reads out a short story by playwright Joel Tan. The enviro-political opinions of a postcolonial crocodile are as good a premise for a short story as any, I suppose, but the film isn’t particularly well integrated with the rest of the show.

Soil, Sinew and Bone, a visually arresting game of digital mirrors composed of rural footage from Screen Archive South East, equates modern agriculture and warfare. That there is an historical connection is undeniable: the chemist Franz Haber received the Nobel Prize in Chemistry in 1918 for his invention of the Haber–Bosch process, a method of synthesising ammonia from nitrogen and hydrogen. That ammonia, a fertiliser, can be used in the manufacture of explosives, is an irony familiar to any GCSE student, though it’s by no means obvious why agriculture should be left morally tainted by it.

Alas, Blandy can’t resist the sirens of overstatement. We eat, he says “while others scratch for existence in the baked earth.” Never mind that since 1970, hunger in the developing world has more than halved, and that China saw its hunger level fall from a quarter of its vast population to less than a tenth by 2016 — all overwhelmingly thanks to Haber-Bosch.

Defenders of the artist’s right to be miserable in face of history will complain that I am taking “Atomic Light” far to literally — to which I would respond that I’m taking it seriously. Bad faith is bad faith whichever way you cut it. If in your voiceover you dub Walt Disney’s Mickey “this mouse of empire”, if you describe some poor soul’s carefully tended English garden as the “pursuit of an unnatural perfection wreathed in poisons”, if you use footage of a children’s tea party to hector your audience about wheat and sugar, and if you cut words and images together to suggest that some jobbing farmer out shooting rabbits was a landowner on the lookout for absconding workers, then you are simply piling straws on the camel’s back.

Thank goodness, then, for Sunspot, Blandy’s fourth, visually much simpler film, that juxtaposes the lives and observations of two real-life solar astronomers, Joseph Hiscox in Los Angeles and Yukiaki Tanaka in Tokyo, who each made drawings of the sun on the day the Hiroshima bomb dropped.

Here’s a salutary and saving reminder that, to make art, you’re best off letting the truth speak for itself.

Moral good taste

Reading Humanly Possible by Sarah Bakewell for the Telegraph, 15 March 2023

In 1362 Western literary culture took a swerve when Petro Petroni, a monk from Siena, told Giovanni Boccaccio of his recent vision: that if Boccaccio didn’t stop collecting non-Christian books (never mind writing them), God was going to kill him. Boccaccio, somewhat rattled, told his friend, the poet Petrarch, and Petrarch, being a bibliophile, was unimpressed: if Boccaccio did decide to take Petroni’s advice, would he mind giving him first pick of the discards?

Humanly Possible is an anecdotal history — witty, warm-hearted (here and there, gratingly matey) — of the Western mind’s seven-hundred-year effort to ignore priestly and sectarian blarney, so as to nurture its own voice, its own conscience, its own good. Humanists believe that we each of us have a spark of good will and that, fed on charity, education and civic effort, all these sparks can together enlighten society.

Bakewell acknowledges that such civilising efforts are traceable in many traditions, and have probably been going on for ever. (And I do mean forever. I remember the relief I felt once, walking through Athens’s Acropolis museum, as I left behind the dead-eyed, fatuously grinning statues of the Archaic period (550 BC) for the dignified, melancholy, humane creatures of the Classical era that followed.)

Bakewell’s is story of spiritual and intellectual triumph, beginning in Italy around 1300 and which, thanks to that bloody Twentieth century — its two world wars and litany of totalitarian atrocity — ends with a hideous and disturbing twist.

Organising Europe’s humanists into a “tradition” is rather like herding cats. Bakewell’s organisational ability deserves applause. Here, bibliomanes like Boccaccio give way to physicians like Vesalius, then, via memoirists like Montaigne and philosophers like Paine and Hume, to novelists like E.M. Forster, with his impassioned plea that we “only connect!” with each other. It’s an epic, spine-tingling, seamless account.

Bakewell, who used to look after early printed books at the Wellcome Library in London, has a melting love of all those joyful 16th-century recitations of human excellence, and does a terrific job of communicating the ethical achievements behind such apparent fripperies as the Third Earl of Shaftesbury’s “Inquiry Concerning Virtue and Merit”, which would replace stern religiosity and unthinking obedience with nothing more than “a kind of moral ‘good taste’”.

I was reminded here of Kenneth Clark, in the 1969 TV series Civilisation, trying to explain why the paintings of Watteau (of all people) represent some sort of ethical high-water mark for western civilisation. In both cases, the point is well made, and well received, but at the same time, both assertions feel a bit underwhelming. All that cultural effort, all that struggle and invention, suffering and heroism, led up to — good taste? (Bakewell, fully aware of the problem, quotes the Cambodian filmmaker Rithy Panh: “There’s also a banality of good; and an everydayness of good.”)

The argument against humanism goes like this: an unbeliever, guided purely by their own conscience, is a person without morals. If society tolerated such people, breakdown would ensue.

And the awful thing is that this argument is not altogether wrong. Sometimes, breakdown is exactly what happens. Josef Stalin didn’t see to the deaths of 20 million of his own people by being an anti-humanist. Not at all: he was one of the best-read men of his generation, obsessed to the point of madness with constant intellectual and (though he wouldn’t have called it this) spiritual self-improvement. All fascism’s John-the-Baptist figures — from Maurice Barres to Martin Heidegger — were humanists maddened by the alienations of heavy industry, looming automation and evident democratic failure.

It’s not that the humanist idea is flawed, so much as it is no defence against our self-fulfilling belief in our own badness (and that’s a switch that’s frighteningly easy to throw: Bakewell mentions Savonarola and “his bonfire of the vanities”, but, generously, turns a blind eye to Greta Thunberg).

In his autobiography, quoted here, the Austrian novelist Stefan Zweig talked of the humanists’ “beautiful error” — believing that “better learning, better reading and better reasoning would be enough to bring about a better world.”

Bakewell is too clear-eyed an historian, and too honest a writer, to gloss over the weakness of this pose. But she’s clear enough about the alternatives, too. They’re all species of harshness in Bakewell’s chatty and persuasive book — one form or another of force or war or slavery. We may feel jolly silly at times, waving Oscar Wilde’s love of curtain fabrics in the face of the world’s barbarity — but that might be as good a weapon as we’ll ever get.

“The solutions are not even in the works”

Reading Pegasus by Laurent Richard and Sandrine Rigaud for New Scientist, 1 February 2023.

Fifty thousand?”

Edward Snowden’s 2013 leaks from the US National Security Agency had triggered a global debate around state surveillance — and even he couldn’t quite believe the scale of the story as it was described to him in the summer of 2021.

Whistle-blowers had handed French investigative journalists Laurent Richard and and Sandrine Rigaud a list of 50,000 phone numbers. These belonged to people flagged for attack by a cybersurveillance software package called Pegasus.

The journalistic investigation that followed is the subject of this non-fiction thriller: a must-read for anyone remotely interested in cryptography and communications, and a dreadful warning for the rest of us. “Regular civilians being targeted with military-grade surveillance weapons — against their will, against their knowledge, and with no recourse — is a dystopian future we really are careening toward,” the authors warn, “if we don’t understand this threat and move to stop it.”

Pegasus offers a fascinating insight into how journalism has evolved to tackle a hyper-connected world. Eye witnesses and whistle-blowers have better access than ever before to sympathetic campaigning journalists from all over the world. But of course, this advantage is shared with the very governments and corporations and organised crime networks that want to silence them.

To drag Pegasus into the light, Laurent’s Forbidden Stories consortium choreographed the activities of more than eighty investigative journalists from seventeen media organisations across four continents and eight languages.

The consortium got together in March 2021 knowing full well that they would have to conclude their investigation by June, by which time Pegasus’ creators at the Israeli company NSO were bound to twig that their brainchild was being hacked.

The bigger the names on that phone list, the harder it would be to keep any investigation under wraps. Early on the name of Jorge Carrasco cropped up: the lead partner in Forbidden Stories’ massive cross-border collaboration to finish the investigations of murdered Mexican journalist Regina Martínez. Then things just got silly: a son of Turkish president Recep Erdogan turned up; and then the names of half the French cabinet. Also the cell number for Emmanuel Macron, the president of France. Laurent Richard recalls, “Macron was the name that made me realise how truly dangerous it was to have access to this list.”

In a pulse-accelerating account that’s never afraid to dip into well-crafted technical detail, the authors explain how Pegasus gains free rein on a mobile device, without ever tipping off the owner to its presence. Needless to say it evolved out of software designed to serve baffled consumers waiting in long queues on tech support call lines. Shalev Hulio and Omri Lavie, who would go on to found NSO and create Pegasus, cut their teeth developing programmes that allowed support technicians to take charge of the caller’s phone.

It was not long before a European intelligence service came calling. Sold and maintained for more than sixty clients in more than forty countries, Pegasus gave security services an edge over terrorists, criminal gangs and paedophiles — and also, as it’s turned out, over whistleblowers, campaigners, political opponents, journalists, and at least one Emirati princess trying to get custody of her children. This book is not a diatribe against the necessary (or at any rate ubiquitous) business of government surveillance and espionage. It is about how, in the contest between ordinary people and the powerful, software is tilting the field wildly in the latter’s favour.

The international journalistic collaboration that was the Pegasus Project sparked the biggest global surveillance scandal since Snowden; it’s led to a European Parliament inquiry into government spyware, legal action from major technology companies, government sanctions against the NSO Group and countless individual legal complaints. But the authors spend little time sitting in their laurels. Pegasus may be dead, but demand for a successor is only growing. In the gap left by NSO, certain governments are making offers to certain tech companies that add zeroes to the fees NSO enjoyed. Nor do the authors expect much to come out of the public debate that has followed their investigation: “The issues… might have been raised,” they concede, “but the solutions are not even in the works.”

Rogues and heroes

Reading Sam Miller’s Migrants: The Story of Us All for the Telegraph, 23 January 2023

The cultural opprobrium attached to immigration has been building at least since Aristotle’s day, according to Sam Miller’s flawed, fascinating stab at a global history of migration.

Today, “having a permanent home and a lifelong nationality are considered normal, as if they were part of the human condition.” On the contrary, says Miller: humankind is the migratory species par excellence, settling every continent bar Antarctica, not once, but many times over.

Mixed feelings about this process have a deep anthropological foundation. Forget national and regional rivalries; those came later, and are largely explanations after the fact. What really upsets settled people is the reminder that, long ago, their kind chose to live an urban life and became less as a consequence: less wily, less tough, less resilient. The emergence of the first cities coincided with the first poems in uneasy praise of wild men: think of Mesopotamian Enkidu, or Greek Heracles. Aristotle, writing in 330 BC, declares that “he who is without a city-state by nature, and not by circumstance, is either a rogue or greater than a human being” — a wonderfully uneasy and double-edge observation that acknowledges a pre-urban past populated by formidable feral heroes.

Athenians suppressed this awareness; they were the first Western people to take pride in being, in Herodotus’s words, “the only Greeks who never migrated.” Ming Dynasty China performed the same flim-flam, a 15th-century administrator declaring: “There exists a paramount boundary within Heaven and Earth: Chinese on this side, foreigners on the other. The only way to set the world in order is to respect this boundary”.

“History books have, on the whole, been written by the sedentary for the sedentary,” says Miller, and naturally reflect a settled people’s chauvinism. The migration stories we learned at school are often wrong. The Vandals who “sacked” Rome in 455, did not, as a general rule, kill or rape or burn.

Alas, neither did they write; nor did the Roma, until the nineteenth century; nor did the (handsomely literate) Chinese of Victorian London. Migrants rarely find time to write, and where first-person accounts are missing, fantasy is bred. Some of it (Asterix) is charming, some of it (Fu Manchu) is anything but.

Miller thinks that humans naturally emigrate, and our unease about this is the result of pastoralism, cities, and other historical accidents.

The trouble with this line of argument is that there are umpteen “natural” reasons why people move about the earth. Humans naturally consume and lay waste to their immediate environment. Humans naturally overbreed. Humans naturally go to war. Why invoke some innate “outward urge”?

Different distances on the human story allow one to tell wildly different stories. If you follow humanity through deep time, our settlement of the almost the entire planet looks very much like manifest destiny and we’ll all surely end up on Mars tomorrow. If on the other hand you trace the movements of people over a few dozen generations, you’ll discover that, absent force majeure, people are homebodies, moving barely a few weeks’ walking distance from their birthplaces.

What is migration, anyway? Not much more than a hundred years ago, women regularly “migrated” (as Miller says, “it might take as long to cross a large English county as it would to fly halfway around the world today”) to marry or to work as governesses, domestic servants and shop workers. And yet they would never have called themselves “migrants”.

Miller, in a praiseworthy bid to tell a global story, adopts the broadest possible definition of migration: one that embraces “slaves and spouses, refugees and retirees, nomads and expats, conquerors and job-seekers.”

Alas, the broader one’s argument, the less one ends up saying. While they’re handsomely researched and stirringly written, I’m not sure our concepts of migration are much enriched by Miller’s brief tilts at historical behemoths like slavery and the maritime spice route.

What emerges from this onion of a book (fascinating digressions around no detectable centre), is, however, more than sufficent compensation. We have here the seed of much more enticing and potentially more influential project: a modern history that treats the modern nation state — pretending to self-reliance behind ever-more-futile barriers — as but a passing political arrangement, and not always a very useful one.

In view of the geopolitical crises being triggered by climate change, we may very soon need (or else be forced by circumstances) to come up with forms of government outside the rickety and brittle nation state. And in that case, peripatetic perspectives like Miller’s may be just what we need.

 

Who’s left in the glen?

Watching Emily Munro’s Living Proof: A climate story for New Scientist, 6 October 2021

Most environmental documentaries concentrate, on the environment. Most films about the climate crisis focus on people who are addressing the crisis.

Assembled and edited by Emily Munro, a curator of the moving image at the National Library of Scotland, Living Proof is different. It’s a film about demobbed soldiers and gamekeepers, architects and miners and American ex-pats. It’s about working people and their employers, about people whose day-to-day actions have contributed to the industrialisation of Scotland, its export of materials and methods (particularly in the field of off-shore oil and gas), and its not coincidental environmental footprint.

Only towards the end of Munro’s film do we meet protestors of any sort. They’re deploring the construction of a nuclear power plant at Torness, 33 miles east of Edinburgh. Even here, Munro is less interested in the protest itself, than in one impassioned, closely argued speech which, in the context of the film, completes an argument begun in Munro first reel (via a public information film from the mid-1940s) about the country’s political economy.

Assembled from propaganda and public information films, promotional videos and industrial reports, Living Proof is an archival history of what Scotland has told itself about itself, and how those stories, ambitions and visions have shaped the landscape, and effected the global environment.

Munro is in thrall to the changing Scottish industrial landscape, from its herring fisheries to its dams, from its slums and derelict mine-heads to the high modernism of its motorways and strip mills. Her vision is compelling and seductive. Living Proof is also — and this is more important — a film which respects its subjects’ changing aspirations. It tells the story of a poor, relatively undeveloped nation waking up to itself and trying to do right by its people.

It will come as no surprise, as Glasgow prepares to host the COP26 global climate conference, to hear that the consequences of those efforts have been anything but an unalloyed good. Powered by offshore oil and gas, home to Polaris nuclear missiles, and a redundancy-haunted grave for a dozen heavy industries (from coal-mining to ship-building to steel manufacture), Scotland is no-one’s idea of a green nation.

As Munro’s film shows, however, the environment was always a central plank of whatever argument campaigners, governments and developers made at the time. The idea that the Scots (and the rest of us) have only now “woken up to the environment” is a pernicious nonsense.

It’s simply that our idea of the environment has evolved.

In the 1940s, the spread of bog water, as the Highlands depopulated, was considered a looming environmental disaster, taking good land out of use. In the 1950s automation promised to pull working people out of poverty, disease and pollution. In the 1960s rapid communications were to serve an industrial culture that would tread ever more lightly over the mine-ravaged earth.

It’s with the advent of nuclear power, and that powerful speech on the beach at Torness, that the chickens come home to roost. That new nuclear plant is only going to employ around 500 people! What will happen to the region then?

This, of course, is where we came in: to a vision of a nation that, if cannot afford its own people, will go to rack and ruin, with (to quote that 1943 information film) “only the old people and a few children left in the glen”.

Living Proof critiques an economic system that, whatever its promises, can cannot help but denude the earth of its resources, and pauperise its people. It’s all the more powerful for being articulated through real things: schools and roads and pharmaceuticals, earth movers and oil rigs, washing machines and gas boilers.

Reasonable aspirations have done unreasonable harm to the planet. That’s the real crisis elucidated by Living Proof. It’s a point too easily lost in all the shouting. And it’s rarely been made so well.

The tools at our disposal

Reading Index, A History of the, by Dennis Duncan, for New Scientist, 15 September 2021

Every once in a while a book comes along to remind us that the internet isn’t new. Authors like Siegfried Zielinski and Jussi Parikka write handsomely about their adventures in “media archaeology”, revealing all kinds of arcane delights: the eighteenth-century electrical tele-writing machine of Joseph Mazzolari; Melvil Dewey’s Decimal System of book classification of 1873.

It’s a charming business, to discover the past in this way, but it does have its risks. It’s all too easy to fall into complacency, congratulating the thinkers of past ages for having caught a whiff, a trace, a spark, of our oh-so-shiny present perfection. Paul Otlet builds a media-agnostic City of Knowledge in Brussels in 1919? Lewis Fry Richardson conceives a mathematical Weather Forecasting Factory in 1922? Well, I never!

So it’s always welcome when an academic writer — in this case London based English lecturer Dennis Duncan — takes the time and trouble to tell this story straight, beginning at the beginning, ending at the end. Index, A History of the is his story of textual search, told through charming portrayals of some of the most sophisticated minds of their era, from monks and scholars shivering among the cloisters of 13th-century Europe to server-farm administrators sweltering behind the glass walls of Silicon Valley.

It’s about the unspoken and always collegiate rivalry between two kinds of search: the subject index (a humanistic exercise, largely un-automatable, requiring close reading, independent knowledge, imagination, and even wit) and the concordance (an eminently automatable listing of words in a text and their locations).

Hugh of St Cher is the father of the concordance: his list of every word in the bible and its location, begun in 1230, was a miracle of miniaturisation, smaller than a modern paperback. It and its successors were useful, too, for clerics who knew their bibles almost by heart.

But the subject index is a superior guide when the content is unfamiliar, and it’s Robert Grosseteste (born in Suffolk around 1175) who we should thank for turning the medieval distinctio (an associative list of concepts, handy for sermon-builders), into something like a modern back-of-book index.

Reaching the present day, we find that with the arrival of digital search, the concordance is once again ascendant (the search function, Ctl-F, whatever you want to call it, is an automated concordance), while the subject index, and its poorly recompensed makers, are struggling to keep up in an age of reflowable screen text. (Sewing embedded active indexes through a digital text is an excellent idea which, exasperatingly, has yet to catch on.)

Running under this story is a deeper debate, between people who want to access their information quickly, and people (especially authors) who want people to read books from beginning to end.

This argument about how to read has been raging literally for millennia, and with good reason. There is clear sense in Socrates’ argument against reading itself, as recorded in Plato’s Phaedrus (370 BCE): “You have invented an elixir not of memory, but of reminding,” his mythical King Thamus complains. Plato knew a thing or two about the psychology of reading, too: people who just look up what they need “are for the most part ignorant,” says Thamus, “and hard to get along with, since they are not wise, but only appear wise.”

Anyone who spends too many hours a day on social media will recognise that portrait — if they have not already come to resemble it.

Duncan’s arbitration of this argument is a wry one. Scholarship, rather than being timeless and immutable, “is shifting and contingent,” he says, and the questions we ask of our texts “have a lot to do with the tools at our disposal.”

The Art of Conjecturing

Reading Katy Börner’s Atlas of Forecasts: Modeling and mapping desirable futures for New Scientist, 18 August 2021

My leafy, fairly affluent corner of south London has a traffic congestion problem, and to solve it, there’s a plan to close certain roads. You can imagine the furore: the trunk of every kerbside tree sports a protest sign. How can shutting off roads improve traffic flows?

The German mathematician Dietrich Braess answered this one back in 1968, with a graph that kept track of travel times and densities for each road link, and distinguished between flows that are optimal for all cars, and flows optimised for each individual car.

On a Paradox of Traffic Planning is a fine example of how a mathematical model predicts and resolves a real-world problem.

This and over 1,300 other models, maps and forecasts feature in the references to Katy Börner’s latest atlas, which is the third to be derived from Indiana University’s traveling exhibit Places & Spaces: Mapping Science.

Atlas of Science: Visualizing What We Know (2010) revealed the power of maps in science; Atlas of Knowledge: Anyone Can Map (2015), focused on visualisation. In her third and final foray, Börner is out to show how models, maps and forecasts inform decision-making in education, science, technology, and policymaking. It’s a well-structured, heavyweight argument, supported by descriptions of over 300 model applications.

Some entries, like Bernard H. Porter’s Map of Physics of 1939, earn their place thanks purely to their beauty and for the insights they offer. Mostly, though, Börner chooses models that were applied in practice and made a positive difference.

Her historical range is impressive. We begin at equations (did you know Newton’s law of universal gravitation has been applied to human migration patterns and international trade?) and move through the centuries, tipping a wink to Jacob Bernoulli’s “The Art of Conjecturing” of 1713 (which introduced probability theory) and James Clerk Maxwell’s 1868 paper “On Governors” (an early gesture at cybernetics) until we arrive at our current era of massive computation and ever-more complex model building.

It’s here that interesting questions start to surface. To forecast the behaviour of complex systems, especially those which contain a human component, many current researchers reach for something called “agent-based modeling” (ABM) in which discrete autonomous agents interact with each other and with their common (digitally modelled) environment.

Heady stuff, no doubt. But, says Börner, “ABMs in general have very few analytical tools by which they can be studied, and often no backward sensitivity analysis can be performed because of the large number of parameters and dynamical rules involved.”

In other words, an ABM model offers the researcher an exquisitely detailed forecast, but no clear way of knowing why the model has drawn the conclusions it has — a risky state of affairs, given that all its data is ultimately provided by eccentric, foible-ridden human beings.

Börner’s sumptuous, detailed book tackles issues of error and bias head-on, but she left me tugging at a still bigger problem, represented by those irate protest signs smothering my neighbourhood.

If, over 50 years since the maths was published, reasonably wealthy, mostly well-educated people in comfortable surroundings have remained ignorant of how traffic flows work, what are the chances that the rest of us, industrious and preoccupied as we are, will ever really understand, or trust, all the many other models which increasingly dictate our civic life?

Borner argues that modelling data can counteract misinformation, tribalism, authoritarianism, demonization, and magical thinking.

I can’t for the life of me see how. Albert Einstein said, “Everything should be made as simple as possible, but no simpler.” What happens when a model reaches such complexity, only an expert can really understand it, or when even the expert can’t be entirely sure why the forecast is saying what it’s saying?

We have enough difficulty understanding climate forecasts, let alone explaining them. To apply these technologies to the civic realm begs a host of problems that are nothing to do with the technology, and everything to do with whether anyone will be listening.

The coasts of Britain and why to avoid them

A piece for the Financial Times to mark National Maritime Museum Cornwall’s Monsters of the Deep exhibition, 14 March 2020 

In February this year, even as Stuart Slade was in Falmouth assembling the exhibits for a new museum show on sea creatures, a 60ft fin whale was washed up and died on a nearby beach. “Nothing compares to seeing such an animal up close,” he says. “You come away awed, and full of wonder, and just a little bit afraid.”

The objects filling Slade’s gallery for the mysterious Monsters of the Deep exhibition are by turns terrifying, wonderful and funny — sometimes all three. Some are real, some reconstructed, some, like the worrisomely convincing corpse of a mermaid, are assembled out of parts to entertain or gull the public.

The show, at the National Maritime Museum Cornwall, marks the moment Falmouth’s “local” museum learns to punch well above its weight, embracing global phenomena and potentially difficult themes, such as the discomfortingly large role imagination plays in how we see the real world.

Hoaxes, which you might think would be something of an embarrassment here, prove central to the exhibition’s vision. In any event, they can’t be ignored, not while Falmouth’s very own sea monster, Morgawr (first sighted in 1975), could be prowling the bay. According to a report from around the time in the Falmouth Packet: “Mrs Scott said she would ‘never forget the face on that thing’ as long as she lived.”

It transpired that Morgawr was mischievously fabricated from coordinated fictional “sightings”, but the lines between fact and fiction tend to be blurred whenever sea monsters are involved. On September 25 1808, on Stronsay in the Orkney Islands, a large carcass was washed ashore. Edinburgh’s leading natural history society, the Wernerian, decided it was a new species, probably a sea serpent.

By the time the London anatomist Sir Everard Home realised it was more likely a decayed basking shark, the Beast’s reputation had firmly rooted itself in local folklore. Google (if you dare) pictures of dead basking sharks. Their jaws, dorsal and caudal fins disappear first, leaving them with tiny heads and long necks, like ancient reptiles transported from some deep corner of the Mesozoic.

Most sea monsters are real animals, misidentified under extreme circumstances. In 1493, Christopher Columbus “quite distinctly saw three mermaids” off the coast of Haiti. “They are not so beautiful as they are said to be,” he remarked in his journal, “for their faces had some masculine traits.” They were in fact manatees: four-metre long marine mammals with prehensile upper lips and widely spaced eyes (it had been a long voyage).

Most useful to the Falmouth show are the depictions of sea monsters in art, because it’s here that we get to grips with the key question: why have creatures that do not exist persisted in our imaginations since we first put pigment to cave wall?

Some believe sea monsters are a folk memory of creatures long extinct. Native Australian legends of the fearsome Bunyip (head of an emu, body of a dog, tail of a horse) might just scrape by as descriptions of extinct Australian marsupials such as the diprotodon or Palorchestes. But what are we to make of Mishipeshu, which terrified generations of Anishinaabe in the Great Lakes region of Canada? This was an underwater panther, for whom, needless to say, no fossil records exist.

These forms of wonder and fear change across time. The Kraken is supposed to be an octopus, at least according to the Victorian poet Alfred Lord Tennyson. But in 13th-century Greenland, the Kraken was more like a giant crab. Slade, who has been head of public programming in Falmouth for 15 years, says: “The conversations I’ve had about this exhibition tend all to go the same way. People point out that only 5 per cent of the oceans has been explored. From there, it’s just a hop, skip and a jump to saying, ‘There must be something else out there that hasn’t yet been discovered.’”

The inference is mistaken but not obviously so. Back in 1893, the biologist Thomas Henry Huxley wrote in The Times: “There is not an a priori reason that I know of why snake-bodied reptiles, from fifty feet long and upwards, should not disport themselves in our seas as they did in those of the cretaceous epoch which, geologically speaking, is a mere yesterday.”

Palaeontologist Darren Naish, who is lead curator of the Falmouth exhibition, is willing to entertain Huxley’s theory: “His was the right attitude to take at the time, because the life of the deep oceans was only just being discovered.” (Monsters of the Deep makes much of the groundbreaking research expedition led by HMS Challenger, which between 1872 and 1876 discovered a staggering 4,700 new species of marine life.) “Large fossil dinosaurs and early whales, and some amazing gigantic living animals, had been discovered only relatively recently,” Naish points out. “The whale shark, the world’s biggest fish, was a mid 19th-century discovery.”

In an effort to make the new findings comprehensible, folkloric sea monsters were associated with ancient reptiles. Accounts invariably began with the observation that, for years and years, people have talked about giant serpents in the oceans and long-bodied monsters, then go on to point to the actual fossil evidence that such things were a reality. The earliest artists’ impressions have the plesiosaur (50ft long; extended neck; four oar-shaped flippers on a broad, flat body) as a giant, coiling serpent.

How, then, can we be certain that such beasts aren’t out there? In 1998, Charles Paxton, an aquatic ecologist at the Animal Behaviour Research Group at Oxford university, used a statistical technique to estimate the current diversity of large marine animals, based on their rate of discovery.

Extrapolating on data from 1830 to 1995, Paxton produced a graph showing the rate at which these animals are coming to light. He estimated that at most there are around 50 new large species still waiting to be discovered and, according to his graph, we’re likely to come across one every five-and-a-half years or so.

What will they look like? Cryptozoologists — researchers who aspire to the scientific study of undiscovered animals — have for years held out for the existence of radically novel animals, “living plesiosaurs” that have somehow survived from the time of the dinosaurs. These “cryptid” animals come with brilliant names like the super-otter and the father of all turtles (names invented by Bernard Heuvelmans, whose 1968 book In the Wake of the Sea-Serpents kick-started marine cryptozoology). They are meant to be gargantuan, more than 15 metres long, and unlike any creature known to science.

However, as Naish observes, “Of all the animals that have been discovered in recent decades, none has been radically novel.” Recent discoveries have included the Megamouth Shark and a couple of new types of beaked whale (the most recent was spotted by local whalers in Hokkaido, Japan last September). And as each new whale or shark is discovered, the chances of there still being a complete outlier in hiding — something really out there in terms of what’s possible — grow infinitesimally small.

So where did they all go, those writhing sea serpents, gargantuan crabs and city-block-sized squid? An article in an 1875 edition of the West Briton (a local Cornish weekly, still in print) offers clues. It tells the tale of two fishermen, setting nets in Gerrans Bay near Truro, who discovered a serpent “coiled about their floating cork. Upon their near approach, it lifted its head and showed signs of defiance, upon which they struck it forcibly with an oar”. Later, they pursued it and dragged it ashore for a look-see, “after which, it was killed on the rocks and most inconsiderably cast out to sea”.

Which is to say, if sea monsters existed, we must already have killed them. It’s something we’re worryingly good at. “It’s difficult to be tremendously optimistic about the persistence of ecosystems and many animal species,” says Naish, looking to the future of the oceans. “I find it hard to think that marine mammals and ray-finned fishes and sharks will persist into the future.” Instead, the sea monsters of tomorrow are going to be small and numerous, as the oceans, ever warmer and more acidic, fill with cephalopods, jellies, nematode worms and algae.

We can, however, look forward to some new invasions from the deep before that. With whole ecosystems shifting poleward as the planet warms, lionfish, sea snakes, crown-of-thorns starfish and at least three species of shark (hammerhead, ragged tooth and blacktip) are already heading for UK beaches.

Meanwhile, octopuses and squid will fill the niches vacated by over-harvested fish. Their life cycles are so short that they’ll be able to adapt faster than anything with a backbone. Right now, squid are multiplying crazily in British waters, although they’ll eventually lose out to the true inheritors of the oceans: the jellyfish.

In 2009, in the Sea of Japan, the giant Nomura jellyfish (up to two metres in diameter and weighing 200kg) began clogging and bursting fishing nets. This was deemed worthy of headlines at the time, but the jellyfish — most little more than gelatinous bags containing digestive organs and gonads — had barely begun their campaign of conquest. These days, they’re just as likely to be found blocking the seawater intake valves of British nuclear power plants.

For the future, Naish envisions a massively simplified food chain dominated by fast-growing things that can survive in warm seas where there’s very little oxygen. “You’re talking about a vast biomass but made of small detritus feeders living on algae,” he says. Given a few billion years of natural selection, some jellies may evolve into colonial super-organisms quite big enough to stride about on. “I like the idea of giant colonial jellies — like enormous but squishy islands,” he muses, “or big serpentine things that move slowly, drifting along on the weaker, less oxygenated currents of the future.”

That’s not all. A new surface fauna may evolve from mid-water and deep-water plankton, says Naish, “in which case your large surface-dwelling animals would be weird, spiny and translucent. A sea full of translucent, floating crustaceans; I see some merit in that idea.”

Mind reeling, I walk out of the exhibition past a Jenny Haniver. That’s the carcass of a ray or a skate that someone has cut and folded and dried, so that it ends up looking like a fairy, or a mermaid, demon, or dragon. No one really knows what they’re for. In some places, they’re said to have magic powers; most often they were simply made as curios for sailors. You find these sorts of things all over the world, from Mexico to Japan.

Morgawr may be a fake, and the Stronsay Beast just a hillock of rotten fish meat. Still, the monsters of the deep live. And for as long as human beings tread the earth, they cannot die.

Human/nature

Was the climate crisis inevitable? For the Financial Times, 13 September 2019

Everything living is dying out. A 2014 analysis of 3,000 species, confirmed by recent studies, reveals that half of all wild animals have been lost since 1970. The Amazon is burning, as is the Arctic.

An excess of carbon dioxide in the atmosphere, meanwhile, has not only played havoc with the climate but also reduced the nutrient value of plants by about 30 per cent since the 1950s.

And we’re running out of soil. In the US, it’s eroding 10 times faster than it’s being replaced. In China and India, the erosion is more than three times as bad. Five years ago, the UN Food and Agriculture Organization claimed we had fewer than 60 years of harvests left if soil degradation continued at its current rate.

Why have we waited until we are one generation away from Armageddon before taking such problems seriously?

A few suggestions: first, the environment is far too complicated to talk about — at least on the tangled information networks we have constructed for ourselves.

Second, we’re lazy and we’re greedy, like every other living thing on the planet — though because most of us co-operate with each other, we are arguably the least greedy and least lazy animals around.

Where we fall down is in our tendency to freeload on our future selves. “Discounting the future” is one of our worst habits, and one that in large part explains why we leave even important, life-and-death actions to the last minute.

Here’s a third reason why we’re dealing so late with climate change. It’s the weirdest, and maybe the most important of the three. It’s that we know we are going to die.

Thinking about environmental threats reminds us of our own mortality, and death is a prospect so appalling we’ll do anything — anything — to stop thinking about it.

“I used to wonder how people could stand the really demonic activity of working behind those hellish ranges in hotel kitchens, the frantic whirl of waiting on a dozen tables at one time,” wrote Ernest Becker in his Pulitzer-winning meditation The Denial of Death in 1973.

“The answer is so simple that it eludes us: the craziness of these activities is exactly that of the human condition. They are ‘right’ for us because the alternative is natural desperation.”

Psychologists inspired by Becker have run experiments to suggest it’s the terror of death that motivates consciousness and all its accomplishments. “It raised the pyramids in Egypt and razed the Twin Towers in Manhattan,” is the memorable judgment of the authors of 2015’s best-selling book The Worm at the Core.

This hardly sounds like good news. But it may offer us, if not a solution to the current crisis, at least a better, healthier and more positive way of approaching it.

No coping mechanism is infallible. We may be profoundly unwilling to contemplate our mortality, and to face up to the slow-burn, long-term threats to our existence, but that anxiety can’t ultimately be denied. Our response is to bundle it into catastrophes — in effect to construe the world in terms of crises to make everyday existence bearable.

Even positive visions of the future assume the necessity for cataclysmic change: why else do we fetishise “disruption”? “The concept of progress is to be grounded in the idea of the catastrophe,” as the German philosopher Walter Benjamin put it.

Yes, we could have addressed climate change much more easily in the 1970s, when the crisis wasn’t so urgent. But the fact is, we’re built for urgent action. A flood. A drought. A famine. We know where we are in a catastrophe. It may be that our best is yet to come.

Will our best be enough? Will we move quickly and coherently enough to save ourselves from the catastrophes attendant on massive climate change? That’s a hard question to answer.

The earliest serious attempts at modelling human futures were horrific. One commentator summed up Thomas Malthus’s famous 1798 Essay on the Principle of Population as “150 pages of excruciatingly detailed travellers’ accounts and histories . . . of bestial life, sickness, weakness, poor food, lack of ability to care for young, scant resources, famine, infanticide, war, massacre, plunder, slavery, cold, hunger, disease, epidemics, plague, and abortion.”

Malthus, an English cleric driven up the wall by positive Enlightenment thinkers such as Godwin and Condorcet, set out to remind everybody that people were animals. Like animals, their populations were bound eventually to exceed the available food supply. It didn’t matter that they dressed nicely or wrote poetry. If they overbred, they would starve.

We’ve been eluding this Malthusian trap for centuries, by bolting together one cultural innovation after another. No bread? Grow soy. No fish? Breed insects. Eventually, on a finite planet, Malthus will have his revenge — but when?

The energy thinker Vaclav Smil’s forthcoming book Growth studies the growth patterns of everything from microorganisms to mammals to entire civilisations. But the Czech-Canadian academic is chary about breaking anything as complicated as humanity down to a single metric.

“In the mid-1980s,” he recalls, “people used to ask me, when would the Chinese environment finally collapse? I was writing about this topic early on, and the point is, it was never going to collapse. Or it’s constantly collapsing, and they’re constantly fixing parts of it.”

Every major city in China has clean water and improving air quality, according to Smil. A few years ago people were choking on the smog.

“It’s the same thing with the planet,” he says. “Thirty years ago in Europe, the number-one problem wasn’t global warming, it was acid rain. Nobody mentions acid rain today because we desulphurised our coal-fired power plants and supplanted coal with natural gas. The world’s getting better and worse at the same time.”

Smil blames the cult of economics for the way we’ve been sitting on our hands while the planet heats up. The fundamental problem is that economics has become so divorced from fundamental reality,” he says.

“We have to eat, we have to put on a shirt and shoes, our whole lives are governed by the laws that govern the flows of energy and materials. In economics, though, everything is reduced to money, which is only a very imperfect measure of those flows. Until economics returns to the physical rules of human existence, we’ll always be floating in the sky and totally detached from reality.”

Nevertheless, Smil thinks we’d be better off planning for a good life in the here and now, and this entails pulling back from our current levels of consumption.

“But we’re not that stupid,” he says, “and we may have this taken care of by people’s own decision making. As they get richer, people find that children are very expensive, and children have been disappearing everywhere. There is not a single European country now in which fertility will be above replacement level. And even India is now close to the replacement rate of 2.1 children per family.”

So are we out of the tunnel, or at the end of the line? The brutal truth is, we’ll probably never know. We’re not equipped to know. We’re too anxious, too terrified, too greedy for the sort of certainty a complex environment is simply not going to provide.

Now that we’ve spotted this catastrophe looming over our heads, it’s with us for good. No one’s ever going to be able to say that it’s truly gone away. As Benjamin tersely concluded, “That things ‘just go on’ is the catastrophe.”