Newspapers and the “Manufacture of Consent”

When we think about newspapers, various associations come to mind. Examples include Jeff Bezos purchasing The Washington Post or William Randolph Hearst inspiring Citizen Kane. Newspapers have, to some extent, devolved into a vehicle for propaganda, as described in Walter Lippmann’s Public Opinion.

Lippmann is famously quoted for advocating the “manufacture of consent.”

Max Weber describes the economic function of newspapers:

The newspaper as an institution came into the service of commerce at an astonishingly late date.

The newspaper, as an institution, is not a product of capitalism. It brought together in the first place political news and then mainly all sorts of curiosities from the world at large. The advertisement, however, made its way into the newspaper very late. It was never entirely absent but originally it related to family announcements, while the advertisement as a notice by the merchant, directed toward finding a market, first becomes an established phenomenon at the end of the 18th century—in the journal which for a century was the first in the world, the “Times.” Official price bulletins did not become general until the 19th century; originally all the exchanges were closed clubs, as they have remained in America virtually down to the present. Hence in the 18th century, business depended on the organized exchange of letters. Rational trading between regions was impossible without secure transmission of letters. This was accomplished partly by the merchant guilds and in part by butchers, wheelwrights, etc. The final stage in the rationalization of transmission of letters was brought about by the post, which collected letters and in connection therewith made tariff agreements with commercial houses. In Germany, the family of Thurn and Taxis, who held the postal concession, made notable advances in the rationalization of communication by letter. Yet the volume of correspondence is in the beginning surprisingly small. In 1633, a million letters were posted in all England while today a place of 4,000 population will equal the number.

Max Weber, General Economic History, Collier Books, 1966 (Third Printing), page 220.

Herbert Hunt provides a useful overview of the newspaper as a political tool in his introduction to Honoré de Balzac’s Lost Illusions (French: Illusions perdues).

The first half of the nineteenth century witnessed the rapid rise to power of the periodical press. Journalism had been active — though dangerous to those engaged in it — during the Revolutionary period. Napoleon had kept the press under his thumb, as Giroudeau points out on page 235. The ‘freedom’ of the press was one of the most controversial issues both under the Restoration and the July Monarchy. Under Louis XVIII and Charles X the struggle between those who, like the Liberals and Bonapartists, wanted to keep the Revolutionary principles and gains intact, and the Conservatives of various hues, especially the ‘Ultras’, who wanted to put the political clock back, was an affair of major importance; likewise, under Louis-Philippe, the conflict between the spirit of stagnation and the parties in favour of ‘movement’. Balzac’s contention is that the majority of journalists under these three monarchs, instead of recognizing that they were called to a serious, even sacred mission, turned the Press into an instrument for self-advancement, prostituted principles to intrigue and used journalism merely as a means of acquiring money, position and power. He is reluctant to admit that there were great, responsible press organs, like Le Journal des Débats, Le Conservateur, Le Constitutionnel and, from 1824, Le Globe, which stood firm on principle; he is above all aware of the vogue which the petits journaux enjoyed after the fall of Napoleon, and of the role they played as political privateers.

The petits journaux were so-called because they were produced in smaller format than the important dailies or weeklies, which were more or less grave, staid and ponderous. They proliferated in Paris once the fall of the Empire had given a relative, though still precarious liberty to the Press — precarious because it was constantly threatened by the increasingly reactionary governments of the time. The politicians of the Right found it difficult to keep the newspapers under control even by such means as stamp-duty, caution-money, fines, suspensions and suppressions, the object of these being mainly to put obstacles in the way of would-be founders of hostile periodicals. The ‘little papers’, short-lived as they often proved to be, were much given to journalistic sharp-shooting. They preferred satire, personal attack, sarcasm and scandal-mongering to serious argument or the affirmation of ideals. They were mostly Opposition journals and were a constant thorn in the flesh of the Government. Balzac’s aim was to expose their addiction to ‘graft’, intrigue, blackmail and the misuse of the feuilleton, namely the bottom portion of the first page or other pages generally reserved for critical articles and frequently devoted to the malicious task of slashing literary reputations. Andoche Finot — the prototype of such later newspaper magnates as Émile de Girardin and Armand Dutacq, pioneers in 1836 in the founding of cheap dailies which relied on advertisement and serialized novels as a chief source of income — acquires a large share in a big daily and hands on to the equally unprincipled Lousteau the editorship of the ‘little paper’ he already owns. Balzac probably had Le Figaro chiefly in mind, a periodical which was constantly going bankrupt or being suppressed but kept popping up again under different editors. Hector Merlin’s royalist Drapeau Blanc, edited by Martainville, really existed, having been founded in 1819; so did Le Réveil. Other examples of ‘little papers’ before 1830 were Le Nain Jaume (Bonapartist), Le Diable Boiteux and Le Corsaire (both Liberal), Le Voleur, La Mode, La Silhouette, and, under Louis-Philippe, not only the phoenix-like Figaro, but also La Caricature, Le Charivari (ancestor of our English Punch), and once more Le Corsaire: a few among many. Louis-Philippe and his Cabinets were easy prey for these stinging gad-flies whose unremitting satire and innuendo remind one of the present-day Canard Enchaîné.

It is an amusing thought that, in the late twenties and early thirties, Balzac had himself been a contributor to these disreputable rags and sometimes had a hand in the running of them; for instance he had helped Philipon to found La Caricature. Throughout his career he contributed many novels in serial form to the more important newspapers, notably those founded by Girardin and Dutacq — La Presse and Le Siècle. But by the time he was writing A Great Man in Embryo he had left the petits journaux far behind him. He himself tried his luck as a newspaper-proprietor and editor: he bought La Chronique de Paris in 1836 and founded La Revue Parisienne in 1840. Both of these ventures failed. We can well imagine therefore what a large amount of bile was accumulating inside him. On the whole, reviews of his works appearing in periodicals had been hostile if not harsh. He suffered much from the disparagement of editors and critics such as Sainte-Beuve and Jules Janin respectively. He was always quarrelling with Émile de Girardin. And so he took his revenge. He had already made a preliminary attack on the periodical press in The Skin. And he followed up his attack of 1839 with his Monograph of the Paris Press (1842).

Honoré de Balzac, Lost Illusions, translated and introduced by Herbert Hunt, Penguin Books, 1971, pages xiv-xvi.

Balzac’s novel is very concerned with all aspects of journalism. For example, chapter 17 is titled “How a news-sheet is edited” and chapter 18 is a symposium on newspapers. Chapter 18 quotes a German guest who states, “I thank god there are no newspapers in my country.” (page 312). Another participant states, “In corporate crimes no one is implicated.” “A newspaper can behave in the most atrocious manner and no one on the staff considers that his own hands are soiled.” (page 314).

‘The influence and power of newspapers are only just dawning,’ said Finot. ‘Journalism is in its infancy; it will grow up. In ten years from now, everything will be subject to publicity. Thought will enlighten the world…’

Honoré de Balzac, Lost Illusions, Penguin Books, 1971, page 313.

Newspapers are an evil,’ said Claude Vignon. ‘An evil which could be utilized, but the Government wants to fight it. There’ll be a conflict. Who will go under? That’s the question.’

Honoré de Balzac, Lost Illusions, Penguin Books, 1971, page 313.

We should heed Vignon’s warning. Nazi Germany’s three main newspapers confirm this danger.

The Interconnectedness of Everything and How It Should Influence Our Thinking

Max Weber, considered to be the father of modern sociology alongside Émile Durkheim, wrote a classic of economic history, General Economic History (GermanWirtschaftsgeschichte).

Weber concludes chapter 25 (“Free Wholesale Trade”) with:

The railway is the most revolutionary instrumentality known to history, for economic life in general and not merely for commerce, but the railway was dependent on the age of iron; and it also like so many other things, was the plaything of princely and courtier interests.

Max Weber, General Economic History, Collier Books, Third Printing, 1966, page 221.

Zooming out you may intuit that these narrow gauge explanations are inadequate but much better than nothing. We are faced with the problem of connecting railroads and the Iron Age to larger transformations from which these railroads were born.

Your high school history book might mention Abraham Darby III; to quote Wikipedia:

He built the largest cast iron structure of his era: the first cast-iron bridge ever built, as a crossing over the Severn near Coalbrookdale. The bridge made it possible for the village of Ironbridge to grow up around it, with the area being subsequently named Ironbridge Gorge.

He was the third of four men of the same name, all English ironmasters, from several generations of a Quaker family that played a pivotal role in the Industrial Revolution. Without James Watt, there would be no steam engine; without which, there would be no railroad. Watt in turn improved upon the Newcomen atmospheric engine, while consulting other scientists.

Despite the brilliance of Watt’s engine, however, without the financial backing of Matthew Boulton, it might have been a failure.

If you read Watt and Boulton’s correspondence, Watt is always thinking locally, whereas Boulton is already talking about selling the invention globally. This is the ecosystem in which Weber’s economic observations reside.

Weber leads up to the point of the importance of railroads by discussing the evolution of land transport.

Land transport also remained as before. The post produced no change; it merely forwarded letters and small packages, but did not concern itself with large scale production, which was decisive for economic life.

Only the roads underwent an extraordinary improvement, through the construction of turnpikes. In this the French government under Sully took the lead, while England leased the roads to private enterprisers who collected tolls for their use. The building of the turnpikes wrought a revolution in commercial life comparable to no other before the appearance of the railways. There is no comparison between the present density of road traffic and that of this period. In 1793, 70,000 horses went through the little town of Lüneburg while as late as 1846 only 40,000 were used in freight transport in all Germany. The costs of land carriage amounted to ten or twenty times the freight on the railways at a later time, and were three to four times as high as the charges for inland shipping at the same period. A half billion ton-kilometers was the highest figure for transportation for the movement on land in Germany, while in 1913, 67 billions were carried on the railroads.

Max Weber, General Economic History, Collier Books, Third Printing, 1966, page 221.

Weber also connects railroads with the atmosphere of speculation:

Such speculation underwent an enormous expansion with the building of railroads; these provided the paper which first unchained the speculative urge. Under the head of goods, grains, and a few colonial products available in large volume, and then other goods, were drawn into the circle of exchange speculation during the 19th century.

Max Weber, General Economic History, Collier Books, Third Printing, 1966, page 219-220.

Everything is connected to everything else, and without realizing this truth and sensing the underlying transformations, you cannot get a clear signal from history.

Economics-Watching: BRICS Currency Creates Dilemma for the Dollar

by Christopher Whalen, from China Daily

The term “BRICS currency” typically refers to a hypothetical or proposed unified currency for the BRICS grouping. It’s not a single, physical currency currently in use, but rather a concept for a potential future monetary system that some suggest will reduce the dominance of the U.S. dollar in international trade and finance.

Is BRICS currency cooperation about immediate de-dollarization or long-term financial sovereignty? The answer is that BRICS cooperation may include reducing long-term dependence on the dollar as a means of exchange. The dollar is involved in more than half of all trade and 80 percent of all foreign exchange transactions. BRICS currency cooperation aims to gradually reduce the group’s dollar dependency, but challenges remain.

The BRICS concept came about not because the dollar is unsuitable as a means of exchange or unit of account, but rather because of the use of the dollar by Washington as a weapon. As I note in my book, Inflated: Money, Debt and the American Dream, the special role of the dollar in U.S. finance allows the U.S. government to impose harsh compliance and reporting requirements on foreign nationals and institutions. The U.S. is an arbitrary hegemon and does not follow reciprocity with other countries.

The global role of the dollar is an anomaly, the byproduct of two world wars had left the other antagonists broke by the time the Bretton Woods Agreement was signed in July 1944.

Choosing the fiat paper dollar as the default global reserve currency more than seven decades ago reflected the fact that the United States was one of the victors and possessed the wealth that gave Washington unchallenged economic leadership. Prior to World War I, the United Kingdom’s pound sterling was the global standard, but importantly, this paper currency was backed by gold — the only money that is not debt. The dollar, too, was backed by gold — until 1933, when the Franklin Roosevelt administration confiscated gold in private hands to prevent his government from collapsing.

Pound notes started to circulate in England in 1694, shortly after the establishment of the Bank of England. The paper pound helped to fuel the expansion of the British Empire, in large part because the only competing form of money was physical gold. When Britain and other nations left the gold standard in the 1930s, it was due to the deflation caused by the Great Depression rather than a deliberate choice.

The 19th-century rule attributed to English journalist and businessman Walter Bagehot says that in times of crisis, lend freely at a high rate against good collateral. Yet since the currency devaluation and gold seizures of 1933, fiat currencies and below-market interest rates have been the rule. In a global scheme in which the government occupies the prime position, the operative term remains “financial repression”, whereby governments control markets and artificially suppress rates of return on debt. For this reason, the dollar is losing its role as a store of value to gold.

The fact that the dollar continues to trade strongly versus other currencies reflects the reality that as the main means of exchange globally, the dollar cannot be easily replaced. One reason for this continued support for the dollar is that the trade in petroleum and other commodities is so large that it requires an equally large currency to accommodate it. Also, neither the Europeans nor the Japanese, the only two possible alternatives, are willing to risk the external deficits or inflation that the U.S. suffers as the host for the global currency.

What global currency will replace the fiat paper dollar? None. As this article is being written, gold is the second-largest reserve asset for central banks after the dollar. “The initiation in 2002 of the Shanghai Gold Exchange was of great strategic significance, both for gold and the global monetary system,” notes veteran gold fund manager Henry Smyth in an interview in The Institutional Risk Analyst. “Now it is completely clear what happened.”

Smyth and many other observers see the creation of the SGE in 2002 as the return of gold to the international monetary system. But while gold is growing in importance as a reserve asset for many countries, it does not mean that the role of the dollar as a global means of exchange or unit of account is about to change.

The dollar will remain the dominant asset. And even then, displacing the dollar will require a major change in the international monetary system, a change that is already underway.

The author is the chairman of Whalen Global Advisors LLC in New York and the author of Inflated: Money, Debt and the American Dream published by Wiley Global (2025).

China: Deep History

Winston Churchill says somewhere (if we paraphrase) that the further back you are able to look, the more secure your ability to analyze the present and the future. Without these ‘historical smarts’, your sense of direction is very feeble. Let us use the novel, Lost Illusions, by Honoré de Balzac as a back door into historical smarts.

This novel was originally published in three parts between 1837 and 1843 and is set mostly in the 1820s, primarily in provincial France. It is unique because it starts with technology and commerce.

At the time when this story begins, the Stanhope press and inking-rollers were not yet in use in small provincial printing-offices. Angoulême, although its paper-making industry kept it in contact with Parisian printing, was still using those wooden presses from which the now obsolete metaphor ‘making the presses groan’ originated. Printing there was so much behind the times that the pressmen still used leather balls spread with ink to dab on the characters. The bed of the press holding the letter-filled ‘forme’ to which the paper is applied was still made of stone and so justified its name ‘marble’. The ravenous machines of our times have so completely superseded this mechanism — to which, despite its imperfections, we owe the fine books produced by the Elzevirs, the Plantins, the Aldi and the Didots — that it is necessary to mention this antiquated equipment which Jérôme-Nicolas Séchard held in superstitious affection; it has its part to play in this great and trivial story.

Not only do we get this conceptual framework about printing technology, but later on in the novel, Balzac gives us a further insight into paper-making and textiles, including a long discussion of China.

In England, where four-fifths of the population use cotton to the exclusion of linen, they make nothing but cotton paper. The cotton paper is very soft and easily creased to begin with, and it has a further defect: it is so soluble that if you seep a book made of cotton paper in water for fifteen minutes, it turns to a pulp, while an old book left in water for a couple of hours is not spoilt. You could dry the old book, and the pages, though yellow and faded, would still be legible, the work would not be destroyed.

“There is a time coming when legislation will equalize our fortunes, and we shall all be poor together; we shall want our linen and our books to be cheap, just as people are beginning to prefer small pictures because they have not wall space enough for large ones. Well, the shirts and the books will not last, that is all; it is the same on all sides, solidity is drying out. So this problem is one of the first importance for literature, science, and politics.

“One day, in my office, there was a hot discussion going on about the material that the Chinese use for making paper. Their paper is far better than ours, because the raw material is better; and a good deal was said about this thin, light Chinese paper, for if it is light and thin, the texture is close, there are no transparent spots in it. In Paris there are learned men among the printers’ readers; Fourier and Pierre Leroux are Lachevardiere’s readers at this moment; and the Comte de Saint-Simon, who happened to be correcting proofs for us, came in in the middle of the discussion. He told us at once that, according to Kempfer and du Halde, the Broussonetia furnishes the substance of the Chinese paper; it is a vegetable substance (like linen or cotton for that matter). Another reader maintained that Chinese paper was principally made of an animal substance, to wit, the silk that is abundant there. They made a bet about it in my presence. The Messieurs Didot are printers to the Institute, so naturally they referred the question to that learned body. M. Marcel, who used to be superintendent of the Royal Printing Establishment, was umpire, and he sent the two readers to M. l’Abbe Grozier, Librarian at the Arsenal. By the Abbe’s decision they both lost their wages. The paper was not made of silk nor yet from the Broussonetia; the pulp proved to be the triturated fibre of some kind of bamboo. The Abbe Grozier had a Chinese book, an iconographical and technological work, with a great many pictures in it, illustrating all the different processes of paper-making, and he showed us a picture of the workshop with the bamboo stalks lying in a heap in the corner; it was extremely well drawn.

“Lucien told me that your father, with the intuition of a man of talent, had a glimmering of a notion of some way of replacing linen rags with an exceedingly common vegetable product, not previously manufactured, but taken direct from the soil, as the Chinese use vegetable fibre at first hand. I have classified the guesses made by those who came before me, and have begun to study the question. The bamboo is a kind of reed; naturally I began to think of the reeds that grow here in France.

Labor is very cheap in China, where a workman earns three halfpence a day, and this cheapness of labor enables the Chinese to manipulate each sheet of paper separately. They take it out of the mould, and press it between heated tablets of white porcelain, that is the secret of the surface and consistence, the lightness and satin smoothness of the best paper in the world. Well, here in Europe the work must be done by machinery; machinery must take the place of cheap Chinese labor. If we could but succeed in making a cheap paper of as good a quality, the weight and thickness of printed books would be reduced by more than one-half. A set of Voltaire, printed on our woven paper and bound, weighs about two hundred and fifty pounds; it would only weigh fifty if we used Chinese paper. That surely would be a triumph…

In 2025, we are to some extent, back to China, going from the proto-industrial world to our industrial and even digital world.

To educate oneself on all of this, you should look at the supreme scholarly achievement of the 20th century, namely Professor Joseph Needham’s masterpiece, Science and Civilisation in China.

India: Deep History

In his lectures, Professor Amartya Sen, the Harvard Nobel Prize in Economics winner, mentions Sir Mortimer Wheeler, Director-General of the Archaeological Survey of India. Wheeler wrote, while reporting on the excavation of the Indus Valley Civilization (of India), that the plumbing and sewerage were advanced, in some ways surpassing modern equivalents.

Sen’s larger point is that history is characterized by phases of rise and fall and not just classes and class struggles à la Marx.

Consider the following depiction of the East India Company, from The Anarchy: The Relentless Rise of the East India Company (also subtitled The East India Company, Corporate Violence, and the Pillage of an Empire) by William Dalrymple.


On 28 August 1608, Captain William Hawkins, a bluff sea captain with the Third Voyage, anchored his ship, the Hector, off Surat, and so became the first commander of an EIC vessel to set foot on Indian soil.

India then had a population of 150 million — about a fifth of the world’s total — and was producing about a quarter of global manufacturing; indeed, in many ways it was the world’s industrial powerhouse and the world’s leader in manufactured textiles. Not for nothing are so many English words connected with weaving — chintz, calico, shawl, pyjamas, khaki, dungarees, cummerbund, taffetas — of Indian origin. It was certainly responsible for a much larger share of world trade than any comparable zone and the weight of its economic power even reached Mexico, whose textile manufacture suffered a crisis of ‘de-industrialisation’ due to Indian cloth imports. In comparison, England then had just 5 per cent of India’s population and was producing just under 3 per cent of the world’s manufactured goods. A good proportion of the profits on this found its way to the Mughal exchequer in Agra, making the Mughal Emperor, with an income of around £100 million,* by far the richest monarch in the world.

The Mughal capitals were the megacities of their day: ‘They are second to none either in Asia or in Europe,’ thought the Jesuit Fr Antonio Monserrate, ‘with regards either to size, population, or wealth. Their cities are crowded with merchants, who gather from all over Asia. There is no art or craft which is not practised there.’ Between 1586 and 1605, European silver flowed into the Mughal heartland at the astonishing rate of 18 metric tons a year, for as William Hawkins observed, all nations bring coyne and carry away commodities for the same’. For their grubby contemporaries in the West, stumbling around in their codpieces, the silk-clad Mughals, dripping in jewels, were the living embodiment of wealth and power — a meaning that has remained impregnated in the word ‘mogul’ ever since.

By the early seventeenth century, Europeans had become used to easy military victories over the other peoples of the world.

* Over £10,000 million today.

Think of the larger point: what you just read is the story of Indian de-industrialization and its negative results. Ask yourself whether American de-industrialization is something of an echo of this, as manufacturing is offshored.

Navigating through Sources

Consciousness and the Novel: Connected Essays by the famous British novelist David Lodge is a classic work published by Harvard in 2004.

In this Lodge book, the author mentions a famous British society-watcher, Charles Masterman. In 1909, Masterman published his best-known study, The Condition of England, which tells us that England at that time experienced a greater inflow of migrants into London than in previous centuries taken together.

[Charles Frederick Gurney Masterman PC MP (24 October 1873 – 17 November 1927) was a British radical Liberal Party politician, intellectual and man of letters. He worked closely with such Liberal leaders as David Lloyd George and Winston Churchill in designing social welfare projects, including the National Insurance Act 1911. During the First World War, he played a central role in the main government propaganda agency.]

We then notice that one recurrent topic in various movie versions of the E. M. Forster novel Howards End (1910, set in those years) is the “horrifying” trend where great mansions and stately estates (Howards End and Wickham Place, say, in the novel) are all being demolished and replaced by ugly “flats.”

There must be, one thinks, a direct link between all the massive migrations into London at the time and all the proliferating flats at the “expense” of beautiful and historical villas. (This “demolish” trend is also part of the story of the classic novel A Handful of Dust by Evelyn Waugh, 1934)

In the predecessor to Downton Abbey called Upstairs, Downstairs, the story ends in 1930 with a sign outside the great “house” at Eaton Place offering flats coming soon, as the demand for housing (think of San Francisco today) is so massive that sellers can make a fortune selling out to developers, move into one of the flats being created, and live off the sale for the rest of their lives and “duck” the higher “Lloyd George taxes.” (In Downton Abbey, the dowager played by Maggie Smith repeatedly lashes out at Prime Minister Lloyd George as a kind of financial traitor.)

We see from this simple example how students should learn to “jump” between books and movies and TV miniseries to get a stronger focus on what’s being depicted on screens and pages and not just “swim along” at the surface level without any “drilling down.”

Education is largely the struggle or habit where students learn to bring pattern and structure out of “chaos,” thus giving narratives some overall shape.

This reminds one of the opening lines of Beryl Markham’s 1942 Africa memoir:

“How is it possible to bring order out of memory? I should like to begin at the beginning, patiently, like a weaver at his loom. I should like to say, ‘This is the place to start; there can be no other.’ ”

from West with the Night by Beryl Markham

This is a similar impetus: to bring order out of memory or others’ memories in books and movies from various times and places.

Movies as a Parallel University: The Case of Romantic Imperialism

When we think of romantic imperialism, we think of Rudyard Kipling’s poems, Winston Churchill’s The River War and perhaps Teddy Roosevelt’s macho “strenuous life” romanticized militarism (which the neocons somewhat knowingly aped to get the U.S. to invade Iraq in 2003). We should also recall British movie classics like The Four Feathers and “deflationary” versions of these jingoistic notions in The Man Who Would be King. During the 1930s, the Hungarian brothers Alexander & Zoltan Korda created many classic empire-celebration films in London, such as 1935’s Sanders of the River.

The Nigerian writer Chinua Achebe’s attack on Joseph Conrad’s Heart of Darkness as a kind of toxic “othering” of all Africans is a culmination of these imperial and anti-imperial tendencies.

Lastly, Maupassant’s classic Bel-Ami represents Algeria as a “colonial badlands” for French domination, killing, despoiling, profiteering, and later culminates in Meursault’s random murder of an Arab again in Algeria in Camus’ classic The Stranger. This literary concatenation also fits into this set of colonial imperial phenomena.

Niall Ferguson (the famous Harvard/Stanfordempire enthusiast”) finds his forerunner in the 1940 classic movie Beyond Tomorrow. The following “row” takes place on Christmas Eve between Chadwick (the Niall Ferguson imperialist) and Melton who sees empire as land-grabbing which you can dress up and prettify any way you like (“a grab is a grab,” he says):

Allan Chadwick: I tell you England’s territorial expansion had quite a different significance.

George Melton: No matter how thin you slice it, a grab is a grab.

Allan Chadwick: Grab?

That’s a specious term. England carried civilization into the wilderness. What was Australia before she redeemed it from the Aborigines?

Allan Chadwick: The truth is there isn’t an acre of the Empire that isn’t proud to fly the British flag.

The quick irritated exchange from Beyond Tomorrow is a good example of this eternal argument, allowing you to then “jump off” from this movie-as-university to do more exploring.

Precursors of Trumpism in American Culture: The Great Gatsby

Donald Trump’s brand of “nativist populism” is prefigured very clearly in the 1925 classic American novel The Great Gatsby by F. Scott Fitzgerald, the subject of several film adaptations. The anxiety in the Trump movement is called “The Great Replacement” (i.e., white people supplanted by minoritiesAmerican non-whites will “link up” with Mexicans and Chinese, stealing the place, property and role of white people). It has never occurred to these people that most of the world’s population is non-white. In The Great Gatsby, Tom Buchanan, Daisy’s husband, represents this anxiety.

He links these fears to the anxiety that “the sun is getting hotter.” (That is, he’s being threatened cosmically too, not only by racial demography.) His fears are not of “climate change,” based on something rational, but obsessive and maniacal “threat assessments.” They mirror the “irrational clusters of threats” of Trump and his voters, who “want to be paid in advance forever for their being white, Christian and American.” They demand “racial tenure.” This Tom Buchanan syndrome may be considered a type of “globalization backlash.”

Tom Buchanan lays it out in the novel:

“Civilization’s going to pieces,” broke out Tom violently. “I’ve gotten to be a terrible pessimist about things. Have you read The Rise of the Colored Empires by this man Goddard?…Well, it’s a fine book, and everybody ought to read it. The idea is if we don’t look out the white race will be — will be utterly submerged. It’s all scientific stuff; it’s been proved…This fellow has worked out the whole thing. It’s up to us, who are the dominant race, to watch out or those other races will have control of things…The idea is that we’re Nordics. I am, and you are, and we’ve produced all the things that go to make civilization—oh, science and art, and all that. Do you see?”

(The Great Gatsby, Chapter 1)

“…Nowadays people begin by sneering at family life and family institutions, and next they’ll throw everything overboard and have marriage between black and white.”

Flushed with his impassioned gibberish, he saw himself standing alone on the last barrier of civilization.

“We’re all white here,” murmured Jordan.

Angry as I was, as we all were, I was tempted to laugh whenever he opened his mouth. The transition from libertine to prig was so complete.

(The Great Gatsby, Chapter 7)

The Brexit phenomenon in the United Kingdom is not identical but does overlap (i.e., the “left behind” people in England want “historical ethnic-national tenure” and a kind of re-segregation).

Education: Disease, History and Lit

The Italian writer Giovanni Boccaccio lived through the plague as it ravaged the city of Florence in 1348. The experience inspired him to write The Decameron.

The Plague of 1665 in England was a major upheaval affecting Isaac Newton’s life.

The 1984 movie, A Passage to India (David Lean) set in 1920s India, has a scene where the ever-present lethal threat of cholera is discussed as Doctor Aziz lies sick of a fever.

The W. Somerset Maugham novel, The Painted Veil (2006 movie) is also about cholera in the Chinese countryside in the 1920s.

Manzoni’s 1827 The Betrothed, the most famous classic novel of Italian literature, centers on the plague to drive the story.

Overview:

Etymologically, the term “pest” derives from the Latin word “pestis” (pest, plague, curse). Hardly any disease had such cultural and historical relevance as the bubonic plague. Throughout the centuries, the plague was the most terrifying infectious contagious disease which generated a series of demographic crises. The plague epidemics influenced the evolution of society biologically and culturally speaking. The Black Death was one of the most devastating pandemics in human history, is estimated to have killed 30%–60% of Europe’s population, reducing the world’s population from an estimated 450 million to between 350 and 375 million in 1400. This has been seen as having created a series of religious, social and economic upheavals, which had profound effects on the course of European history. It took 150 years for Europe’s population to recover.

The plague returned at various times, killing more people, until it left Europe in the 19th century. Modern epidemiology (Dr. John Snow, London) has its roots in cholera management and water sanitation as well as waste management.

Education involves seeing disease as a major protagonist in all history and not as a footnote.

The classic Plagues and Peoples should accordingly be studied by every student: Plagues and Peoples is a book on epidemiological history by William Hardy McNeill, published in 1976.

It was a critical and popular success, offering a radically new interpretation of the extraordinary impact of infectious disease on cultures and world history itself.

Early View Alert: Water Resources Research

from the American Geophysical Union’s journals:

Research Articles

Modeling the Snow Depth Variability with a High-Resolution Lidar Data Set and Nonlinear Terrain Dependency

by T. Skaugen & K. Melvold

Summary: Using airborne laser, 400 million snow depth measurements at Hardangervidda in Southern Norway have been collected. The amount of data has made in-depth studies of the spatial distribution of snow and its interaction with the terrain and vegetation possible. We find that the terrain variability, expressed by the square slope, the average amount of snow, and whether the terrain is vegetated or not, largely explains the variation of snow depth. With this information it is possible to develop equations that predict snow depth variability that can be used in environmental models, which again are used for important tasks such as flood forecasting and hydropower planning. One major advantage is that these equations can be determined from the data that are, in principle, available everywhere, provided there exists a detailed digital model of the terrain.

[Archived PDF article]

Phosphorus Transport in Intensively Managed Watersheds

by Christine L. Dolph, Evelyn Boardman, Mohammad Danesh-Yazdi, Jacques C. Finlay, Amy T. Hansen, Anna C. Baker & Brent Dalzell

Abstract: When phosphorus from farm fertilizer, eroded soil, and septic waste enters our water, it leads to problems like toxic algae blooms, fish kills, and contaminated drinking supplies. In this study, we examine how phosphorus travels through streams and rivers of farmed areas. In the past, soil lost from farm fields was considered the biggest contributor to phosphorus pollution in agricultural areas, but our study shows that phosphorus originating from fertilizer stores in the soil and from crop residue, as well as from soil eroded from sensitive ravines and bluffs, contributes strongly to the total amount of phosphorus pollution in agricultural rivers. We also found that most phosphorus leaves farmed watersheds during the very highest river flows. Increased frequency of large storms due to climate chaos will therefore likely worsen water quality in areas that are heavily loaded with phosphorus from farm fertilizers. Protecting water in agricultural watersheds will require knowledge of the local landscape along with strategies to address (1) drivers of climate chaos, (2) reduction in the highest river flows, and (3) ongoing inputs and legacy stores of phosphorus that are readily transported across land and water.

[Archived PDF of article]

Detecting the State of the Climate System via Artificial Intelligence to Improve Seasonal Forecasts and Inform Reservoir Operations

by Matteo Giuliani, Marta Zaniolo, Andrea Castelletti, Guido Davoli & Paul Block

Abstract: Increasingly variable hydrologic regimes combined with more frequent and intense extreme events are challenging water systems management worldwide. These trends emphasize the need of accurate medium- to long-term predictions to timely prompt anticipatory operations. Despite in some locations global climate oscillations and particularly the El Niño Southern Oscillation (ENSO) may contribute to extending forecast lead times, in other regions there is no consensus on how ENSO can be detected, and used as local conditions are also influenced by other concurrent climate signals. In this work, we introduce the Climate State Intelligence framework to capture the state of multiple global climate signals via artificial intelligence and improve seasonal forecasts. These forecasts are used as additional inputs for informing water system operations and their value is quantified as the corresponding gain in system performance. We apply the framework to the Lake Como basin, a regulated lake in northern Italy mainly operated for flood control and irrigation supply. Numerical results show the existence of notable teleconnection patterns dependent on both ENSO and the North Atlantic Oscillation over the Alpine region, which contribute in generating skillful seasonal precipitation and hydrologic forecasts. The use of this information for conditioning the lake operations produces an average 44% improvement in system performance with respect to a baseline solution not informed by any forecast, with this gain that further increases during extreme drought episodes. Our results also suggest that observed preseason sea surface temperature anomalies appear more valuable than hydrologic-based seasonal forecasts, producing an average 59% improvement in system performance.

[Archived PDF of article]

Landscape Water Storage and Subsurface Correlation from Satellite Surface Soil Moisture and Precipitation Observations

by Daniel J. Short Gianotti, Guido D. Salvucci, Ruzbeh Akbar, Kaighin A. McColl, Richard Cuenca & Dara Entekhabi

Abstract: Surface soil moisture measurements are typically correlated to some degree with changes in subsurface soil moisture. We calculate a hydrologic length scale, λ, which represents (1) the mean-state estimator of total column water changes from surface observations, (2) an e-folding length scale for subsurface soil moisture profile covariance fall-off, and (3) the best second-moment mass-conserving surface layer thickness for a simple bucket model, defined by the data streams of satellite soil moisture and precipitation retrievals. Calculations are simple, based on three variables: the autocorrelation and variance of surface soil moisture and the variance of the net flux into the column (precipitation minus estimated losses), which can be estimated directly from the soil moisture and precipitation time series. We develop a method to calculate the lag-one autocorrelation for irregularly observed time series and show global surface soil moisture autocorrelation. λ is driven in part by local hydroclimate conditions and is generally larger than the 50-mm nominal radiometric length scale for the soil moisture retrievals, suggesting broad subsurface correlation due to moisture drainage. In all but the most arid regions, radiometric soil moisture retrievals provide more information about ecosystem-relevant water fluxes than satellite radiometers can explicitly “see”; lower-frequency radiometers are expected to provide still more statistical information about subsurface water dynamics.

[Archived PDF of article]

Process-Guided Deep Learning Predictions of Lake Water Temperature

by Jordan S. Read, Xiaowei Jia, Jared Willard, Alison P. Appling, Jacob A. Zwart, Samantha K. Oliver, Anuj Karpatne, Gretchen J. A. Hansen, Paul C. Hanson, William Watkins, Michael Steinbach & Vipin Kumar

Abstract: The rapid growth of data in water resources has created new opportunities to accelerate knowledge discovery with the use of advanced deep learning tools. Hybrid models that integrate theory with state-of-the art empirical techniques have the potential to improve predictions while remaining true to physical laws. This paper evaluates the Process-Guided Deep Learning (PGDL) hybrid modeling framework with a use-case of predicting depth-specific lake water temperatures. The PGDL model has three primary components: a deep learning model with temporal awareness (long short-term memory recurrence), theory-based feedback (model penalties for violating conversation of energy), and model pre-training to initialize the network with synthetic data (water temperature predictions from a process-based model). In situ water temperatures were used to train the PGDL model, a deep learning (DL) model, and a process-based (PB) model. Model performance was evaluated in various conditions, including when training data were sparse and when predictions were made outside of the range in the training data set. The PGDL model performance (as measured by root-mean-square error (RMSE)) was superior to DL and PB for two detailed study lakes, but only when pretraining data included greater variability than the training period. The PGDL model also performed well when extended to 68 lakes, with a median RMSE of 1.65 °C during the test period (DL: 1.78 °C, PB: 2.03 °C; in a small number of lakes PB or DL models were more accurate). This case-study demonstrates that integrating scientific knowledge into deep learning tools shows promise for improving predictions of many important environmental variables.

[Archived PDF of article]

Adjustment of Radar-Gauge Rainfall Discrepancy Due to Raindrop Drift and Evaporation Using the Weather Research and Forecasting Model and Dual-Polarization Radar

by Qiang Dai, Qiqi Yang, Dawei Han, Miguel A. Rico-Ramirez & Shuliang Zhang

Abstract: Radar-gauge rainfall discrepancies are considered to originate from radar rainfall measurements while ignoring the fact that radar observes rain aloft while a rain gauge measures rainfall on the ground. Observations of raindrops observed aloft by weather radars consider that raindrops fall vertically to the ground without changing in size. This premise obviously does not stand because raindrop location changes due to wind drift and raindrop size changes due to evaporation. However, both effects are usually ignored. This study proposes a fully formulated scheme to numerically simulate both raindrop drift and evaporation in the air and reduces the uncertainties of radar rainfall estimation. The Weather Research and Forecasting model is used to simulate high-resolution three-dimensional atmospheric fields. A dual-polarization radar retrieves the raindrop size distribution for each radar pixel. Three schemes are designed and implemented using the Hameldon Hill radar in Lancashire, England. The first considers only raindrop drift, the second considers only evaporation, and the last considers both aspects. Results show that wind advection can cause a large drift for small raindrops. Considerable loss of rainfall is observed due to raindrop evaporation. Overall, the three schemes improve the radar-gauge correlation by 3.2%, 2.9%, and 3.8% and reduce their discrepancy by 17.9%, 8.6%, and 21.7%, respectively, over eight selected events. This study contributes to the improvement of quantitative precipitation estimation from radar polarimetry and allows a better understanding of precipitation processes.

[Archived PDF of article]

The Role of Collapsed Bank Soil on Tidal Channel Evolution: A Process-Based Model Involving Bank Collapse and Sediment Dynamics

by K. Zhao, Z. Gong, F. Xu, Z. Zhou, C. K. Zhang, G. M. E. Perillo & G. Coco

Abstract: We develop a process-based model to simulate the geomorphodynamic evolution of tidal channels, considering hydrodynamics, flow-induced bank erosion, gravity-induced bank collapse, and sediment dynamics. A stress-deformation analysis and the Mohr-Coulomb criterion, calibrated through previous laboratory experiments, are included in a model simulating bank collapse. Results show that collapsed bank soil plays a primary role in the dynamics of bank retreat. For bank collapse with small bank height, tensile failure in the middle of the bank (Stage I), tensile failure on the bank top (Stage II), and sectional cracking from bank top to the toe (Stage III) are present sequentially before bank collapse occurs. A significant linear relation is observed between bank height and the contribution of bank collapse to bank retreat. Contrary to flow-induced bank erosion, bank collapse prevents further widening since the collapsed bank soil protects the bank from direct bank erosion. The bank profile is linear or slightly convex, and the planimetric shape of tidal channels (gradually decreasing in width landward) is similar when approaching equilibrium, regardless of the consideration of bank erosion and collapse. Moreover, the simulated width-to-depth ratio in all runs is comparable with observations from the Venice Lagoon. This indicates that the equilibrium configuration of tidal channels depends on hydrodynamic conditions and sediment properties, while bank erosion and collapse greatly affect the transient behavior (before equilibrium) of the tidal channels. Overall, this contribution highlights the importance of collapsed bank soil in investigating tidal channel morphodynamics using a combined perspective of geotechnics and soil mechanics.

[Archived PDF of article]

A Physically Based Method for Soil Evaporation Estimation by Revisiting the Soil Drying Process

by Yunquan Wang, Oliver Merlin, Gaofeng Zhu & Kun Zhang

Abstract: While numerous models exist for soil evaporation estimation, they are more or less empirically based either in the model structure or in the determination of introduced parameters. The main difficulty lies in representing the water stress factor, which is usually thought to be limited by capillarity-supported water supply or by vapor diffusion flux. Recent progress in understanding soil hydraulic properties, however, have found that the film flow, which is often neglected, is the dominant process under low moisture conditions. By including the impact of film flow, a reexamination on the typical evaporation process found that this usually neglected film flow might be the dominant process for supporting the Stage II evaporation (i.e., the fast falling rate stage), besides the generally accepted capillary flow-supported Stage I evaporation and the vapor diffusion-controlled Stage III evaporation. A physically based model for estimating the evaporation rate was then developed by parameterizing the Buckingham-Darcy’s law. Interestingly, the empirical Bucket model was found to be a specific form of the proposed model. The proposed model requires the in-equilibrium relative humidity as the sole input for representing water stress and introduces no adjustable parameter in relation to soil texture. The impact of vapor diffusion was also discussed. Model testing with laboratory data yielded an excellent agreement with observations for both thin soil and thick soil column evaporation experiments. Model evaluation at 15 field sites generally showed a close agreement with observations, with a great improvement in the lower range of evaporation rates in comparison with the widely applied Priestley and Taylor Jet Propulsion Laboratory model.

[Archived PDF of article]

Floodplain Land Cover and Flow Hydrodynamic Control of Overbank Sedimentation in Compound Channel Flows

by Carmelo Juez, C. Schärer, H. Jenny, A. J. Schleiss & M. J. Franca

Abstract: Overbank sedimentation is predominantly due to fine sediments transported under suspension that become trapped and settle in floodplains when high-flow conditions occur in rivers. In a compound channel, the processes of exchanging water and fine sediments between the main channel and floodplains regulate the geomorphological evolution and are crucial for the maintenance of the ecosystem functions of the floodplains. These hydrodynamic and morphodynamic processes depend on variables such as the flow-depth ratio between the water depth in the main channel and the water depth in the floodplain, the width ratio between the width of the main channel and the width of the floodplain, and the floodplain land cover characterized by the type of roughness. This paper examines, by means of laboratory experiments, how these variables are interlinked and how the deposition of sediments in the compound channel is jointly determined by them. The combination of these compound channel characteristics modulates the production of vertically axised large turbulent vortical structures in the mixing interface. Such vortical structures determine the water mass exchange between the main channel and the floodplain, conditioning in turn the transport of sediment particles conveyed in the water, and, therefore, the resulting overbank sedimentation. The existence and pattern of sedimentation are conditioned by both the hydrodynamic variables (the flow-depth ratio and the width ratio) and the floodplain land cover simulated in terms of smooth walls, meadow-type roughness, sparse-wood-type roughness, and dense-wood-type roughness.

[Archived PDF of article]

Identifying Actionable Compromises: Navigating Multi-city Robustness Conflicts to Discover Cooperative Safe Operating Spaces for Regional Water Supply Portfolios

by D. F. Gold, P. M. Reed, B. C. Trindade & G. W. Characklis

Summary: Cooperation among neighboring urban water utilities can help water managers face challenges stemming from climate change and population growth. Water utilities can cooperate by coordinating water transfers and water restrictions in times of water scarcity (drought) so that water is provided to areas that need it most. In order to successfully implement these policies, however, cooperative partners must find a compromise that is acceptable to all regional actors, a task complicated by asymmetries in resources and risks often present in regional systems. The possibility of deviations from agreed upon actions is another complicating factor that has not been addressed in water resources literature. Our study focuses on four urban water utilities in the Research Triangle region of North Carolina who are investigating cooperative drought mitigation strategies. We contribute a framework that includes the use of simulation models, optimization algorithms, and statistical tools to aid cooperating partners in finding acceptable compromises that are tolerant modest deviations in planned actions. Our results can be used by regional utilities to avoid or alleviate potential planning conflicts and are broadly applicable to urban regional water supply planning across the globe.

[Archived PDF of article]

Detecting Changes in River Flow Caused by Wildfires, Storms, Urbanization, Regulation, and Climate across Sweden

by Berit Arheimer & Göran Lindström

Abstract: Changes in river flow may appear from shifts in land cover, constructions in the river channel, and climatic change, but currently there is a lack of understanding of the relative importance of these drivers. Therefore, we collected gauged river flow time series from 1961 to 2018 from across Sweden for 34 disturbed catchments to quantify how the various types of disturbances have affected river flow. We used trend analysis and the differences in observations versus hydrological modeling to explore the effects on river flow from (1) land cover changes from wildfires, storms, and urbanization; (2) dam constructions with regulations for hydropower production; and (3) climate-change impact in otherwise undisturbed catchments. A mini model ensemble, consisting of three versions of the S-HYPE model, was used, and the three models gave similar results. We searched for changes in annual and daily stream flow, seasonal flow regime, and flow duration curves. The results show that regulation of river flow has the largest impact, reducing spring floods with up to 100% and increasing winter flow by several orders of magnitude, with substantial effects transmitted far downstream. Climate changed the total river flow up to 20%. Tree removal by wildfires and storms has minor impacts at medium and large scales. Urbanization, on the contrary, showed a 20% increase in high flows also at medium scales. This study emphasizes the benefits of combining observed time series with numerical modeling to exclude the effect of varying weather conditions, when quantifying the effects of various drivers on long-term streamflow shifts.

[Archived PDF of article]

Assessing the Feasibility of Satellite-Based Thresholds for Hydrologically Driven Landsliding

by Matthew A. Thomas, Brian D. Collins & Benjamin B. Mirus

Summary: Soil wetness and rainfall contribute to landslides across the world. Using soil moisture sensors and rain gauges, these environmental conditions have been monitored at numerous points across the Earth’s surface to define threshold conditions, above which landsliding should be expected for a localized area. Satellite-based technologies also deliver estimates of soil wetness and rainfall, potentially offering an approach to develop thresholds as part of landslide warning systems over larger spatial scales. To evaluate the potential for using satellite-based measurements for landslide warning, we compare the accuracy of landslide thresholds defined with ground- versus satellite-based soil wetness and rainfall information. We find that the satellite-based data over-predict soil wetness during the time of year when landslides are most likely to occur, resulting in thresholds that also over-predict the potential for landslides relative to thresholds informed by direct measurements on the ground. Our results encourage the installation of more ground-based monitoring stations in landslide-prone settings and the cautious use of satellite-based data when more direct measurements are not available.

[Archived PDF of article]

Modeling the Translocation and Transformation of Chemicals in the Soil-Plant Continuum: A Dynamic Plant Uptake Module for the HYDRUS Model

by Giuseppe Brunetti, Radka Kodešová & Jiří Šimůnek

Abstract: Food contamination is responsible for thousands of deaths worldwide every year. Plants represent the most common pathway for chemicals into the human and animal food chain. Although existing dynamic plant uptake models for chemicals are crucial for the development of reliable mitigation strategies for food pollution, they nevertheless simplify the description of physicochemical processes in soil and plants, mass transfer processes between soil and plants and in plants, and transformation in plants. To fill this scientific gap, we couple a widely used hydrological model (HYDRUS) with a multi-compartment dynamic plant uptake model, which accounts for differentiated multiple metabolization pathways in plant’s tissues. The developed model is validated first theoretically and then experimentally against measured data from an experiment on the translocation and transformation of carbamazepine in three vegetables. The analysis is further enriched by performing a global sensitivity analysis on the soilplant model to identify factors driving the compound’s accumulation in plants’ shoots, as well as to elucidate the role and the importance of soil hydraulic properties on the plant uptake process. Results of the multilevel numerical analysis emphasize the model’s flexibility and demonstrate its ability to accurately reproduce physicochemical processes involved in the dynamic plant uptake of chemicals from contaminated soils.

[Archived PDF of article]

Physical Controls on Salmon Redd Site Selection in Restored Reaches of a Regulated, Gravel-Bed River

by Lee R. Harrison, Erin Bray, Brandon Overstreet, Carl J. Legleiter, Rocko A. Brown, Joseph E. Merz, Rosealea M. Bond, Colin L. Nicol & Thomas Dunne

Abstract: Large-scale river restoration programs have emerged recently as a tool for improving spawning habitat for native salmonids in highly altered river ecosystems. Few studies have quantified the extent to which restored habitat is utilized by salmonids, which habitat features influence redd site selection, or the persistence of restored habitat over time. We investigated fall-run Chinook salmon spawning site utilization and measured and modeled corresponding habitat characteristics in two restored reaches: a reach of channel and floodplain enhancement completed in 2013 and a reconfigured channel and floodplain constructed in 2002. Redd surveys demonstrated that both restoration projects supported a high density of salmon redds, 3 and 14 years following restoration. Salmon redds were constructed in coarse gravel substrates located in areas of high sediment mobility, as determined by measurements of gravel friction angles and a grain entrainment model. Salmon redds were located near transitions between pool-riffle bedforms in regions of high predicted hyporheic flows. Habitat quality (quantified as a function of stream hydraulics) and hyporheic flow were both strong predictors of redd occurrence, though the relative roles of these variables differed between sites. Our findings indicate that physical controls on redd site selection in restored channels were similar to those reported for natural channels elsewhere. Our results further highlight that in addition to traditional habitat criteria (e.g., water depth, velocity, and substrate size), quantifying sediment texture and mobility, as well as intragravel flow, provides a more complete understanding of the ecological benefits provided by river restoration projects.

[Archived PDF of article]

Mountain-Block Recharge: A Review of Current Understanding

by Katherine H. Markovich, Andrew H. Manning, Laura E. Condon & Jennifer C. McIntosh

Abstract: Mountain-block recharge (MBR) is the subsurface inflow of groundwater to lowland aquifers from adjacent mountains. MBR can be a major component of recharge but remains difficult to characterize and quantify due to limited hydrogeologic, climatic, and other data in the mountain block and at the mountain front. The number of MBR-related studies has increased dramatically in the 15 years since the last review of the topic was conducted by Wilson and Guan (2004), generating important advancements. We review this recent body of literature, summarize current understanding of factors controlling MBR, and provide recommendations for future research priorities. Prior to 2004, most MBR studies were performed in the southwestern United States. Since then, numerous studies have detected and quantified MBR in basins around the world, typically estimating MBR to be 5–50% of basin-fill aquifer recharge. Theoretical studies using generic numerical modeling domains have revealed fundamental hydrogeologic and topographic controls on the amount of MBR and where it originates within the mountain block. Several mountain-focused hydrogeologic studies have confirmed the widespread existence of mountain bedrock aquifers hosting considerable groundwater flow and, in some cases, identified the occurrence of interbasin flow leaving headwater catchments in the subsurface—both of which are required for MBR to occur. Future MBR research should focus on the collection of high-priority data (e.g., subsurface data near the mountain front and within the mountain block) and the development of sophisticated coupled models calibrated to multiple data types to best constrain MBR and predict how it may change in response to climate warming.

[Archived PDF of article]

An Adjoint Sensitivity Model for Steady-State Sequentially Coupled Radionuclide Transport in Porous Media

by Mohamed Hayek, Banda S. RamaRao & Marsh Lavenue

Abstract: This work presents an efficient mathematical/numerical model to compute the sensitivity coefficients of a predefined performance measure to model parameters for one-dimensional steady-state sequentially coupled radionuclide transport in a finite heterogeneous porous medium. The model is based on the adjoint sensitivity approach that offers an elegant and computationally efficient alternative way to compute the sensitivity coefficients. The transport parameters include the radionuclide retardation factors due to sorption, the Darcy velocity, and the effective diffusion/dispersion coefficients. Both continuous and discrete adjoint approaches are considered. The partial differential equations associated with the adjoint system are derived based on the adjoint state theory for coupled problems. Physical interpretations of the adjoint states are given in analogy to results obtained in the theory of groundwater flow. For the homogeneous case, analytical solutions for primary and adjoint systems are derived and presented in closed forms. Numerically calculated solutions are compared to the analytical results and show excellent agreements. Insights from sensitivity analysis are discussed to get a better understanding of the values of sensitivity coefficients. The sensitivity coefficients are also computed numerically by finite differences. The numerical sensitivity coefficients successfully reproduce the analytically derived sensitivities based on adjoint states. A derivative-based global sensitivity method coupled with the adjoint state method is presented and applied to a real field case represented by a site currently being considered for underground nuclear storage in Northern Switzerland, “Zürich Nordost,” to demonstrate the proposed method. The results show the advantage of the adjoint state method compared to other methods in term of computational effort.

[Archived PDF of article]

Hydraulic Reconstruction of the 1818 Giétro Glacial Lake Outburst Flood

by C. Ancey, E. Bardou, M. Funk, M. Huss, M. A. Werder & T. Trewhela

Summary: Every year, natural and man-made dams fail and cause flooding. For public authorities, estimating the risk posed by dams is essential to good risk management. Efficient computational tools are required for analyzing flood risk. Testing these tools is an important step toward ensuring their reliability and performance. Knowledge of major historical floods makes it possible, in principle, to benchmark models, but because historical data are often incomplete and fraught with potential inaccuracies, validation is seldom satisfactory. Here we present one of the few major historical floods for which information on flood initiation and propagation is available and detailed: the Giétro flood. This flood occurred in June 1818 and devastated the Drance Valley in Switzerland. In the spring of that year, ice avalanches blocked the valley floor and formed a glacial lake, whose volume is today estimated at 25×106 m3. The local authorities initiated protection works: A tunnel was drilled through the ice dam, and about half of the stored water volume was drained in 2.5 days. On 16 June 1818, the dam failed suddenly because of significant erosion at its base; this caused a major flood. This paper presents a numerical model for estimating flow rates, velocities, and depths during the dam drainage and flood flow phases. The numerical results agree well with historical data. The flood reconstruction shows that relatively simple models can be used to estimate the effects of a major flood with good accuracy.

[Archived PDF of article]

The Representation of Hydrological Dynamical Systems Using Extended Petri Nets (EPN)

by Marialaura Bancheri, Francesco Serafin & Riccardo Rigon

Abstract: This work presents a new graphical system to represent hydrological dynamical models and their interactions. We propose an extended version of the Petri Nets mathematical modeling language, the Extended Petri Nets (EPN), which allows for an immediate translation from the graphics of the model to its mathematical representation in a clear way. We introduce the principal objects of the EPN representation (i.e., places, transitions, arcs, controllers, and splitters) and their use in hydrological systems. We show how to cast hydrological models in EPN and how to complete their mathematical description using a dictionary for the symbols and an expression table for the flux equations. Thanks to the compositional property of EPN, we show how it is possible to represent either a single hydrological response unit or a complex catchment where multiple systems of equations are solved simultaneously. Finally, EPN can be used to describe complex Earth system models that include feedback between the water, energy, and carbon budgets. The representation of hydrological dynamical systems with EPN provides a clear visualization of the relations and feedback between subsystems, which can be studied with techniques introduced in nonlinear systems theory and control theory.

[Archived PDF of article]

A Regularization Approach to Improve the Sequential Calibration of a Semidistributed Hydrological Model

by A. de Lavenne, V. Andréassian, G. Thirel, M.-H. Ramos & C. Perrin

Abstract: In semidistributed hydrological modeling, sequential calibration usually refers to the calibration of a model by considering not only the flows observed at the outlet of a catchment but also the different gauging points inside the catchment from upstream to downstream. While sequential calibration aims to optimize the performance at these interior gauged points, we show that it generally fails to improve performance at ungauged points. In this paper, we propose a regularization approach for the sequential calibration of semidistributed hydrological models. It consists in adding a priori information on optimal parameter sets for each modeling unit of the semi-distributed model. Calibration iterations are then performed by jointly maximizing simulation performance and minimizing drifts from the a priori parameter sets. The combination of these two sources of information is handled by a parameter k to which the method is quite sensitive. The method is applied to 1,305 catchments in France over 30 years. The leave-one-out validation shows that, at locations considered as ungauged, model simulations are significantly improved (over all the catchments, the median KGE criterion is increased from 0.75 to 0.83 and the first quartile from 0.35 to 0.66), while model performance at gauged points is not significantly impacted by the use of the regularization approach. Small catchments benefit most from this calibration strategy. These performances are, however, very similar to the performances obtained with a lumped model based on similar conceptualization.

[Archived PDF of article]

Proneness of European Catchments to Multiyear Streamflow Droughts

by Manuela I. Brunner & Lena M. Tallaksen

Summary: Droughts lasting longer than 1 year can have severe ecological, social, and economic impacts. They are characterized by below-average flows, not only during the low-flow period but also in the high-flow period when water stores such as groundwater or artificial reservoirs are usually replenished. Limited catchment storage might worsen the impacts of droughts and make water management more challenging. Knowledge on the occurrence of multiyear drought events enables better adaptation and increases preparedness. In this study, we assess the proneness of European catchments to multiyear droughts by simulating long discharge records. Our findings show that multiyear drought events mainly occur in regions where the discharge seasonality is mostly influenced by rainfall, whereas catchments whose seasonality is dominated by melt processes are less affected. The strong link between the proneness of a catchment to multiyear events and its discharge seasonality leads to the conclusion that future changes toward less snow storage and thus less snow melt will increase the probability of multiyear drought occurrence.

[Archived PDF of article]

Equifinality and Flux Mapping: A New Approach to Model Evaluation and Process Representation under Uncertainty

by Sina Khatami, Murray C. Peel, Tim J. Peterson & Andrew W. Western

Abstract: Uncertainty analysis is an integral part of any scientific modeling, particularly within the domain of hydrological sciences given the various types and sources of uncertainty. At the center of uncertainty rests the concept of equifinality, that is, reaching a given endpoint (finality) through different pathways. The operational definition of equifinality in hydrological modeling is that various model structures and/or parameter sets (i.e., equal pathways) are equally capable of reproducing a similar (not necessarily identical) hydrological outcome (i.e., finality). Here we argue that there is more to model equifinality than model structures/parameters, that is, other model components can give rise to model equifinality and/or could be used to explore equifinality within model space. We identified six facets of model equifinality, namely, model structure, parameters, performance metrics, initial and boundary conditions, inputs, and internal fluxes. Focusing on model internal fluxes, we developed a methodology called flux mapping that has fundamental implications in understanding and evaluating model process representation within the paradigm of multiple working hypotheses. To illustrate this, we examine the equifinality of runoff fluxes of a conceptual rainfall-runoff model for a number of different Australian catchments. We demonstrate how flux maps can give new insights into the model behavior that cannot be captured by conventional model evaluation methods. We discuss the advantages of flux space, as a subspace of the model space not usually examined, over parameter space. We further discuss the utility of flux mapping in hypothesis generation and testing, extendable to any field of scientific modeling of open complex systems under uncertainty.

[Archived PDF of article]

Role of Extreme Precipitation and Initial Hydrologic Conditions on Floods in Godavari River Basin, India

by Shailesh Garg & Vimal Mishra

Abstract: Floods are the most frequent natural calamity in India. The Godavari river basin (GRB) witnessed several floods in the past 50 years. Notwithstanding the large damage and economic loss, the role of extreme precipitation and antecedent moisture conditions on floods in the GRB remains unexplored. Using the observations and the well-calibrated Variable Infiltration Capacity model, we estimate the changes in the extreme precipitation and floods in the observed (1955–2016) and projected future (2071–2100) climate in the GRB. We evaluate the role of initial hydrologic conditions and extreme precipitation on floods in both observed and projected future climate. We find a statistically significant increase in annual maximum precipitation for the catchments upstream of four gage stations during the 1955–2016 period. However, the rise in annual maximum streamflow at all the four gage stations in GRB was not statistically significant. The probability of floods driven by extreme precipitation (PFEP) varies between 0.55 and 0.7 at the four gage stations of the GRB, which declines with the size of the basins. More than 80% of extreme precipitation events that cause floods occur on wet antecedent moisture conditions at all the four locations in the GRB. The frequency of extreme precipitation events is projected to rise by two folds or more (under RCP 8.5) in the future (2071–2100) at all four locations. However, the increased frequency of floods under the future climate will largely be driven by the substantial rise in the extreme precipitation events rather than wet antecedent moisture conditions.

[Archived PDF of article]

Research Letters

Combined Effect of Tides and Varying Inland Groundwater Input on Flow and Salinity Distribution in Unconfined Coastal Aquifers

by Woei Keong Kuan, Pei Xin, Guangqiu Jin, Clare E. Robinson, Badin Gibbes & Ling Li

Abstract: Tides and seasonally varying inland freshwater input, with different fluctuation periods, are important factors affecting flow and salt transport in coastal unconfined aquifers. These processes affect submarine groundwater discharge (SGD) and associated chemical transport to the sea. While the individual effects of these forcings have previously been studied, here we conducted physical experiments and numerical simulations to evaluate the interactions between varying inland freshwater input and tidal oscillations. Varying inland freshwater input was shown to induce significant water exchange across the aquifer-sea interface as the saltwater wedge shifted landward and seaward over the fluctuation cycle. Tidal oscillations led to seawater circulations through the intertidal zone that also enhanced the density-driven circulation, resulting in a significant increase in the total SGD. The combination of the tide and varying inland freshwater input, however, decreased the SGD components driven by the separate forcings (e.g., tides and density). Tides restricted the landward and seaward movement of the saltwater wedge in response to the varying inland freshwater input in addition to reducing the time delay between the varying freshwater input signal and landward-seaward movement in the saltwater wedge interface. This study revealed the nonlinear interaction between tidal fluctuations and varying inland freshwater input will help to improve our understanding of SGD, seawater intrusion, and chemical transport in coastal unconfined aquifers.

[Archived PDF of article]