“Yes, tea banished the fairies.”

“There have been numerous explanations for why the fairies disappeared in Britain – education, advent of electrical lighting, evangelical religion. But one old man in the village of Alves, Moray, Scotland knew the real reason in 1851: tea drinking. Yes, tea banished the fairies.”

The historian Owen Davies wrote this in referring to a passage from an old source. He didn’t mention where it came from. In doing a search on Google Books, multiple results came up. The earliest is supposedly on page 624 of the 1850 Family Herald – Volumes 8-9, but there is no text for it available online. Several other books from the 1850s to 1880s reprinted it. (1)

Below is the totality of what Davies shared. It might have originally been part of a longer passage, but it is all that I could find from online sources. It’s a short account and intriguing.

“How do you account,” said a north country (3) minister of the last age (the late Rev. Mr. M’Bean, of Alves,) to a sagacious old elder of his session, “for the almost total disappearance of the ghosts and fairies that used to be common in your young days?” “Tak’ my word for’t, minister,” replied the old man, “it’s a’ owing to the tea; whan the tea cam’ in, ghaists an’ fairies gaed out. Weel do I mind whan at a’ our neebourly meetings — bridals, christenings, lyke-wakes, an’ the like — we entertained ane anither wi’ rich nappy ale; an’ when the verra dowiest o’ us used to get warm i’ the face, an’ a little confused i’ the head, an’ weel fit to see amaist onything when on the muirs on yer way hame. But the tea has put out the nappy; an’ I have remarked that by losing the nappy we lost baith ghaists and fairies.”

Will Hawkes noted that, “‘nappy’ ale meant strong ale.” In response to Davies, James Evans suggested that, “One thing which I haven’t seen mentioned here is that there is an excellent chance that the beer being produced in this region, at this time, was mildly hallucinogenic.” And someone following that asked, “Due to ergot?” Now that makes it even more intriguing to consider. There might have been good reason people used to see more apparitions. Whether or not ergot was involved, we do know that in the past all kinds of herbs were put into groot or gruit ales for nutritional and medicinal purposes but also maybe for the affect they had on the mind and mood. Herbs, instead of hops, used to be what distinguished ale from beer.

Let me make some connections. Alcohol is a particular kind of drug. Chuck Pezeshki argues that, “alcohol is much more of a ‘We’ drug when used in moderation, than an ‘I’ drug” (Leadership for Creativity Isn’t all Child’s Play). He adds that, “There’s a reason for the old saying ‘when the pub closes, the revolution starts!’” Elsewhere, he offers the contrast that, “Alcohol is on average is pro-empathetic, sugar anti-empathetic” (The Case Against Sugar — a True Psychodynamic Meta-Review).

Alcohol from grains and grapes defined much of civilization, more so than any other mind-altering substance. Or that was the case once farming became more systematized and productive during the Axial Age, such that something had to be done with the more consistent surplus yields at a time when storage and preservation was a problem. And with increased surpluses came increased size and concentration of populations, i.e., mass urbanization of ever larger cities and ever more expansive empires.

That brought with it a need to solve the problem of unsafe drinking water, easily resolved by making alcohol that kills microbes. Whatever the case may have been in prior ages, at least by the early medieval period drunkenness had become a defining feature of many cultures in the Western world:

“Indeed the general rowdiness of English drinking at this period, which precluded priests from participating, is attested in the twelfth century historian William of Malmesbury. In describing the character of the Angles before the Norman conquest of AD 1066 he says that all drank together, throughout the night and day, that intoxication was common, making men effeminate, and that the people would eat until they were sick and drink until they threw up (these last two habits being passed on to the conquerors). Another tradition of theirs not mentioned until after the conquest is that of the wassail, or toast, first found in the poet Layamon around AD 1200.”

It makes sense. Until the modern era of reforms focused on public health, a large part of the population was dependent on alcoholic beverages for safe drinking. Even common wells were easily contaminated because people with unclean hands were constantly handling the bucket that was repeatedly dropped back down into the water.  No one back then understood that it was bacteria that made them sick, but they did understand that alcohol kept them healthy. The health component was emphasized by adding in nutritional and medicinal herbs.

So they developed a drinking culture. Everyone, not only men but women and children as well, drink alcohol on a daily basis. Much of it was low in alcohol level (e.g., small beer), but much of it had greater amounts. Many people spent their lives in a mild state of near constant inebriation. Throw in some mildly mind-altering herbs and it would’ve shaped the entire mentality and culture. Going back to the ancient world, alcohol had been associated with spirituality, religion, ritual, worship, ecstasy, and the supernatural. The specific practice of groot ales is described in the earliest records. Such a cultural habit that probably extended over millennia had to have had a major impact on their sense of identity and reality, in supporting an overt expression of the bundled mind, what Julian Jaynes describes as the bicameral mind.

Think about it. Tea, coffee, cocoa, and sugar weren’t introduced into the West until colonialism; i.e., early modernity. It took a while for them to become widespread. They first were accessible only to those in the monied classes, including the intellectual and artistic elite but also the clerical elite. Some have argued that these stimulants are what fueled the Enlightenment Age. And don’t forget that tea played a key role in instigating the American Revolution. Changes in diet often go hand in hand with changes in culture (see below). One might note that this was the precise era when the elite began talking about the early spread of a mental health epidemic referred to as melancholia and acedia, what we would now understand as depression.

There are those like Terrence McKenna who see psychedelics as having much earlier played a central role in shaping the human mind. This is indicated by the wide use of psychedelics by indigenous populations all over the planet and by evidence of their use among ancient people. Psychedelics and entheogens preceded civilization and it seems that their use declined as civilization further developed. What replaced psychedelics over time were the addictive stimulants. That other variety of drug has a far different affect on the human mind and culture.

The change slowly began millennia ago. But the full takeover of the addictive mentality only seems to have come fully into its own in recent centuries. The popularizing of caffeinated drinks in the 19th century is a key example of the modernizing of the mind. People didn’t simply have more active imaginations in the past. They really did live in a cultural worldview where apparitions were common, maybe in the way that Julian Jaynes proposed that in the Bronze Age people heard communal voices. These weren’t mere hallucinations. It was central to the lived reality of their shared culture.

In traditional societies, alcohol was used for social gatherings. It brought people together and, maybe combined with other substances, made possible a certain experience of the world. With the loss of that older sense of communal identity, there was the rise of the individual mindset isolated by addictive stimulants. This is what has fueled all of modernity. We’ve been buzzing ever since. Stimulants broke the spell of the fairies only to put us under a different spell, that of the demiurgic ego-consciousness.

“The tea pots full of warm water,” as Samuel Tissot put in his 1768 An Essay on Diseases Incident to Literary and Sedentary Persons, “I see upon their tables, put me in mind of Pandora’s box, from whence all sorts of evils issue forth, with this difference however, that they do not even leave the hopes of relief behind them; but, on the contrary, by inducing hypochondriac complaints, diffuse melancholy and despair.” (2) That is to say the modern mind was transformed and, according to some, not in a good way. The living world became inanimate, no longer bustling with animistic beings, and presumably the last of the bicameral voices went silent.

This change didn’t come easily, though. The bundled mind is probably the evolutionary norm of the human species. It’s our default mode. That is because humans evolved with with psychedelic plants and rotting alcoholic fruit (some other species have been observed stashing fruit to eat later maybe with the intention to become inebriated). The stimulants, although around previously, were far less common until agriculture. The targeting of alcohol and psychedelics with criminalization was, as previously noted, social control as substance control; but it could be added that it was also mind control. Still, archaic habits die hard. If stimulants help people survive the unnatural stresses of the capitalist work week, it is alcohol that many turn to on the weekends to forget those stresses and become themselves again.

Immediately preceding U.S. Prohibition, there was a health epidemic involving moral panic and culture war. It was mixed up with ideological conflict over proper social norms, social roles, and social identities. In particular, there was an obsession over masculinity and the fear of emasculation or else of libidinous dissipation (e.g., an anti-masturbation campaign focused on young boys). Addiction and alcoholism were seen to play a role. This was at a time when large number of northern European immigrants established a heavy beer-drinking culture. Going back to the Greeks and Romans, there had long been a cultural belief that, according to Galenic humoral theory, beer was ‘cold’ and hence effeminizing. All alcohol was deemed to be potentially ‘cold’, but specifically beer.

This is the origins of the class bias toward wine, although wine too sometimes was seen as ‘cold’. It might also contribute to why stimulants were so important to the male-dominated work culture of capitalism. To be a man, meant to be assertive and aggressive, to be mentally focused and intense, to always be on one’s game. The sedating and slowing affect of alcohol is the opposite of the ideal attributes of a successful alpha male as part of the capitalist elite.

* * *

(1) The Country Gentleman – Vol. XII No. 23 (1858), William Hopkin’s “The Cruise of the Betsey” from Fraser’s Magazine for Town and Country – Volume 58 (1858) and from Littell’s Living Age – Volume 59 (1858), John William Kirton’s One Thousand Temperance Anecdotes [&c.] (1868), John William Kirton’s A Second Thousand of Temperance Anecdotes (1877), The Church of England Temperance Chronicle – No. 42 Vol. VIII (1880), and The Guernsey Magazine – Vol. X No. 12 (1882).

(2) “There is another kind of drink not less hurtful to studious men than wine; and which they usually indulge in more freely; I mean warm liquors [teas], the use of which is become much more frequent since the end of the last century. A fatal prejudice insinuated itself into physic about this period. A new spirit of enthusiasm had been excited by the discovery of the circulation: it was thought necessary for the preservation of health to facilitate it as much as possible, by supplying a great degree of fluidity to the blood, for which purpose it was advised to drink a large quantity of warm water. Cornelius Bontekoe, a Dutch physician, who died afterwards at Berlin, first physician to the elector of Brandenburgh, published in 1679 a small treatise in Dutch, upon tea, coffee, and chocolate, in which he bestows the most extravagant encomiums on tea, even when taken to the greatest excess, as far as one or two hundred cups in a day, and denies the possibility of its being hurtful to the stomach. This error spread itself with surprising rapidity all over the northern part of Europe; and was attended with the most grievous effects. The æra of its introduction is marked by an unhappy revolution in the account of the general state of health at that time. The mischief was soon noticed by accurate observers. M. Duncan, a French physician settled at Rotterdam, published a small work in 1705, wherein we find, amidst a great deal of bad theory, some useful precepts against the use of hot liquors (I). M. Boerhaave strongly opposed this pernicious custom; all his pupils followed his example, and all our eminent physicians are of the same opinion. The prejudice has at last been prevented from spreading, and within these few years seems to have been rather less prevalent (m); but unfortunately it subsists still among valetudinarians, who are induced to continue these pernicious liquors, upon the supposition that all their disorders proceed from a thickness of blood. The tea-pots full of warm water I see upon their tables, put. me in mind of Pandora’s box, from whence all sorts of evils issue forth, with this difference however, that they do not even leave the hopes of relief behind them; but, on the contrary, by inducing hypochondriac complaints, diffuse melancholy and despair. […]

“The danger of these drinks is considerably increased, as I have before observed, by the properties of the plants infused in them; the most fatal of these when too often or too freely used, is undoubtedly the tea, imported to us since near two centuries past from China and Japan, which has so much increased diseases of a languid nature in the countries where it has been introduced, that we may discover, by attending to the health of the inhabitants of any city, whether they drink tea or not; and I should imagine one and the greatest benefits that could accrue to Europe, would be to prohibit the . importation of this famous leaf, which contains no essential parts besides an acrid corrosive gum, with a few astringent particles (o), imparting to the tea when strong, or when the infusion has stood a long time and grown cold, a styptic taste, slightly felt by the tongue, but which does not prevent the pernicious effects of the warm water it i$ drenched in. These effects are so striking, that I have often seen very strong and healthy men, seized with faintness, gapings, and uneasiness, which lasted for some hours after they had drank a few cups of tea fasting, and sometimes continued the whole day. I am sensible that these bad effects do not shew themselves so plainly in every body, and that there are some who drink tea every day, and remain still in good health; but these people drink it with moderation. Besides, the non-existence of any danger cannot be argued from the instances «f some few who have been fortunate enough to escape it.

“The effects of coffee differing from’ those of tea, it cannot be placed in the same class ; for coffee, although made with- warm water, is not so pernicious for this reason, as it is on account of its being a powerful stimulus, producing strong irritations in the fibres by its bitter aromatic oil This oil combined as it is with a kind of very nourishing meal, and of easy digestion, would make this berry of great consequence, in pharmacy, as one of the bitter stomachics, among which it would be the most agreeable, as well as one of the most active. This very circumstance is sufficient to interdict the common use of it, which must be exceedingly hurtful. A continual irritation of the fibres of the stomach must at length destroy their powers; the mucus is, carried off, the nerves are irritated and acquire singular spasms, strength fails, hectic fevers come on with a train of other diseases, the cause of which is industriously concealed, and is so much the more difficult to eradicate, as this sharpness united with an oil seems not only to infect the fluids, but even to adhere to the vessels themselves. On the contrary, when seldom taken, it exhilerates, breaks down the slimy substances in the stomach, quickens its action, dispels the load and pains of the head, proceeding from interrupted digestions, and even clears the ideas and sharpens the understanding, if we may credit the accounts of men of letters, who have therefore used it very freely. But let me be permitted to ask, whether Homer, Thucydides, Plato, Xenophon, Lucretius, Virgil, Ovid, Horace, Petronius, to which I may venture to add Corneille and Moliere, whose masterpieces will ever be the delight of the remotest posterity, let me ask, I say, whether they drank coffee? Milk rather takes off from the irritation occasioned by coffee, but still does not entirely prevent all its pernicious effects, for even this mixture has some disadvantages peculiar to itself. Men of learning, there fore, who are prudent, ought in general to keep coffee as their favourite medicine, but should never use it as a common drink. The custom is so much the more dangerous, as it soon degenerates into a habit of necessity, which few men have the resolution to deprive themselves of. We are sensible of the poison, and swallow (32) it because it is palatable.”

(3) The original passage that inspired this post asserted that the source of the anecdote was “a north country minister of the last age”. One might wonder what was considered the last age. A generation earlier? The century before? Anyway, maybe more significant was that, as the storyteller was of the north country, presumably the ‘old man’ who responded was also of the north country. At the very least, that meant north of London, but probably referring to the rural north from the Midlands to the borderlands, maybe all of it being part of the vast north to the literary imagination of the southern elite and urbanites.

This brings us to another point about ‘nappy ale’, in probably referring to entheogenic groot ale. Often non-herbal beer made out of hops apparently first most strongly took hold in southern England, although with exceptions along the northern coast and major waterways. Yet ale remained the most popular and widespread alcoholic drink until the colonial era began in early modernity. The takeover of hops beer, at least in Merry Ol’ England (i.e., London and surrounding region), happened to coincide with the English Civil War. That conflict is what some consider the first modern revolution and class war. Also, it was primarily a contest for power between southern Cavaliers and northern Roundheads. One is then reminded of how important taverns were in the colonies during the American Revolution.

Joshua Thomas Ravenhill writes, “Brewing with hops had been established in the Low Countries by the early fourteenth century. As has been demonstrated by Milan Pajic, Doche aliens, mainly from Holland and Zeeland, were the first to start brewing beer in England by at least the end of the fourteenth century, and there were alien beerbrewers in England by at least 1399. It took a long time for beer to become as popular as ale in the capital. Bennett and Bich Luu, writing of London, argue that it was not until the reign of Elizabeth I that natives were drinking beer in greater quantities than ale and that large numbers of natives brewed beer” (The Experiences of Aliens in Later Medieval London and the Negotiation of Belonging, 1400 – 1540).

By the way, the use of hops in general only began to significantly take hold, in England and Europe, during the 14th century. That was the period of pre-Reformation religious hereticism, political radicalism, class war, and peasants’ revolts. The English Peasants’ Revolt is the earliest conflict that sometimes gets called the first modern revolution because the ideological rhetoric was showing the signs of what later would become more well articulated. That was the major shifting point for the decline of feudalism. The earliest of the enclosures began around then, only becoming a government-sanctioned enclosure movement with the Glorious Revolution following the English Civil War. So, the two major English internal conflicts that get called modern or modern-like revolutions bookended the period of the rise of hops beer in replacing groot ale.

As the original passage maybe indicates, it was the lower class of rural former peasantry that had been still drinking groot ale into the first centuries of the modern period, such that it was still a clear cultural memory in 1850. In Europe where hops were used earlier, “Ale made with gruit was a drink for the poor and the sick” (Richard W. Unger, Beer in the Middle Ages and the Renaissance). Hops was often a trade good and the poor were more dependent on local ingredients. Yet the groot tradition barely hung on in a few isolated locations: “In rural western Norway in the 1950s brewers still used pors, that is bog myrtle. The survival of the practice was certainly exceptional since in the sixteenth, but especially in the seventeenth centuries, there were campaigns in central Europe to get rid of grut or pors.”

Communal peasant-like identities also lasted longer in rural areas. Possibly, it was the fact that groot ale supported such a mentality that it became a target of the reformers of morality, land, and agriculture. Not only how land and property is structured will structure the mind but also what is grown on the land and who owns the land to decide what is grown there. That is what the enclosure movement was all about, in order to promote the enclosure of mind and identity in the form of the hyper-individualistic self as capitalist, consumer, and worker; as opposed to being defined by the communal reality of feudal villages and the commons (The Enclosure of the Mind).

Hops beer could be industrially mass-produced, as opposed to the more local and often female-dominated home production of groot ale. But also it was about productivity. Two things. Hallucinogens, even if mild, don’t put people in the mindset for long grueling hours of labor that requires intense focus, such as working in a factory or mine, particularly the night shift or double shifts. Hallucinogens mixed into alcohol is even less supportive of profitable efficiency of workers. But even alcohol alone, if only hops beer, is not all that beneficial for the bottom line. In early capitalism, workers were sometimes paid with beer; but that practice quickly ended. Interestingly, early modern politics was also rife with alcohol. Candidates would give out beer and, in places like the American South, election day was basically a drunken carnival (Edmund S. Morgan, Inventing the People, pp. 202-203; Winter Season and Holiday Spirit). That was likewise suppressed as time went on.

* * *

6/26/21 – Update: There is some relevant and interesting info to be added. It’s not merely that stimulants replaced psychedelics. The shift is a bit more complex. As tea and coffee became more common drinks, so did beer brewed with hops that replaced the archaic practice of gruit ales brewed with herbs. That used to be the distinction between beer and ale, whether or not it included hops. Also, the distinction was that those herbs were often stimulating, aphrodisiacal, and psychotropic. For example, some of the same herbal ingredients are used in absinthe.

The use of hops in brewing beer is first recorded in Northern France in 800s. It didn’t spread to England until the 1400s, then began to catch on in the 1500s, and became useful for beer preservation as colonial trade expanded in the following centuries. Even with the advantages, gruit ales without hops remained common, particularly in rural areas. At a time when most alcohol produced was consumed personally or sold locally, there was little need to preserve beer with hops. It wasn’t until the mass production later in the industrial age that hops became king. But that was already being felt by the 19th century when the fairies were disappearing.

There was motivation for this. There is an obvious benefit for modern capitalism in the use of stimulants. Hops, on the other hand, is a depressive and lowers sex drive. The more that stimulants are used, the more that depressives are needed to wind down at the end of the day. That is opposed to the gruit ales, often lower in alcohol, that were imbibed all day long to maintain a mild buzz without the constant up and down cycle of addiction to stimulants and depressives. The Church, by the 1500s, had already caught on that hops would make for a more passive population in subduing people’s sinful nature; similar to why they used diet for social control (i.e., banning red meat before and during Carnival).

The increasing use of hops coincided with the rise of modernity, the enclosure movement, mass urbanization, colonial trade, capitalism, and industrialization. This also included land reforms and agricultural improvements that led to grain surpluses. So, with industrial farming and industrial breweries, beer could be produced in vast amounts, preserved with hops, and then shipped where needed. It was a marriage made in heaven. Meanwhile, the workers were forced to suck down the caffeine to keep up with the new grueling factory work. The older tradition of alewives making gruit ale at home probably was decimated with the moral panic of witch persecutions. Yet home brewing continued in many places into the early 20th century before finally making a more recent comeback.

The following are some links and excerpts:

Alcohol in the 17th Century: Age of Discovery
by David J. Hanson

1673: A group of citizens petitioned Parliment for legislation to prohibit brandy, coffee, rum, tea and chocolate. It was because  ‘these greatly hinder the consumption of Barley, Malt, and Wheat, the product of our land.’ Parliment did not take action.58 (Bickerdyke, J. The Curiosities of Ale and Beer. London: Spring, 1965, p. 118.)

Alcohol in the 18th Century: European Expansion
by David J. Hanson

1700-1730: ‘Housewives in the northern colonies [of what is now the US] brewed beer every few days, since their product had a short shelf life.’6 (Blocker, J. Kaleidoscope in Motion. Drinking in the United States, 1400-2000. In: Holt. M. (Ed.) AlcoholOxford: Berg, 2006. Pp. 225-240. P. 227.) […]

1790″ “Parliament made it illegal to pay wages in liquor.54” (Magee, M. 1000 Years of Irish Whiskey. Dublin: O’Brien, 1980, p. 76.)

Alcohol in the 19th Century (And Emergence of Temperance)
by David J. Hanson

People had accepted drunkenness as part of life in the eighteenth century.2 (Austin, G. Alcohol in Western Society from Antiquity to 1800. Santa Barbara, CA: ABC-Clio, 1985, p. xxv.) But the nineteenth century brought a change in attitudes as a result of increasing industrialization. This created the need for a reliable and punctual work force.3 (Porter, R. Introduction. In: Sournia, J.-C. A History of Alcoholism. Oxford: Basil Blackwell, 1990, p. xii.) Employers wanted self-discipline instead of self-expression. They wanted task orientation in place of relaxed conviviality. It followed that drunkenness was a threat to industrial efficiency and growth. […]

People blamed alcohol for problems caused by industrialization and urbanization. Thus, they blamed it for problems such as urban crime, poverty and high infant mortality. However, gross overcrowding and unemployment contributed greatly to these problems.9 (Porter, R. Introduction. In: Sournia, J.-C. A History of Alcoholism. Oxford: Basil Blackwell, 1990, p. 21)

People also blamed alcohol for more and more personal, social and religious/moral problems. […]

1804: As early as 1804, temperance organizations began in the Netherlands.15 (Garrelsen, H., and van de Goor, I. The Netherlands. In: Heath. Pp. 190-200. P. 191.)

British physician Thomas Trotter suggested that chronic drunkenness was a disease.16 (Plant, M. The United kingdom. In: Heath, D. Pp. 289-299. P. 291.) […]

Post-1865: After the American Civil War (1861-1865) beer replaced whiskey as preferred beverage of working men.62 (Rorabaugh, W. The Alcoholic Republic. NY: Oxford U Press, 1979.) […]

1886: Coca-Cola [i.e., cocaine] was a temperance beverage.93 (Blocker, J., et al. Alcohol and Temperance in Modern History: An International Encyclopedia. Vol. 1. Santa Barbara, CA: ABC-CLIO, 2003, xxxi-xiv.)

What the Hell is a Gruit Ale?
from American Craft Beer

Most don’t know that the herbal collections making up gruit were the original “hops” – at least before gruit’s use began to dwindle in a large way during the 15th and 16th centuries. Many factors went into its disappearance, including the passing of the German beer purity law, Reinheitsgebot, which originally stated that water, barley, and hops were the only ingredients that could be used in beer production.

Another explanation for the disuse of gruit is based in religion – since some herbs used were known to have stimulating and even aphrodisiac effects, switching to a sedative substance like hops satisfied a Puritan need to keep people from enjoying themselves (sound familiar?).

Beer Without Hops: History of Gruit Ales
from 2nd Kitchen

Gruit ale’s are much stronger than beer made with hops, causing narcotic, aphrodisiacal, and psychotropic effects. While this led to its recreational use popularity, it also led to its downfall.

Hops is an anaphradesiacal herb – meaning it lowers sexual drive. This is offset by the alcohol in beer. However, gruit doesn’t react this way and instead includes chemicals known as alkaloids.

Alkaloids are known to cause a chemical reaction with receptors in the brain similar to that of THC found in Marijuana and Absinthe. Many times gruit and absinthe share common ingredients such as wormwood and exhibit similar effects.

Gruit Ales: Beer Before Hops
by Andy Sparhawk

Gruit Ales: The Original War on Drugs

Despite gruit beers being alcoholic in nature, it is likely the effects of the herb mix contributed to its recreational effects, popularity and downfall. Each of the main herbs is considered much stronger in effect, psychotropic even, than beer’s modern substitute, Humulus lupulus, writes Buhner. “It is important to keep in mind the properties of gruit ale: it is highly intoxicating – narcotic, aphrodisiacal, and psychotropic when consumed in sufficient quantities,” Buhner explains. “The hopped ale that took its place is quite different.”

Gruit beers were favored by many in medieval Europe dating back prior to the predominant use of hops, writes Buhner, but the narcotic effects of the herbs, kept closely guarded by the church or lordships made the blend a target. A bitter battle between the religions, regions and businessmen made the attack against gruit beers reminiscent of the war on drugs. “Hops, when they began to be suggested for use as a primary additive, in both Germany and England, were bitterly resisted,” explained Buhner. (Stephen Harrod Buhners’ book “Sacred and Herbal Healing Beers”)

The war between ingredients played out over the course of two centuries,” writes Buhner, “simultaneously with the Protestant Reformation.”

As part of the Reformation, “Protestant religious intolerance of Catholic indulgence that was the genesis of the temperance movement.” Buhner goes on to explain, “The Protestant reformists were joined by merchants and competing royals to break the financial monopoly of the Church. The result was ultimately the end of a many-thousand-years’ tradition of herbal beer making in Europe and the limiting of beer and ale into one limited expression of beer production — that of hopped ales or what we call beer today.”

Gruit, or Brewing Without Hops
from Home Brewing

Before the beer purity laws which swept Europe in the 1500s, beer was made with many different admixtures, and Gruit was one variety which was popular. Recipes for gruit were different depending on which herbs grew locally. According to GruitAle.com, gruit usually included the following herbs: Yarrow (Achillea millefolium), Bog Myrtle (Myrica Gale), and Marsh Rosemary (Ledum palustre). This claim is also supported by the book Sacred & Healing Herbal Beers, by Stephen Harrod Buhner. This book contains many ancient recipes for beer, including a section on gruit. Additional herbs which have been found in gruit recipes are Juniper berries, Mugwort, Wormwood, Labrador Tea, Heather, Licorice, and some others.

There are a few factors to consider when comparing the inebriatory qualities of gruit in comparison to more commonly made beer. It is held amongst those experienced in gruit inebriation that gruit rivals hopped beer on many accounts. One factor is that hops create a sedentary spirit in the imbiber. Amongst those knowledgeable about herbs, hops tea is well known as a catalyst for dreams, and creates drowsiness for the beer drinker. Hops is also an anaphradesiacal herb – meaning that it lessens sexual desire. While the alcohol in beer can lessen inhibitions – which may result in bawdier activities in many – the anaphradesiacal effect of the hops does counter act this to some degree. Gruit, on the other hand, does not counter this effect and also has a unique inebriatory effect due to the chemical composition of the herbs involved in its manufacture. One of noticeable aspect of this chemical composition is the Thujone content.

Thujones are chemicals known as alkaloids, which cause an additional form of inebriation when imbibed in beer. According to Jonathan Ott’s book, Pharmaecotheon I, Thujones act upon some of the same receptors in the brain as tetrahydrocannabinol (THC, as found in Marijuana), and are also present in the spirit known as Absinthe. Gruit and Absinthe sometimes share the same herbs in their manufacture, such as Wormwood, Anise seed, and Nutmeg, but it is the herb Yarrow (Achilles Millefolium) that contains the lion’s share of thujones in the gruit concoction.

Yarrow is an herb with many uses and plays a profound part in history and myth. According to Buhner, its use can be traced back 60,000 years. Through many different cultures, from Dakota to ancient Romans, Yarrow has been used to staunch serious wounds – it is even rumored to have been used by Achilles (hence the name Achilles Millefolium, the thousand leaved plant of Achilles). According to Buhner, the plants aphrodisiacal qualities are also documented in the Navaho culture. As an inebriant, it has been used in the Scandinavian countries and in North America as well.

Bog Myrtle (Myrica gale) and Wild Rosemary (Ledum glandulosum) also have many uses in the realm of herbalism, but not nearly as many as Yarrow. Both tend to have inebriation enhancing effects in beer, but also tend to cause a headache and probably a wicked hangover, if too much is drunk. The use of Bog Myrtle in ale was continued through the 1940s in Europe and the 1950s in outlying areas of England and the Scandinavian countries – Wild Rosemary probably through the 18th century.

For Centuries, Alewives Dominated the Brewing Industry
by Addison Nugent

BEER HAS BEEN AN ESSENTIAL aspect of human existence for at least 4,000 years—and women have always played a central role in its production. But as beer gradually moved from a cottage industry into a money-making one, women were phased out through a process of demonization and character assassination. […]

Professional brewsters and alewives had several means of identifying themselves and promoting their businesses. They wore tall hats to stand out on crowded streets. To signify that their homes or taverns sold ale, they would place broomsticks—a symbol of domestic trade—outside of the door. Cats often scurried around the brewsters’ bubbling cauldrons, killing the mice that liked to feast on the grains used for ale.

If all of this sounds familiar, it’s because this is all iconography that we now associate with witches. While there’s no definitive historical proof that modern depictions of witches were modeled after alewives, some historians see uncanny similarities between brewsters and anti-witch propaganda. One such example exists in a 17th-century woodcut of a popular alewife, Mother Louise, who was well-known in her time for making excellent beer.

While the relationship between alewives and witch imagery has still yet to be proven, we do know for sure that alewives and brewsters had a bad reputation from the jump. Beyond the cheating that some of their counterparts engaged in, brewsters also had to deal with the bad rap their entire gender suffered because of original sin. […]

Brewsters’ bad reputation didn’t help their case when wealthier, more socially-connected men started taking up the trade. After the devastation of the Black Plague, people began drinking a lot more ale, doing so in public alehouses instead of at home. This also marked a shift in people’s relationship with beer, which moved from being just a necessity and occasional indulgence to something closer to what we have today. Men suddenly saw they could make a real profit off of what was once seen as a semi-lucrative side gig for women. So they built taverns that were bigger and cleaner than the makeshift ones that alewives provided, and people flocked to them to revel and conduct business alike. Over time, alewives grew to be seen not only as tricky, but also dirty and their beer unsanitary.

Women continued to make low-alcohol ale for their family’s daily consumption after the Industrial Revolution increased production methods, which made buying beer cheaper and easier than making it at home. But that died in the 1950s and 1960s, when marketing campaigns branded beer as a “manly drink.” Companies such as Schlitz, Heineken, and Budweiser depicted beer as a means of unwinding after a long day of work, often featuring women serving their suited-up husbands cold bottles of brew.

The long battle between ale and beer
by Martyn Cornell

For those of you still with me: here’s a quote on ale and beer from 1912, less than a century ago, from a book called Brewing, by Alfred Chaston Chapman:

“At the present day the two words are very largely synonymous, beer being used comprehensively to include all classes of malt liquor, whilst the word ale is applied to all beers other than stout and porter.”

Why weren’t stout and porter called ales? This is a reflection, 200 years on, of the origin of porter (and brown stout) in the brown beers made by the beer brewers of London, rivals of the ale brewers for 500 years, ever since immigrants from the Low Countries began brewing in England with hops.

“Obadiah Poundage”, the aged brewery worker who wrote a letter to the London Chronicle in 1760 about the tax on “malt liquors” (the general term used for ale and beer as a class in the 18th century), is usually mined for the light he threw on the history of porter, but he is also very revealing on the continuing difference between ale and beer. In Queen Anne’s reign, about 1710, Poundage said, the increase in taxes on malt (caused by the expense of the War of the Spanish Succession) caused brewers to look to make a drink with less malt and more hops: “Thus the drinking of beer became encouraged in preference to ale … but the people not easily weaned from their heavy sweet drink, in general drank ale mixed with beer.”

This ale seems to have been brown ale (and the beer brown beer), for Poundage says that it was the gentry, “now residing in London more than they had done in former times”, who “introduced the pale ale, and the pale small beer they were habituated to in the country; and either engaged some of their friends, or the London brewers to make for them these kinds of drinks.” The pale ale “was sold by the victualler at 4d per quart and under the name of two-penny.” It was the need to counter the success of this pale ale that “excited the brown beer trade to produce, if possible, a better sort of commodity, in their way, than heretofore had been made”, an effort that “succeeded beyond expectation” with the development of what became known as porter, because of its popularity with London’s many street porters. But while the “brown beer trade” developed into the porter brewers, the ale brewers continued to find a market.

Indeed, outside London and the south of England, beer does not seem to have been that popular until Queen Anne’s time at the earliest. Daniel Defoe, writing in his Tour through the Eastern Counties of England, published in 1722, about the great hop fair at Stourbridge, just outside Cambridge, on the banks of the Cam, said:

“As to the north of England, they formerly used but few hops there, their drink being chiefly pale smooth ale, which required no hops, and consequently they planted no hops in all that part of England, north of the Trent; nor did I ever see one acre of hop ground planted beyond Trent in my observation; but as for some years past, they not only brew great quantities of beer in the north, but also use hops in the brewing their ale much more than they did before; so they all come south of Trent to buy their hops; and here being vast quantities brought, it is great part of their back carriage into Yorkshire and Northamptonshire, Derbyshire, Lancashire, and all those counties; nay, of late since the Union, even to Scotland itself.”

It looks to have taken a century for the habit of putting hops in ale to spread north: in 1615, Gervase Markham published The English Huswife, in which he declared:

“The generall use is by no means to put any hops into ale, making that the difference betwixt it and beere, that the one hath hops the other none; but the wiser huswives do find an error in that opinion, and say the utter want of hops is the reason why ale lasteth so little a time, but either dyeth or soureth, and therefore they will to every barrell of the best ale allow halfe a pound of good hops.”

Fourteen years after Defoe’s report on North of England pale ale, the first edition of the London and Country Brewer, by the Hertfordshire farmer William Ellis, succinctly summed up the difference between ale and beer in the 1730s:

“For strong brown ale brewed in any of the winter months and boiled an hour, one pound is but barely sufficient for a hogshead, if it be tapped in three weeks or a month. If for pale ale brewed at that time, and for that age, one pound and a quarter of hops; but if these ales are brewed in any of the summer months there should be more hops allowed.

“For October or March brown beer, a hogshead made from eleven bushels of malt boiled an hour and a quarter, to be kept nine months, three pounds and a half ought to be boiled in such drink at the least. For October or March pale beer, a hogshead made from fourteen bushels, boiled an hour and a quarter and kept twelve months, six pounds ought to be allowed to a hogshead of such drink and more if the hops are shifted in two bags, and less time given the wort to boil.”

Going on Ellis’s figures, early 18th century ale contained up to 60 per cent more hops than Gervaise Markham’s “huswives” used in ale brewing a century earlier, but still only around a quarter as much hops as the beer. This, Ellis said, was because “Ale … to preserve in its mild Aley Taste, will not admit of any great Quantity of Hops.”

* * *

Some relevant selections from previous posts:

The Agricultural Mind

Addiction, of food or drugs or anything else, is a powerful force. And it is complex in what it affects, not only physiologically and psychologically but also on a social level. Johann Hari offers a great analysis in Chasing the Scream. He makes the case that addiction is largely about isolation and that the addict is the ultimate individual. It stands out to me that addiction and addictive substances have increased over civilization. Growing of poppies, sugar, etc came later on in civilization, as did the production of beer and wine (by the way, alcohol releases endorphins, sugar causes a serotonin high, and both activate the hedonic pathway). Also, grain and dairy were slow to catch on, as a large part of the diet. Until recent centuries, most populations remained dependent on animal foods, including wild game. Americans, for example, ate large amounts of meat, butter, and lard from the colonial era through the 19th century (see Nina Teicholz, The Big Fat Surprise; passage quoted in full at Malnourished Americans). In 1900, Americans on average were only getting 10% of carbs as part of their diet and sugar was minimal.

Something else to consider is that low-carb diets can alter how the body and brain functions. That is even more true if combined with intermittent fasting and restricted eating times that would have been more common in the past. Taken together, earlier humans would have spent more time in ketosis (fat-burning mode, as opposed to glucose-burning) which dramatically affects human biology. The further one goes back in history the greater amount of time people probably spent in ketosis. One difference with ketosis is cravings and food addictions disappear. It’s a non-addictive or maybe even anti-addictive state of mind. Many hunter-gatherer tribes can go days without eating and it doesn’t appear to bother them, and that is typical of ketosis. This was also observed of Mongol warriors who could ride and fight for days on end without tiring or needing to stop for food. What is also different about hunter-gatherers and similar traditional societies is how communal they are or were and how more expansive their identities in belonging to a group. Anthropological research shows how hunter-gatherers often have a sense of personal space that extends into the environment around them. What if that isn’t merely cultural but something to do with how their bodies and brains operate? Maybe diet even plays a role. […]

It is an onslaught taxing our bodies and minds. And the consequences are worsening with each generation. What stands out to me about autism, in particular, is how isolating it is. The repetitive behavior and focus on objects resonates with extreme addiction. As with other conditions influenced by diet (shizophrenia, ADHD, etc), both autism and addiction block normal human relating in creating an obsessive mindset that, in the most most extreme forms, blocks out all else. I wonder if all of us moderns are simply expressing milder varieties of this biological and neurological phenomenon. And this might be the underpinning of our hyper-individualistic society, with the earliest precursors showing up in the Axial Age following what Julian Jaynes hypothesized as the breakdown of the much more other-oriented bicameral mind. What if our egoic consciousness with its rigid psychological boundaries is the result of our food system, as part of the civilizational project of mass agriculture?

The Spell of Inner Speech

This person said a close comparison was being in the zone, sometimes referred to as runner’s high. That got me thinking about various factors that can shut down the normal functioning of the egoic mind. Extreme physical activity forces the mind into a mode that isn’t experienced that often and extensively by people in the modern world, a state of mind combining exhaustion, endorphins, and ketosis — a state of mind, on the other hand, that would have been far from uncommon before modernity with some arguing ketosis was once the normal mode of neurocogntivie functioning. Related to this, it has been argued that the abstractions of Enlightenment thought was fueled by the imperial sugar trade, maybe the first time a permanent non-ketogenic mindset was possible in the Western world. What sugar (i.e., glucose), especially when mixed with the other popular trade items of tea and coffee, makes possible is thinking and reading (i.e., inner experience) for long periods of time without mental tiredness. During the Enlightenment, the modern mind was borne out of a drugged-up buzz. That is one interpretation. Whatever the cause, something changed.

Also, in the comment section of that article, I came across a perfect description of self-authorization. Carla said that, “There are almost always words inside my head. In fact, I’ve asked people I live with to not turn on the radio in the morning. When they asked why, they thought my answer was weird: because it’s louder than the voice in my head and I can’t perform my morning routine without that voice.” We are all like that to some extent. But for most of us, self-authorization has become so natural as to largely go unnoticed. Unlike Carla, the average person learns to hear their own inner voice despite external sounds. I’m willing to bet that, if tested, Carla would show results of having thin mental boundaries and probably an accordingly weaker egoic will to force her self-authorization onto situations. Some turn to sugar and caffeine (or else nicotine and other drugs) to help shore up rigid thick boundaries and maintain focus in this modern world filled with distractions — likely a contributing factor to drug addiction.

The Crisis of Identity

Prior to talk of neurasthenia, the exhaustion model of health portrayed as waste and depletion took hold in Europe centuries earlier (e.g., anti-masturbation panics) and had its roots in humor theory of bodily fluids. It has long been understood that food, specifically macronutrients (carbohydrate, protein, & fat), affect mood and behavior — see the early literature on melancholy. During feudalism food laws were used as a means of social control, such that in one case meat was prohibited prior to Carnival because of its energizing effect that it was thought could lead to rowdiness or even revolt (Ken Albala & Trudy Eden, Food and Faith in Christian Culture).

There does seem to be a connection between an increase of intellectual activity with an increase of carbohydrates and sugar, this connection first appearing during the early colonial era that set the stage for the Enlightenment. It was the agricultural mind taken to a whole new level. Indeed, a steady flow of glucose is one way to fuel extended periods of brain work, such as reading and writing for hours on end and late into the night — the reason college students to this day will down sugary drinks while studying. Because of trade networks, Enlightenment thinkers were buzzing on the suddenly much more available simple carbs and sugar, with an added boost from caffeine and nicotine. The modern intellectual mind was drugged-up right from the beginning, and over time it took its toll. Such dietary highs inevitably lead to ever greater crashes of mood and health. Interestingly, Dr. Silas Weir Mitchell who advocated the ‘rest cure’ and ‘West cure’ in treating neurasthenia and other ailments additionally used a “meat-rich diet” for his patients (Ann Stiles, Go rest, young man). Other doctors of that era were even more direct in using specifically low-carb diets for various health conditions, often for obesity which was also a focus of Dr. Mitchell.

Are you a hedgehog or a fox?

Are you a hedgehog or a fox? “The fox knows many things, but the hedgehog knows one big thing,” wrote Archilochus. Isaiah Berlin further developed this distinction.

I’m definitely a fox. I lack the capacity of knowing one big thing like the hedgehog. My mind is too unfocused and wandering for that. Some people are clever like foxes. That doesn’t describe me. I’m a fox simply because of not being a hedgehog. If I was capable of knowing one big thing, I’d gladly have pursued such a path. But the limitations of my mind don’t allow it. I long ago gave up on trying to find one big thing.

At first, I was thinking that most people are hedgehogs. The reason is because most people have something they believe in or adhere to and they feel no inclination to think beyond what they think they know, what is considered acceptable in mainstream thought or what is deemed to be true within the group they identify with. On the other hand, most people aren’t particularly consistent in their thinking either. I doubt that the average person knows one big thing, even if they tend to follow those who claim to know one big thing.

The average person in their lifetime is likely to jump from one big thing to another, depending on what is popular at the moment or depending on what is said by those perceived as authority figures. There is no single big thing they are likely to hold onto as something they personally know and care about. What they follow is more situational and that could be considered a more fox-like trait. So, I’d say most people don’t entirely fall into either category, which is maybe to be expected. As often is the case, the average person isn’t found at either of these extremes.

Whatever you are, it is more important the kind of person you listen to. As discussed in the BBC video (linked above), foxes are more accurate in their predictions than hedgehogs. There is more to life than predictions, though. In other research, pessimists were more accurate in assessing present conditions (depressive realism), even if they weren’t so talented at predicting changing conditions. A pessimistic hedgehog could still be right about his or her one big thing, as it applies to the present situation. In terms of decision-making, it matters whether you’re looking for immediate results or long-term planning.

Based on context, either a hedgehog or a fox could lead you astray. Then again, not all things are equal. A pessimistic fox might be the best balance, at least in terms of truth-seeking, whereas an optimistic hedgehog might be the least reliable on all accounts. Still, if you want to get shit done, an optimistic hedgehog probably will be the most motivated to act (and inspiring others to follow) according to their beliefs and hence change conditions so as to manifest them. Their delusions might make them more effective under the right conditions, in order to create new realities. But effective sometimes can be dangerous when a society is under the spell of authoritarian true believers.

I’ll stick to being a fox. It’s what I’m good at. Also, it’s a more interesting way of looking at the world. To be a fox is to be curious, if nothing else. That won’t stop me from looking for some one big thing, not that I’ll likely ever find it. I’ll run from one place to another, sniffing around the hedgehogs, as they are fascinating creatures. And who knows, a few of them might be onto something.

* * *

See previous post:

Fox and Hedgehog, Apollo and Dionysus

A Century of Dietary and Nutritional Trends

At Optimizing Nutrition, there is a freaking long post with a ton of info: Do we need meat from animals? Let me share some of charts showing changes over the past century. As calories have increased, the nutrient content of food has been declining. Also, with vegetable oils and margarine shooting up, animal fat and dietary cholesterol intake has dropped.

Carbs are a bit different. They had increased some in the early 20th century. That was in response to meat consumption having declined in response to Upton Sinclair’s muckraking of the meat industry with his book The Jungle. That was precisely at the time when industrialization had made starchy carbs and added sugar more common. For perspective, read Nina Teicholz account of the massive consumption of animal foods, including nutrient-dense animal fat and organ meats, among Americans in the prior centuries:

“About 175 pounds of meat per person per year! Compare that to the roughly 100 pounds of meat per year that an average adult American eats today. And of that 100 pounds of meat, more than half is poultry—chicken and turkey—whereas until the mid-twentieth century, chicken was considered a luxury meat, on the menu only for special occasions (chickens were valued mainly for their eggs). Subtracting out the poultry factor, we are left with the conclusion that per capita consumption of red meat today is about 40 to 70 pounds per person, according to different sources of government data—in any case far less than what it was a couple of centuries ago.” (The Big Fat Surprise, passage quoted in Malnourished Americans).

What we forget, though, is that low-carb became popular for a number of decades. In the world war era, there was a lot of research on the ketogenic diet. Then around the mid-century, low-carb diets became common and carb intake fell. Atkins didn’t invent the low-carb diet. Science conferences on diet and nutrition, into the 1970s, regularly had speakers on low-carb diets (either Gary Taubes or Nina Teicholz mentions this). It wasn’t until 1980 that the government began seriously promoting the high-carb diet that has afflicted us ever since. Carb intake peaked out around 2000 and dropped a bit after that, but has remained relatively high.

The inflammatory omega-6 fatty acids combined with all the carbs has caused obesity, as part of metabolic syndrome. That goes along with the lack of nutrition that has caused endless hunger as Americans have been eating empty calories. The more crap you eat, the more your body hungers for nutrition. And all that crap is designed to be highly addictive. So, Americans eat and eat, the body hungering for nutrition and not getting it. Under natural conditions, hunger is a beneficial signal to seek out what the body needs. But such things as sugar have become unlinked from nutrient-density.

Unsurprisingly, Americans have been getting sicker and sicker, decade after decade. But on a positive note, recently there is a slight drop in how many carbs Americans are eating. This is particularly seen with added sugar. And it does seem to be making a difference. There is evidence that the diabetes epidemic might finally be reversing. Low-carb diets are becoming popular again, after almost a half century of public amnesia. That is good. Still, the food most American have access to remains low quality and lacking in nutrition.












image3-1.png

Carnivore Is Vegan

“I’m going to tell your audience something that not many farmers would ever admit. This happens on all farms. If you like eating avocados, for a farmer to grow avocados financially, especially biodynamically, where we’re enhancing the ecosystem and helping nature, we have to grow at least 20 to 40 acres of avocado, and we have to be able to sell those directly to our market, to our consumer.

“So here I am, farming 20 to 40 acres. That’s going to require me to kill at least 35 to 40,000 gophers to protect those trees. Humming birds, accidentally when I spray non-synthetically-derived organic spray, accidentally killing bees, accidentally killing ladybugs, and intentionally killing ground squirrels. So there are 50 to 100,000 deaths that happen just to grow avocados.

“And my point is that none of us are getting out of this without blood on our hands. It’s just at what point and how connected are you to the process, but that doesn’t excuse you from the reverence and the responsibility of life.”
~ Rich Roll, vegan farmer and influencer (clip & full video)

“A lot of animals are killed in all kinds of agriculture. I’ll never forget the first time I saw a combine harvester go through an organic soybean field and kill all the animals that had made that field their home. Among the many animals that died that day were baby bunnies that were skinned by the blades and were then eaten alive by hawks. The hawks followed the harvester through the field looking for an easy meal. I knew that the farmer had contracted his crop to an organic tofu company and that most of the people eating this food would be vegans and vegetarians. The irony of this situation was enough to stop me from going vegan for many years afterwards. I would frequently bring up this anecdote when I would argue with vegan friends. It still annoys me when my fellow vegans act as though their lifestyle is 100% cruelty free and that no animals die in the process of making their food. It speaks to an ignorance of the realities of rural life.”
 ~ Charlie Knoles, self-identified vegan, meditation teacher, B.S. in Environmental Biology

Which diet causes the most harm? And which the least? The least harm principle is central to veganism; as it is to some religions, from Seventh Day Adventism to Buddhism (ahimsa). Some vegans go so far as to suggest that this principle is more of a philosophy, worldview, and lifestyle than it is necessarily, primarily, and entirely a diet. Indeed, others go even further in treating it as a religion or as central to their religious or spiritual practice. For the sake of argument, we are going to use that definition. Veganism is about the consequences that the diet and everything else directly and indirectly causes or otherwise contributes to and is complicit in. So, we can’t know what is vegan merely by what kinds of foods a particular eating pattern includes or excludes. And hence we can’t know which diet is most ‘vegan’ in causing the least harm by isolating diet from all the rest.

The etymology of ‘diet’ connects the word back to the meaning of ‘lifestyle’ or ‘way of life’. For veganism, this implies empathy, compassion, loving-kindness, and moral concern; in relation to the larger living world. As a lifelong environmentalist, I take quite seriously the vegan ideal and critique. I’m a bleeding-heart liberal, an animal-loving and tree-hugging sensitive male, not to mention having a streak of radical leftism. The political views of many vegans overlap with my own. Yet I’ve never been a vegan, although I briefly was vegetarian when younger, as my brothers (and their families) still are vegetarian. For whatever reason, the fair number of self-identified vegans I’ve known over the decades never swayed me to eliminate all animal foods and products, much less aspire to the broader vegan identity. Let me explain why.

Even limiting ourselves to a dietary ideology alone, we have to consider the broader context. Diets are supported, promoted, and made possible by the entire network of food system, agriculture, land management, resource usage, environmental practices, ecosystems, petrochemicals, transportation, industry, processing, packaging, economics, trade, markets, sellers, monied interests, lobbyist organizations, public policies, official dietary recommendations, institutionalized ideologies, funding of scientific research, etc. The majority of harms along with other costs are indirect and hidden and externalized onto others, sometimes privatized (e.g., poor rural housing next to chemical-sprayed farm fields) and at other times socialized (e.g., chemicals getting into the water supply to be cleaned up by a public water plant).

I’ve long been obsessed with externalized costs and the moral hazard that follows. This is a particular problem when ideology and money are mixed. Diet has been enmeshed in ideology for millennia (e.g., religious food laws) and the food system has long been central to most major economies, such as how the United States became so wealthy and profitable primarily through agriculture. Veganism magnifies this confluence. There is no other dietary ideology that is more dogmatic or more dependent on agriculture. So, to assess veganism in its mainstream form is to analyze how modern food production is shaped by and conforms to modern ideology; and how in turn it bolsters the ancient ideological impulse within food systems. It’s not only what diet does or does not cause the most harm but also how we perceive and understand harm or fail to do so.

“I’ve watched enough harvests to know that cutting a wheat field amounts to more decapitated bunnies under the combine than you would believe.”
~ Barbara Kingsolver

“As I was thinking about the vegan conclusion, I remembered my childhood on the farm and where our food comes from and how it is produced. Specifically, I remembered riding on farm equipment and seeing mice, gophers, and pheasants in the field that were injured or killed every time we worked the fields. Therefore, I realized that animals of the field are killed in large numbers annually to produce food for humans.”
~ Stephen L. Davis

“When I inquired about the lives lost on a mechanized farm, I realized what costs we pay at the supermarket. One Oregon farmer told me that half of the cottontail rabbits went into his combine when he cut a wheat field, that virtually all of the small mammals, ground birds, and reptiles were killed when he harvested his crops. Because most of these animals have been seen as expendable, or not seen at all, few scientific studies have been done measuring agriculture’s effects on their populations.”
~ Ted Kerasote

If veganism means the overall avoidance or lessening of the death, suffering, and exploitation particularly of animals and other sentient life (including humans), then it is rationally and morally plausible that an animal-based diet, including carnivore and maybe even lacto-ovo-vegetarianism, is potentially the most vegan diet around; assuming it is organically-grown and locally-sourced, sustainably-managed and regeneratively-farmed, pasture-raised and wild-caught. Besides hunting and gathering, pastoralism as a food system and way of life kills the fewest animals, fewer than agriculture by far. For every life taken by a meat-eater (e.g., a single pasture-raised chicken or cow) or egg-and-dairy-eater, a vegan might kill hundreds or thousands (coyotes, foxes, deer, rodents, snakes, birds, insects, spiders, etc). That isn’t even to include the vast spectrum of species and entire ecosystems annihilated in the original creation of farmland.

Over an entire year, a single human can on a carnivore diet or a single small family on an omnivore diet could survive on the meat, organs, fat, marrow, bone broth, etc from a single cow: 570lb beef at 605,000 cals, 280lb fat and bone, 32lb offal/carcass shrink (Dr. Zoe Harcombe PhD, Should We Be Vegan). That would allow for around a couple pounds of fatty beef and organ meats per day every day, 365 days per year (on days that I do strict carnivore and beef only, I typically eat about 2-3 lbs). Or one could eat two pigs instead, each producing upwards of 270lb pork, bacon, and pork belly; not to mention a ton of lard to use for cooking, including for plant foods. But if one prefers chicken (3.3lb each but with less fat and calories), that would mean the death of 228 animals, according to Dr. Harcombe; not that many people are likely to eat a chicken-exclusive diet. Of course, those on animal-based diets could get much of their diet from eggs and dairy as well, neither of which necessarily requires killing any animals.

Furthermore, whatever one’s choice of animal foods, all of it could be locally, sustainably, and regeneratively raised; even on open land with wildlife habitat and wildlife grazing. Compare that to the ecological devastation of industrial agriculture (and all of the industrial system that goes with it) that is a major force behind our present ongoing mass extinction. Farming directly kills 7.3 billion wild animals globally or 114 per hectare of cropland farmed, excluding the deaths of insects and spiders (from honeybee population collapse caused by insecticides to monarch butterfly population collapse caused by fencerow-to-fencerow farming), not to mention the wiping out of microbial life in the soil. But that isn’t even to take into account the even larger indirect death count from the entire industrial food system that vegans and vegetarians are dependent on (The Farming Truth Project, Hypoxic Dead Zones and Agriculture). To put it in full context:

“18.04 animals die in the production of 10,000 grams of plant-based protein. This is in comparison to only 3.68 deaths for 10,000 grams of animal-based protein. […] 18.35 animals die to produce 1,000 servings of plant-based food. This is in comparison to only 8.31 deaths for 1,000 servings of animal-based foods. […] Plant products kill 2.96 times more animals per calorie, 4.9 times more per gram of protein, and 2.21 times more per serving than animal products. Plant foods are over twice as deadly as animal foods. […] 114 animals die per hectare of crop land farmed versus only 46 animals dying per hectare of pastureland for livestock. […] a vegan kills 1.16 times more animals with the amount of servings realistically consumed compared with an omnivore” (The Farming Truth Project, Vegans Kill More Animals – Here’s Proof; also see: Introduction: Ways that Animals are Killed in Crop Production; & How Many Die For Your Food: Calculating the Death Toll of Crop Production vs. Livestock Production).

For even further context, a cow only needs about an acre of land for pasture (there are approximately 2.5 acres per hectare); 25-35 pigs can also be kept on a mere acre; and 50 chickens could be raised on an acre, such as putting them on the pasture after the cows to eat the maggots from the cow manure. That is all the land required for someone on a carnivore diet. A vegan, on the other hand, depends on two acres, almost a hectare (William Swanson, How Much Land Does It Take To Feed One Person – Online Calculator). If we calculate from the above data, two acres would kill about 88 animals every year. Yet on two acres of carnivory, one could easily raise enough food for an entire family with a relatively small number of animal deaths, especially if one of those acres was used to raise a dairy cow and egg-laying hens. So, even if a carnivore or omnivore also eats some other meat and animal foods besides beef, they would be hard put to kill as many animals as is the case on the vegan diet.

All in all, someone on a fully carnivore diet would kill the least of all, particularly as a carnivore diet is typically low-carb and so tends toward less hunger/cravings and hence less snacking. That would be even more true for meat from animals raised on pasture. Whether meat-eating or meat-abstaining, the death count is at least partly known and so false claims of unintentionality is no justifiable rationalization. There is no avoidance of moral culpability. This is not about being clever but about what is genuinely least harmful and most environmentally sustainable, as human and non-human health are intertwined. Rather than a pose of moral righteousness, our concern should be with what brings the greater overall good.

It’s no small point that the people with nutrient-dense animal foods are overall healthier, whereas the vegans require additional nutritional fortification and supplementation which would contribute further to their land usage, environmental externalized costs, and harm to life. If veganism was the healthiest and most sustainable diet, why has there never been a vegan society in all of human existence? Even in equatorial regions plant foods have limited growing seasons. The hunter-gatherer Hadza, for example, only have fruit and honey available a few months of the year. As another example, the Piraha living in the lush and abundant Amazon forest depend for their diet 90% on fish.

I did do a carnivore diet for a couple of months as an experiment, although I wasn’t strict about it. For a while now, I’ve been back on a diet that tends toward ketogenic, paleo, and traditional foods. My food sourcing is important to me with an emphasis on locally produced, seasonally available, organic, and pasture-raised. This means I regularly shop at the nearby farmers market. So, despite not being carnivore at present, I am heavily biased toward animal foods with plenty of meat and eggs, along with some dairy. The plant foods I eat are also almost entirely from the farmers market, in particular the fermented veggies I enjoy. That translates as eating a greater proportion of plant foods when available in the warm time of the year and more animal foods in winter. Not only is this diet extremely healthy but also highly ethical and environmentally sustainable.

Raising animals on pasture avoids all of the problems associated with industrial agriculture and factory farming. It is actually a net gain for local ecosystems, the biosphere, and the human species. The health of the soil actually improves with pasture and atmospheric carbon is captured — indeed, grasslands draw down more carbon than do farm fields or forests. Run-off, erosion, and pollution are also eliminated. On top of that, pasture provides habitat for wildlife, as opposed to mass farming and monoculture that destroys habitat and displaces wildlife, not to mention poisons, starves and slaughters immense numbers of wildlife. If you’re pro-life in the broadest sense, the last thing in the world you’d want to be is vegan, as it is inherently and inevitably dependent on industrial agriculture and mass transportation.

Vegan arguments against harm to animals don’t apply to a pasture-raised and wild-caught carnivore diet or any local animal-based diet combined with locally and seasonally available plant foods. (By the way, today was the beginning of wild mulberry season — delicious! I was knocked right out of ketosis and was glad for it. That is the reason plants evolved the highly addictive drug called sugar, so that we would eat their fruit and spread their seeds, not so that one day agriculture would make possible industrially-produced and health-destroying high fructose corn syrup.)

Veganism creates a similar disconnect as seen with right-wing ‘pro-lifers’ who oppose abortion. As I’ve pointed out, countries that ban abortions don’t decrease the rate of abortions and sometimes increase them. The main change is whether abortions are legal and safe or illegal and unsafe. But anti-abortionists refuse to accept responsibility for the consequences of the policies they support. Similarly, vegans also refuse to accept responsibility for the deaths and destruction that their diet incurs. Whether one intentionally or unintentionally causes harm, the harm is equally real. This is how symbolic ideology that makes people feel good trumps practical concerns about what actually makes the world a better place.

“What do plants eat? They eat dead animals; that’s the problem. For me that was a horrifying realization. You want to be an organic gardener, of course, so you keep reading ‘Feed the soil, feed the soil, feed the soil…’

“All right. Well, what does the soil want to eat? Well, it wants manure, and it wants urine, and it wants blood meal and bone meal. And I…could not face that. I wanted my garden to be pure and death-free. It didn’t matter what I wanted: plants wanted those things; they needed those things to grow.”
~ Lierre Keith

“There is no place left for the buffalo to roam. There’s only corn, wheat, and soy. About the only animals that escaped the biotic cleansing of the agriculturalists are small animals like mice and rabbits, and billions of them are killed by the harvesting equipment every year. Unless you’re out there with a scythe, don’t forget to add them to the death toll of your vegetarian meal. They count, and they died for your dinner, along with all the animals that have dwindled past the point of genetic feasibility.”
~ Lierre Keith

There is no reason the world’s population couldn’t live according to the meat-based diet I and many others follow; or else some other version of an animal-based diet such as the Paleo diet or the traditional Mediterranean diet, but also lacto-ovo-vegetarianism. Plant-based advocates ask for evidence that eating meat and other animal foods is sustainable. Are these people utterly disconnected from reality? Ruminants have been around for 50 million years. Chickens and other fowl descended from dinosaurs. And fish can be traced back 530 million years. Animals eating other animals has been going on for over 800 million years. Humans began eating meat, animal fat, and marrow 2.6 million years ago. The overall biomass hasn’t changed much over time. Also, cows don’t increase total atmospheric methane because the grasslands they graze on capture methane. It’s a freaking natural cycle! It’s been going on for as long as life has existed. Isn’t that long enough to prove sustainability?

Besides, very little of the arable land available can be used for farming plant foods. But most of it can be used for grazing. Also, grazing animals for food can be done alongside keeping the land open for wild animals as well. Keep in mind that, in North America, there once were more buffalo roaming the continent than there are now cows and the vast herds of buffalo were what kept the prairies healthy. Even in countries that don’t have good farmland, animals can always be raised locally on pasture or open land, from mountains and valleys to grasslands and deserts. There is no country in the world that lacks land for grazing. If not cows, then chicken, ostriches, pigs, goats, sheep, camels, or whatever else; not to mention traditional ways of raising fish in ponds (a major resource of the ancient Romans and early medieval Europeans).

Let’s put this in perspective, 90% of usable land in North America can only be used for wildlife and livestock, not farming. In other places (Africa, India, Australia, etc), it’s even higher at 95% of the usable land. A point of confusion is that some major global organizations, like the United Nations, only speak of animal farming in terms of pastures and meadows that are only two-thirds of the land in use (Food and Agriculture Organization of the United Nations, Sustainable Food and Agriculture) and, in the United States, a little over one-third of total land is pasture (Dave Merrill & Lauren Leatherby, Here’s How America Uses Its Land). But none of this includes savannahs, shrublands, tundra, forests, wetlands, mountains, rough foothills, rocky islands, arid areas, and deserts where one can sometimes graze cattle but certainly graze animals other than cows; such as chickens, goats, pigs, camels, alpacas, etc with much of it falling under the category of ‘rangeland’ that by itself is half the earth’s land surface (World Wildlife Fund, New Data Shows Rangelands Make Up Half the World’s Land Surface – and Present a Severely Underutilized Opportunity to Address the Climate and Biodiversity Crises); along with hunting, trapping, and fishing of wild game.

Yet even when only including agricultural lands and ignoring non-agricultural lands and waters that could potentially be used for immense and sustainable food production, one study still found that, “The vegan diet, surprisingly, fed fewer people than two of the omnivore diets and both of the other vegetarian diets, suggesting food choices that make use of grazing and forage land as well as cropland could feed more people than those that completely eliminate animal-based food from our diets” (Kristen Satre Meyer, Which Diet Makes the Best Use of Farmland? You Might Be Surprised.). So, all of the animal-based diets were proven more environmentally sustainable than the strictly plant-exclusive diet. The study’s analysis did conclude that reducing meat was more sustainable for agricultural lands, the few percentages of all land. It was designed to be biased against animal foods, and yet the animal-based diets still showed their merit. Now add in the animal foods from half the earth’s land surface and all of the earth’s water.

Oceans, seas, lakes, ponds and rivers aside, there are so many kinds of lands and so many ways they could be implemented for local and global food production. Conventional industrial farming of bathing GMO monocrops in chemicals, with its erosion and pollution, is not going to be the future. As an odd example, think of the traditional pig farming on Okinawa, a small rocky island, where the pig pen was traditionally underneath the house where human waste and excrement was fed to the pigs — how does one describe that kind of efficient and effective land use? Not that it’s being suggested that Americans should follow this specific example, although it does demonstrate how animal foods can be increased in ways that can’t be as easily done with plant foods. We are surrounded by lands unused and underutilized. The amount of wasted land in the average suburb could be used to raise a large part of the foods needed for those living there. We Americans have come to take for granted how much land we not only waste but use destructively, such as the chemical-drenched ecological deserts of suburban yards and greenspaces. Many suburbs are built on farmland. Why are we so insane as to build housing on arable land? We should be emphasizing and incentivizing residential concentration, not sprawl.

What plant-based environmentalists ignore is that deforestation is rarely done for cattle grazing, particularly not deforestation of rainforests that have poor soil for grazing. The cause of that deforestation is primarily for other reasons, from logging to mining, but half of it is for croplands to produce palm oil and soy. Cows are only put on such poor soil as an afterthought when there is nothing else to do with the land. In the US, it’s interesting to note how no one is talking about the deforestation of farmland: “As forests have been cleared from farmland, a long-term decline in grazed forestland of 186 million acres has taken place since the start of the MLU series” (Daniel Bigelow, A Primer on Land Use in the United States). We could replant a lot of trees on farmland, and that would healthier for the soil and provide habitat for wildlife, but then it could only be used for grazing.

Government agencies in the United States (EPA, USDA, etc), fortunately, do categorize the other kinds of grazing lands: grassland pasture and range, including shrub and brushland; and forest land grazed (EPA, Definitions of Land Use Categories). For whatever reason, these vast tracts of non-agricultural lands never come up in terms of animal production within mainstream environmentalist arguments, critiques, and debates. Many of the present farmland in places like California couldn’t be used for agriculture at all, if not for the massive redistribution of water from elsewhere. Yet this otherwise dry landscape is perfectly fine for grazing that requires no irrigation.

Critics of an animal-based diet like to blame cattle for using excessive water, but the reality is 94% of the water used is from greenwater; i.e., rain that falls on the land where the cattle are kept; and that is factoring in factory-farmed animals that spend 80% of their lives on pasture (M. M. Mekonnen & A. Y. Hoekstra, The Green, Blue and Grey Water Footprint of Farm Animals and Animal Products). The point being that cattle are not the reason rivers and aquifers are being drained. If one wants to complain about water-intensive farming, the target of one’s ire should be favorite crops like cotton, rice, potatoes, onion, garlic, sugarcane, sugar beets, almonds, walnuts, avocados, olives, raisins, grapes, applies, apricots, cherries, peaches, nectarines, pears, plums, prunes, figs, kiwis, bananas, grapefruit, lemons, oranges, dates, jojoba, etc. Imagine a vegan environmentalist trying to avoid those environmentally unsustainable crops, along with other problematic crops such as soy, corn, and spinach (Quynh Nguyen, 5 Least Sustainable Vegan (Plant-Based) Foods).

The amount of land unused or underutilized for animal food production and procurement is immense. That is not the case for agricultural land that is already being pushed to its most extreme capacity. So, considering only 3% of land is permanent crops (Hannah Ritchie & Max Roser, Land Use), are we going to try to feed the global population with just a few percentages of the available land and ignore the rest? And are we going to ignore the 71% of the earth’s surface that is water and that produces fish and seafood? In ever more intensively farming, we are destroying what is left of the arable land and polluting the water. We’ve already lost most of the earth’s top soil, mostly over the past century; whereas regenerative pasture can actually increase top soil.

“Roughly sixty percent of insects in plant agricultural areas, in China, Europe, and North America, have disappeared. This includes all insects, not just insects that eat crops. Tilling, harvesting, and chemicals kill. Mono-crops, fields with a single kind of plant, don’t provide habitat to animals that need a variety of plant species to survive.

“Of the top five crops raised in the US for human uses, corn, soybeans, rice, wheat, and cotton… all are protected by destroying animal species endemic to the areas they grow in. Of these crops, 75% of corn is grown for either ethanol fuel, corn oils, and corn syrups. Human uses. 95% plus of soybeans are processed to extract oils for human uses, and the waste product after the oils are extracted is fed to livestock. Rice is almost exclusively human use. Most wheat is ground for flour. Cotton is grown for fibers to make cloth.

“Of crops grown exclusively for animal feeds, natural or improved pasture is actually one of the few crops that provide habitat for wild species. Alfalfa is a perennial crop so land is tilled far less often, and has such long roots that it needs very little supplemental watering.

“Can farmers grow crops without killing animals? With the present world population, the necessity for industrial scale agriculture, I don’t see how. But it is easy to see that plant agriculture kills far more animals per pound of nutrition than raising animals.”
 ~ Todd Elliot, former rancher, B.S. in Animal Science from Utah State University

Farmland, in the first place, is created by killing numerous species and destroying ecosystems and replacing them with an ecological desert; not to mention the need for constant killing of any wildlife that attempts to return to the land. “Land conversion from natural ecosystems to agriculture has historically been the largest cause of greenhouse gas emissions), linked to loss of biomass and carbon in biomass above and below ground. Today, land conversion to agriculture continues to be a major driver of biodiversity loss and land degradation” (Food and Agriculture Organization of the United Nations, Sustainable Food and Agriculture). That is insanity! Industrial agriculture and factory farming makes no sense, except from a capitalist model of private profit and externalized/socialized costs. A local animal-based diet — if not carnivore or omnivore, then ovo-lacto-vegetarian — is the only way to feed the world’s population, maintain optimal health, avoid the greatest harm to animals, and ensure environmental sustainability.

Veganism didn’t exist prior to modern agriculture, industrialization, and mass transport. Grazing animals, on the other hand, has been the mainstay of the human diet for millions of years. There is no traditional diet that wasn’t centered on animal foods, the source of the most energy-dense and nutrient-dense foods, guaranteeing every essential and conditionally essential nutrient, many of which are missing or insufficient on a plant-exclusive diet. And when done low-carb as was typical of traditional societies, ketosis allows people to eat less food and go for longer periods of time without eating. Many people on animal-based diets do regular fasting (OMAD, intermittent, and extended). In ketosis, I easily skip meals or go several days without food and it doesn’t bother me. Since ketosis allows for smaller intake of food, that is an additional decreased impact on the environment.

The standard American diet (SAD) that is plant-based is neither healthy for the individual nor healthy for the environment. Keep in mind that almost all junk foods are vegan: potato chips, crackers, cookies, candy, pop, etc (the main ingredients being potatoes, wheat, corn, rice, sugar, and seed oils). This vegan junk food is mass farmed, mass produced, and mass shipped, not to mention mass subsidized. Even most healthier plant-based foods, including whole foods, that vegans rely upon are shipped from distant regions and countries with very little regulation for the health of environment and workers — think about the environmentally-unsustainable and water-wasting Californian agriculture that provides much of the produce for plant-based diets, particularly in winter. Veganism contributes to pollution and the need for heavily-subsidized infrastructure.

The human health aspect, though, is no small issue. Someone on an animal-based diet requires no supplements or fortified foods to avoid nutritional deficiencies. Vegans, on the other hand, have to carefully supplement to avoid serious health problems. All of those supplements and fortified foods are industrially-produced and that contributes to pollution and environmental degradation. On top of that, those who don’t include sufficient animal foods in their diet, even when they supplement, still tend to have metabolic diseases. Keep in mind that metabolic diseases are the single greatest healthcare cost. And the industrial production of medical supplies and pharmaceuticals is one of the largest sources of pollution and trash. Healthcare alone has a higher carbon footprint than animal farming.

What is ethical about any of this? Good intentions are not good enough. We can’t separate ourselves from the world we live in. It’s a fantasy that we can live apart from the natural cycle of life and death. Trying to force that fantasy upon the world, some might call that a nightmare. A diet is part of an ecosystem, all contained within a living biosphere. In pretending to be separate, we cause even more death and suffering. Mass extinction was always inherent to agriculture. “The end,” as Lierre Keith said, “was written into the beginning.” There is no avoiding this, as long as we continue down this path of exploitative civilization. We can embrace that ending, though, and seek a new beginning.

“Agriculture is the biggest mistake in human history,” as put by George Armelagos. And on the same note, Jared Diamond wrote that, “Archaeologists studying the rise of farming have reconstructed a crucial stage at which we made the worst mistake in human history. Forced to choose between limiting population or trying to increase food production, we chose the latter and ended up with starvation, warfare, and tyranny.” So, are we doomed? Only if we choose to be. Agriculture as we know it can’t continue. Can it be done differently? Others have offered more optimistic answers.

If we hope to find another way before it’s too late, we must look for inspiration in the traditional food systems that still survive. And there most definitely is hope. We already know of ways to reverse the damage and rehabilitate the land. No doubt further understandings will be gained over time that will allow even greater results. But the key is that more animals, wild and domestic, will be needed to make possible this course of action. That is to say, in place of ecological deserts of monocultural farming, we need to return to the environmental norm of biodiversity within thriving ecosystems.

“The persistence of human life on this planet depends on soil ecosystems. Ultimately, I don’t care what diet you eat as long as it leads to the enrichment of organic matter in the earth and mycorrhizal networks. Show me the plant-based diet that does this. Without ruminants ecosystems will collapse. Tilling of the soil for mono-crop agriculture is the enemy (and releases massive amounts of carbon) not cows, Bison and other animals.”
~Dr. Paul Saladino

“If we took 75% of the world’s trashed rangeland, we could restore it from agriculture back to functioning prairies — with their animal cohorts — in under fifteen years. We could further sequester all of the carbon that has been released since the beginning of the industrial age. So I find that a hopeful thing because, frankly, we just have to get out of the way. Nature will do the work for us. This planet wants to be grassland and forest. It does not want to be an agricultural mono-crop.”
~ Lierre Keith

“Viewing this global scene, as I have been doing for many years, I will stake my life on it that humanity’s best hope lies in one simple idea that no scientist can sensibly argue against – that management in this 21st century should be holistic and no longer reductionist. And Holistic Management of course includes recognizing that only livestock with Holistic Planned Grazing (or better process when developed) can address global desertification, annual burning of billions of hectares of grasslands and savannas, and regenerate the world’s dying soils and soil life essential to addressing climate change. […]

“Reductionist management, without using livestock managed on the land in a way that addresses global desertification and climate change, will inevitably lead to the doomsday predictions of Wallace-Wells. Billions of people dead and hundreds of cities destroyed and worse in the relatively near future no matter how many hopeful measures we might take.”
~ Allan Savory

* * *

Are animals killed in the process of farming vegan foods? Is it possible for a vegan to ensure that no animals were harmed in the production of their food without growing it themselves?
from Quora

Dan Eady: “Intensive farming practices such as wheat cropping introduced to natural environments kills far more than just animals it destroys entire ecosystems. Many species of plant and animal life are wiped out or displaced as the cropping practice begins. This new environment is then usually favourable to a much smaller number and less diverse number of species. So animals such as rodents attempt to colonise the changed environment but are then killed through human control methods or inadvertently through the growth and harvest practices employed through human activity upon the crop.”

Tariq Hossenbux: “As many of the other answers state, billions of insects and animals are killed when crops are conventionally grown. Millions of snakes, groundhogs and other small creatures. Wheat farmers routinely poison mice, and pesticides kill countless insects.

“What is really interesting though is that using a field for cattle pasture land may actually result in less total animal deaths and also preserve the native plant life. Many migrating insects depends on particular weeds to eat, and crop farmers often use excessive amounts of herbicide wiping them out. This one of the reasons for the decline of Monarch Butterfly populations in North America.”

Dan Hunter: “Yes, animals get killed when you grow crops. Other answers have mentioned running animals over when plowing and mowing, but if you just think about the fact you are converting a natural environment into cropland you soon realize that a lot of animals just lost their homes. So not only does crop production kill animals, it often kills all the succeeding generations of animals on that land.

“To illustrate the idea think about american bison and barbed wire. Before the farmers got to the prairies there were herds of buffalo so large they could take days to pass through a location. Wherever they went they ate the grass and trampled what they didn’t eat. As soon as the first plow made it through the Cumberland gap and onto the prairies the buffalo was doomed. If the market hunters had not shot the buffalo into near extinction the farmers with their plows and wire fencing would have sealed their fate because the fencing to protect the crops would have meant no migration of the buffalo to fresh pasture and certain starvation for them.

‘You can also look up the fate of the prairie chicken and the black footed ferret. These were also destroyed by wheat farming. Many farms were created by draining wetlands. This means loss of habitat for animals like beaver, muskrats, ducks, geese, frogs, etc. It does not really matter if the farmers are large agribusiness or if the are small farm holders. The result is the same.”

Kamia Taylor: “All of the previous answers talk about what gets killed by tilling before that ground can be planted. But if we got back even further, massive amounts of native prairies, wetlands and forests are still being destroyed, along with every living thing that called that area home, from birds, amphibians, mammals, insects and more — all so that more corn, SOY (a vegan’s favorite go-to food) and grains can be planted there — not to mention rainforests being torn apart ruthlessly for the production of palm kernel and other oils, coffee, cacao.

“In addition, massive numbers of animals are being killed off (over 200 Tule Elk died just recently) so that water they would have had access to is diverted to support, as an example, almond farming for vegan almond milk. Most people have have never planted anything have no idea just how much water vegetables and fruits use to come to maturity.

“So not matter what you do, whether you are vegetarian or omnivorous, you ARE going to impact the rest of the planet negatively to feed yourself. The good new is that when you die, you can be cremated and become compost to feed the next generation.”

Belinda Mellor: “Besides the small animals, of which there are millions killed, there is also deforestation in order to grow crops such as coffee, tea, palm oil, bananas, sugar, coconuts… some of these have been devastating. For instance, as a family we considered spending a year on an island that had a fairly sizeable coconut industry, and were advised that we would need vitamin tablets, as getting fresh fruit and vegetables was difficult – everywhere had been stripped. That was historical destruction, but just today I read about the rescue of an orang-utan stranded in a tiny ‘island’ of forest cut off by palm oil planting. She was lucky, many of her kind have perished, killed by logging machinery. And don’t forget all the birds that are not just accidentally killed, but are culled for fear of them eating crops: in Australia it’s all-out war on some parrot species for that reason.”

* * *

Here is another argument comes up, but usually only shows up in brief comments. The following is a good response in explaining why the argument makes no sense: “No, the majority of this agriculture is for human consumption, not to feed livestock” (from the comments section of Karen Lindquist’s The Least-Harm Fallacy of Veganism). I’ll first share the comment to which the second comment is a response.

Ira
September 27, 2019 at 1:40 am

“Yeah, I agree. Agriculture is very destructive, and we should localize. However:

“Is not the majority of this agriculture to feed livestock? And how could we feed pigs and chickens without it? They aren’t ruminants.

“Think about what would happen if we kept our meat consumption the same, but released the 70.4% of cows, 98.3% of pigs, and 99.9% of chickens in the US that live on factory farms to open grasslands? How could we possibly do this without bulldozing every last tree?”

Karin Lindquist
October 8, 2019 at 2:15 pm

“No, the majority of this agriculture is for human consumption, not to feed livestock. Livestock get the left-overs, the crop failures, and the stuff that didn’t grade to top-quality grade for use in every part of the term “human consumption” from being made into biofuel to vegetable oil to clothing. Animals also get the by-products that come from the conversion of these crops to various products for humans because the landfills would be overflowing if animals couldn’t take them, making that an environmental disaster in and of itself (as if landfills aren’t already an environmental disaster already), and because those animals turn those waste products into nutritional edible food. More here: https://www.ethicalomnivore.org/are-farm-animals-starving-the-planet-of-food-humans-cant-even-eat/

“Why would anyone be dumb enough to release a large number of animals that aren’t even adapted to live in such an environment? They’d die out very quickly, either from starvation because they don’t know how to forage on their own for food or they just can’t live in such an environment, or by predation. (It seems that you’ve never been on open grasslands before; trees on open grasslands are very rare. You only find trees in forests or savannahs.) The better solution to that problem you propose is via gradual phasing out of such systems and moving towards regenerative, well-managed pastured-based systems that produce and maintain the breeds and types of animals that are adapted to such a system. No “bulldozing every last tree” required. If you want a good example of what that kind of system looks like, look at operations like Polyface Farms and Brown Ranch in North Dakota. Great examples of stacked enterprises with a pasture-based system that is most certainly replicable, and FAR more efficient than any degenerative, monoculture CAFO operation.

“Think outside the box!!!! All isn’t as it appears.”

Also see:
What Livestock Eat
The Farming Truth Project

* * *

Carnivore Is Vegan:
Bad Vegan Logic: Accidental Deaths vs Intentional Deaths – Carnivore is Vegan
A Carnivore Diet is More Vegan than a Vegan Diet – Carnivore is Vegan
Vegans Use Slave Cows to Make Fertilizer
Dairy is 2000 X’s More Ethical Than Almond Milk
Stir-Fry Genocide: Mushrooms Are Not Vegan

Field Deaths in Plant Agriculture
by Bob Fischer and Andy Lamey

There’s no such thing as a green vegan
by Mary Harrington

There’s no such thing as vegan food
by Claire Taylor

Millennial veganism
by Joanna Blythman

But are you truly vegan?
by Matthew Evans

Australia’s vegan lie revealed: How plant-based diets still result in hundreds of thousands of animal deaths a year
by Lauren Ferri

Ordering the vegetarian meal? There’s more animal blood on your hands
by Mike Archer

The Least Harm Principle May Require that Humans Consume a Diet Containing Large Herbivores, Not a Vegan Diet
by Stephen L. Davis

The Least-Harm Fallacy of Veganism
by Karin Lindquist

Are Farm Animals Starving the Planet of Food… Humans Can’t Even Eat?
by Karin Lindquist

Want an ethical diet? It’s not as simple as going vegan, says farmer Matthew Evans
from ABC News

The Vegetarian Myth: Food, Justice and Sustainability
Chapter 1: Why This Book?
by Lierre Keith

The Hidden Cost of Veganism – Lierre Keith #143
from ReWild Yourself

Lierre Keith & The Agripocalypse
by Lawrence Rosenberg

Any ‘planetary diet’ must also work for the poorest and most vulnerable
by Andrew Salter

Eating Local Meat is Actually More Sustainable than Veganism
from Heartland Fresh Family Farm

Why vegetarianism will not save the world
by Ian MacKenzie

If you care about the planet, eat more beef
by Danielle Smith

Ruminants are more important to the world than you might have thought!
by Troy Downing

Report: Cut red-meat eating by 80 percent to save the planet?
by Anne Mullens and Bret Scher

Can vegetarians save the planet? Why campaigns to ban meat send the wrong message on climate change
by Erin Biba

EAT-Lancet report’s recommendations are at odds with sustainable food production
by Sustainable Food Trust

Report urging less meat in global diet ‘lacks agricultural understanding’
from FarmingUK

War on burgers continues with false environmental impact claims
by Amanda Radke

Testimony before the Committee on Agriculture, Nutrition and Forestry U.S. Senate
by Frank Mitloehner

Sorry, But Giving Up on Meat Is Not Going to Save The Planet
by Frank M. Mitloehner

Don’t Blame Cows For Climate Change
by Sylvia Wright

Cattle and methane: More complicated than first meets the (rib) eye
by Stephan Lewandowsky and Asa Wahlquist

Beef’s ‘Sustainability’ Involves More Than Greenhouse Gases
by Jesse Bussard

Is Agriculture Feeding the World or Destroying It? Dr. Frank Mitloehner Discusses Ag, Climate Change
from Farms.com

Environmental Hoofprint Matters — Frank Mitloehner, UC Davis
from Farm To Table Talk

Sustainable Dish Episode 83: The Truth About Greenhouse Gas Emissions in Livestock Production with Frank Mitloehner
with Diana Rodgers

UN admits flaw in report on meat and climate change
by Alastair Jamieson

Can Dietary Changes Limit Greenhouse Gas Emissions?
by Wyatt Bechtel

Scientist: Don’t blame cows for climate change
by Paul Armstrong

Climate change policy must distinguish (long-lived) carbon dioxide from (short-lived) methane–Oxford study
by Susan MacMillan

Alan Savory @ PV1 – The role of livestock in a new agriculture that can save city-based civilization
by Julia Winter

Effective Livestock Grazing And A Regenerative Future
by Allan Savory

Climate Change – Cause and Remedy
by Allan Savory

Climate Change Best Addressed Planting Trees, Or Regenerating Grasslands?
by Allan Savory

Fate Of City-Based Civilization In The Hands Of Farmers
by Allan Savory

How We Can Offer Hope For Our Grandchildren In A Floundering, Leaderless World
by Allan Savory

Hope For The Future – First Real Hope In Centuries.
by Allan Savory

Response To “Goodbye – And Good Riddance – To Livestock Farming”
by Daniela Ibarra-Howell

Why Homo Sapiens Are A Keystone Predator In Rewilding Projects
by Caroline Grindrod

Red meat bounds down the carbon neutral path
by Shan GoodwinShan Goodwin

Can cows cause more climate change than cars?
by Frédéric Leroy

Climate, Food, Facts
from Animal Agriculture Alliance

The Worst Mistake in the History of the Human Race
by Jared Diamond

Was agriculture the greatest blunder in human history?
by Darren Curnoe

Could Veganism Cause Extinctions?
by Patrice Ayme

It takes 21 litres of water to produce a small chocolate bar. How water-wise is your diet?
by Brad Ridoutt

Dietary Dictocrats of EAT-Lancet
Like water fasts, meat fasts are good for health.
Fasting, Calorie Restriction, and Ketosis
Ketogenic Diet and Neurocognitive Health
The Agricultural Mind

Last Edit and Revision: 8/19/22

Corporate Media Slowly Catching Up With Nutritional Studies

“The change in dietary advice to promote low-fat foods is perhaps the biggest mistake in modern medical history.”
 ~ Dr. Aseem Malhotra, cardiologist and expert on heart disease

“Fundamental problems were 2-fold. First, acceptance of weak associational epidemiological data as proof of causation. Second promotion of diet-heart hypothesis/lo fat diet to the public ahead of definitive proof of outcomes. Diet-heart hypothesis then became incontestable dogma.”
~ Tim Noakes, emeritus professor, scientist, and expert on low-carb diets

We’re All Guinea Pigs in a Failed Decades-Long Diet Experiment
by Markham Heid, Vice

The US Department of Agriculture, along with the agency that is now called Health and Human Services, first released a set of national dietary guidelines back in 1980. That 20-page booklet trained its focus primarily on three health villains: fat, saturated fat, and cholesterol.

Recently, research has come out strongly in support of dietary fat and cholesterol as benign, rather than harmful, additions to person’s diet. Saturated fat seems poised for a similar pardon.

“The science that these guidelines were based on was wrong,” Robert Lustig, a neuroendocrinologist at the University of California, San Francisco, told VICE. In particular, the idea that cutting fat from a person’s diet would offer some health benefit was never backed by hard evidence, Lustig said.

Just this week, some of Lustig’s colleagues at UCSF released an incendiary report revealing that in the 1960s, sugar industry lobbyists funded research that linked heart disease to fat and cholesterol while downplaying evidence that sugar was the real killer.

Nina Teicholz, a science journalist and author of the The Big Fat Surprise, said a lot of the early anti-fat push came from the American Heart Association (AHA), which based its anti-fat stance on the fact that fat is roughly twice as calorie-dense as protein and carbohydrates.

“[The AHA] had no clinical data to show that a low-fat diet alone would help with obesity or heart disease,” Teicholz told VICE. But because fat was high in calories, they adopted this anti-fat position, and the government followed their lead. Surely the 1960s research rigged by the Sugar Association, which was published in the prestigious New England Journal of Medicine, added to our collective fat fears.

By the 1990s, when Teicholz says the epidemiological data started piling up to show that a low-fat, high-carb diet did not help with weight loss or heart disease—calories be damned—much of the damage was already done. The US public was deep in what nutrition experts sometimes call the “Snackwell phenomenon”—a devotion to low-fat and low-calorie processed snack foods, which people pounded by the bagful because they believed them to be healthy.

“This advice [to avoid fat] allowed the food industry to go hog-wild promoting low-fat, carb-heavy packaged foods as ‘light’ or ‘healthy,’ and that’s been a disaster for public health,” Lustig said.

The stats back him up. Since the US government first published a set of national nutrition guidelines in 1980, rates of obesity and related diseases like diabetes have more than doubled. “Childhood diabetes was basically unheard of, and now it’s an epidemic,” Lustig said.

Overseas, national health authorities followed America’s lead on fat. The results have been similarly grim. Earlier this year, a UK nonprofit called the National Obesity Forum (NOF) published a blistering condemnation of its government’s diet and nutrition policies. […]

Teicholz said it’s hard to overstate the effect of national health authorities’ pro-carb, anti-fat stance. A whole generation of health professionals accepted—and passed on to their patients—the government’s guidance to avoid fat and cholesterol. Many still do.

“Both professional and institutional credibility are at stake,” she said when asked why more doctors and policymakers aren’t making noise about the harms caused by the government’s dietary guidance. She also mentioned food industry interests, the potential for “massive class-action lawsuits,” and the shame of copping to nearly a half-century of bad diet advice as deterrents for USDA and other health authorities when it comes to admitting they were wrong. […]

But one thing is clear: Dietary fat was never the boogeyman health authorities made it out to be.

“I think most of us would be 90 percent of the way to a really healthy diet if we just cut out processed foods,” UCSF’s Lustig said. “We wouldn’t need diet guidelines if we ate real food.”

We get what we pay for.

“Switzerland had the highest rate of return for empty wallets and Denmark for wallets with money in them. European countries overall, including Russia, got high marks for honesty.

“China had the lowest rate of return for empty wallets and Peru for wallets with money. I am disappointed that the United States is so far down on the list.”

I don’t feel disappointed about the US ranking. Or at least I don’t feel surprised. On many measures, the US often ranks around the middle. We are a middling country. Yes, above average, but middling. We lead the pack among the mediocre countries.

You see this with measures of culture of trust, democracy, freedom of press, health outcomes, education quality, etc. We tend to be far above the worst countries and well below the best (although specific US states often rank near the bottom in international comparisons). This has been the state of the nation for many decades now. It’s not exactly a new trend, this slipping down the international rankings.

But when older Americans were younger, the US was often the top ranking country in the world on numerous measures. Hence, the disappointment some Americans experience in remembering the country that once was. Sadly, that country hasn’t existed for a while now. We took American ‘greatness’ for granted and lost our sense of aspiration. Without the Soviet Union to compete against, Americans became morally and physically flabby.

Consider height. Americans used to be the tallest population on the planet. Now we share a ranking of 32nd with Israel, Hungary, Poland, Greece, Italy, French Polynesia, Grenada, and Tonga. Height is one of those indicators of the general health of a society and correlates with such things as inequality (and, by the way, high inequality in turn correlates with worse outcomes even for the wealthy, compared to the wealthy in low inequality countries). Trust also falls as inequality rises.

Still, we Americans on average are taller than 82 other countries. So, not bad. But still a major drop compared to the past. We are a declining society in many ways, specifically relative to other countries that are advancing. It’s a sign of the times. It’s also unsurprising that the United States is declining as a global superpower as well. A government’s power is built on the health of the population and the success of the society.

Anyone who dismisses the public good is naive. This is why the most effective social democracies that massively invest the in the public good now lead the world in nearly every ranking. The United States has chosen the opposite path, shifting wealth to the top for short term gains for a minority of the population, not to mention wasting our resources on military adventurism and imperial expansionism. We get what we pay for.

A ranking of countries by civic honesty

Gary Taubes On Biological Homeostasis

Gary Taubes wrote, “This is Curt Richter talking about his diabetic rat experiments published in 1941. It raises an obvious question: Could Richter’s rats have been smarter than the expert committees of the American Diabetes Association? I’m just saying….”

We found that when pancreatectomized rats with marked diabetes were offered a carbohydrate, a fat, and a protein in separate containers, in place of the mixed diet, they refused the carbohydrate and ate large amounts of fat and protein (7). As a result they lost their symptoms of diabetes, i.e., their blood sugar fell to its normal level, they gained weight, ate less food, and drank only normal amounts of water.

More from the same paper:

“When placed on a self-selection diet, the rats were no longer forced to take carbohydrate, except for the small amount contained in the yeast. They stopped eating sucrose and ate large amounts of olive oil; and as a consequence, the water was no longer drained from the tissues. The thirst disappeared, energy from the sugar was no longer lost; so the food intake decreased to its normal level again. With the return to the McCollum diet, the symptoms were reversed again.

“The dietary selections made by the diabetic rats closely agree with the diets determined empirically by clinicians from human diabetics in the preinsulin period. It was generally agreed that patients did much better on a high fat than on a low fat diet. Since the insulin increases the ability to utilize carbohydrate, the need for a high fat diet is no longer present. In preliminary experiments we have found that, when treated with insulin, our diabetic rats stop taking olive oil in such large amounts and eat sucrose.”

* * *

Increased Fat and Decreased Carbohydrate Appetite of Pancreatectomized Rats
by Curt P. Richter and Edward C. H. Schmidt, Jr.

BERNARD, who in 1859 first stated his concept of the constancy of the internal environment, described various physiological mechanisms, part responses of the organism, responses of individual organs, which contribute to the maintenance of this constancy. We have recently found that behavior mechanisms, responses of the total organism, may also serve to maintain the constancy of the internal environment. The existence of these behavior mechanisms became established in experiments in which certain physiological mechanisms had been excluded. Thus, after adrenalectomy had removed the chief physiological means of regulating sodium metabolism, it was found that the animal itself made an effort to maintain the sodium balance by seeking and ingesting large amounts of sodium chloride (1). Similarly, after parathyroidectomy had removed the physiological mechanisms for the maintenance of a constant calcium balance, the animals themselves made an effort to correct the calcium loss by ingesting large amounts of calcium solution (2).

Catching Up On Lost Time – The Ancestral Health Symposium, Food Reward, Palatability, Insulin Signaling and Carbohydrates… Part II(C)
by Gary Taubes

It was well known at the time (although it may have been forgotten since then), as I discussed in Good Calories, Bad Calories, that animals can be made to like one food more than another, and so eat more of the one than the other, by interventions that influenced their underlying physiologic/metabolic/hormonal states. Here’s how I illustrated this in GC,BC:

Throughout the first half of the twentieth century, a series of experimental observations, many of them from [Curt] Richter’s laboratory [at Johns Hopkins University], raised questions about what is meant by the concepts of hunger, thirst and palatability, and how they might reflect metabolic and physiological needs. For example, rats in which the adrenal glands are removed cannot retain salt and will die within two weeks on their usual diet from the consequences of salt depletion. If given a supply of salt in their cages, however, or given the choice of drinking salt water or pure water, they will chose to either eat or drink the salt and, by doing so, keep themselves alive indefinitely. These rats will develop a “taste” for salt that did not exist prior to the removal of their adrenal glands. Rats that have had their parathyroid glands removed will die within days of tetany, a disorder of calcium deficiency. If given the opportunity, however, they will drink a solution of calcium lactate rather than water—not the case with healthy rats—and will stay alive because of that choice. They will appear to like the calcium lactate more than water. And rats rendered diabetic voluntarily choose diets devoid of carbohydrates, consuming only protein and fat. “As a result,” Richter said, “they lost their symptoms of diabetes, i.e., their blood sugar fell to its normal level, they gained weight, ate less food and drank only normal amounts of water.

In short, change underlying physiologic/hormonal conditions and it will affect what an animal chooses to eat and so seems to like or find rewarding. The animal’s behavior and perceptions will change in response to a change in homeostasis – in the hormonal milieu of the cells in the body.

It’s quite possible that all those foods we seem to like, or even the ones we find rewarding but don’t particularly like, as Dr. Guyenet argues, and that subsequently cause obesity (not necessarily the same thing) are those foods that somehow satisfy an underlying metabolic and physiological demand. This in turn might induce our brains to register them as more palatable or rewarding, but the initial cause would be the effect in the periphery. The nutrient composition of the food, in this case, would be the key—what it’s doing in the body, not necessarily the brain.

Good Calories, Bad Calories
by Gary Taubes
pp. 331-332

This idea that energy expenditure increases to match consumption, and that the ability to do this differs among individuals, also serves to reverse the cause-and-effect relationship between weight and physical activity or inactivity. Lean people are more active than obese people, or they have, pound for pound, a higher expenditure of energy, *89 because a greater proportion of the energy they consume is made available to their cells and tissues for energy. By this conception, lean people become marathon runners because they have more energy to burn for physical activity; their cells have access to a greater proportion of the calories they consume to use for energy. Less goes to making fat. That’s why they’re lean. Running marathons, however, will not make fat people lean, even if they can get themselves to do it, because their bodies will adjust to the extra expenditure of energy, just as they would adjust to calorie-restricted diets.

Our propensity to alter our behavior in response to physiological needs is what the Johns Hopkins physiologist Curt Richter called, in a heralded 1942 lecture, “total self-regulatory functions.” Behavioral adaptation is one of the fundamental mechanisms by which animals and humans maintain homeostasis. Our responses to hunger and thirst are manifestations of this, replenishing calories or essential nutrients or fluids. Physical activity, as Richter suggested, is another example of this behavioral regulation, in response to an excess or dearth of calories. “We may regard the great physical activity of many normal individuals, the play activity of children, and perhaps even the excessive activity of many manic patients, as efforts to maintain a constant internal balance by expending excessive amounts of energy,” he explained. “On the other hand, the low level of activity seen in some apparently normal people, the almost total inactivity seen in depressed patients, again may be regarded as an effort to conserve enough energy to maintain a constant internal balance.”

p. 457-460

This is where physiological psychologists provided a viable alternative hypothesis to explain both hunger and weight regulation. In effect, they rediscovered the science of how fat metabolism is regulated, but did it from an entirely different perspective, and followed the implications through to the sensations of hunger and satiety. Their hypothesis explained the relative stability of body weight, which has always been one of the outstanding paradoxes in the study of weight regulation, and even why body weight would be expected to move upward with age, or even move upward on average in a population, as the obesity epidemic suggests has been the case lately. And this hypothesis has profound implications, both clinical and theoretical, yet few investigators in the field of human obesity are even aware that it exists.

This is yet another example of how the specialization of modern research can work against scientific progress. In this case, endocrinologists studying the role of hormones in obesity, and physiological psychologists studying eating behavior, worked with the same animal models and did similar experiments, yet they published in different journals, attended different conferences, and thus had little awareness of each other’s work and results. Perhaps more important, neither discipline had any influence on the community of physicians, nutritionists, and psychologists concerned with the medical problem of human obesity. When physiological psychologists published articles that were relevant to the clinical treatment of obesity, they would elicit so little attention, said UCLA’s Donald Novin, whose research suggested that the insulin response to carbohydrates was a driving force in both hunger and obesity, that it seemed as though they had simply tossed the articles into a “black hole.”

The discipline of physiological psychology was founded on Claude Bernard’s notion of the stability of the internal environment and Walter Cannon’s homeostasis. Its most famous practitioner was the Russian Ivan Pavlov, whose career began in the late nineteenth century. The underlying assumption of this research is that behavior is a fundamental mechanism through which we maintain homeostasis, and in some cases—energy balance in particular—it is the primary mechanism. From the mid-1920s through the 1940s, the central figure in the field was Curt Richter of Johns Hopkins. “In human beings and animals, the effort to maintain a constant internal environment or homeostasis constitutes one of the most universal and powerful of all behavior urges or drives,” Richter wrote.

Throughout the first half of the twentieth century, a series of experimental observations, many of them from Richter’s laboratory, raised questions about what is meant by the concepts of hunger, thirst, and palatability, and how they might reflect metabolic and physiological needs. For example, rats whose adrenal glands are removed cannot retain salt, and will die within two weeks on their usual diet, from the consequences of salt depletion. If given a supply of salt in their cages, however, or given the choice of drinking salt water or pure water, they will choose either to eat or to drink the salt and, by doing so, keep themselves alive indefinitely. These rats will develop a “taste” for salt that did not exist prior to the removal of their adrenal glands. Rats that have had their parathyroid glands *132 removed will die within days of tetany, a disorder of calcium deficiency. If given the opportunity, however, they will drink a solution of calcium lactate rather than water—not the case with healthy rats—and will stay alive because of that choice. They will appear to like the calcium lactate more than water. And rats rendered diabetic voluntarily choose diets devoid of carbohydrates, consuming only protein and fat. “As a result,” Richter said, “they lost their symptoms of diabetes, i.e., their blood sugar fell to its normal level, they gained weight, ate less food and drank only normal amounts of water.”

The question most relevant to weight regulation concerns the quantity of food consumed. Is it determined by some minimal caloric requirement, by how the food tastes, or by some other physical factor—like stomach capacity, as is still commonly believed? This was the question addressed in the 1940s by Richter and Edward Adolph of the University of Rochester, when they did the experiments we discussed earlier (see Chapter 18), feeding rats chow that had been diluted with water or clay, or infusing nutrients directly into their stomachs. Their conclusion was that eating behavior is fundamentally driven by calories and the energy requirements of the animal. “Rats will make every effort to maintain their daily caloric intake at a fixed level,” Richter wrote. Adolph’s statement of this conclusion still constitutes one of the single most important observations in a century of research on hunger and weight regulation: “Food acceptance and the urge to eat in rats are found to have relatively little to do with ‘a local condition of the gastro-intestinal canal,’ little to do with the ‘organs of taste,’ and very much to do with quantitative deficiencies of currently metabolized materials”—in other words, the relative presence of usable fuel in the bloodstream.

How The CIA Watched Over The Destruction Of Gary Webb

Having looked into American history over the years, the most disheartening insight is how closely tied the corporate media has been to the intelligence agencies. Released government documents and investigative journalism shows that it goes back to the early Cold War. And it has continued ever since. But it rarely gets talked about, especially in the corporate media. Funny how that works.

“On September 18, the agency released a trove of documents spanning three decades of secret government operations. Culled from the agency’s in-house journal, Studies in Intelligence, the materials include a previously unreleased six-page article titled “Managing a Nightmare: CIA Public Affairs and the Drug Conspiracy Story.” Looking back on the weeks immediately following the publication of “Dark Alliance,” the document offers a unique window into the CIA’s internal reaction to what it called “a genuine public relations crisis” while revealing just how little the agency ultimately had to do to swiftly extinguish the public outcry. Thanks in part to what author Nicholas Dujmovic, a CIA Directorate of Intelligence staffer at the time of publication, describes as “a ground base of already productive relations with journalists,” the CIA’s Public Affairs officers watched with relief as the largest newspapers in the country rescued the agency from disaster, and, in the process, destroyed the reputation of an aggressive, award-winning reporter.”

Traditional Hospitality

I like to watch the historical videos by James John Townsend. He is easygoing and always informative. The channel’s focus is on early America. Many of the videos are about food, but he also talks about how people lived. Some of my favorites are when he references the accounts described in travel journals.

In one video I just watched (the first below), Townsend talks about backwoods hospitality. It was expected that travelers would be welcome at almost any house or cabin, for food and shelter. An extra setting might be placed at the dinner table, just in case someone stopped by. Visits were common, whether from strangers or neighbors, friends, and family. Townsend mentions that a large part of their lives was filled with socializing. And it didn’t matter how poor the household. Almost anyone would be made to feel welcome, sometimes including local or traveling Native Americans.

Some of this custom may even have been learned or modeled after the lifestyle of Native Americans who were in the habit of coming and going as they pleased. There was often an open door policy on the frontier, a detail I’ve come across in reading history books. Even as villages formed, this friendly attitude was maintained. It was expected that what happened in your home was of relevance to the entire community. This nosiness could even be rather imposing, sometimes to the point of being oppressive. Neighbors might come right in, if they thought you were having affair. A bit too ‘friendly’ at times.

The idea of a home as a place of utmost privacy is a rather recent invention, as are locks on doors (when I was a kid, my family never locked the door except when we went on vacations). What is interesting is that some of the commenters point out that their own grandparents used to maintain similar practices of hospitality. For example, keeping a candle in the window was a way of signalling that visitors were welcome. Something similar was done with lamps to signal refuge for Jews escaping the Nazis. Likewise, candles and lamps were used as a signal on the Underground Railroad. In some parts of the world, especially in rural areas, welcoming strangers and those in need remains a living tradition.

That exemplifies how much the modern world has changed in industrialized societies. The kind of hospitality that would have existed for millennia, would have been the norm under most circumstances, that has faded from living memory for most of us Westerners. Something has been lost, a sense of community and common humanity, of interdependence and basic kindness, and we don’t usually even think about what has been lost, if we ever notice it’s missing.

This isn’t to romanticize the past, as Townsend also points out that the frontier was a dangerous place where not everyone was friendly. Still, it was a far friendlier place than the world we live in today. People had to be open and welcome in relation to others because survival depended upon it. Isolation wasn’t a choice. Yet the daily lived experience of community doesn’t exist for many people at this point, much less the attitude of hospitality toward strangers. We go about our lives as if we don’t really need anyone else.

I’ve known family members complain when others in the family stopped by unannounced. On the other hand, my mother remembers her family getting in the car on a weekly basis to make random visits to friends and family, and they were always welcome, often with a massive meal being prepared on the spot with no advanced notice. That was probably a carryover from my mother’s family having lived in what was the frontier not many generations before.

* * *

American People Keep Going Further Left

About a decade ago, I wrote a long piece analyzing all the polling data I could find across decades. It was obvious that the vast majority of Americans not only were quite far left but had been so for a long time and were going even further left over time. It wasn’t a new phenomenon. The leftward trend can be followed back into last century.

This shouldn’t be surprising when one looks at the politics of the early to mid-20th century. The politics were even more radical in my grandparents’ early life and it remained that way into my parents’ childhood. There was massive labor organizing with pitched battles. Communists were found everywhere, especially among the working class and minorities, including in the Deep South. The top tax rate was as high as it has ever been and the taxing the rich paid for numerous social programs, job programs, infrastructure rebuilding, etc. Everything from college education to housing was heavily subsidized.

Why don’t we know this? Because it has been written out of the history books used in both public and private schools — with the textbook industry being big business. Because the corporate media is the propaganda wing of plutocracy. And because the ruling elite in both parties have gone to immense effort to constantly push the Overton window to the right. It is only in our enforced ignorance through indoctrination from a young age that the American public is made to feel divided and impotent. The majority of Americans are told the public policies they support are too left-wing, too radical, too fringe. It is one of the most effective propaganda campaigns in world history.

Even now, the forces of corruption are pushing for lesser evilism one more time. Yet each time it pushes politics further right into ever greater evil. The corporate control of the government grows. And the main welfare system in our country is the socialism for the rich. We Americans haven’t yet fought back because we’ve been told we were part of a minority, that we don’t matter. But what if we Americans decided to fight for democracy once again? Then who would stop us? If they tried, it would be revolution. There is no time for democracy like the present. We should not accept anything less.

This is our country. This is our government. It’s time we take it back. That would make America great again, like it was in the radical era generations ago and in the revolutionary era upon which our country was founded. That is as American as it gets, the common people fighting against corrupt power. It’s an American tradition. Let’s honor that tradition. [Fill in your favorite quote from Thomas Jefferson writing about watering the tree of liberty, Abraham Lincoln speaking about justice and equality, Martin Luther King Jr. preaching about the arc of the moral universe, or whatever other great American figure you prefer.]

* * *

Surprise! The “Center” in US Politics Is Very Progressive
by Robert Reich, Common Dreams

On the economy,76 percent of Americans favor higher taxes on the super-rich, including over half of registered Republicans. Over 60 percent favor a wealth tax on fortunes of $50 million or more. Even Fox News polls confirm these trends.

What about health care? Well, 70 percent want Medicare for All, which most define as Medicare for anyone who wants it. Sixty percent of Republicans support allowing anyone under 65 to buy into Medicare.

Ninety-two percent want lower prescription drug pricesOver 70 percent think we should be able to buy drugs imported from Canada.

On family issues, more than 80 percent  of Americans want paid maternity leave. Seventy-nine percent of voters want more affordable child care, including 80 percent of Republicans.

Meanwhile, 60 percent of Americans support free college tuition for those who meet income requirements.

Sixty-two percent think climate change is man-made and needs addressing.

Eighty-four percent think money has too much influence in politics. In that poll, 77 percent support limits on campaign spending, and that includes 71 percent of Republicans.

AOC, Sanders, and Warren Are the Real Centrists Because They Speak for Most Americans
by Mehdi Hasan, The Inercept

The Green New Deal is extremely popular and has massive bipartisan support. A recent survey from the Yale Program on Climate Change Communication and George Mason University found that a whopping 81 percent of voters said they either “strongly support” (40 percent) or “somewhat support” (41 percent) the Green New Deal, including 64 percent of Republicans (and even 57 percent of conservative Republicans).

What else do Ocasio-Cortez, Warren, and Sanders have in common with each other — and with the voters? They want to soak the rich. Ocasio-Cortez suggested a 70 percent marginal tax rate on incomes above $10 million — condemned by “centrist” Schultz as “un-American” but backed by a majority (51 percent) of Americans. Warren proposed a 2 percent wealth tax on assets above $50 million — slammed by “moderate” Bloomberg as Venezuelan-style socialism, but supported by 61 percent of voters, including 51 percent of Republicans. (As my colleague Jon Schwarz has demonstrated, “Americans have never, in living memory, been averse to higher taxes on the rich.”)

How about health care? The vast majority (70 percent) of voters, including a majority (52 percent) of Republicans, support a single-payer universal health care system, or Medicare for All. Six in 10 say it is “the responsibility of the federal government” to ensure that all Americans have access to health care coverage.

Debt-free and tuition-free college? A clear majority (60 percent) of the public, including a significant minority (41 percent) of Republicans, support free college “for those who meet income levels.”

A higher minimum wage? According to Pew, almost 6 in 10 (58 percent) Americans support increasing the federal minimum wage from $7.25 an hour to (the Sanders-recommended) $15 an hour.

Gun control? About six out of 10 (61 percent) Americans back stricter laws on gun control, according to Gallup, “the highest percentage to favor tougher firearms laws in two or more decades.” Almost all Americans (94 percent) back universal background checks on all gun sales — including almost three-quarters of National Rifle Association members.

Abortion? Support for a legal right to abortion, according to a June 2018 poll by NBC News and the Wall Street Journal, is at an “all-time high.” Seven out of 10 Americans said they believed Roe v. Wade “should not be overturned,” including a majority (52 percent) of Republicans.

Legalizing marijuana? Two out of three Americans think marijuana should be made legal. According to a Gallup survey from October 2018, this marks “another new high in Gallup’s trend over nearly half a century.” And here’s the kicker: A majority (53 percent) of Republicans support legal marijuana too!

Mass incarceration? About nine out of 10 (91 percent) Americans say that the criminal justice system “has problems that need fixing.” About seven out of 10 (71 percent) say it is important “to reduce the prison population in America,” including a majority (52 percent) of Trump voters.

Immigration? “A record-high 75 percent of Americans,” including 65 percent of Republicans and Republican-leaning independents, told Gallup in 2018 that immigration is a “good thing for the U.S.” Six in 10 Americansoppose the construction of a wall on the southern border, while a massive 8 in 10 (81 percent) support a pathway to citizenship for undocumented immigrants living in the United States.

Ocasio-Cortez’s Socialism Can Work in the Midwest — With a Rebrand
by Eric Levitz, Intelligencer

Both Medicare for All and single-payer health care enjoy majority support in recent polling from the Kaiser Family Foundation. Data for Progress (DFP), a progressive think tank, used demographic information from Kaiser’s poll to estimate the level of support for Medicare for All in individual states. Its model suggests that, in a 2014 turnout environment — which is to say, one that assumes higher turnout for Republican constituencies — a majority of voters in Indiana, Michigan, Ohio, Iowa, Wisconsin, and Pennsylvania would all support a socialist takeover of the health-insurance industry (so long as you didn’t put the idea to them in those terms).

Now, it is true that support for Medicare for All is malleable when pollsters introduce counterarguments. But even if we stipulate that support for the policy is somewhat weaker than it appears, there is little doubt that any Democrat running on Medicare for All in a purple district will have a more mainstream position on health-care policy than the national Republican Party. Polls consistently find that an overwhelming majority of the American public — one that includes most Republican voters — supports higher federal spending on health care, and opposes cuts to Medicaid (just 12 percent of the public supports cutting that program). Every major GOP health-care plan introduced in the past decade runs counter to those preferences; the ones introduced in the last year would have slashed Medicaid spending by nearly $1 trillion.

The most radical economic policy on Ocasio-Cortez’s platform — a federal job guarantee — meanwhile, actually polls quite well in “flyover country.” In a survey commissioned by the Center for American Progress, a supermajority of voters agreed that “for anyone who is unemployed or underemployed, the government should guarantee them a job with a decent wage doing work that local communities need, such as rebuilding roads, bridges, and schools or working as teachers, home health-care aides, or child-care providers.”

Critically, support for this premise was almost exactly as strong among rural-dwelling demographic groups as it was among urban ones: According to DFP’s modeling, CAP’s proposal boasts roughly 69 percent support in urban zip codes, and 67 percent in rural ones.

There are a lot of reasonable, technocratic objections to the job guarantee as a policy. But polling suggests that there is majoritarian support for a massive public-jobs program of some kind — and that framing said program as “guaranteed jobs” might be politically effective.

Other items on Ocasio-Cortez’s platform poll similarly well. A bipartisan majority of voters have espoused support for “breaking up the big banks” in recent years, while nearly 70 percent of Americans want the government to take “aggressive action” on climate change, according to Reuters/Ipsos.

“Housing as a human right” might sound radical, but in substance, it’s anything but: The Department of Housing and Urban Development believesit could end homelessness with an additional $20 billion a year in funding; other experts put that price tag even lower. I don’t think the question, “Should the government raise taxes for the rich by $20 billion, if doing so would end all homelessness in the U.S.?” has been polled, but I would be surprised if it didn’t poll well, even in the Midwest.

Similarly, on the question of immigration enforcement, Ocasio-Cortez’s position is likely more palatable when rendered in concrete terms than in abstract ones. Many white Midwesterners might recoil at phrases like “abolish ICE” or “open borders.” But if one asks the question, “Should the government concentrate its immigration-enforcement resources on combating violent criminals and gang activity, instead of going after law-abiding day laborers?” I suspect you’d find more support for the democratic socialist point of view.

The palatability of Ocasio-Cortez’s policy platform reflects two important realities: Actually existing “democratic socialism” — which is to say, the brand championed by its most prominent proponents in elected office — is almost indistinguishable from left-liberalism; and left-liberal policies are already quite popular in the United States.

If all Americans voted for the party whose positions on economic policy best matched their own stated preferences, then the Republican Party would not be competitive in national elections. The GOP’s strength derives entirely from the considerable appeal of white identity politics with constituencies that happen to wield disproportionate power over our political system.