The Ground of Our Being Touches Us

“The foot feels the foot when the foot feels the ground.”

That quote by Ernest Wood is often misattributed to the Buddha. And it does express a Buddhist-like thought. Take the notion of co-touching from the Samyukta Nikaya:

“Who touches not is not touched. Touching he is touched.”

That being “touched by touch” (Thag vs.783) is a part of dependent co-arising and the bundle theory of mind, both central tenets of Buddhism (Robert Alvarado, The Foot Feels the Foot When It Feels the Ground). The separate, autonomous, and self-willed egoic-consciousness is not fundamentally real.

Everything that exists does so not as a ‘thing’ but as a feeling, a process, a movement, and a relationship. The self or any part of the self (e.g., the foot) emerges in awareness through interaction with the experienced world and perceived other. This is the sensory and social world as the ground of our being.

One theory in social science suggests that humans develop a theory of mind about others first before internalizing it as a self concept. So, the self is the introjected other. It’s similar to Lev Vygotsky’s private speech or self-talk, as a precursor to inner speech, that is the child’s imitation of adults talking to the child (Speaking Is Hearing).

We all begin life by first talking to ourselves as an other. And we carry this into adulthood. When you talk to yourself, who are you talking to and who is doing the talking? The other forever defines us, as if if the ground were to leave a print on our foot.

To the mind, the developing mind most of all, the world around us provides affordances (James J. Gibson) for actions and other behavior (The Embodied Spider; & “…just order themselves.”). These are known more for what they make possible and allow than for what they supposedly are, their socially constructed thingness.

We never know the world except as our experience of the world, since there is no self to know or experience without the world. The world is the primal self. The self is in and of the world. There is nowhere else to be.

That is why there is no foot in and of itself within awareness, no Platonic ideal of a ‘foot’, not without the ground that affords the foot the capacity to express it’s instinctual nature of footness. If one were to be so cruel as to completely bind an infant’s foot so that it could never move, it would shrivel up into crippled paralysis with little if any sensation.

The very sense of self would be constrained and the lesser for it. To emphasize this point, consider that the infant that is not touched at all simply dies. A foot is the touch and movement of the foot in relation to the ground. We aren’t separate from the world, not outside it, but immersed in it and an extension of it.

We need touch. We are touch. We touch by being touched.

* * *

There are two ways you can demonstrate such truths to yourself. The first method is to practice meditation and mindfulness for years, preferably under the guidance of a religious or spiritual teacher, guru, etc. That is arguably worth the effort. But it does require commitment, effort, and sacrifice; and, admittedly, most of us feel too lazy to try.

If you just want to get a small taste of it, sit or stand completely still while softly and unblinkingly gazing at an unchanging visual field (e.g., an indoor wall). Give it a minute or so and your entire vision will go blank, not even go black but simply to disappear as an experience. Sensory perception is dependent on movement and change, either in the environment or from our bodies.

There is another self-experiment one can do. The above quote about the foot feeling the ground can be taken literally. Take your shoes off and actually feel the ground. Walk around your lawn. Maybe even go for a jog, if you have somewhere nice and safe for your tender feet. Try that for a few weeks or a few months. Being barefoot is the normal state of humanity, quite likely what Ernest Wood was doing when he had the above thought.

Feel what it’s like to not have your feet bound and numbed in tight shoes, thick soles, and synthetic materials. Feel what it’s like to be electromagnetically grounded, physically connected, and sensorily in relationship to the earth. The affordance of the earth is far different than the affordance (or rather unaffordance) of modern footwear.

Each will elicit different ways of inhabiting one’s body, moving in the world, and perceiving reality; maybe even altering one’s very sense of identity. Then contemplate all the thousands of other ways we are disconnected, distracted, and numbed from direct sensory experience of the natural world. That is how the isolated self is socially constructed, supported, and maintained. These rigid boundaries of self are the defensive walls of egoic consciousness.

* * *

It’s interesting to consider the fact that Buddhism arose in an environment where most people in the past walked around barefoot, since is its a warmer clime and industrialization took hold much more slowly. That is true of other religious traditions, like Hinduism and animism, that question or refute or simply never acknowledge the ego-self. Would the bundle theory of mind even occur in a society where everyone had worn shoes for centuries or millennia?

Even Western philosophers like David Hume who have written about the bundle theory of mind, as some argue, likely learned of it from Christian missionaries having returned from the East. These were ideas that apparently never originated in the West or, if they did, it was so long ago they were forgotten; maybe back when Europe was still tribal and animistic, back when footwear would’ve been more akin to a moccasin that doesn’t desensitize the foot.

Shoes are only needed in colder regions, such as  Europe and North America; and only needed on rough ground, such as plowed fields. Maybe that is a causal or contributing factor to such a strong tradition of egoic individualism developing there. The European and American traditions of Christianity fear and disparage connection to nature. Maybe a long history of wearing shoes has predisposed people to that experience and worldview, identity and way of being.

That reminds one of the WEIRD cultural bias (Western, Educated, Industrialized, Rich, Democratic) that correlates to a unique profile of personality traits and social behaviors. One researcher and writer on the topic, Joseph Henrich (The WEIRDest People in the World), argues for various causes for this development in the modern West, from Catholic marriage laws to literacy rates. But maybe footwear should be added to this list.

Written laws and written texts are examples of media that are made possible by media technology (e.g., bound books and moveable type printing presses). Footwear likewise mediates our sensory experience of reality and hence footwear could be considered a media technology. It shapes not only the foot but also the foot-mind-eye axis, as a core dynamic function within the body-mind-world axis — proprioception and perception.

Besides Joseph Henrich, numerous others have theorized about mediated reality: Marshall McLuhan, E. R. Dodds, Bruno Snell, Julian Jaynes, Eric Havelock, etc. But the main focus has been on language, specifically written language. That is important in a literary culture with high rates of literacy. Nonetheless, footwear have been more central and earlier introduced to Western culture. And it’s the modern thick, bulky, and constraining shoe that has become so common over the past few centuries, in relation to our altering the environment so that such protective footwear is needed.

* * *

This post is another ripple in a river of thought. We’ve been slowly building upon a theory about the physical aspects of social constructivism: the infrastructure and apparatuses and systems that shape and confine us, the lifeway patterns and pathway dependencies that predetermine and preclude our individual and collective behaviors, the ideological interpellation that hails us with voices of authority and authorization, the metonymic and metaphorical framings that came with changes in media technology.

This involves agricultural system, food laws, and dietary ideology; land reform as moral reform and substance control as social control. One can show an increasing shift, across recent millennia (particularly starting in the axial Age but speeding up in modernity), from non-addictive psychedelics and evolutionarily-consistent foods to addictive sedatives, stimulants, and high-carb foods (alcohol, opium, cocaine, tea, coffee, sugar cane, grains, etc).

All of these things, it can be argued, rigidified psychological and social boundaries. Yet no single factor alone would likely have made possible and probable the emergence of the post-bicameral, post-axial, and post-traditional hyper-individualistic Jaynesian egoic consciousness of the body-mind as isolated-subject and container-object. It was also the continuing development along each technological line that forced the transformation.

Footwear has been around for millennia, whereas more recent is the invention of shoes that are highly-restrictive, thick-soled, and synthetically non-conductive.  Similarly, language existed for millennia prior to writing, bound books, printing presses, e-books, email, texting, etc. Even written language operated far back in the archaic world but only in a minimalistic fashion, primarily as bureaucratic accounting, before it ever developed into literacy as we know it. There are still other tools of identity formation like transitional objects (teddy bears, pacifiers, etc) that were or are not common in premodern or non-WEIRD societies.

The development and accrual of changes formed slowly, if the results sometimes only fully erupted following a triggering point (e.g., Bronze Age collapse). Those eruptions allowed for a destabilizing or destruction of some former pathway dependencies, in order to lay down new foundations, but always using the material of what came before. Still, some pathway dependencies were so entrenched they remained; if reshaped, restructured, and repurposed (e.g., written text).

This area of study also overlaps with with issues of physical health, mental health, and public health. Specifically, there is an interesting history of how dietary systems and food laws (e.g., Christianized Galenic humoralism) were used to enforce identity, culture, and social order. There has been an ongoing change in what is eaten that during modernity has led to disease epidemics, health crises, and moral panics. The relationship between diet and identity might’ve been more well appreciated in the past.

These contemplations are also mixed up with the study of archaic and ancient societies, along with the anthropological literature on animistic tribes. This particularly focuses on the transitional period from the late Bronze Age and it’s collapse to the Axial Age and the resultant post-Axial world. During the Bronze Age, there was what Julian Jaynes called the bicameral mind, a type of bundled mind, with voice-hearing traditions. Growth of size and complexity of the Mediterranean empires in the late Bronze Age is what caused their collapse, as overwhelmed by decades of natural disasters, refugee crises, and marauders.

That is what cleared the board to make way for the Axial Age, although the changes had already begun in the Bronze Age (e.g., written laws). One of the changes that didn’t happen until the Axial Age was the systematization of agriculture where former weedy farm fields became the focus of more intensive and controlled farming. This increased dependable surplus yields and so provided more agricultural foods in the diet, but it also meant better pest control, including eliminating most of the ergot that would take over unmanaged fields.

Ergot, as a psychedelic, was inevitably consumed on a more regular basis prior to this ancient agricultural reform, often unintentionally but sometimes on purpose as part of rituals. Interestingly, coinciding with lessening it in the food supply was also the appearance of cultivars of addictive substances like opium, sugar cane, etc. In Europe, there was a ‘regression’ after the fall of the Roman Empire. Some knowledge and practice of agricultural management was forgotten, as fields returned to being weedy again. Following that was what appears to have been regular mass ergot intoxications and sometimes deadly dancing manias, what is called ergotism or St. Anthony’s Fire.

Later agricultural reforms eliminated ergot again. Yet other psychedelics persisted in European culture. Medieval church imagery often portrays fly agaric ‘magic’ mushrooms. Such imagery continued into early modernity, as seen in Christmas cards.

* * *

Related to dietary practices and the food system, there is another connection that could be made. There were also agricultural differences between East and West. One study sought to discern agricultural differences as linked to socio-cultural and socio-cognitive differences. Yes, it’s true that Westerners grow more wheat and Easterners more rice; and it’s true that these agricultural systems require different relational patterns and practices. Wheat farming can be done by a single man with a plow, but rice farming requires numerous people working together and is more labor intensive in requiring twice as many hours of work. Furthermore, rice-growing communities have to collectively build and cooperatively maintain infrastructure (dikes and canals) for water management and irrigation.

Some have speculated that this constructs, encourages, and enforces divergent cultural identities and ways of thinking. This might be what underlies the stereotypical contrast between Eastern and Western thought. The former focuses more holistically, interdependently, and concretely on environment, background, and relationships; and the latter focuses more analytically, atomistically, and abstractly on the individual, foreground, and action. Also, descendants of rice-growers are more loyal to friends and family; while descendants of wheat-growers have more successful patents for new inventions. The thing is we don’t need to stop there with a simple hypothesis of causal link, since we can control some of the potential confounders by making a comparison within a single country, though still other confounders remained uncontrolled.

Wheat and other cereal crops (e.g., millet) are also grown in parts of Asia, specifically in northern China; while southern Chinese are rice farmers. Multiple studies have been done in comparing and contrasting the personalities, cultures, social practices, etc of these two agricultural populations. Even in the East, wheat farmers are more individualistic and rice farmers more communal. But also the same divide is seen in thinking styles with the Asian wheat farmers, as with European wheat farmers, in being more likely to use linear thought in focusing on isolated objects and subjects in the foreground while not noticing much about the overall context.

To return to the topic at hand, it might be useful to look at other aspects of what differentiates the two. Are Chinese wheat farmers more likely than Chinese rice farmers to wear shoes or boots more often and to wear shoes or boots with thicker soles and narrow enclosed toe boxes, as opposed to wearing thin, open-toed sandals or going barefoot? One suspects that would be the case.

It wouldn’t only be that the dirt clods of wheat fields are harder on the feet than the soft mud of rice patties. The colder climate of northern China would require wearing thicker shoes for a large part of the year for protection against coldness, discomfort, and frostbite. Interestingly, a similar pattern is seen in Europe as well with the concentration of wheat farming countries in the north with their long history of Protestant-style individualism, as contrasted to southern European Catholicism and communalism. A better and more comparable example is the United States.

Wheat-farming, of course, has been practiced in the northern states for a long time; but also rice-farming has been common in a large swath of the Deep South, what is called the Rice Belt. Similar to southern China, “even when the correlations were examined only within the Deep and Peripheral South, the correlations of collectivism with cotton and rice production remained strong” (Dov Cohen, Patterns of Individualism and Collectivism Across the United States). That is strong supporting evidence. To link it back to the main topic, for most of Southern history in the US, going barefoot was far more common. That has contributed to greater hookworm rates, as this parasite tends to enter through the sole of the foot from infested soil. Also, note that wheat-farming and industrialization has been concentrated in the northern states, as was the case in northern Europe. Industrialization, by the way, is the letter ‘I’ in the WEIRD acronym; and maybe the letter ‘W’ for Western could equally represent wheat-farming.

When we think of farming cultures and practices as affecting identity, personality, and mentality, we rarely think about what people are physically wearing as being causally significant or even relevant. But consider that the person in a colder climate is not only more likely to have restrictive, binding, and thick footwear but also restrictive, binding, and thick clothing and outdoor gear. Maybe it’s no mere coincidence that, for example, many animistic tribes with their extremes of a bundled mind tend to go barefoot entirely and often to barely wear any clothing at all, other than maybe a breech cloth (e.g., Piraha). Even among farming societies, some where heavy, cumbersome clothing and others lighter and looser (e.g., the sarong common in the East).

It is interesting how much our society, particularly among intellectuals and scholars (i.e., the literary elite), is obsessed with language and, most of all, written language. We have the most literary culture that has ever existed since language was invented. And it’s precisely populations with high rates of literacy that are the most WEIRD, to the extent that brain scans shows it alters the development of brain structure and neurocognition (see Joseph Henrich).

As such, we Weirdos see everything through language and text (e.g., this post here), and so that is the primary lens through which we understand the world and humanity. There is an obsession with the study of language, from text to new media: philology, postmodernism, linguistic relativity, metaphor theory, etc. So, language and the media of language gets disproportionate credit and blame for much of the changes, problems, and advancements in society. The differences between the cultures and mentalities of East and West are often placed within a linguistic frame.

But even when language isn’t the focus, what we emphasize is often something else that is equally less tangible. When farming is studied, what researchers tend to isolate out as causal are how people relate and act within different agricultural systems, the kind of thing that is harder to measure objectively. Oddly, it almost never occurs to them to think about the most basic and concrete factors like what is grown and eaten in affecting the body-mind, despite the vast knowledge we’ve accrued in nutrition studies. Diets determine nutritional profiles and biological functioning, one of the most powerful affects on neurocognitive development.

Or consider how one of the most transformative changes in all of human existence was the agricultural revolution in general, no matter if wheat or rice or whatever else. It increased size and concentration of human populations, increased size and concentration of domesticated animal populations, and increased contact between humans and animals. It also increased pathogen exposure and parasite load, both of which research shows to raise the measures of social conservatism and authoritarianism, insularity and collectivism, which are not only correlated to social behaviors but also altered personality traits (low openness, high conscientiousness, etc) and brain structure (e.g., larger amygdala).

Pathogen and parasite levels do follow a regional pattern as well, more near the Equator and less the further away; although this can’t entirely explain the agricultural differences. Thomas “Talhelm’s study found that Chinese students who lived just south or just north of the rice-wheat divide were as different from each other as students from the far south and the far north. And he noted rice-producing Japan scores uniformly high on the collectivist scale, even though the country is cooler and wealthier than most of China” (Bryan Walsh, In China, Personality Could Come Down to Rice Versus Wheat). Even rice-growing islands in the north fall in line with the southern pattern of behavior and personality (X. Dong, T. Talhelm, & X. Ren, Teens in Rice County Are More Interdependent and Think More Holistically Than Nearby Wheat County). Do people in all Chinese rice-growing populations, whether south or north, have similar footwear?

On the other hand, Walsh writes, “The rice theory isn’t foolproof. It’s almost certain that none of the young Chinese college students participating in Talhelm’s study have any direct experience with wheat or rice farming, which raises the question of how these psychological values are transmitted.” Maybe it’s not entirely about who is growing which crop and how it is grown, as part of a socio-cultural order. Instead, it’s possible that more important is who is eating which crop. Chinese, in general, are less individualistic than Westerners, no matter which region they live in. The simplest explanation could be that, as part of a national food system, all Chinese on average eat more rice and less wheat than Westerners. It might be about nutritional differences in each crop (e.g., gluten).

Then again, it could be something else not directly related to the crop or diet. Different kinds of farming in different environments and climates will incur different public health conditions and hence different physical health of individuals. The contrast between rice and wheat farmers goes far beyond merely how people socially organize within an agricultural system or what they eat within a food system. After all, what kind of footwear one wears or does not wear depends entirely on the agricultural system and all that is involved with it, such as infrastructure, housing, etc. Enclosed footwear, for example, could be protective against parasites and pathogens when the ground is covered in human and animal feces. So, it would depend also on the animal side of the farming equation.

None of the studies that we’ve seen, however, have ever been concerned with or curious about these kinds of confounding factors. This is a vast cultural blind spot. We forget that we are embodied minds that are co-extensive with the physical world around us, not to mention bundled minds in a bundled world. It rarely, if ever, occurs to us to think about something so simple as what we are wearing. Yet footwear, like a thousand other unrecognized factors, potentially has immense impact on us.

Spend some time observing people with modern synthetic shoes. Most of them walk stiffly, awkwardly, and often flat-footed; not to mention demonstrating an unbalanced and ungrounded way of physically holding themselves. Obviously, many people aren’t comfortable in their own bodies, absolute the opposite of barefoot indigenous people. Maybe simple things like footwear affect us far more profoundly than we are aware of, to the point of affecting our ability to grasp the Buddhist truth that to touch is to be touched.

Also, the combination of other unacknowledged factors could create a greater influence than any single factor alone. It would be a cumulative effect over a lifetime. So, yes, shoes will stunt and distort the bone and soft tissue development of the feet. There would be a lack of musculature and mobility that would make one prone to injury, not only in the feet but also from the stress caused in how it would throw off the movement of other joints, particularly the knees and hips. The feet are the foundation of the body, the contact and connecting point between body and world, and hence the mediating point in the sense of the embodied and enworlded self.

Diet and nutrition could exacerbate problems related to the feet and everything influenced by it. Dr. Weston A. Price, for example, observed that populations with deficiencies in fat-soluble vitamins had worse bone development: thin bones, asymmetrical features, narrow shoulders, narrow chests, caved-in chests, narrow jaws, crowded teeth, etc. But this also affected the bones in the feet: pigeon toes, flat feet, etc. This probably would make the feet narrower like all the rest.

Combine that with further squishing the feet even narrower into confining shoes. Most modern people are being crippled from a young age. We modern Westerners feel morally superior than the premodern Chinese who bound the feet of girls, and yet we also bind the feet of not only girls but also boys and then continue to do so into adulthood. No doubt, the premodern Chinese bound girls feet precisely because it alters behavior and is used for the social constructivism of particular personality traits and social roles, maybe not unlike hobbling a horse to make it more calm and controllable.

What might our own practice of foot-binding have on the entire population? When we personally observe or scientifically study our fellow humans, we tend to look to their faces, heads, arms, and upper bodies; in terms of their gaze, expression, tone of voice, gestures, etc. That is what we think most as defining who a person is, whereas their lower body of hips, legs, and feet is secondary as almost a mere extension of the upper body. We might be wiser to spend more time looking down to the literal ground of our being.

Wheat versus Rice:

Enclosure of the Mind

“[T]he chief matter . . . being now not the fruits of the earth, and the beasts that subsist on it, but the earth itself; as that which takes in, and carries with it all the rest.
~ John Locke, Two Treatises of Government, 1689

“As long as we keep ourselves busy tilling the earth, there is no fear of any of us becoming wild.”
~ Michel Guillaume Jean de Crevecoeur, Letters From an American Farmer, 1782

“Inclosure came and trampled on the grave
Of labour’s rights and left the poor a slave …
And birds and trees and flowers without a name
I sighed when lawless law’s enclosure came.”
~ John Clare, The Mores, 1820

“Strangely enough they have a mind to till the soil and the love of possession is a disease with them. These people have made many rules that the rich may break but the poor may not. They take their tithes from the poor and weak to support the rich and those who rule.
“They claim this mother of ours, the earth, for their own and fence their neighbors away; they deface her with their buildings and their refuse. The nation is like a spring freshet that overruns its banks and destroys all that are in its path.”
~ Sitting Bull, Speech at the Powder River Council, 1877

The time has arrived when we should definitely make up our minds to recognize the Indian as an individual and not as a member of a tribe. The General Allotment Act is a mighty pulverizing engine to break up the tribal mass. It acts directly upon the family and the individual.
~ Teddy Roosevelt, Address to Congress about Dawes Act, 1901

The early modern period saw the legal push for land enclosure, privatization, and consolidation. It became a powerful force in the 18th century, which destroyed the ancien regime, destabilized the social order, and precipitated revolts and eventually revolution. This was central to Enlightenment thought in the creation or exacerbation of Jaynesian consciousness, post-bicameral nostalgia, Platonic/Cartesian anxiety, atomistic individualism, capitalist realism, social Darwinism, and WEIRD culture. In a short period of time, land reform, agricultural improvements, and technological advancements led to the first dependable grain surpluses, particularly the increase of wheat production, the sudden availability and affordability of white flour, and the industrial development of the high-carb standard American diet (SAD). Also, with colonial trade, tobacco, tea and sugar replaced local smoking herbs and herb-infused beer. Heading into the 19th century and continuing into the next, all of this combined might have contributed to the disappearance of the fairies and the emergence of a crisis of identity, followed by moral panic along with the rise of widespread mental illness and drug addiction and other diseases of civilization, which continues to worsen, not to mention increasing rates of such things as autism — all of it central to what one could call the agricultural mind, exacerbated by mass urbanization, industrialization, and big ag.

This is an ongoing line of speculation, but the land enclosure angle is somewhat new. We’ve previously written about the enclosure movement, privatization and the loss of the Commons, as it obviously is one of the most central changes in recent history, arguably key to understanding nearly all other changes in modernity. It coincided not only with capitalism, corporatism, and industrialization but also colonial imperialism and its vast trade network. There really is no way of comphrehending what all the fuss was about, from the English Peasants’ Revolt to the English Civil War to the American Revolution, without knowing how feudalism was forcefully and violently dismantled not by the peasants and serfs but by aristocrats and monarchs. Other economic practices and systems were seen as more profitable or otherwse attractive. Eliminating the feudal system of parishes and commons, for example, eliminated all of the inconvenient social obligations and traditional roles of noblesse oblige that constrained power according to the authorizng precedence of living tradition and custom. Part of the complaint of some aristocrats, including the more radical-minded like Thomas Jefferson, was that the ancien regime was perceived as oppressively confining to everyone, including the aristocracy. But to destroy that old order meant creating something radically new in its place, which would involve new subjectivities, identities, and roles.

That was the self-enforced task set before the Enlightenment thinkers and later reformers. Individuality and independence was praised, but some at the time admitted to or hinted at the fact that these were not natural law and human birthright. They had to be artificially created. First off, let’s set down a distinction: “Like social constructionism, social constructivism states that people work together to construct artifacts. While social constructionism focuses on the artifacts that are created through the social interactions of a group, social constructivism focuses on an individual’s learning that takes place because of his or her interactions in a group” (Wikipedia). Another way of thinking about this was described by Richard M. Doyle: “The philosopher Louis Althusser used the language of “interpellation” to describe the function of ideology and its purchase on an individual subject to it, and he treats interpellation as precisely such a “calling out.” Rather than a vague overall system involving the repression of content or the production of illusion, ideology for Althusser functions through its ability to become an “interior” rhetorical force that is the very stuff of identity, at least any identity subject to being “hailed” by any authority it finds itself response-able to” (Darwin’s Pharmacy). A social artifact, once socially constructed, offers an affordance that unconsciously enforces the authorization of social constructivism through the interpellation of calling out a particular behavioral subjectivity we become identified with in responding. So, to give a concrete example, we are enacting the propertied self when, after seeing a no trespassing sign, we don’t cross a fence. We’ve been hailed by the authorization of an implicit ideological realism that makes a claim over us, constraining not only our behavior but more importantly our identity. But that response has to be taught, modeled, and internalized — fences and walls, like roads and sidewalks, become the infrastructure emblazoned upon the mind.

This civilizing process was more starkly apparent at the beginning of modernity because so much of what we take for granted, within this dominant ideological realism, did not yet exist. To establish private landholdings was necessary to form the structure for the propertied self, far beyond mere self-ownership in not being a slave (i.e., liberty). The danger, to the emerging capitalist class, was that there were competing structures of identity with the communal self and bundled mind that continued to assert itself. Consider the elite intellectual William Godwin (1756–1836) who saw “associations as constructing their members’ subjectivities, not merely directing their energies incorrectly,” writes Robert Anderson. “In this sense, then, associations are analagous to what Louis Althusser calls Ideological State Apparatuses which provide material rituals and practice, which subjects recognize themselves. Unlike Althusser’s state apparatuses, which hail subjects as individuals, political associations, in Godwin’s view, construct a “common mass” subject, in which subjects are undifferentiated one from another. Since, as Sayer and Corrigan argue, the construction of subjectivity is central to the success of a nation-state, this function of political associations is no trivial matter” (“Ruinous Mixture”: Godwin, Enclosure and the Associated Self). Those like Godwin thought collectivities were a bad thing, since individualistic propertied elites such as himself represented the ideal in his utopian ideology. During this same era, George Washington warned of the threat of politcal parties and one wonders if he had similar worries on his mind, considering his treatment of the collective action of Shays’ Rebellion. Robert Anderson explains what this entails:

“The Enclosure Movement, which yokes the realms of the subject and of property, gives some historical grounding for Julia Kristeva’s theory of the abject, which describes the psychic imperatives that drive the subject to distinguish itself from a “common mass.” This force, I am suggesting, determines the movement towards the enclosure of both the commons and the “self.” It concerns an anxiety about the “clean and proper” (“le propre”) boundaries of the self (“le propre”). The subject is constructed through a process of exclusion and boudnary-defense which involves an attempt ot ensure the singularity and integrity of the self within its boundaries, and an attempt to protect those boundaries of the self—not merely the self, but the boundaries themselves. Abjection names the proces of “exclusion” through wich “‘I’ expell myself” from indifferentiation and wildness/animality. The abject, then, threatens to “engulf” the subject because it is a reminder of what it must push aside in order to live. We can se this at work in Young’s claim that enclosure transformed the country from “boundless wilds and uncultivated wastes” into “well-peopled” “inclosures . . . cultivated in a most husband-like manner . . . and yielding an hundred times times the produce.” It is to guard against the “ruinous Effects of a Mixture of opposite Interests” and the “untidiness” of common and use-rights, that enclosure takes place. It cleans and distinguishes le propre—the self, the property—from the “improper.” In his chapter on “The Principles of Property,” Godwin argues that property performs this very function. In spite of the great injustices it causes, the right to property is so “sacred” that no exertion or sacrifice to protect it can be too great (2.440-50). It creates an “essential” “sphere” which protects man from outside intervention, thereby freeing up a space for the operation of “private judgment,” which is necessary for the improvement of man” (2.433). This improvement is threatened if the self is not protected from being “resolve[d] . . . into one common mass” (1.289). Abjection, then, is the psychological engine for improvement.

“The history of enclosure bears out Kristeva’s argument that abjection is ultimately a reliance on the law, which “shapes the body into a territory protected by the “differentiations of proper-clean and improper-dirty” (72). Thompson reveals the extent to which “reasons of improvement” had acquired the status of legal terminology, in particular as a justification for the enclosure of the commons (“Custom” 134-60 passim). A. W. B. Simpson’s A History of Land Law articulates the historical change from “communal rights” of the commons to individual rights, which both made possible and were produced by the enclosure: “[t]he tenurial system converted the villagers [who used the land as common village property] into tenants, and the theory of the law placed the freehold of most of the lands of the manor in the lord. . . . Thus a theory of individual ownership supplants earlier more egalitarian notions” of property. And with this change, common rights came to be seen as having originated “in the grant of the lord,” rather than as a “customary rights associated with the communal system of agriculture practiced in primitive village communities.” In cases where enclosure was contested, however, court rulings often reversed the implicit chronology of “improvement” to suggest that enclosure was the natural state of property rather than an innovation.”

This demonstrates how the conservative authority of hierarchical individualism usurped the role of traditional authority of the ancestral commons, the latter a vestige of archaic authorization of the bicameral mind. The historical revisionism of the conservative project of individualistic privatization hints at the underlying reactionary mind that fuels the radical transformation through the invented tradition of ideological realism dressed up in robes from the wardrobe of moral imagination, proclaiming it has always been this way and putting a narratized spell of historical amnesia upon Jaynesian consciousness — and so individuality erases the evidence of its own origins, like the scaffolding removed from a cathedral after being built by thousands of laboerers over centuries. The threat of collective action of worker associations, labor unions, etc is not that they represent something radically and entirely new but that they are old impulses/habits carried over from the lingering habitus of the ancien regime and traditional communities that keep challenging the radical modernity of reactionary conservatism. The conservative counterrevolution is itself revolutionary, as it is also authoritarian. As noted many times before, the ideology of independence of hyper-individualism is inseparable from dependence of authoritarianism (as violently oppressive militarism, totalitarianism, imperialism, and statism) — concentrated and centralized power, concentrated and centralized land ownership, concentrated and centralized psychic energy (withdrawn form the common world-self and enclosed). It requires concerted political effort and monopolization of violence to break apart communal land and identity. The capitalist self of hyper-individualism began with the wealthy elite precisely because they were the initial beneficiaries of the enclosure movement. They were enclosing not only land but their own minds and selves from the ancient common mass of the lingering traces of the bicameral mind. Many were thinking about these issues.

Thomas Jefferson and Thomas Paine’s land reform proposals are as much, if not more, about selfhood and social identity as they are about economics (the elimination of entail and primogeniture was intended as a direct attack on aristocracy). Neither trusted an elite to control all land and all benefits from land but they (fatalistically?) accepted that the enclosure movement was irreversible or necessary for the society that was being created, even as they acknowledged the loss of freedom as demonstrated by Native Americans who could act freely precisely because they were acting within a commons (Benjamin Franklin also made such observations about the greater indigenous freedom and its attraction). These specific founders wanted to make all individuals either land owners (Jefferson’s yeoman farmers as republican citizens) or beneficiaries of land ownership (Paines’s citizens dividend), in both cases a response to the enclosure movement as it encroached in on the New World through land consolidation. Self-development had been limited to the elite, but what if self-development could be made available to all. The most radical challenge of Enlightenment thought was that all of humanity, even women and the poor and non-Europeans, shared a common human nature and that self-cultivated individuality was a universal potential, while others saw it as a necessary demand and obligation (develop an individual self or be punished). Like these two, Adam Smith thought inequality opposed a free society of individual citizens. And for this reason, Smith worried that, as opposed to agriculture, the new industrial labor might dumb down the population and so public education was necessary. Without land to cultivate as part of Jeffersonian republicanism, society would have to teach other methods of self-cultivation. Godwin likewise was concerned with education motivated by a belief that every individual should independently research, analyze, and assess everything for themselves; such deification of individualism being an impossible ideal, of course; but that apparently was of no great concern to him beause he was of a less practical bent, as opposed to Jefferson and Paine’s aspirations to offer real world solutions. From Godwin’s perspective, the point was to create and enforce individualism, including actively destroying collectivities, and then everything else would presumably fall into place.

Godwin opposed the commoners re-creating the ancient practice of the commons for the very reason it was such a natural and deeply entrenched impulse within the shared psyche. Later on, it would be the same reason collective adoptions had to be illegalized to destroy Shaker communities, collective land ownership had to be constrained to weaken Hutterite communities, and collective labor unions had to be busted to shatter working class communites. Individualism isn’t created only one time in the past but must be constantly re-created through the policies and actions of government, the punishment and coercion of law, and the encouragement of incentives and subsidies. Individualism is such a weak, unstable, and unnatural state that it would break apart without constantly being shored up and defended. The modern psyche is ever seeking to return to its origins in the bundled mind of bicameralism, animism, or some other variant. The inherent failure of individualism is regularly affirmed by how individualist realism is entirely dependent on collectivist institutions of state governments, state-created corporate charters, etc — such as giving greater rights, privileges, benefits, power, autonomy, and representation to corporate persons than to most individual humans. We are suffused with an authoritarian collectivism that is the actual system behind the charade of individualism. As with Edmund Burke, Godwin’s fear of combinations, mixings, and associations — the undifferentiated masses — expressed a fear of the impure and disorderly; like an obsessive-compulsive child forever lining up her toys and panicking whenever anyone touches them. This is the demand for principled consistency in the WEIRD mind, but the only principle is order for the sake of order, as demonstration of hierarchical power to assert the authority that authorizes ideological realism. It must be an enforced order because the ancient organic orders of tribe, kinship, village, commons, etc or the grassroots organizing of communities and workers can’t be trusted because it can’t be controlled hierarchically through centralized authority and concentrated power. When the last traces of bicameral voices have been silenced, conservatives see hierarchy as the only authority left to command authorization, be it the hierarchy of Klan, church, military, or something similar.

Hierarchy, though, can only accomplish this if it has been narratized and internalized, by way of the interpellation of symbolic conflation where an ideological realism recedes from consciousness in becoming the calcified frame of thought and perception. This was what made the enclosure movement essential in reifying an abstract ideology. It had to be imprinted upon not only the human psyche but the land itself, the literal ground of psyche as our embodied sense of place. The early land reforms rigidified boundaries, regimented land ownership, and systematized infrastructure — roads were straightened and waterways channelized. As the echoes of the living bicameral voices of ancestral spirits were transformed into the written word as the “dead hand” of corpses (i.e., widespread literacy), the soil became mere dust and land mere property with the earth being mapped and bounded. Some traditions such as Quaker living constitutionalism sought to hold onto the remnants, as part of the memory of a former British communalism. The living landscape invoked by Australian Aborigines maybe was not so different than the English practice of beating the bounds and wassailing that reinforced a collective enclosure of a shared subjectivity. Once the commons were gone, there were no bounds of the commons left to be ritually beat as a community nor communal lands inabited by spirits to be wassailed. Land reform was social reform and moral reform. Godwin’s described education of the mind like the cultivation of enclosed land, which reminds one that Lockean land rights were defined not merely by use but by cultivation or improvement of enclosed land (including John Locke’s consitutional defense of slavery; propertied self going hand in hand with the embodied self literally being property to be owned; though Locked suggested a vague qualification about how much could be enclosed, which meant the rich could accumulate vast tracts of land as long as theoretically somewhere there is still land available for others), wherease the pre-Lockean land rights of Roger Williams acknowledged that any use of even non-enclosed land proved (demonstrated and expressed) ownership, which might simply have been an invocation of the old Charter of the Forest, “guaranteeing the right to commoning (recovered in 1217), which in turn recognized subsistence rights, e.g., the right to widow’s estovers (wood needed for housing repairs, implements, etc.), and to subsistence usufructs (the temporary use of another person’s land)” (Carolyn Lesjak, 1750 to the Present: Acts of Enclosure and Their Afterlife); some of the practices continuing into 19th cenury American property law and still barely hanging on today in certain Western countries.

It is intriguing to think about how recent this happened, but first consider where it began. “In the Middle Ages, fifty per cent or more of the land was commons, accessible to everybody,” says Mark Vernon (Spiritual Commons). Then the enclosures began. “Overall, the pace of enclosure rose dramatically after the 1760s as landowners turned to parliament for the legitimization of their claims,” writes Nina Mcquown. “Michael Turner estimates that more than twenty percent of the area of England was enclosed by act of parliament between 1750 and 1819, the vast majority of these acts occurring after 1760 (32). A high concentration—twenty-one percent of the whole of acreage enclosed by parliament—was enclosed in the decades between 1770 and 1780 and in the years of high grain prices during the Napoleonic wars (Yelling 16).11 Although enclosure continued until the end of the nineteenth century, by 1815 only small and discontinuous patches of common fields remained” (“Rank Corpuscles”: Soil and Identity in Eighteenth Century Representations). Then some further details from Gary Snyder: “between 1709 and 1869 almost five million acres were transferred to private ownership, one acre in every seven. After 1869 there was a sudden reversal of sentiment called the ‘open space movement’ which ultimately halted enclosures and managed to preserve, via a spectacular lawsuit against the lords of fourteen manors, the Epping Forest.” To put that in context, following the Englsh Civil War, the Glorious Revoluion reinstated the monarchy in 1688, but there now was a powerful Parliament. That Parliament would be the agent of change, beginning to take strong actions in that next century. Not only were the commons privatized for the colonies were legally constructed as for-profit corporations, along with the creation of quasi-governmental corporations like the East India Company. This led to the complaints by the colonists in demanding the king stand up to Parliament, but the monarchy no longer held the reigns of power. Capitalism was now running the show.

Even then the Charter of the Forest as the founding document of the Commons, having been established in 1217, didn’t officially end until 1971. It almost made it to the end of the Cold War and a new millennia. One might suspect the Commons seemed too communist to be allowed to survive. If it had been maintained, the people might have gotten the wrong idea about who the country belonged to. Even as the politics of it is more than relevant, what made the enclosure movement a revolutionary moment was the transformation of the Western mind. The real issue was the enclosure of the common identity and moral imagination. That is why, as colonial imperialism took hold and expanded, the rhetoric so heavily focused on the symbolic ‘wilderness’ left remaining. Though the “percentage of wastelands—forests, fens, sheep walks, and moors—enclosed and improved during the period of parliamentary enclosure was relatively small,” writes Nina McQuown, they “loomed large in the imaginations of the propagandists responsible for encouraging the expansion of both enclosure and the innovative agricultural practice that it was thought to support.” Carolyn Lesjak writes that, “If enclosure in the 16th century was largely “by agreement” and, in fact, condemned by both the church and the government, who sided with the commoners’ claims regarding “common rights,” by the 1750s the government had taken the lead and over the course of the period from 1750-1830 passed over 4000 Acts of Enclosure, resulting in over 21% of the land (approximately 6.8 million acres) being enclosed (see Ellen Rosenman’s BRANCH essay on “Enclosure Acts and the Commons”). By the end of the century, virtually all the open fields in Britain were gone.” Everything had to be cultivated, even what was deemed useless. All material was to be fodder for improvement and progress, at least in the new mythos. “After the 1760s,” McQuown explains, as the “British improvers turned the logic and language of colonialism inward, towards the wastes,” they also turned inward to colonizing the uncultivated mind.

This makes one realize how false it is to blame everything on the later political revolutions and freethinking radicals. The enclosure movement actually began much earlier around the 14th century, around the time of the English Peasants’ Revolt. Even Parliaments’ legal justifications and enforcement happened generations before the Boston Massacre and Boston Tea Party. This reform of land, self, and mind unsurprisingly preceded and then overlapped with the early modern revolutions. John Adams famously wrote that, “What do We mean by the Revolution? The War? That was no part of the Revolution. It was only an Effect and Consequence of it. The Revolution was in the Minds of the People, and this was effected, from 1760 to 1775, in the course of fifteen Years before a drop of blood was drawn at Lexington. The Records of thirteen Legislatures, the Pamphlets, Newspapers in all the Colonies ought be consulted, during that Period, to ascertain the Steps by which the public opinion was enlightened and informed concerning the Authority of Parliament over the Colonies.” His only error was limiting his scope to the colonies and not pushing it further back. Enclosure of land became reform of mind became revolution of society became rupture of history. The cultivation of farming that once followed astrological cycles of return (i.e., revolution) had ground down the bones of the dead into dust. Humanity was uprooted from the past and temporally dislocated in an abstract narrative, as cyclical time became linear and nostalgia became a disease. The colonists surely experienced this most clearly in how the early waves of colonists largely consisted of the most destitute landless peaseants, many recently evicted from the commons and feudal villages, often arriving as slave-like indentured servants and convict labor — one can imagine the desperation and despair they felt, as being sent to the early colonies was practically a death sentence.

The colonial era may seem like a distant time from the present, but we can sense how the world we now live in was shaped then. Most Westerners remain landless peasants. The commons that once defined a communal experience of reality only remain like the shadows of a nuclear blast, the traces of a living world that remains our ancient inheritance, however cut off we have become. It may seem the egoic boundaries of our individualism have toughened into place like scars, like the crust of parched earth. We feel tired and anxious from the constant effort of mainaintiaing the walls of our mind, to keep the self separate from the world. It takes only a moment’s lapse when our guard is let down before we begin to sense what we have lost. An aching tenderness remains below. We are so hungry for connection that simply stepping into the commons of a forested park can feel like a spiritual experience for many people today. Yet such moments are mere glimpses too often quickly forgotten again. We have no shared experience, no living memory to draw from. We have no solid ground to stand upon. And the path to a different world that existed in the past has been gated shut. Or so it seems. But is that true? Where else could we be but in the world? Nature knows no boundaries nor does the human psyche, if we root down deep enough into our own soil. There is no sense of self without a sense of place for we mould ourselves out of the clay, as we breathe the dust of our ancestors.

Landscape is memory, and memory in turn compresses to become the rich black seam that underlies our territory.

Alan Moore, Coal Country, from Spirits of Place

Ever place has its own… proliferation of stories and every spatial practice constitutes a form of re-narrating or re-writing a place… Walking [into a place] affirms, suspects, tries out, transgresses, respects… haunted places are the only ones people can live in.

Michel de Certeau, The Practice of Everyday Life

* * *

Southern United States: An Environmental History
by Donald E. Davis, pp. 136-7

Without question, Kentucky’s early reputation as a hunter’s paradise influenced public opinion about all those residing in the uplands during the early settlement period. In a world that equated agricultural improvements with civilization, Native Americans living in the backcountry were seen by most Anglo-Europeans as representing the lowest evolutionary stage of human development—hunter. So needing a rationale for conquering and subdividing the largely forested frontier, Kentucky and other Ohio Valley Native Americans became hunters in the minds of most Europeans, even though they were also accomplished agriculturalists. Not surprisingly, after frontier settlers had later adopted many of the same subsistence techniques and hunting practices of their Shawnee, Creek, and Cherokee neighbors, they too were ridiculed by authorities for their “backward” and “primitive” ways. The British military commander for North America, Thomas Cage, was already of the opinion in 1772 that white backcountry settlers “differ little from Indians in their manner of life” [Davies 1972-1981, V. 203]. Perhaps more to the point is frontier historian Stephen Aron, who, in paraphrasing Cage’s letter to the Earl of Hillsborough, wrote that backcountry residents “dressed like Indians, comported themselves like Indians, and indiscriminately consorted with one another like Indians [Aron 1996, 14]. Hunting was blamed as the principal cause of the problem by both religious reformers and the ruling elite, who in their missionary visits and public appeals, tried to promote the latest agricultural reforms among the backwoods populace. Agreeing with the reformers, Crevecoeur, the celebrated author of Letters from an American Farmer, wrote that “as long as we keep ourselves busy tilling the earth, there is no fear of any of us becoming wild; it is the chase and the food it procures that have this strange effect” [Crevecoeur 1957, 215].

“Ruinous Mixture”: Godwin, Enclosure and the Associated Self
by Robert Anderson

In this argument, I turn on its head Godwin’s claim that the right to private property “flows from the very nature of man.” While Godwin argues that the right to property is “founded” on the “right of private judgment” which “flows from the very nature of man” (2.169-70), I will argue that this argument runs counter to his notion that private property “unavoidably suggests some species of law” to guarantee it (2.439). To be more specific, I argue that Godwin’s defense of the “sacred” and “essential” “sphere” surrounding the self (1.1.70, 1.257), which is necessary to potect it from being “resolved . . . into the common mass” (1.289), draws upon the conceptual framework which informs the rhetoric of the Enclosure Movement. In particular, I note his argument that cutting off the individual from the “common mass” is necessary for “improvement”—another term for enclosure. [….]

Part of his “extensive plan of freedom” involved the socialization of the self and (ideally) property and the rejection of all restraints on individual liberty; his “reprobation,” I argue, stems from this same defense of private judgment, which can be said to serve the conservative interests of the powers that be.

  1. The Subject of the Commons

Political associatons came of age in the latter part of the eighteenth century in response to the upheavals wrought by the industrial revoltion. Associations were contesting the state’s efforts to regulate subjectivites. Albert Goodwin recounts that in 1790 in the industrial center of Sheffeld, for example, “the master scissorsmiths,” apprehensive of the collective power of striking scissor grinders, “called a general meeting of the town’s merchants and manufacturers ‘to ooppose the unlawful combinations of the scissor grinders and the combinations of all other workmen.'” The same anxiety about the collective strength of the poor which led the Sheffield city leaders to oppose combinations also led to attempts to eradicate collective landholding arrangements by enclosing the commons. Following the passage of the Private Enclosure Act of 6 June 1791, in whch 6,000 acres of commons were redistrbuted among the wealthy “local land-holders, tithe-owners and large freeholders,” an angry mob, comprising both peasants and industrial laborers, rioted, threatening to destroy “the lives and properties of the freeholders who had approved the enclosure” (165-67). The fact that the mob opposing enclosure included industrial laborers as well as peasant farmers whose land was being appropriated reveals the close connections between enclosure and industrial capitalism. Sayer and Corrigan make the connecton between enclosure, capitalism, and subjectivity in this period more explicit.

But the great catastrophe which above all pervades the eighteenth century is the acceleration of the great “freeing” of labour (and thus making labour-power) that divides wage-labouring from generalized poverty; the long movement from service to employment, from provision to production/consumption, from political theatre to the individualism . . . of the vote: enclosures.” (96)

As Marx argues, enclosure ensures that workers, expropriated from their means of subsistence, are thrust into relations of dependence on the capitalists.

Goodwin goes on to relate that the response of the commoners and laborers also took forms more organized and intellectual than rioting. “When ‘5 or 6 Mechanicks’ began to meet . . . to discuss ‘the enormous high prices of Provisions,'” they initiated the creation of political societies, associatons, for the (self-) education of the working classes (166). They attempted, in the words of one charter, “to persuade their benighted brethren to defend themselves against private and publiic exploitation by the assertion of their natural rights” (qtd. in Goodwin 167). Political societes provided laborers with an organized forum—an institution—to exert influence on the opinions of their fellow laborers, and by extension, on society at large. Godwin opposes political associations on just this account. The “interference of an organized society” to influence “opinion” is “pernicious” (2.2280. “[E]ach man must be taught to enquire and think for himself,” uninfluenced by either “sympahy or coercion,” guided only by “reason.” The “creeds” of politcal associations, on the other hand, encourage “each man to identify his creed with that of his neighbour” (1.288). He goes on to argue that sympathy, like a disease, is especially contagious among undisciplined laborers: “While the sympathy of opinion catches from man to man, especially among persons whose passions have been little used to the curb of judgment, actns may be determined upon, which the solitary reflections of all would have rejected” (1.294). Like the unenclosed commons, sympathy threatens the distinctions upone which general improvements is predicated: the “mind of one man is essentally diistinct from the mind of another. If each do not preserve his individuality, the judgment of all will be feeble, and the progress of our common understandng inexpressibly retarded” (1.236).

1790, the year the Sheffeld master scissorsmiths moved to oulaw the combinations of “grinders” and “workmen,” was also the year in which Edmund Burke published his Reflections on the Revolution in France. Burke reserved his greates hostility—and fear—for the “confusion” of the “swinish multitdude” (314). Reflections reveals the extent to which concerns about the collective power of the masses, the upheavals of the industrial revolution, and anxiety about the French Revolution are intertwined. The “French Revolution,” he argues, was brought about “by the most absurd and ridiculous . . . by the most contemptible instruments. Everything seems out of nature in this strange chaos of levity and ferocity, and all sorts of crimes jumbled together with all sorts of follies.” And further, it is a “monstrous tragi-comic scene, the most opposite passions necessarily succeeded, and sometimes mix with each other in the mind; alternate contempt and indignation; alternate laughter and tears; alternate scorn and horror.” Burke’s concern about the inappropriate mixture driving the French Revolution invokes a common rhetoric for disparaging forms of life among peasants and the laboring population. It appears, as I will argue, in condemnations of the “waste” and the “ruinous . . . Mixture of opposite Interests” in the subsistence economy of the commons, and in Godwin’s critique of the tumult of political associations—both of which are seen as threats to individual integrity and “progress.” It also appears in his analysis of the “mechanism of the human mind.”

“Rank Corpuscles”: Soil and Identity in Eighteenth Century Representations
by Nina Patricia Budabin McQuown

The teleology of improvement could even stretch towards man’s transcendence of matter itself. This idea is amply represented in a notorious reverie from Godwin’s first edition of An Enquiry Concerning Political Justice (1793), where Godwin projects the complete domination of matter—not only the matter of the soil, but also and especially the matter of the body—as the eventual outcome of human progress, beginning with its progress in agriculture. His logic traces a line from improved agriculture to a human transcendence of appetite, illness, and death: “[t]hree fourths of the habitable globe is now uncultivated. The parts already cultivated are capable of immeasurable improvements” (2: 861), he offers, and if we can gain control “over all other matter,” Godwin suggests,

“why not over the matter of our own bodies? If over matter at ever so great a distance, why not over matter which . . . we always carry about with us, and which is in all cases the medium of communication between that principle and the external universe? In a word, why may not man one day be immortal?” (2: 862)

Godwin’s questions are only the most succinct statement of the radical hope that is at the center of late eighteenth-century bourgeois liberalism, which, as Kramnick has argued, linked agricultural improvement to “middle-class disdain for the past, for history, and for custom” (Kramnick, “Eighteenth-Century Science” 9). For reformist thinkers, in all areas of human ambition, improvement was articulated as a break with the past and an optimistic orientation towards the future.

Even so, reformers relied on an analogy between human self-ownership and landownership that draws on inherited parallels between human bodily-economy and the social system.5 Reformers saw an obvious parallel between agriculturally improved land and the human subject, who, cut off by self-reliance from the prejudice of contemporaries as well as the inherited prejudices of the past, could “cultivate” himself towards perfection, so that, as Robert Anderson puts it, “[t]he moral economy and political economy merge in the social and semantic fields covered by ‘improvement’” (630). In the works of both Godwin and Priestley, both subjectivity and soil are divided into discrete properties whose content is to be determined by one and only one owner, protected by the integrity of the individual conscience from absorption into the “common mass” of human thought and opinion (620).6 Enclosure of both self and soil meant divestment from the influence of history—those ancient patriarchs and their prejudices—as much as from the influence of the rights of commonage. If earlier authors imagined the soil as disseminating ownership of England’s past, bearing it physically into the bodies of nationals, later eighteenth-century reformist authors often render the soil as a failed medium for the transmission of historical experience and lingering subjectivities. Such failure is, paradoxically, reinscribed as improvement. Priestley destroys the “foundation” for the prejudicial thought of the past, and Charlotte Smith, as we will see in the conclusion to this chapter, insists on a failure of communication between the present and an incomprehensible past that is buried well below reach of the ploughshare, and is in any case unworthy of transmission. Smith and Priestley deny the relevance of the past to the present because both prefer to build on a different foundation.

This chapter examines late eighteenth-century reformist representations of the soil primarily in the field of agricultural writing. It offers an analysis, first, of Arthur Young’s writing in support of the enclosure of waste soils in several works of the 1770s and 1780s. In contrast to the revolutionary rhetoric of Priestley, Godwin, and Smith, Arthur Young is usually thought of as a political conservative for his response to the French Revolution.7 Yet to call Young a conservative is to fail to appreciate common ground he shared with the likes of progressives such as Godwin and Priestley in his advocacy for enclosure and against tithes and poor rates. Moreover, in the field of agriculture at least, Young was hardly an advocate for the careful and conservative restoration of the edifice of the past. For Young, the waste spaces of Britain must be rendered into an inviting blankness empty and available enough to rival the magnetism of America’s putatively untouched interior. We start by acknowledging the ways that his arguments for the enclosure of wastelands require the figuration of Britain as Locke’s tabula rasa, ripe for human improvement, and move on to a specific discussion of Young’s descriptions of moor soils as the prototypical waste, where we find him forcibly unearthing and dispersing the evidence of other histories and interests in the soil in order to make the past available for improvement towards a progressively more fertile future. In Young’s improvement and enclosure propaganda, we can see that eighteenth-century agricultural writing does not, like Dryden’s translation of the Georgics and Defoe’s Tour in this dissertation’s chapter two, simply mediate, reframe, or cover up relics that it cannot fit into an acceptable narrative of British history, or, like Powell and Philips, allow the concept of recirculation through the soil to provide an alternative, inarticulate, and immediate relation to the past. Nor does Young, like Smollett or Tull, suggest sequestration from the violating agency of decay. Instead, Young offers an improvement that actively un-earths the past. The coherence of Young’s improved Britain is based not on a hermeneutics of repression, where fragmented and conflicting histories are buried out of sight, but on the agricultural improver’s active recycling of the past into fertile soil that will produce a better future. His texts acknowledge the tangles of historical and legal relics and material and customary restraints in and on the soil in order to enact their exhumation and dispersal. By claiming and controlling the power of putrefaction to break down and disseminate relics, Young’s improver takes over the soil’s work of decay. He releases the value of the past for the production of future goods.

In fact, Young’s program—which became the program of the new Royal Agricultural Society in 1793—was so successful that by the end of the eighteenth century, the landscape of Britain was entirely changed. With private enclosures replacing open fields formerly held in common, it was divided into subdivisions set apart by hedgerows, ditches, walls, and straight(er) roads. Where Godwin imagined a mind that could be enclosed and cultivated like soil through improvements, the poet John Clare asserted that by the first decades of the nineteenth century, that the British landscape had indeed come to imitate the private boundaries of the individual conscience. In this poem on the enclosure of his native village in Northhamptonshire, “The Moors,” for example, Clare shows,

“Fence meeting fence in owners’ little bounds
Of field and meadow, large as garden grounds,
In little parcels little minds to please,
With men and flocks imprisoned, ill at ease.”
(46-49)

For Clare as for others, the consonance of a private landscape and a private subjectivity came with a sense of loss, both of individual rights, and of continuity with the past, whose paths “are stopt—the rude philistine’s thrall / Is laid upon them and destroyed them all” (64-5). The sense that improvement had turned out to mean the parceling up of experience into discrete and discontinuous blocks led, for Godwin, to his eventual anxiety that the possibility of future progress had also been lost. How can men whose lives are so strongly separated engage in the communication that leads to human perfection? This anxiety motivates Godwin’s An Essay on Sepulchres (1809), a text in which Godwin ultimately abandons his advocacy for a historical soil, and proposes that dirt—literally the dust of the buried corpses of great men—could be the foundation of improvement by materializing cultural and historical continuity. Godwin’s Essay proposes a different kind of soil-fertility, land that fruits out in knowledge, experience, and sentiment instead of only food. Yet Godwin’s essay is unable to imagine an immediate and therefore openended relation between human bodies and the dust of the dead. He strives to secure stable access to corpses that are also subjects, with particular memories and ideas to represent to their living interlocutors. Intent on controlling the legacy that the past leaves for the future, Godwin can only approach the dead through the medium of their representations—both the texts they leave behind and the monuments he wants to erect at their gravesites. Ultimately, his Essay offers less a plan for the stable continuity of experience across generations, than a revelation of the limits of what representations and mediums can accomplish when they refuse the immediate agency of soil.

The Early Modern 99%
by Harry C. Merritt

Reverberations of battle are the soundtrack to developments in England at the time, where King Charles I would be executed the following year and his kingdom transformed into a commonwealth. During the course of the film, the educated and principled Whitehead is forced into labor together with the alcoholic Jacob and the simpleton Friend by O’Neill, a rogue Irishman seeking self-enrichment. […]

Not just England was in turmoil at this time — much of Europe and the growing number of territories it ruled across the globe experienced extraordinary upheaval during the sixteenth and seventeenth centuries.

Though the “General Crisis of the Seventeenth Century” thesis originally developed by Marxist historian Eric Hobsbawm has since been challenged and amended, a number of broad themes can still be distilled. Religious dissent and political radicalism challenged the authority of both the Catholic Church and monarchs who ruled by the grace of God. Conflicts like the Thirty Years War descended into endless nihilistic pillage and slaughter before lending themselves to the creation of the modern state system. The ruthless quest for precious metals and profits fueled the conquest of Native American peoples and the establishment of the Atlantic slave trade.

Perhaps one of the most powerful conceptualizations of this period can be found in Peter Linebaugh and Marcus Rediker’s book The Many-Headed Hydra: The Hidden History of the Revolutionary Atlantic. According to Linebaugh and Rediker, the ruling classes imagined themselves to be the latter-day incarnation of Hercules, laboring to bring order to a chaotic world. The embodiment of their enemy was the mythological Hydra, whose many heads represented its multifarious elements: religious dissenters, radical commoners, rebellious African slaves, fiercely independent Native Americans, and freethinking women.

In the Americas and on the Atlantic, “the plebian commonism of the Old [World]” encountered “the primitive communism of the New World” and formed a hybrid, alternative vision that set itself against the emergent order of modernity. Late in A Field in England, a hallucinating Whitehead declares, “I am my own master”; this realization is precisely what the ruling classes feared most in the Hydra.

Despite its multitudes, the Hydra was ultimately unsuccessful at challenging the emerging capitalist, colonialist order of modernity. In the centuries since, it would be difficult to imagine a group that parallels the Hydra in its diversity, utopianism, and in the threat it poses to the ruling classes — that is, until today. The emergence of the 99% as a social grouping that has come to be dreaded and despised by members of the 1% reproduces the dynamics and the discourse of that era.

While a new era of globalization erodes the economic security of the vast majority of the US, the 1% and their political supporters insist that they work harder than the rest of us and thus their ownership of nearly half of the world’s wealth is for the greater good. Recently, we have been treated to numerous declarations from members of the 1% suggesting that they are under threat from the 99%.

These shrill cries about impending repression — invoking Nazism seems popular — reveal the degree to which the 1% identify with one another and fear the masses. Like the Hydra, the 99% is a rhetorical construction rather than a social formation with clear class consciousness. Its very diversity constitutes its greatest weakness. The repeated spread, defeat, and resurrection of movements like Occupy Wall Street and Spain’s indignados resemble the scattered but persistent revolts of the Hydra. Today’s Occupy activists should recall that a revolutionary conspiracy by a group of New York City laborers — black and white, slave and free — emerged in 1741 out of a waterfront tavern just blocks from today’s Zuccotti Park. With goals that are simultaneously utopian and practical, these movements appeal to both the basic needs and the deepest desires of common people around the globe. […]

Alain Badiou sees “the invariant features of every real mass movement: egalitarianism, mass democracy, the invention of slogans, bravery, the speed of reactions” embodied in both Thomas Müntzer’s movement of the 1500s and in Tahrir Square of the 2010s. As disparate groups occupy public spaces from Cairo to Madrid to New York, asserting their rights and presenting an alternative vision of their societies, we should not forget the members of the Hydra who fought against the exploitation of the ruling classes in favor of another world during the early modern period.

Some will argue that our present time is too distant to draw many practical lessons from this period. But that does not mean we cannot look to its events, personages, and symbols for inspiration. By coincidence, the rainbow flag used by today’s LGBT and peace activists bears a striking resemblance to the rainbow flag Thomas Müntzer once used to rally the German peasantry — a fitting symbol in any period for uniting a diverse coalition and insisting that another world is possible.

The Effect of Land Allotment on Native American Households During the Assimilation Era
by Christian Dippel and Dustin Frye

Toward the end of the 19th century, with the conclusion of the Indian Wars and the closing of the frontier, reformers and the U.S. government turned their attention towards the cultural assimilation of Native Americans, ninety percent of whom were living on the reservations created in the previous decades. This is signified by the famous 1892 quote: “kill the Indian in him, and save the man.”1 Assimilation efforts were centered on land allotment policies that broke up tribally owned reservation lands into individually owned homestead-sized land allotments. As the Bureau of Indian Affairs (BIA) commissioner noted: “if there were no other reason [for allotment], the fact that individual ownership of property is the universal custom among civilized people of this country would be a sufficient reason for urging the handful of Indians to adopt it.” Allotment was the cornerstone of federal Indian policy beginning with the passing of the General Allotment (or ‘Dawes’) Act in 1887 until it ended with the passing of the Indian Reorganization (or ‘Howard-Wheeler’) Act (IRA) in 1934 (Carlson, 1981, p18).

When a reservation was opened for allotment, all families on the reservation were given allotments, and these allotments were held in a trust managed by the local Indian agent (the BIA’s local superintendents in charge of a reservation). Trust-status meant allottees could not sell or collateralize an allotment. In order to obtain full ‘fee-simple’ legal title for their allotment, allottees had to be declared “competent” by the BIA agent (Carlson, 1981; Banner, 2009; Otis, 2014). In short, Indian allotment was designed as a conditional transfer program aimed at cultural assimilation. The first ‘treatment arm’ was an unconditional transfer program: receiving an allotment gave the allottee the unconditional right to use the land for their own purposes, as well as the right to leasing rents. The second treatment arm was only obtained conditional on proving one’s “competence.” Allotment’s conditional transfer arm (full title) was worth almost 20 times annual per capita incomes in our data, orders of magnitude larger than modern-day conditional transfer programs. Our paper is an investigation into how individual households responded to the incentives created by this program.

We hypothesize that individual allottees responded to the allotment policy’s incentive structure by signalling cultural assimilation to the BIA agents in order to be able to obtain full property rights over their allotments. First evidence of this comes from an annual panel of reservation aggregate data from the BIA’s annual reports from 1911 to 1934. In addition to schooling, these data include very direct measures of assimilation or assimilation-signalling, namely the number of “church-going Indians” and of those “wearing civilized dress.” We combine these data with the universe of Indian allotments, which the Bureau of Land Management (BLM) has digitized with geo-location and issuance year. In a within-reservation over-time comparison, we find that school-attendance, the number of church-going Indians and the number of those wearing civilized dress increased in lock-step with the expansion of allotment, even after controlling for potential changes in school and clerical infrastructure.

1 Quote from a speech by Capt. Richard Pratt, founder of the first Indian boarding school. Appendix-Figure A1 shows one of the many “before/after” pictures one finds in association with the Assimilation Era.

Metaphorical Space and Enclosure in Old English Poetry
by Benjamin S. Waller

A Language Older Than Words
by Derrick Jensen, pp. 101-6

Only recently—especially after teaching at a university for a few years— have I come to understand why the process of schooling takes so long. Even when I was young it seemed to me that most classroom material could be presented and assimilated in four, maybe five, years. After you learn fractions and negative numbers in first or second grade, what new principles are taught in math until algebra in junior high? It’s the same with science, art, history, reading, certainly writing. Nearly everything I learned those years—and this was true for my friends as well—was gleaned through books and conversations outside class. It’s true to the point of cliché that most of the “crap” we learn in high school, as Simon and Garfunkel put it, is a bland stew of names, dates, and platitudes to be stored up the night before each test, then forgotten the moment the test is handed in.

During high school, I believed the primary purpose of school was to break children of the habit of daydreaming. If you force them to sit still long enough, eventually they tire even of sinking turn-around fadeaways at the buzzer to win NBA championships. Having sat in the back of the class lining rockets over the left field fence for the better part of thirteen years, I was ready to move on.

I’ve since come to understand the reason school lasts thirteen years. It takes that long to sufficiently break a child’s will. It is not easy to disconnect children’s wills, to disconnect them from their own experiences of the world in preparation for the lives of painful employment they will have to endure. Less time wouldn’t do it, and in fact, those who are especially slow go to college. For the exceedingly obstinate child there is graduate school.

I have nothing against education; it’s just that education— from the Greek root educere, meaning to lead forth or draw out, and originally a midwife’s term meaning to be present at the birth of—is not the primary function of schooling. I’m not saying by all this that Mrs. Calloway, my first-grade teacher, was trying to murder the souls of her tiny charges, any more than I’ve been trying to say that individual scientists are necessarily hell-bent on destroying the planet or that individual Christians necessarily hate women and hate their bodies. The problem is much worse than that, it is not merely personal nor even institutional (although the institutions we’ve created do mirror the destructiveness of our culture). It is implicit in the processes, and therefore virtually transparent.

Take the notion of assigning grades in school. Like the wages for which people later slave—once they’ve entered “the real world”—the primary function of grades is to offer an external reinforcement to coerce people to perform tasks they’d rather not do. Did anyone grade you when you learned how to fish? What grades did you get for pretending, shooting hoops, playing pinball, reading good books, kissing (“I’m sorry, dear, but you receive a C”), riding horses, swimming in the ocean, having intense conversations with close friends? On the other hand, how often have you returned, simply for the joy of it, to not only peruse your high school history textbook, but to memorize names and dates, and, once again for the joy of it, to have a teacher mark, in bright red, your answers as incorrect?

Underlying tests as given in school are the presumptions not only that correct answers to specific questions exist, but that these answers are known to authority figures and can be found in books. Tests also generally discourage communal problem solving. Equally important is the presumption that a primary purpose of school is to deliver information to students. Never asked is the question of how this information makes us better people, or better kissers, for that matter. Systematically—inherent in the process—direct personal experience is subsumed to external authority, and at every turn creativity, critical thought, and the questioning of fundamental assumptions (such as, for example, the role of schooling on one’s socialization) are discouraged.

If you don’t believe me, pretend for a moment you’re once again in school. Pretend further that you have before you the final test for a final required class. If you fail this test, you fail the class. While you may have enjoyed the process of schooling, and may even have enjoyed this class, you enjoyed neither enough to warrant repetition. Pretend the test consists of one essay question, and pretend you know the instructor well enough to understand that if you mimic the instructor’s opinions you’ll get a higher grade. If you disagree with the instructor—pretend, finally, that you do— you’ll be held to a higher standard of proof. What do you do? Do you speak your mind? Do you lead with your heart? Do you take risks? Do you explore? Do you write the best damn essay the school has ever seen, then return next year to retake the class? Or do you join with thousands—if not millions—of students who face this dilemma daily and who astutely bullshit their way through, knowing, after all, that c stands for Credit?

Grades, as is true once again for wages in later life, are an implicit acknowledgment that the process of schooling is insufficiently rewarding on its own grounds for people to participate of their own volition. If I go fishing, the time on the water— listening to frogs, smelling the rich black scent of decaying cattails, holding long conversations with my fishing partner, watching osprey dive to emerge holding wriggling trout—serves me well enough to make me want to return. And even if I have a bad day fishing, which, as the bumper sticker proclaims, is supposed to be “better than a good day at work,” I still receive the reward of dinner. The process and product are their own primary rewards. I fish; I catch fish; I eat fish. I enjoy getting better at fishing. I enjoy eating fish. No grades nor dollars are required to convince me to do it. Only when essential rewards disappear does the need for grades and dollars arise.

It could be argued that I’m missing the point, that the product of the years of homework and papers and tests are not the physical artifacts, nor the grades, nor the bits of information, but instead the graduates themselves. But that’s my point exactly, and we must ask ourselves what sort of product is that, from what sort of process.

A primary purpose of school—and this is true for our culture’s science and religion as well—is to lead us away from our own experience. The process of schooling does not give birth to human beings—as education should but never will so long as it springs from the collective consciousness of our culture—but instead it teaches us to value abstract rewards at the expense of our autonomy, curiosity, interior lives, and time. This lesson is crucial to individual economic success (“I love art,” my students would say, “but I’ve got to make a living”), to the perpetuation of our economic system (What if all those who hated their jobs quit?), and it is crucial, as should be clear by now, to the rationale that causes all mass atrocities.

Through the process of schooling, each fresh child is attenuated, muted, molded, made—like aluminum—malleable yet durable, and so prepared to compete in society, and ultimately to lead this society where it so obviously is headed. Schooling as it presently exists, like science before it and religion before that, is necessary to the continuation of our culture and to the spawning of a new species of human, ever more submissive to authority, ever more pliant, prepared, by thirteen years of sitting and receiving, sitting and regurgitating, sitting and waiting for the end, prepared for the rest of their lives to toil, to propagate, to never make waves, and to live each day with never an original thought nor even a shred of hope.

In Letters From an American Farmer, Michel Guillaume Jean de Crévecoeur noted: “There must be in the Indians’ social bond something singularly captivating, and far superior to be boasted of among us; for thousands of Europeans are Indians, and we have no examples of even one of those Aborigines having from choice become Europeans.”

Benjamin Franklin was even more to the point: “No European who has tasted Savage Life can afterwards bear to live in our societies.” It was commonly noted that at prisoner exchanges, Indians ran joyously to their relatives while white captives had to be bound hand and foot to not run back to their captors.

It is small wonder, then, that from the beginning, whenever we have encountered an indigenous culture, we have had the Lord our God— replaced now by economic exigency—tell us that “thou shalt smite them; and utterly destroy them; thou shalt make no covenant with them, nor shew mercy unto them.” What seems at first aggression is in fact self preservation, a practical staunching of what would otherwise be an unmanageable and embarrassing flow of desertions.

The same self-preservation motivated my father’s actions when I was a child. To preserve the person that he had become, he had to smite and utterly destroy all who reminded him of what could have been, and of the person he once was, far beyond conscious memory, before his parents, too, out of self-preservation destroyed him. So he lashed out with fist, foot, voice, penis, all so he could forget, all so we could never know, ourselves, that alternatives to fear existed. Had he been able to destroy the stars to so destroy me, he would have done it. Had he been able to destroy the stars, as even now we are destroying the seas and forests and grasslands and deserts, he would have succeeded, I am sure, in destroying me.

In the eighteenth century, de Crévecoeur wrote, “As long as we keep ourselves busy tilling the earth, there is no fear of any of us becoming wild.” Though the wild outside diminishes each day, as do intact cultural alternatives, the fear of these alternatives remains. The fear shall remain so long as we live the way we do, and so long as there are alternatives we must avoid. The alternatives shall remain so long as there is life. We should not be surprised, then, that our culture as a whole must destroy all life and that we as individuals must not dwell upon the horrors we visit not only upon others but upon ourselves, that we dwell instead upon the daily earning of our bread, and beyond that pile upon ourselves project after project to keep ourselves always occupied, always unconscious of the fact that we do not have to live this way, always blindered to alternatives. For if we looked we might see, if we saw we might act, and if we acted we might take responsibility for our own lives. If we did that, what then?

Henry Adams on the Bundled Mind

Of all forms of pessimism, the metaphysical form was, for a historian, the least enticing. Of all studies, the one he would rather have avoided was that of his own mind. He knew no tragedy so heartrending as introspection, and the more, because-as Mephistopheles said of Marguerite–he was not the first. Nearly all the highest intelligence known to history had drowned itself in the reflection of its own thought, and the bovine survivors had rudely told the truth about it, without affecting the intelligent. One’s own time had not been exempt. Even since 1870 friends by scores had fallen victims to it. Within five-and-twenty years, a new library had grown out of it. Harvard College was a focus of the study; France supported hospitals for it; England published magazines of it. Nothing was easier than to take one’s mind in one’s hand, and ask one’s psychological friends what they made of it, and the more because it mattered so little to either party, since their minds, whatever they were, had pretty nearly ceased to reflect, and let them do what they liked with the small remnant, they could scarcely do anything very new with it. All one asked was to learn what they hoped to do.

Unfortunately the pursuit of ignorance in silence had, by this time, led the weary pilgrim [i.e., himself] into such mountains of ignorance that he could no longer see any path whatever, and could not even understand a signpost. He failed to fathom the depths of the new psychology, which proved to him that, on that side as on the mathematical side, his power of thought was atrophied, if, indeed, it ever existed. Since he could not fathom the science, he could only ask the simplest of questions: Did the new psychology hold that the νΧή–soul or mind–was or was not a unit? He gathered from the books that the psychologists had, in a few cases, distinguished several personalities in the same mind, each conscious and constant, individual and exclusive.

The fact seemed scarcely surprising, since it had been a habit of mind from earliest recorded time, and equally familiar to the last acquaintance who had taken a drug or caught a fever, or eaten a Welsh rarebit before bed; for surely no one could follow the action of a vivid dream, and still need to be told that the actors evoked by his mind were not himself, but quite unknown to all he had ever recognized as self. The new psychology went further, and seemed convinced that it had actually split personality not only into dualism, but also into complex groups, like telephonic centres and systems, that might be isolated and called up at will, and whose physical action might be occult in the sense of strangeness to any known form of force.

Dualism seemed to have become as common as binary stars. Alternating personalities turned up constantly, even among one’s friends. The facts seemed certain, or at least as certain as other facts; all they needed was explanation.

This was not the business of the searcher of ignorance, who felt himself in no way responsible for causes. To his mind, the compound νΧή took at once the form of a bicycle-rider, mechanically balancing himself by inhibiting all his inferior personalities, and sure to fall into the sub-conscious chaos below, if one of his inferior personalities got on top. The only absolute truth was the sub-conscious chaos below, which every one could feel when he sought it.

Whether the psychologists admitted it or not, mattered little to the student who, by the law of his profession, was engaged in studying his own mind. On him, the effect was surprising. He woke up with a shudder as though he had himself fallen off his bicycle. If his mind were really this sort of magnet, mechanically dispersing its lines of force when it went to sleep, and mechanically orienting them when it woke up–which was normal, the dispersion or orientation? The mind, like the body, kept its unity unless it happened to lose balance, but the professor of physics, who skipped on a pavement and hurt himself, knew no more than an idiot what knocked him down, though he did know–what the idiot could hardly do–that his normal condition was idiocy, or want of balance, and that his sanity was unstable artifice. His normal thought was dispersion, sleep, dream, inconsequence; the simultaneous action of different thought-centres without central control. His artificial balance was acquired habit. He was an acrobat, with a dwarf on his back, crossing a chasm on a slack-rope, and commonly breaking his neck.

By that path of newest science, one saw no unity ahead–nothing but a dissolving mind-and the historian felt himself driven back on thought as one continuous Force, without Race, Sex, School, Country, or Church.

The Education of Henry Adams
Chapter XXIX
“The Abyss of Ignorance” (1902)
pp. 432-434

(Credit to Ron Pavellas for bringing this passage to my notice.)

Who are we hearing and talking to?

“We are all fragmented. There is no unitary self. We are all in pieces, struggling to create the illusion of a coherent ‘me’ from moment to moment.”
~ Charles Fernyhough

“Bicamerality hidden in plain sight.”
~ Andrew Bonci

“What I tell you in the dark, speak in the daylight; what is whispered in your ear, proclaim from the roofs.”
~ Matthew 10:27

“illusion of a completed, unitary self”
Bundle Theory: Embodied Mind, Social Nature
The Mind in the Body
Making Gods, Making Individuals
The Spell of Inner Speech
Reading Voices Into Our Minds
Verbal Behavior
Keep Your Experience to Yourself

Epigenetic Memory and the Mind

Epigenetics is fascinating, even bizarre by conventional thought. Some worry that it’s another variety of determinism, just not located in the genes. I have other worries, if not that particular one.

How epigenetics work is that a gene gets switched on or off. The key point is that it’s not permanently set. Some later incident, conditions, behavior, or whatever can switch it back the other way again. Genes in your body are switched on and off throughout your lifetime. But presumably if no significant changes occur in one’s life some epigenetic expressions remain permanently set for your entire life.

Where it gets fascinating is that it’s been proven that epigenetics gets passed on across multiple generations and no one is certain how many generations. In mice, it can extend at least upwards of 7 generations or so, as I recall. Humans, of course, haven’t been studied for that many generations. But present evidence indicates it operates similarly in humans.

Potentially, all of the major tragedies in modern history (violence of colonialism all around the world, major famines in places like Ireland and China, genocides in places like the United States and Rwanda, international conflicts like the world wars, etc), all of that is within the range of epigenetis. It’s been shown that famine, for example, switches genes for a few generations that causes increased fat retention and in the modern world that means higher obesity rates.

I’m not sure what is the precise mechanism that causes genes to switch on and off (e.g., precisely how does starvation get imprinted on biology and become set that way for multiple generations). All I know is it has to do with the proteins that encase the DNA. The main interest is that, once we do understand the mechanism, we will be able to control the process. This might be a way of preventing or managing numerous physical and psychiatric health conditions. So, it really will mean the opposite of determinism.

This research reminds me of other scientific and anecdotal evidence. Consider the recipients of organ transplants, blood and bone marrow transfusions, and microbiome transference. This involves the exchange of cells from one body to another. The results have shown changes in mood, behavior, biological functioning, etc

For example, introducing a new microbiome can make a skinny rodent fat or a fat rodent skinny. But also observed are shifts in fairly specific memories, such as an organ transplant recipient craving something the organ donor craved. Furthermore, research has shown that genetics can jump from the introduced cells to the already present cells, which is how a baby can potentially end up with the cells of two fathers if a previous pregnancy was by a different father, and actually it’s rather common for people to have multiple DNAs in their body.

It intuitively makes sense that epigenetics would be behind memory. It’s easy to argue that there is no other function in the body that has this kind and degree of capacity. And that possibility would blow up our ideas of the human mind. In that case, some element of memories would get passed on multiple generations, explaining certain similarities seen in families and larger populations with shared epigenetic backgrounds.

This gives new meaning to the theories of both the embodied mind and the extended mind. There might also having some interesting implications for the bundle theory of mind. I wonder too about something like enactivism which is about the human mind’s relation to the world. Of course, there are obvious connections of this specific research with neurological plasticity and of epigenetics more generally with intergenerational trauma.

So, it wouldn’t only be the symptoms of trauma or else the benefits of privilege (or whatever other conditions that shape individuals, generational cohorts, and sub-populations) being inherited but some of the memory itself. This puts bodily memory in a much larger context, maybe even something along the lines of Jungian thought, in terms of collective memory and archetypes (depending on how long-lasting some epigenetic effects might be). Also, much of what people think of as cultural, ethnic, and racial differences might simply be epigenetics. This would puncture an even larger hole in genetic determinism and race realism. Unlike genetics, epigenetics can be changed.

Our understanding of so much is going to be completely altered. What once seemed crazy or unthinkable will become the new dominant paradigm. This is both promising and scary. Imagine what authoritarian governments could do with this scientific knowledge. The Nazis could only dream of creating a superman. But between genetic engineering and epigenetic manipulations, the possibilities are wide open. And right now, we have no clue what we are doing. The early experimentation, specifically research done covertly, is going to be of the mad scientist variety.

These interesting times are going to get way more interesting.

* * *

Could Memory Traces Exist in Cell Bodies?
by Susan Cosier

The finding is surprising because it suggests that a nerve cell body “knows” how many synapses it is supposed to form, meaning it is encoding a crucial part of memory. The researchers also ran a similar experiment on live sea slugs, in which they found that a long-term memory could be totally erased (as gauged by its synapses being destroyed) and then re-formed with only a small reminder stimulus—again suggesting that some information was being stored in a neuron’s body.

Synapses may be like a concert pianist’s fingers, explains principal investigator David Glanzman, a neurologist at U.C.L.A. Even if Chopin did not have his fingers, he would still know how to play his sonatas. “This is a radical idea, and I don’t deny it: memory really isn’t stored in synapses,” Glanzman says.

Other memory experts are intrigued by the findings but cautious about interpreting the results. Even if neurons retain information about how many synapses to form, it is unclear how the cells could know where to put the synapses or how strong they should be—which are crucial components of memory storage. Yet the work indeed suggests that synapses might not be set in stone as they encode memory: they may wither and re-form as a memory waxes and wanes. “The results are really just kind of surprising,” says Todd Sacktor, a neurologist at SUNY Downstate Medical Center. “It has always been this assumption that it’s the same synapses that are storing the memory,” he says. “And the essence of what [Glanzman] is saying is that it’s far more dynamic.”

Memory Transferred Between Snails, Challenging Standard Theory of How the Brain Remembers
by Usha Lee McFarling

Glanzman’s experiments—funded by the National Institutes of Health and the National Science Foundation—involved giving mild electrical shocks to the marine snail Aplysia californica. Shocked snails learn to withdraw their delicate siphons and gills for nearly a minute as a defense when they subsequently receive a weak touch; snails that have not been shocked withdraw only briefly.

The researchers extracted RNA from the nervous systems of snails that had been shocked and injected the material into unshocked snails. RNA’s primary role is to serve as a messenger inside cells, carrying protein-making instructions from its cousin DNA. But when this RNA was injected, these naive snails withdrew their siphons for extended periods of time after a soft touch. Control snails that received injections of RNA from snails that had not received shocks did not withdraw their siphons for as long.

“It’s as if we transferred a memory,” Glanzman said.

Glanzman’s group went further, showing that Aplysia sensory neurons in Petri dishes were more excitable, as they tend to be after being shocked, if they were exposed to RNA from shocked snails. Exposure to RNA from snails that had never been shocked did not cause the cells to become more excitable.

The results, said Glanzman, suggest that memories may be stored within the nucleus of neurons, where RNA is synthesized and can act on DNA to turn genes on and off. He said he thought memory storage involved these epigenetic changes—changes in the activity of genes and not in the DNA sequences that make up those genes—that are mediated by RNA.

This view challenges the widely held notion that memories are stored by enhancing synaptic connections between neurons. Rather, Glanzman sees synaptic changes that occur during memory formation as flowing from the information that the RNA is carrying.

Reading Voices Into Our Minds

Each of us is a multitude. There is no single unified self. Our thoughts are a conversation. The voices of family echo in our minds when we first leave home and long after our loved ones have died. Then there are all the television, movie, and commercial characters that invade our consciousness with their catchphrases, slogans, and taglines. And we can’t forget how songs get stuck on cognitive repeat or emerge as a compulsion to sing.

Yet another example are the intimate voices imagined as you read novels, a form of inner speech that can carry on after you have put down a book. These can be the most powerful voices. There is nothing that compares to the long periods of time spent with compelling fiction. The voice of characters in a novel are heard within your own head as you read. You can return to this experience again and again, until the characters have become internalized and their words inscribed upon your psyche. Their voices becomes your own voices.

This chorus of voices is constantly playing in the background, a caucophony of thoughts vying for your attention. But occasionally they rise into the spotlight of your consciousness. Even then, it rarely occurs to any of us how strange those voices are, except when some particular voice insistently refuses to go away and maybe even seems to have a mind of its own. Then we might begin to question the distinction between them and us and question what kind of being we are that can contain both.

There is an argument that novels help us develop theory of mind. But maybe in the process novels, along with certain other modern media, result in a particular kind of mind or minds. We come to identify or otherwise incorporate what we empathize with. The worlds we inhabit long enough eventually inhabit us. And what we’ve heard through out our lives can have a way of continuing to speak to us, layers upon layers of voices that for some of us can speak clearly.

* * *

Fictional characters make ‘experiential crossings’ into real life, study finds
by Richard Lea

It’s a cliche to claim that a novel can change your life, but a recent study suggests almost a fifth of readers report that fiction seeps into their daily existence.

Researchers at Durham University conducted a survey of more than 1,500 readers, with about 400 providing detailed descriptions of their experiences with book. Nineteen per cent of those respondents said the voices of fictional characters stayed with them even when they weren’t reading, influencing the style and tone of their thoughts – or even speaking to them directly. For some participants it was as if a character “had started to narrate my world”, while others heard characters talking, or imagined them reacting to things going on in everyday life.

The study, which was carried out in collaboration with the Guardian at the 2014 Edinburgh international book festival, also found that more than half of the 1,500 respondents said that they heard the voices of characters while reading most or all of the time, while 48% reported a similar frequency of visual or other sensory experiences during reading.

According to one of the paper’s authors, the writer and psychologist Charles Fernyhough, the survey illustrates how readers of fiction are doing more than just processing words for meaning – they are actively recreating the worlds and characters being described.

“For many of us, this can involve experiencing the characters in a novel as people we can interact with,” Fernyhough said. “One in seven of our respondents, for example, said they heard the voices of fictional characters as clearly as if there was someone in the room with them.”

When they asked readers to describe what was happening in detail, the researchers found people who described fictional characters remaining active in their minds after they had put the book down, and influencing their thoughts as they went about their daily business – a phenomenon Fernyhough called “experiential crossing”.

The term covers a wide range of experiences, from hearing a character’s voice to feeling one’s own thoughts shaped by a character’s ideas, sensibility or presence, he continued. “One respondent, for example, described ‘feeling enveloped’ by [Virginia Woolf’s] character Clarissa Dalloway – hearing her voice and imagining her response to particular situations, such as walking into a Starbucks. Sometimes the experience seemed to be triggered by entering a real-world setting similar to one in the novel; in other situations, it felt like seeing the world through a particular character’s eyes, and judging events as the character would.”

The characters who make the leap into readers’ lives are typically “powerful, vivid characters and narrators”, Fernyhough added, “but this will presumably vary hugely from person to person”.

* * *

 

Ian Cheng on Julian Jaynes

Down an Internet Rabbit Hole With an Artist as Your Guide
by Daniel McDermon

The art of Ian Cheng, for example, is commonly described in relation to video games, a clear influence. But the SI: Visions episode about him touches only lightly on that connection and on Mr. Cheng’s career, which includes a solo exhibition earlier this year at MoMA PS1. Instead, viewers go on a short but heady intellectual journey, narrated by Mr. Cheng, who discusses improv theater and the esoteric theories of the psychologist Julian Jaynes.

Jaynes, Mr. Cheng said, posits that ancient people weren’t conscious in the way that modern humans are. “You and I hear an internal voice and we perceive it to be a voice that comes from us,” Mr. Cheng says in the video. But Jaynes argued that those voices might well have been perceived as other people.

In that theory, Mr. Cheng explained in an interview, “The mind is actually composed of many sub-people inside of you, and any one of those people is getting the spotlight at any given time.” It’s a model of consciousness that is echoed in the film “Inside Out,” in which an adolescent girl’s mind comprises five different characters.

This conception of consciousness and motivation helped him build out the triad of digital simulations that were shown at MoMA PS1. In those works, Mr. Cheng created characters and landscapes, but the narrative that unfolds is beyond his control. He has referred to them as “video games that play themselves.”

“Lack of the historical sense is the traditional defect in all philosophers.”

Human, All Too Human: A Book for Free Spirits
by Friedrich Wilhelm Nietzsche

The Traditional Error of Philosophers.—All philosophers make the common mistake of taking contemporary man as their starting point and of trying, through an analysis of him, to reach a conclusion. “Man” involuntarily presents himself to them as an aeterna veritas as a passive element in every hurly-burly, as a fixed standard of things. Yet everything uttered by the philosopher on the subject of man is, in the last resort, nothing more than a piece of testimony concerning man during a very limited period of time. Lack of the historical sense is the traditional defect in all philosophers. Many innocently take man in his most childish state as fashioned through the influence of certain religious and even of certain political developments, as the permanent form under which man must be viewed. They will not learn that man has evolved,4 that the intellectual faculty itself is an evolution, whereas some philosophers make the whole cosmos out of this intellectual faculty. But everything essential in human evolution took place aeons ago, long before the four thousand years or so of which we know anything: during these man may not have changed very much. However, the philosopher ascribes “instinct” to contemporary man and assumes that this is one of the unalterable facts regarding man himself, and hence affords a clue to the understanding of the universe in general. The whole teleology is so planned that man during the last four thousand years shall be spoken of as a being existing from all eternity, and with reference to whom everything in the cosmos from its very inception is naturally ordered. Yet everything evolved: there are no eternal facts as there are no absolute truths. Accordingly, historical philosophising is henceforth indispensable, and with it honesty of judgment.

What Locke Lacked
by Louise Mabille

Locke is indeed a Colossus of modernity, but one whose twin projects of providing a concept of human understanding and political foundation undermine each other. The specificity of the experience of perception alone undermines the universality and uniformity necessary to create the subject required for a justifiable liberalism. Since mere physical perspective can generate so much difference, it is only to be expected that political differences would be even more glaring. However, no political order would ever come to pass without obliterating essential differences. The birth of liberalism was as violent as the Empire that would later be justified in its name, even if its political traces are not so obvious. To interpret is to see in a particular way, at the expense of all other possibilities of interpretation. Perspectives that do not fit are simply ignored, or as that other great resurrectionist of modernity, Freud, would concur, simply driven underground. We ourselves are the source of this interpretative injustice, or more correctly, our need for a world in which it is possible to live, is. To a certain extent, then, man is the measure of the world, but only his world. Man is thus a contingent measure and our measurements do not refer to an original, underlying reality. What we call reality is the result not only of our limited perspectives upon the world, but the interplay of those perspectives themselves. The liberal subject is thus a result of, and not a foundation for, the experience of reality. The subject is identified as origin of meaning only through a process of differentiation and reduction, a course through which the will is designated as a psychological property.

Locke takes the existence of the subject of free will – free to exercise political choice such as rising against a tyrant, choosing representatives, or deciding upon political direction – simply for granted. Furthermore, he seems to think that everyone should agree as to what the rules are according to which these events should happen. For him, the liberal subject underlying these choices is clearly fundamental and universal.

Locke’s philosophy of individualism posits the existence of a discreet and isolated individual, with private interests and rights, independent of his linguistic or socio-historical context. C. B. MacPhearson identifies a distinctly possessive quality to Locke’s individualist ethic, notably in the way in which the individual is conceived as proprietor of his own personhood, possessing capacities such as self-reflection and free will. Freedom becomes associated with possession, which the Greeks would associate with slavery, and society conceived in terms of a collection of free and equal individuals who are related to each through their means of achieving material success – which Nietzsche, too, would associate with slave morality.  […]

There is a central tenet to John Locke’s thinking that, as conventional as it has become, remains a strange strategy. Like Thomas Hobbes, he justifies modern society by contrasting it with an original state of nature. For Hobbes, as we have seen, the state of nature is but a hypothesis, a conceptual tool in order to elucidate a point. For Locke, however, the state of nature is a very real historical event, although not a condition of a state of war. Man was social by nature, rational and free. Locke drew this inspiration from Richard Hooker’s Laws of Ecclesiastical Polity, notably from his idea that church government should be based upon human nature, and not the Bible, which, according to Hooker, told us nothing about human nature. The social contract is a means to escape from nature, friendlier though it be on the Lockean account. For Nietzsche, however, we have never made the escape: we are still holus-bolus in it: ‘being conscious is in no decisive sense the opposite of the instinctive – most of the philosopher’s conscious thinking is secretly directed and compelled into definite channels by his instincts. Behind all logic too, and its apparent autonomy there stand evaluations’ (BGE, 3). Locke makes a singular mistake in thinking the state of nature a distant event. In fact, Nietzsche tells us, we have never left it. We now only wield more sophisticated weapons, such as the guilty conscience […]

Truth originates when humans forget that they are ‘artistically creating subjects’ or products of law or stasis and begin to attach ‘invincible faith’ to their perceptions, thereby creating truth itself. For Nietzsche, the key to understanding the ethic of the concept, the ethic of representation, is conviction […]

Few convictions have proven to be as strong as the conviction of the existence of a fundamental subjectivity. For Nietzsche, it is an illusion, a bundle of drives loosely collected under the name of ‘subject’ —indeed, it is nothing but these drives, willing, and actions in themselves—and it cannot appear as anything else except through the seduction of language (and the fundamental errors of reason petrified in it), which understands and misunderstands all action as conditioned by something which causes actions, by a ‘Subject’ (GM I 13). Subjectivity is a form of linguistic reductionism, and when using language, ‘[w]e enter a realm of crude fetishism when we summon before consciousness the basic presuppositions of the metaphysics of language — in plain talk, the presuppositions of reason. Everywhere reason sees a doer and doing; it believes in will as the cause; it believes in the ego, in the ego as being, in the ego as substance, and it projects this faith in the ego-substance upon all things — only thereby does it first create the concept of ‘thing’ (TI, ‘Reason in Philosophy’ 5). As Nietzsche also states in WP 484, the habit of adding a doer to a deed is a Cartesian leftover that begs more questions than it solves. It is indeed nothing more than an inference according to habit: ‘There is activity, every activity requires an agent, consequently – (BGE, 17). Locke himself found the continuous existence of the self problematic, but did not go as far as Hume’s dissolution of the self into a number of ‘bundles’. After all, even if identity shifts occurred behind the scenes, he required a subject with enough unity to be able to enter into the Social Contract. This subject had to be something more than merely an ‘eternal grammatical blunder’ (D, 120), and willing had to be understood as something simple. For Nietzsche, it is ‘above all complicated, something that is a unit only as a word, a word in which the popular prejudice lurks, which has defeated the always inadequate caution of philosophers’ (BGE, 19).

Nietzsche’s critique of past philosophers
by Michael Lacewing

Nietzsche is questioning the very foundations of philosophy. To accept his claims means being a new kind of philosopher, ones who ‘taste and inclination’, whose values, are quite different. Throughout his philosophy, Nietzsche is concerned with origins, both psychological and historical. Much of philosophy is usually thought of as an a priori investigation. But if Nietzsche can show, as he thinks he can, that philosophical theories and arguments have a specific historical basis, then they are not, in fact, a priori. What is known a priori should not change from one historical era to the next, nor should it depend on someone’s psychology. Plato’s aim, the aim that defines much of philosophy, is to be able to give complete definitions of ideas – ‘what is justice?’, ‘what is knowledge?’. For Plato, we understand an idea when we have direct knowledge of the Form, which is unchanging and has no history. If our ideas have a history, then the philosophical project of trying to give definitions of our concepts, rather than histories, is radically mistaken. For example, in §186, Nietzsche argues that philosophers have consulted their ‘intuitions’ to try to justify this or that moral principle. But they have only been aware of their own morality, of which their ‘justifications’ are in fact only expressions. Morality and moral intuitions have a history, and are not a priori. There is no one definition of justice or good, and the ‘intuitions’ that we use to defend this or that theory are themselves as historical, as contentious as the theories we give – so they offer no real support. The usual ways philosophers discuss morality misunderstands morality from the very outset. The real issues of understanding morality only emerge when we look at the relation between this particular morality and that. There is no world of unchanging ideas, no truths beyond the truths of the world we experience, nothing that stands outside or beyond nature and history.

GENEALOGY AND PHILOSOPHY

Nietzsche develops a new way of philosophizing, which he calls a ‘morphology and evolutionary theory’ (§23), and later calls ‘genealogy’. (‘Morphology’ means the study of the forms something, e.g. morality, can take; ‘genealogy’ means the historical line of descent traced from an ancestor.) He aims to locate the historical origin of philosophical and religious ideas and show how they have changed over time to the present day. His investigation brings together history, psychology, the interpretation of concepts, and a keen sense of what it is like to live with particular ideas and values. In order to best understand which of our ideas and values are particular to us, not a priori or universal, we need to look at real alternatives. In order to understand these alternatives, we need to understand the psychology of the people who lived with them. And so Nietzsche argues that traditional ways of doing philosophy fail – our intuitions are not a reliable guide to the ‘truth’, to the ‘real’ nature of this or that idea or value. And not just our intuitions, but the arguments, and style of arguing, that philosophers have used are unreliable. Philosophy needs to become, or be informed by, genealogy. A lack of any historical sense, says Nietzsche, is the ‘hereditary defect’ of all philosophers.

MOTIVATIONAL ANALYSIS

Having long kept a strict eye on the philosophers, and having looked between their lines, I say to myself… most of a philosopher’s conscious thinking is secretly guided and channelled into particular tracks by his instincts. Behind all logic, too, and its apparent tyranny of movement there are value judgements, or to speak more clearly, physiological demands for the preservation of a particular kind of life. (§3) A person’s theoretical beliefs are best explained, Nietzsche thinks, by evaluative beliefs, particular interpretations of certain values, e.g. that goodness is this and the opposite of badness. These values are best explained as ‘physiological demands for the preservation of a particular kind of life’. Nietzsche holds that each person has a particular psychophysical constitution, formed by both heredity and culture. […] Different values, and different interpretations of these values, support different ways of life, and so people are instinctively drawn to particular values and ways of understanding them. On the basis of these interpretations of values, people come to hold particular philosophical views. §2 has given us an illustration of this: philosophers come to hold metaphysical beliefs about a transcendent world, the ‘true’ and ‘good’ world, because they cannot believe that truth and goodness could originate in the world of normal experience, which is full of illusion, error, and selfishness. Therefore, there ‘must’ be a pure, spiritual world and a spiritual part of human beings, which is the origin of truth and goodness. Philosophy and values But ‘must’ there be a transcendent world? Or is this just what the philosopher wants to be true? Every great philosophy, claims Nietzsche, is ‘the personal confession of its author’ (§6). The moral aims of a philosophy are the ‘seed’ from which the whole theory grows. Philosophers pretend that their opinions have been reached by ‘cold, pure, divinely unhampered dialectic’ when in fact, they are seeking reasons to support their pre-existing commitment to ‘a rarefied and abstract version of their heart’s desire’ (§5), viz. that there is a transcendent world, and that good and bad, true and false are opposites. Consider: Many philosophical systems are of doubtful coherence, e.g. how could there be Forms, and if there were, how could we know about them? Or again, in §11, Nietzsche asks ‘how are synthetic a priori judgments possible?’. The term ‘synthetic a priori’ was invented by Kant. According to Nietzsche, Kant says that such judgments are possible, because we have a ‘faculty’ that makes them possible. What kind of answer is this?? Furthermore, no philosopher has ever been proved right (§25). Given the great difficulty of believing either in a transcendent world or in human cognitive abilities necessary to know about it, we should look elsewhere for an explanation of why someone would hold those beliefs. We can find an answer in their values. There is an interesting structural similarity between Nietzsche’s argument and Hume’s. Both argue that there is no rational explanation of many of our beliefs, and so they try to find the source of these beliefs outside or beyond reason. Hume appeals to imagination and the principle of ‘Custom’. Nietzsche appeals instead to motivation and ‘the bewitchment of language’ (see below). So Nietzsche argues that philosophy is not driven by a pure ‘will to truth’ (§1), to discover the truth whatever it may be. Instead, a philosophy interprets the world in terms of the philosopher’s values. For example, the Stoics argued that we should live ‘according to nature’ (§9). But they interpret nature by their own values, as an embodiment of rationality. They do not see the senselessness, the purposelessness, the indifference of nature to our lives […]

THE BEWITCHMENT OF LANGUAGE

We said above that Nietzsche criticizes past philosophers on two grounds. We have looked at the role of motivation; the second ground is the seduction of grammar. Nietzsche is concerned with the subject-predicate structure of language, and with it the notion of a ‘substance’ (picked out by the grammatical ‘subject’) to which we attribute ‘properties’ (identified by the predicate). This structure leads us into a mistaken metaphysics of ‘substances’. In particular, Nietzsche is concerned with the grammar of ‘I’. We tend to think that ‘I’ refers to some thing, e.g. the soul. Descartes makes this mistake in his cogito – ‘I think’, he argues, refers to a substance engaged in an activity. But Nietzsche repeats the old objection that this is an illegitimate inference (§16) that rests on many unproven assumptions – that I am thinking, that some thing is thinking, that thinking is an activity (the result of a cause, viz. I), that an ‘I’ exists, that we know what it is to think. So the simple sentence ‘I think’ is misleading. In fact, ‘a thought comes when ‘it’ wants to, and not when ‘I’ want it to’ (§17). Even ‘there is thinking’ isn’t right: ‘even this ‘there’ contains an interpretation of the process and is not part of the process itself. People are concluding here according to grammatical habit’. But our language does not allow us just to say ‘thinking’ – this is not a whole sentence. We have to say ‘there is thinking’; so grammar constrains our understanding. Furthermore, Kant shows that rather than the ‘I’ being the basis of thinking, thinking is the basis out of which the appearance of an ‘I’ is created (§54). Once we recognise that there is no soul in a traditional sense, no ‘substance’, something constant through change, something unitary and immortal, ‘the way is clear for new and refined versions of the hypothesis about the soul’ (§12), that it is mortal, that it is multiplicity rather than identical over time, even that it is a social construct and a society of drives. Nietzsche makes a similar argument about the will (§19). Because we have this one word ‘will’, we think that what it refers to must also be one thing. But the act of willing is highly complicated. First, there is an emotion of command, for willing is commanding oneself to do something, and with it a feeling of superiority over that which obeys. Second, there is the expectation that the mere commanding on its own is enough for the action to follow, which increases our sense of power. Third, there is obedience to the command, from which we also derive pleasure. But we ignore the feeling the compulsion, identifying the ‘I’ with the commanding ‘will’. Nietzsche links the seduction of language to the issue of motivation in §20, arguing that ‘the spell of certain grammatical functions is the spell of physiological value judgements’. So even the grammatical structure of language originates in our instincts, different grammars contributing to the creation of favourable conditions for different types of life. So what values are served by these notions of the ‘I’ and the ‘will’? The ‘I’ relates to the idea that we have a soul, which participates in a transcendent world. It functions in support of the ascetic ideal. The ‘will’, and in particular our inherited conception of ‘free will’, serves a particular moral aim

Hume and Nietzsche: Moral Psychology (short essay)
by epictetus_rex

1. Metaphilosophical Motivation

Both Hume and Nietzsche1 advocate a kind of naturalism. This is a weak naturalism, for it does not seek to give science authority over philosophical inquiry, nor does it commit itself to a specific ontological or metaphysical picture. Rather, it seeks to (a) place the human mind firmly in the realm of nature, as subject to the same mechanisms that drive all other natural events, and (b) investigate the world in a way that is roughly congruent with our best current conception(s) of nature […]

Furthermore, the motivation for this general position is common to both thinkers. Hume and Nietzsche saw old rationalist/dualist philosophies as both absurd and harmful: such systems were committed to extravagant and contradictory metaphysical claims which hinder philosophical progress. Furthermore, they alienated humanity from its position in nature—an effect Hume referred to as “anxiety”—and underpinned religious or “monkish” practises which greatly accentuated this alienation. Both Nietzsche and Hume believe quite strongly that coming to see ourselves as we really are will banish these bugbears from human life.

To this end, both thinkers ask us to engage in honest, realistic psychology. “Psychology is once more the path to the fundamental problems,” writes Nietzsche (BGE 23), and Hume agrees:

the only expedient, from which we can hope for success in our philosophical researches, is to leave the tedious lingering method, which we have hitherto followed, and instead of taking now and then a castle or village on the frontier, to march up directly to the capital or center of these sciences, to human nature itself.” (T Intro)

2. Selfhood

Hume and Nietzsche militate against the notion of a unified self, both at-a-time and, a fortiori, over time.

Hume’s quest for a Newtonian “science of the mind” lead him to classify all mental events as either impressions (sensory) or ideas (copies of sensory impressions, distinguished from the former by diminished vivacity or force). The self, or ego, as he says, is just “a kind of theatre, where several perceptions successively make their appearance; pass, re-pass, glide away, and mingle in an infinite variety of postures and situations. There is properly no simplicity in it at one time, nor identity in different; whatever natural propension we may have to imagine that simplicity and identity.” (Treatise 4.6) […]

For Nietzsche, the experience of willing lies in a certain kind of pleasure, a feeling of self-mastery and increase of power that comes with all success. This experience leads us to mistakenly posit a simple, unitary cause, the ego. (BGE 19)

The similarities here are manifest: our minds do not have any intrinsic unity to which the term “self” can properly refer, rather, they are collections or “bundles” of events (drives) which may align with or struggle against one another in a myriad of ways. Both thinkers use political models to describe what a person really is. Hume tells us we should “more properly compare the soul to a republic or commonwealth, in which the several members [impressions and ideas] are united by ties of government and subordination, and give rise to persons, who propagate the same republic in the incessant change of its parts” (T 261)

3. Action and The Will

Nietzsche and Hume attack the old platonic conception of a “free will” in lock-step with one another. This picture, roughly, involves a rational intellect which sits above the appetites and ultimately chooses which appetites will express themselves in action. This will is usually not considered to be part of the natural/empirical order, and it is this consequence which irks both Hume and Nietzsche, who offer two seamlessly interchangeable refutations […]

Since we are nothing above and beyond events, there is nothing for this “free will” to be: it is a causa sui, “a sort of rape and perversion of logic… the extravagant pride of man has managed to entangle itself profoundly and frightfully with just this nonsense” (BGE 21).

When they discover an erroneous or empty concept such as “Free will” or “the self”, Nietzsche and Hume engage in a sort of error-theorizing which is structurally the same. Peter Kail (2006) has called this a “projective explanation”, whereby belief in those concepts is “explained by appeal to independently intelligible features of psychology”, rather than by reference to the way the world really is1.

The Philosophy of Mind
INSTRUCTOR: Larry Hauser
Chapter 7: Egos, bundles, and multiple selves

  • Who dat?  “I”
    • Locke: “something, I know not what”
    • Hume: the no-self view … “bundle theory”
    • Kant’s transcendental ego: a formal (nonempirical) condition of thought that the “I’ must accompany every perception.
      • Intentional mental state: I think that snow is white.
        • to think: a relation between
          • a subject = “I”
          • a propositional content thought =  snow is white
      • Sensations: I feel the coldness of the snow.
        • to feel: a relation between
          • a subject = “I”
          • a quale = the cold-feeling
    • Friedrich Nietzsche
      • A thought comes when “it” will and not when “I” will. Thus it is a falsification of the evidence to say that the subject “I” conditions the predicate “think.”
      • It is thought, to be sure, but that this “it” should be that old famous “I” is, to put it mildly, only a supposition, an assertion. Above all it is not an “immediate certainty.” … Our conclusion is here formulated out of our grammatical custom: “Thinking is an activity; every activity presumes something which is active, hence ….” 
    • Lichtenberg: “it’s thinking” a la “it’s raining”
      • a mere grammatical requirement
      • no proof of an thinking self

[…]

  • Ego vs. bundle theories (Derek Parfit (1987))
    • Ego: “there really is some kind of continuous self that is the subject of my experiences, that makes decisions, and so on.” (95)
      • Religions: Christianity, Islam, Hinduism
      • Philosophers: Descartes, Locke, Kant & many others (the majority view)
    • Bundle: “there is no underlying continuous and unitary self.” (95)
      • Religion: Buddhism
      • Philosophers: Hume, Nietzsche, Lichtenberg, Wittgenstein, Kripke(?), Parfit, Dennett {a stellar minority}
  • Hume v. Reid
    • David Hume: For my part, when I enter most intimately into what I call myself, I always stumble on some particular perception or other, of heat or cold, light or shade, love or hatred, pain or pleasure.  I never can catch myself at any time without a perception, and never can observe anything but the perception.  (Hume 1739, Treatise I, VI, iv)
    • Thomas Reid: I am not thought, I am not action, I am not feeling: I am something which thinks and acts and feels. (1785)