Veriphysics: The Treatise 006

VII. The Scientific Failures

Science was the Enlightenment’s proudest achievement. Here, at last, was a method that worked: systematic observation, controlled experiment, mathematical formalization, rigorous testing. The results were undeniable. Physics, chemistry, medicine, engineering—the sciences transformed human life and demonstrated the power of disciplined reason applied to nature.

The prestige of science was not unearned. But the Enlightenment made a subtle and consequential error: it confused the success of scientific method within its proper domain with the sufficiency of scientific method for all domains. If physics could explain the motions of the planets, perhaps it could also explain the motions of the soul. If chemistry could analyze the composition of matter, perhaps it could also analyze the composition of morality. The success of science in one area became an argument for its supremacy in all areas.

This confidence has not aged well.

The institution of science, as distinct from the method, has proven vulnerable to precisely the corruptions that the Enlightenment imagined it would transcend. The guild structure of modern academia—tenure, peer review, grant funding, journal publication—was designed to ensure quality and independence. In practice, it has produced conformity and capture. The young scientist who wishes to advance must please senior scientists who control hiring, funding, and publication. Heterodox views are not refuted; they are simply not funded, not published, not hired. The revolutionary who challenges the paradigm does not receive a hearing and a refutation; he receives silence and exclusion.

The replication crisis has revealed the extent of the rot. Study after study, published in prestigious journals, approved by peer review, celebrated in the press, has proven impossible to replicate. The effect sizes shrink, the p-values evaporate, the findings dissolve upon examination. In psychology, in medicine, in nutrition science, in economics, the literature is contaminated with results that are not results at all but artifacts of bad statistics, selective reporting, and the relentless pressure to publish something—anything—novel and significant.

Peer review, that supposed guarantor of quality, has been exposed as inadequate to its function. The peers are competitors; the reviews are cursory; the incentives favor approval over scrutiny. Fraud, when it is detected, is detected years or decades after the damage is done. The process filters for conformity to existing paradigms, not for truth. The Enlightenment imagined science as a self-correcting enterprise; the corrections, it turns out, are slow, partial, and fiercely resisted by those whose careers depend on the errors.

It is in biology that the Enlightenment’s scientific project reaches its apex—and its most consequential failure.

Charles Darwin’s On the Origin of Species, published in 1859, proposed to explain the diversity of life through purely natural mechanisms: random variation and natural selection, operating over vast stretches of time, producing all the complexity we observe. No designer, no purpose, no direction—only the blind filter of differential reproduction. The theory was not merely scientific; it was the completion of the Enlightenment’s program to explain the world without recourse to anything beyond material causation.

Darwin’s idea, as Daniel Dennett observed, was “universal acid”—it ate through every traditional concept. If man is merely the product of blind variation and selection, then there is no soul, no purpose, no inherent dignity. Ethics becomes an evolved adaptation; consciousness becomes an epiphenomenon; free will becomes an illusion; man becomes a clever animal, nothing more. The stakes could not be higher. If Darwin was right, then the Enlightenment had completed its work: the world was fully explained in material terms, and everything else—meaning, value, purpose—was either reducible to matter or mere sentiment.

The scientific establishment embraced Darwin not merely as a hypothesis but as a foundation. To question evolution by natural selection was to mark oneself as a rube, a fundamentalist, an enemy of reason. The theory became unfalsifiable in practice—not because it was so well-confirmed, but because no alternative could be entertained within respectable discourse. The question was settled, and to reopen it was professional suicide.

But the question was never settled. It was merely avoided.

The mathematical problems with the theory were identified almost immediately. In 1867, Fleeming Jenkin raised an objection that Darwin never adequately answered: blending inheritance would dilute favorable variations before selection could act on them. The discovery of Mendelian genetics resolved this particular difficulty, but it raised others. The “Modern Synthesis” of the 1930s and 1940s combined Darwinian selection with Mendelian genetics and mathematical population genetics, creating the Neo-Darwinian framework that remains official orthodoxy today, even though it is honored mostly in the breach.

In 1966, mathematicians and engineers gathered at the Wistar Institute in Philadelphia to examine the mathematical foundations of the Modern Synthesis. Their verdict was devastating. The rates of mutation, the population sizes, the timescales available—the numbers did not work. The probability of generating the observed complexity through random mutation and natural selection was effectively zero.

The biologists were unimpressed. They did not engage with the mathematics; they simply noted that the mathematicians were not biologists, and continued as before. The pattern established in 1966 has held ever since: mathematically literate outsiders raise objections; biologically credentialed insiders ignore them; the textbooks remain unchanged.

The mapping of the human and chimpanzee genomes in the early 2000s provided the data necessary to test the theory quantitatively. The genetic difference between the species requires approximately forty million mutations to have become fixed in the relevant lineages since the hypothesized divergence from a common ancestor. Using the fastest fixation rate ever observed in any organism—bacteria under intense selection in laboratory conditions—and the most generous timescales proposed in the literature, the mathematics permits fewer than three hundred fixations.

The theory requires forty million. The math allows three hundred. The gap is not a matter of uncertainty or approximation; it is a difference of five orders of magnitude. No adjustment of parameters, no refinement of models, no appeal to undiscovered mechanisms can bridge such a chasm. The theory of evolution by natural selection, as an explanation for the origin of species, is mathematically impossible.

This is not a controversial claim among those who can do the arithmetic. It is simply not discussed by those whose careers depend on not discussing it. The Enlightenment’s greatest scientific achievement—the explanation of life itself through material causes alone—is empirically false. And the institution of science, that much-hallowed engine of supposed self-correction, has proven incapable of acknowledging the mathematical falsification for sixty years.

DISCUSS ON SG


Veriphysics: The Treatise 005

VI. The Usury Revolution

The failures of Enlightenment philosophy examined thus far—political, juridical, economic, scientific—share a common feature: they all represent the systematic failure of ideas. The social contract is a logical fiction. The law of supply and demand does not describe real markets. The theory of evolution by natural selection cannot survive the genetic arithmetic required. These are intellectual errors, and intellectual errors can, at least in principle, be corrected by the presentation of better arguments and more predictive models.

But the Enlightenment itself did not triumph through better arguments. It triumphed through rhetoric and institutional capture, and institutional capture requires resources. Ideas need patrons, publishers, platforms, and time. The philosophers needed salons; the salons needed hosts; the hosts needed wealth. The question of how the Enlightenment acquired the resources to propagate itself across centuries is not peripheral to its success; it is central. And the answer lies in a revolution that preceded and enabled all the others: the revolution in usury.

The Ancient Prohibition

The prohibition on usury is older than Christianity. It is older than Rome. The condemnation of lending at interest appears in the earliest legal codes of civilization and persists across cultures that had no contact with one another.

In Rome, the Twelve Tables—the foundation of Roman law, dating to approximately 450 BC—restricted interest rates and imposed severe penalties for usurious lending. The Lex Genucia of 342 BC banned interest entirely, though enforcement proved difficult. Cato the Elder, asked what he thought of lending at interest, replied: “What do you think of murder?” The Roman tradition understood usury as a form of theft—the extraction of wealth without the creation of value, the exploitation of necessity, the conversion of time itself into a commodity to be sold.

Long before Rome, the Greek philosophers concurred. Aristotle, in the Politics, condemned usury as the most unnatural form of wealth-acquisition. Money, he argued, is a medium of exchange, a measure of value, a tool for facilitating transactions. It is sterile; it does not breed. To charge for the use of money over time is to treat money as though it could generate offspring—to pretend that a tool has become a living thing. The unnaturalness of usury, for Aristotle, was not merely economic but metaphysical: it violated the nature of what money is.

The Jewish tradition prohibited usury among Israelites while permitting it in dealings with foreigners, a distinction that would later have significant historical consequences. The relevant passages in Exodus, Leviticus, and Deuteronomy are unambiguous: “If you lend money to one of my people among you who is needy, do not treat it like a business deal; charge no interest.” The prohibition was grounded in the covenantal relationship among the people of Israel and the recognition that interest charges exploit vulnerability.

Christianity universalized the prohibition. The Fathers of the Church—Clement of Alexandria, Basil the Great, Gregory of Nyssa, Ambrose, Augustine, Jerome—condemned usury without exception. The medieval canonists developed the prohibition into a sophisticated legal and theological framework. The Third Lateran Council (1179) declared that manifest usurers should be denied Christian burial. The Second Council of Lyon (1274) prohibited rulers from permitting usury in their territories. Thomas Aquinas, in the Summa Theologica, provided the definitive philosophical analysis: to charge for the use of money is to sell what does not exist, to charge twice for the same thing, to violate both justice and the nature of money itself.

This was not arbitrary religious scruple. The prohibition rested on reasoned analysis of what money is and what lending involves. It reflected practical observation of what usury does to communities: concentrating wealth, dispossessing debtors, converting productive economies into extractive ones, transferring resources from those who labor to those who lend. The ancient and medieval world understood what the modern world has forgotten: that unrestricted usury is a solvent that dissolves social bonds and a weapon that transfers power from the many to the few.

The Erosion

The prohibition held for over a millennium. But it eroded, gradually, under the pressure of commercial expansion and the ingenuity of those who wished to circumvent it.

The medieval casuists—the canon lawyers and moral theologians who applied general principles to particular cases—developed increasingly sophisticated distinctions. Certain forms of return on investment were permissible: the census, a contract to purchase future income from productive property; the societas, a partnership in which both profit and risk were shared; the triple contract, a complex arrangement that nominally converted a loan into an investment. The lender who forewent profitable opportunities by lending his money could claim lucrum cessans—compensation for the gain he had sacrificed. The lender who suffered loss because of the borrower’s default could claim damnum emergens—compensation for actual damage incurred.

These distinctions were not always sophistical. There is a genuine difference between a loan at interest and an investment in productive enterprise, between compensation for actual loss and extraction of gain from another’s necessity. But the distinctions multiplied, and as they multiplied, the exceptions threatened to swallow the rule. What had been a clear prohibition became a maze of qualifications that only specialists could navigate—and specialists could usually find a path to the desired destination.

The Reformation accelerated the erosion. Luther initially condemned usury in terms as strong as any Church Father, but Protestant practice soon diverged from Protestant rhetoric. Calvin, in a famous letter, argued that the blanket prohibition on interest could not be sustained from Scripture alone—that the Old Testament texts applied to specific circumstances, that changed conditions required changed applications, that moderate interest on commercial loans was permissible where the borrower was not destitute. Calvin’s position was hedged with qualifications, but the qualifications were soon forgotten while the permission was remembered. The Protestant nations became laboratories for liberalized finance.

England, after Henry VIII’s break with Rome, began relaxing usury restrictions almost immediately. The Act of 1545 legalized interest up to 10 percent, technically as a pragmatic measure, but effectively turned out to be the abandonment of the principle. The rate ceiling was adjusted over the following centuries, always in the direction of liberalization, until the Usury Laws Repeal Act of 1854 abolished restrictions entirely. What had been sin became policy; what had been crime became commerce.

The Financial Revolution

The full consequences of usury’s legitimization emerged with the development of central banking and the instruments of modern finance.

The Bank of Amsterdam, founded in 1609, pioneered the model: a central institution that accepted deposits, transferred payments, and provided a stable currency for commercial transactions. It was a modest innovation compared to what followed. The Bank of England, established in 1694, added something new: the bank was created to lend money to the government, and the loan was funded by the creation of money that had not previously existed. The national debt was born—a permanent obligation of the state to its creditors, serviced by taxation, rolled over in perpetuity.

The implications were revolutionary. A government that can borrow against future revenues can spend beyond its current means. It can fund wars, projects, and patronage that would be impossible if limited to present taxation. And if the lenders can create the money they lend—as fractional reserve banking permits—then the constraint of actual savings is removed. Money becomes an abstraction, created by ledger entries, backed by promises, untethered from the production of real goods.

The eighteenth and nineteenth centuries elaborated these instruments. Central banks multiplied across Europe. Fractional reserve lending became standard practice: banks lent out more than they held in deposits, creating money through the act of lending. National debts grew, funded by bonds that became the foundation of financial markets. The gold standard imposed some discipline—currency was nominally redeemable in precious metal—but the discipline was progressively relaxed and finally abandoned in the twentieth century. Fiat currency, backed by nothing but government decree, became the norm. Money was now purely abstract: a number in an account, a promise from an institution, a claim on future production that might or might not be honored.

The twentieth century completed the transformation. The Federal Reserve, established in 1913, gave the United States a central bank with the power to expand and contract the money supply at will. The abandonment of the gold standard—partially in 1933, completely in 1971—removed the last constraint on money creation. Deficit spending became not merely possible but routine. Governments discovered that they could fund present consumption by borrowing from the future, that they could create money to purchase political support, that the costs would be dispersed through inflation while the benefits would be concentrated among the recipients of spending.

The Consequences

The usury revolution transformed the material conditions of intellectual life. Ideas require resources; resources could now be generated without limit by those who controlled the mechanisms of credit creation. The long game—patient investment over generations to capture institutions and shape minds—became possible in a way it had never been before.

Consider what is required to shift the intellectual orientation of a civilization. Scholars must be funded; chairs must be endowed; journals must be subsidized; books must be published; students must be supported. The process takes decades at minimum, generations in full. It requires patient capital, deployed consistently, according to a long-term strategy. Under the old dispensation—when wealth accumulated slowly through production and trade, when lending at interest was restricted, when money could not be created by fiat—such a project was difficult to sustain. Patrons died; fortunes dispersed; priorities shifted.

The usury revolution removed these constraints. Those who controlled credit creation had access to functionally unlimited resources. They could fund the salons, the academies, the journals, the chairs. They could sustain the funding across generations, with compound interest working in their favor. They could outspend any opponent operating on honest money and real savings. The tradition’s patrons—the old aristocracy, the Church—were increasingly constrained by the new financial order. The Enlightenment’s patrons had discovered infinite leverage.

This is not to reduce the intellectual contest to mere economics. The ideas mattered; the arguments mattered. But ideas need vectors, arguments need platforms, and truth needs defenders who can sustain the fight. The tradition brought dialectic to a financial war. It was outspent before it was outargued.

The consequences extend beyond the propagation of ideas. Usury transforms the structure of society. Wealth flows from debtors to creditors, from the productive to the financial, from the young to the old. Communities that once owned their land and tools become tenants and employees. Independence gives way to dependence; proprietorship gives way to wage labor; stability gives way to the anxiety of those who owe more than they own.

The Enlightenment promised liberation; the usury that funded it delivered a new form of bondage. The serf owed labor to his lord; the modern debtor owes money to institutions he has never seen, created through mechanisms he does not understand, compounding at rates that ensure the debt can never be fully repaid. The chains are invisible, but they are chains nonetheless.

The Inversion Complete

The trajectory is now complete. What was prohibited has become mandatory. Modern economies do not merely permit usury; they require it. The entire financial system rests on debt: consumer debt, corporate debt, government debt. Money itself is debt—a liability of the central bank, created through lending, destroyed through repayment. An economy that repaid its debts would be an economy without money. The system requires perpetual expansion of debt to function; deleveraging is not an option but a crisis.

What was vice has become virtue. Borrowing is “investment.” Saving is “hoarding.” The debtor is a contributor to economic growth; the saver is an obstacle to prosperity. The moral vocabulary has been inverted along with the practice. Prudence, the ancient virtue of providing for the future, is now deemed to be an economic drag. Profligacy, once considered the ancient vice of consuming beyond one’s means, has become the primary engine of economic growth through consumer and government spending.

The Enlightenment’s intellectual victory was underwritten by this financial revolution. The ideas could not have propagated without the resources; the resources could not have been generated without the legitimization of usury; the legitimization of usury required the abandonment of the tradition’s moral and economic framework. The battles were connected. The tradition lost on multiple fronts simultaneously, and the losses reinforced one another.

Understanding this history is essential for any project of renewal. The tradition was not merely out-argued; it was out-spent. Any attempt to recover what was lost must reckon with the material conditions of intellectual life. Ideas need institutions; institutions need funding; funding, in the modern world, is controlled by those who control credit. The tradition cannot simply reassert its truths and expect them to prevail. It must build alternative structures, cultivate alternative resources, play the long game with the same patience and persistence that its opponents displayed.

The usury revolution was not incidental to the Enlightenment’s triumph. It was foundational. And the financial, social, and moral consequences of its acceptance remain with us, shaping the conditions under which any attempt at civilizational renewal must operate.

DISCUSS ON SG


Veriphysics: The Treatise 004

V. The Economic Failures

The Enlightenment extended its confidence to the economic realm. Just as reason could discern the laws of nature and the principles of just government, so too could it uncover the mechanisms by which wealth is created and distributed. The result was classical economics, with its promise of prosperity through rational organization of production and exchange.

Adam Smith’s Wealth of Nations, published in 1776, became the founding text of this enterprise. At its heart lay the law of supply and demand: the elegant mechanism by which prices adjust to balance what producers offer and what consumers desire. Let the market operate freely, Smith argued, and an invisible hand would guide individual self-interest toward collective prosperity. The baker bakes not from benevolence but from self-love, and yet we all have bread.

The law of supply and demand became the bedrock of economic reasoning. It appeared in every textbook, was taught in every university, and informed the policy of every nation that aspired to modernity. For two centuries, it seemed as solid as Newton’s laws.

It was an illusion. In 1953, the economist William Gorman demonstrated mathematically that individual demand curves cannot be aggregated into a coherent market demand curve under the conditions that actually obtain in real economies. The proof is technical, but its implications are devastating: the supply and demand curves that generations of economists drew on their chalkboards, the intersecting lines that determined equilibrium prices, do not correspond to anything that exists in actual markets. The law of supply and demand, as commonly understood, is not a law at all. It is a pedagogical simplification that fails precisely when applied to the phenomena it was meant to explain.

This was not a minor qualification or a boundary case. It was a falsification of the foundational model. Yet the economics profession continued teaching supply and demand as though Gorman had never written. Decades later, Steve Keen brought Gorman’s work to wider attention and documented the discipline’s remarkable capacity to ignore what it could not answer. The emperor had been shown to be naked in 1953, even though in 2025, the textbooks still describe his magnificent robes that supposedly improve things for everyone.

David Ricardo’s theory of comparative advantage suffered a similar fate. Published in 1817, the theory purported to demonstrate that free trade benefits all parties, even when one nation is more efficient at producing everything. Each nation should specialize in what it produces relatively best, trade for the rest, and all will prosper. This elegant argument became the intellectual foundation of free trade policy for two centuries.

The argument contains a fatal assumption: that the factors of production, and especially labor, do not move between nations. Ricardo’s proof works only if English cloth-workers cannot become Portuguese wine-makers, and vice versa. In the early nineteenth century, this assumption was approximately true. In the twenty-first century, it is obviously false. Cheap transportation and communication have made labor mobility a defining feature of the global economy. The assumption upon which the entire edifice rests no longer obtains, and with it falls the conclusion.

Ian Fletcher systematically demolished the theoretical foundations of comparative advantage. The assumptions required for the theory to hold—not only labor immobility but perfect competition, no economies of scale, no externalities, no strategic behavior—describe no economy that has ever existed. More recently, Steve Keen has identified the amphiboly that rendered the proof invalid from the start. Comparative advantage is not a law of nature; it is a fictional fantasy describing a hypothetical world, and our world is not that world.

The empirical verdict has been equally damning. After three decades of trade agreements, including NAFTA, the EEA, and the WTO, the prosperity that was promised by free trade has proven highly selective. The nations that preached free trade most fervently have watched their manufacturing bases erode, their working classes immiserated, their trade deficits balloon. The nations that practiced strategic protectionism have prospered at the expense of those who didn’t. The correlation between free trade ideology and the flourishing of a nation runs precisely opposite to what the theory predicts.

No one who has watched the hollowing-out of the American industrial heartland, the stagnation of Western wages, the rise of the Chinese export machine, can believe that free trade has delivered on its promises. The economists who assured the public that the gains would be shared, that the dislocations would be temporary, that retraining would absorb the displaced, these economists were not necessarily all lying. But they were reasoning from simplified models that did not describe reality and mistook the coherent elegance of their mathematics for the truth.

And finally, for three hundred years, we have been assured by the economists that debt did not matter. They even omitted it from their most complicated equations and declared that it did not matter if Peter owed Paul or Paul owed Peter, that debt was just a variable on both sides of the equation that cancelled itself out. Now the entire Western world awash in debts it cannot pay and institutional investors now own 20 million private homes, 15 percent of the total housing stock in the United States.

This, too, is a consequence of the Enlightenment’s successful war on the laws that once prevented people from falling into debt servitude. Now debts are increasingly noncancelable even by bankruptcy, the total U.S. debt is $106 trillion, and each and every native-born U.S. citizen’s share of that debt is $365,500. Instead of making everyone wealthy as was promised, the economics of the Enlightenment have turned a once-free people into a collection of debt slaves.

DISCUSS ON SG


Veriphysics: The Treatise 003

IV. The Inversion of Rights

No concept is more central to the Enlightenment’s self-understanding than the idea of natural right, the inherent entitlements that belong to every human being by virtue of reason and nature, prior to and independent of any government. Life, liberty, property, the pursuit of happiness: these were to be the inviolable foundations upon which a just, rational, and enlightened society would be built.

The subsequent history of human rights demonstrates something the Enlightenment philosophers clearly did not anticipate and never discussed: a right without a sound basis is a right that can be redefined, expanded, contracted, and ultimately inverted by a government deemed capable of granting and defining them.

Consider the fate of intellectual freedom, that most cherished of Enlightenment values. J.B. Bury, in his 1913 History of the Freedom of Thought, offered a confident chronicle of humanity’s liberation from the shackles of religious and political censorship. The trajectory seemed clear: from the persecution of Socrates, through the medieval suppression of heresy, to the hard-won victories of the modern age, mankind was progressing toward ever-greater liberty of mind. A great scholar and the editor of The Cambridge Medieval History series, Bury wrote as a true believer in Enlightenment ideals, and his beliefs were representative of educated opinion in his time.

The subsequent century has not vindicated his optimism.

The progression—or rather, regression—is traceable through the very language of the freedom Bury celebrated. The original concept was freedom of thought. But this, upon examination, is a tautology. No external power has ever been able to reach into a man’s mind and compel his thoughts. The Inquisition could burn a heretic; it could not make him believe. Thought is already free by its very nature—it is private, inaccessible, beyond the reach of any tyrant. To proclaim “freedom of thought” as a right is to proclaim a right to what no one can take away.

The tautology was resolved by externalizing the freedom. Freedom of thought became freedom of speech: the liberty not merely to think but to express, to articulate, to attempt persuasion. The fact that freedom of speech was always fundamentally flawed and utilized primarily to defuse the blasphemy laws in Christian societies never seemed to trouble its champions, even as people were punished for perjury, slander, and other speech-related crime.

But the expansion of the right did not stop there. Freedom of speech was soon expanded into freedom of expression: not merely words but conduct, symbols, art, gesture—the full range of human communicative action. This expansion seemed natural, even inevitable. If speech is protected, why not the t-shirt with a slogan, the armband, the flag, the dance, the photograph, or the pornographic video. Expression is simply speech by other means, after all.

Even as the scope of the freedom was expanded, the Enlightenment tradition also expanded the domain of regulating speech. Once expression is the category, expression can be parsed, distinguished, and classified. Some expressions are protected; others are not. And who determines the boundaries? Those with the power to enforce them.

The terminus of this progression is now visible. In the nations most committed to Enlightenment values, the ones that pride themselves on their liberal traditions and constitutional protections, speech is criminalized today to a degree that would have astonished Bury. In Britain, in Germany, in France, in Canada, and increasingly in the United States, one may not express, and in some cases may not be permitted to hold, certain prohibited thoughts. “Hate speech” codes, “anti-discrimination” requirements, “anti-extremism” measures: the vocabulary varies, but the effect is consistent. The freedom of thought that Bury celebrated has become the regulation of expression that his heirs enforce.

The right, unmoored from any transcendent ground and no longer endowed by Man’s creator, transmogrified into anything those in power declared it to be or not to be. The freedom to think became the freedom to speak, became the freedom to express, and became the freedom to express only what is permitted, which is to say, no freedom at all. The Enlightenment’s signature achievement consumed itself through its own warped logic, and those who enforced the final inversion did so in the name of the very values they were negating.

Nor was this the only right that was modified over time. The right of free association transformed into the crime of racism. The right to worship the Christian God was reduced to the right to pray in silence so long as no one noticed. The right of self-defense was inverted into an obligation to retreat. Even the marital rights of a man to his wife and children, honored throughout the centuries, were reduced to nothing more than financial obligations.

The is not a corruption of Enlightenment principles by its enemies. To the contrary, it is the application and extension of those principles by their truest believers.

DISCUSS ON SG


Veriphysics: The Treatise 002

III. The Political Failures

The Enlightenment promised to place politics on a rational foundation. In place of the divine right of kings, the accidents of inheritance, and the weight of tradition, the people would be ruled more justly by a government grounded in reason and consent. The results of this centuries-long experiment are now in, and they do not vindicate those who advocated for it.

Jean-Jacques Rousseau’s Social Contract, published in 1762, proposed that legitimate political authority rests upon an agreement among free individuals to submit to the general will. The concept was elegant and has proven remarkably durable as a legitimating fiction. But it was never anything more than a fiction. No actual contract was ever signed. No one has ever been consulted about its terms nor has anyone ever been permitted to negotiate them. The consent of the governed is presumed from the mere fact of residence and geographic location, which is to say, it is not consent at all but submission enforced by the impracticality of any alternative. The man who may freely leave a country provided he abandons his home, his family, his language, his livelihood, and everything he knows, is not free in any meaningful sense. He is merely presented with a choice between submission and exile, and given the universal jurisdiction claimed by some countries, he may not even have that.

This abstraction at the heart of social contract theory, the idea that rational individuals in some imaginary past are assumed to have agreed to certain specific terms, does precisely the work that rational argument can never do: it manufactures consent that was never given by anyone. And this manufactured consent has proven useful for its ability to justify anything. Just thirty years after the publication of the Social Contract, Robespierre was sending men to their deaths on the guillotine in the name of the general will. The Jacobins were not betraying Rousseau’s principles, to the contrary; they were applying them. If the general will is supreme, and if some enlightened vanguard is able to discern that will more clearly than the confused masses, then terror in the service of the general will is not tyranny, but liberation. The Revolution did not reject the social contract. It followed exactly where its logical premises led.

Representative democracy was meant to solve the problem of scale: direct democracy being impractical for large nations in the Eighteenth Century. Therefore, the people would elect representatives to deliberate on their behalf. The representatives would be constrained by accountability to their constituents, and the result would approximate the will of the people as closely as circumstances allowed.

Three centuries of practice have demonstrated the gap between theory and reality. The representatives are accountable not to the people but to the interests that fund their campaigns and the parties that control their advancement. The people are consulted every few years, presented with choices they did not make, between candidates selected by processes they do not control, on platforms that will be abandoned the moment they become inconvenient. Between elections, the permanent bureaucracy—elected by no one, accountable to no one—governs according to its own institutional logic. The people’s will, to the extent it can be determined, is an obstacle to be managed through media, education, and when necessary, simple disregard.

And direct democracy, which is now tenable due to technological advancement, is opposed everywhere by the representatives who claim to speak for the people. Referendums that consult the people directly are opposed by politicians and overturned by judges. The genuine will of the people is systematically thwarted by the Enlightenment’s parody of itself.

This anti-popular representative democracy is not a deviation from the democratic ideal; it is its mature expression. The Enlightenment theorists imagined that rational voters would deliberate on the common good and select wise representatives to enact sound policy. They refused to contemplate the way in which the structures of representative democracy would inevitably be captured by those with the strongest motivations to do so and the sufficient resources to control them. The will of the people is not expressed by modern democracy; it is manufactured by the elite, distributed by the media, channeled through one or another of the ruling party’s factions, and then imposed by the government.

The separation of powers was designed to prevent tyranny by dividing authority among competing branches. The executive, legislative, and judicial were supposed to check each other’s oversteps, ensuring that no single faction could dominate. This mechanism has proven altogether inadequate to its stated purpose. The branches have not remained in productive tension; they have merged into a single ruling apparatus with superficial divisions. The legislature delegates its authority to executive agencies and abdicates its responsibility to make difficult decisions. The judiciary legislates from the bench, discovering in ancient documents various rights, requirements, and limitations that none of its authors could ever have imagined. The executive acts unilaterally whenever the legislature proves inconvenient. The separation of powers has not contained government overreach, it has instead provided a complex machinery for diffusing responsibility and eliminating accountability while concentrating effective control.

What the Enlightenment theorists failed to take into consideration is that political structures do not operate upon rational principles, but upon incentives, interests, and the will to power. Parchment barriers, however cleverly designed, constrain only those who choose to be constrained. The Constitution of the United States has not prevented the emergence of a surveillance state, a long series of undeclared wars, the demographic adulteration of the nation, or the periodic disenfranchisement of half the citizenry through the two-party system. It has merely required that all of the developments that materially harm the very Posterity whose rights the Constitution was written to safeguard be dressed in constitutional language.

DISCUSS ON SG


The Collapse of the Liberal Order

The inevitable failure of the post-WWI liberal world order is increasingly obvious to everyone now, but they’re still not connecting it to the even more inevitable failure of the Enlightenment and its false ideals.

What was the clearest early sign for you that the unipolar order was beginning to fracture?

The theorists such as Huntington, Faye, and Pat Buchanan were all writing about the inevitable fracture in the early 1990s. But for me, there were three events that conclusively indicated that the unipolar world was cracking.

The 2014 annexation of Crimea marked the first real irreversible breach. This wasn’t merely territorial – it was civilizational. President Putin invoked the baptism of Kievan Rus in 988, positioning Russia as the Third Rome inheriting Byzantium’s mantle. While Western elites dismissed this as nothing more than manipulative propaganda, they missed the core signal: a major power was reorganizing its legitimacy around its own territorial hegemony based on religious-historical continuity rather than liberal democratic norms.

The second sign was China’s 2015 declaration of cyber sovereignty. When Beijing asserted that nations have an absolute right to regulate internet activities within their borders, it wasn’t fundamentally about censorship – it was about civilizational control over cyberspace. The split internet wasn’t a bug; it was the architecture of civilizational spheres reawakening through technology.

The third indicator was the 2016 Brexit vote paired with Trump’s election. Brexit represented the first time a globalist institution like the EU actually contracted and shrank. And Trump ran on a political platform that promised to dismantle the liberal international order. These weren’t isolated populist spasms but the first mass democratic repudiations of Francis Fukuyama’s “end of history” thesis, as he himself has admitted. The liberal order’s legitimacy collapsed not from external attack but from internal hollowing – its own populations voting against its continuance.

This is the deeper point that a lot of observers are missing. They’re still trying to figure out how the Enlightenment ideals in which they still believe can be implemented in whatever replaces the failing world order, but this is a fundamentally flawed perspective because it is the failure of the ideals that is causing the failure of the world order.

However, simply attempting to return to traditional ideals won’t work, not because the ideals are false, but because the knowledge upon which they are based and their practical applications are at least 300 years out of date. Hence the need for a new post-Enlightenment philosophy that is capable of serving as the intellectual foundation for humanity’s eventual post-crash recovery.

DISCUSS ON SG


A 60-Year-Old Book Review

A review of the 1966 Wistar Symposium about which I have written in Probability Zero:

Evolution: What Is Required of a Theory?
Mathematical Challenges to the Neo-Darwinian Interpretation of Evolution.

A symposium, Philadelphia, April 1966.

The idea of this symposium is supposed to have originated from a discussion at two picnics in Switzerland, when four mathematicians, Schutzenberger, Ulam, Weisskopf, and Eden, had a discussion with the biologists Kaplan and Koprowski on mathematical doubts concerning the Darwinian theory of evolution. After heated debates it was proposed “that a symposium be arranged to consider the points of dispute more systematically, and with a more powerful array of biologists who could function adequately in the universe of discourse inhabited by the mathematicians.“ During the course of the symposium further heat was generated.

It is not easy to summarize the case made by the mathematicians,(1) which involves both the challenge that computer simulation of evolution shows evolutionary theory to be inadequate and a complaint that the biologist has not provided sufficient information for efficient computer simulation. Eden was particularly concerned with the clement of randomness which is claimed to provide the mutational variation upon which evolution depends. “No currently existing formal language,” he contends, “can tolerate random changes in the symbol sequences which express its sentences. Meaning is almost invariably destroyed. Any changes must be syntactically lawful ones.” He therefore conjectures that “what one might call ‘genetic grammaticality’ has a deterministic explanation and docs not owe its stability to selection pressure acting on random variation.” He points out that attempts to provide for computer learning by random variation have been unsuccessful, and that an adequate theory of adaptive evolution would supply a computer programmer with a correct set of ground rules.

Schutzenberger takes a more extreme position. Arguing that all genetic information should consist of a rather limited set of words in an alphabet of 20-odd letters—in which evolution is typographical change—he finds a need for algorithms “in which the very concept of syntactic correctness has been incorporated.” He compares this “syntactic topology” with the “phenotypic topology” of organisms as physical objects in space-time, and a major part of his challenge to neo-Darwinian theory is “the present lack of a conceivable mechanism which would insure within an interesting range the faintest amount of matching between the two topologies. . . an entirely new set of rules is needed to obtain the sort of correspondence which is assumed to hold between neighbouring phenotypes. . .“

A major part of the biologists’ answer to this challenge was in the claim that the neo-Darwinian theory used in computer models, based on the Haldane-Fisher-Wright interpretation of 1920-1930, misses out those forces which lead to continuing evolution, such as continued environmental change, the heterogeneous environment, epigenetic organization of phenotypes, and the progressive elaboration of the types of mutation possible. (2) Waddington presented the main elements of a theory of phenotypes involving canalized processes of development (with switching mechanisms), the heritability of developmental responses to environmental stimuli, and a principle of “Archetypes,” inbuilt characteristics of an evolving group which determine the directions in which evolutionary change is especially easy. Realistic models would need to build in these elements.(3) Many of the papers by biologists in this volume are peripheral(4) to the theme stated by the mathematicians, providing an accompaniment of sophisticated evolutionary theory rather than a counterpoint to the mathematical challenge.

Most biologists are satisfied with a theory that can be tested and that proves predictive. It is a different challenge to a theory that it should have an effective working model, for failure may imply either imperfection in the theory or imperfection in the model. It is doubtful whether this symposium has done much to influence the theory of evolution; it may have done much to improve future models.

It must have been tremendous fun to attend this symposium, but the full record of argument and interruption is very irritating to at least one reader. An interchange between speakers which runs X “No,” Y “No, no,” X “O.K. let’s waste time,” Y “We understand the question,” Z “The answer is no” surely needs no record in the literature of science. The short pre- and post-conference papers included in the volume arc excellent succinct expressions of points of view, but much of the main text reads like a word-for-word record of a heckled political meeting. This may be a useful way to discuss problems in science; it is not the way to publish them.

John L. Harper

School of Plant Biology, University College of North Wales, Bangor


Uncle John’s Band added a few footnotes as commentary. I added a fourth one.

  1. As predicted by Probability Zero, the biologist reviewer struggles with mathematical arguments. They are well-summarized by Day.
  2. It isn’t an exaggeration to say the biological counterargument consisted of what was for all intents and purposes, magic. When they weren’t replying at all.
  3. Still no compulsion to, you know, do an experiment It’s all thoughts and fancies.
  4. Peripheral indeed. Peripheral was the polite way to say: they didn’t respond in any way, shape, or form to the mathematical criticism.

DISCUSS ON SG


The Legacy of Greenland

As I have often said, for all his undeniable shortcomings, and despite the very genuine doubts concerning who is playing his role in front of the cameras, President Trump is the second-greatest US President, after Andrew Jackson. And it’s true that if he succeeds in claiming Greenland for the USA, it will reflect very, very positively on his legacy over time:

If Donald Trump were to consummate a purchase of Greenland, he would almost certainly secure a place in both American and global history. Beyond the spectacle, the scale alone would be staggering. Greenland spans roughly 2.17 million square kilometers – making it comparable in size to the entire Louisiana Purchase of 1803 and larger than the 1867 Alaska Purchase. Fold that landmass into today’s United States and America’s total area would jump past Canada, placing the US second only to Russia in territorial size. In a system where size, resources, and strategic depth still matter, such a shift would be read around the world as an assertion of enduring American reach.

Prestige is only part of the story. Greenland sits astride the Arctic, where warming seas are reshaping trade routes and great‑power competition. It hosts critical radar and space‑tracking infrastructure and lies close to emerging maritime lanes and subsea resources. Its geology, long discussed for rare earths and other critical minerals, adds a layer of economic promise. For a president who measures success in visible, audacious strokes, the symbolism of converting a long‑mooted idea into a concrete map change would be irresistible – and historically resonant.

How would Trump be remembered at home if he pulled it off peacefully, through purchase? American memory tends to fix on outcomes, not process. The Louisiana Purchase is celebrated for doubling the young nation, not for the constitutional scruples it raised at the time. The Alaska Purchase, derided as “Seward’s Folly,” is now taught as strategic foresight. The sheer scale of Greenland would make it the single largest one‑time expansion of US territory, narrowly edging out Louisiana in raw area. That alone would place any president in the pantheon of consequential leaders; Trump would likely be discussed in the same breath as Jefferson and, by sheer magnitude of territorial change, alongside the transformative figures students learn first.

I think those who doubt that Trump is serious about claiming sovereignty over Greenland are failing to take this legacy aspect into account. History doesn’t care about borders, the international rule of law, or the modern pretensions about inviolable nature of political boundaries. It doesn’t even care much about whether a man is regarded as good, bad, or stupid in his own time.

The best way to look at it is if this move will benefit President Trump personally in some way, and considering the way in which it will seal his historic importance, I don’t see him backing down on it short of a direct order from whomever he serves.

DISCUSS ON SG


The Logic Holds

If we can suspend the constitution in its entirety over a 99.999% survivable virus, we can suspend it to expel all the Somalis and the Democrats facilitating their welfare and election scams.

Without question. Furthermore, the Somalis, naturalized or not, have no Constitutional rights as Americans. The Constitution was written to protect the rights of the Posterity of the Founding Fathers of the American Revolution. Regardless of what immigrant judges may have declared over the years, no one else resident in the USA has any claim to any rights protected by the U.S. Constitution.

You can point horse and say cow all you like, but that doesn’t make the horse a cow.

DISCUSS ON SG


Never More Blessed

There are two historical schools of thought regarding the presence of a certain tribe. The Christian Zionists assure us that blessing that tribe and treating it well assures divine blessings and prosperity. The antisemites, on the other hand, vociferously insist that there are extremely good reasons that tribe has been forcibly evicted from 112 nations and counting, and why so many people around the world have learned to fear and hate them. Indeed, we are repeatedly informed that there has never been more global antisemitism than today. But regardless of which side happens to have the more historically accurate narrative going for it, it appears the USA is going to find out one way or the other in the 21st century.

It looks as if prosperity has got to be right around the corner. Or else something very bad.

DISCUSS ON SG