Propositions, Terms

Brandom puts significant emphasis on Kant and Frege’s focus on whole judgments — contrasted with simple first-order terms, corresponding to natural-language words or subsentential phrases — as the appropriate units of logical analysis. The important part of this is that a judgment is the minimal unit that can be given inferential meaning.

All this looks quite different from a higher-order perspective. Mid-20th century logical orthodoxy was severely biased toward first-order logic, due to foundationalist worries about completeness. In a first-order context, logical terms are expected to correspond to subsentential elements that cannot be given inferential meaning by themselves. But in a higher-order context, this is not the case. One of the most important ideas in contemporary computer science is the correspondence between propositions and types. Generalized terms are interpretable as types, and thus also as propositions. This means that (higher-order) terms can represent instances of arbitrarily complex propositions. Higher-order terms can be thus be given inferential meaning, just like sentential variables. This is all in a formal context rather than a natural-language one, but so was Frege’s work; and for what it’s worth, some linguists have also been using typed lambda calculus in the analysis of natural language semantics.

Suitably typed terms compose, just like functions or category-theoretic morphisms and functors. I understand the syllogistic principle on which Aristotle based a kind of simultaneously formal and material term inference (see Aristotelian Propositions) to be just a form of composition of things that can be thought of as functions or typed terms. Proof theory, category theory, and many other technical developments explicitly work with composition as a basic form of abstract inference. Aristotle developed the original compositional logic, and it was not Aristotle but mid-20th century logical orthodoxy that insisted on the centrality of the first-order case. Higher-order, compositionally oriented logics can interpret classic syllogistic inference, first-order logic, and much else, while supporting more inferentially oriented semantics on the formal side, with types potentially taking pieces of developed material-inferential content into the formal context. We can also use natural-language words to refer to higher-order terms and their inferential significance, just as we can capture a whole complex argument in an appropriately framed definition. Accordingly, there should be no stigma associated with reasoning about terms, or even just about words.

In computer-assisted theorem-proving, there is an important distinction between results that can be proved directly by something like algebraic substitution for individual variables, and those that require a more global rewriting of the context in terms of some previously proven equivalence(s). At a high enough level of simultaneous abstraction and detail, such rewriting could perhaps constructively model the revision of commitments and concepts from one well-defined context to another.

The potential issue would be that global rewriting still works in a higher-order context that is expected to itself be statically consistent, whereas revision of commitments and concepts taken simply implies a change of higher-level context. I think this just means a careful distinction of levels would be needed. After all, any new, revised genealogical recollection of our best thoughts will be in principle representable as a new static higher-order structure, and that structure will include something that can be read as an explanation of the transition. It may itself be subject to future revision, but in the static context that does not matter.

The limitation of such an approach is that it requires all the details of the transition to be set up statically, which can be a lot of work, and it would also be far more brittle than Brandom’s informal material inference. (See also Categorical “Evil”; Definition.)

I am fascinated by the fact that typed terms can begin to capture material as well as purely formal significance. How complete or adequate this is would depend on the implementation.

Higher Order

Before and after early modern mechanism and in contrast to it, Plato, Aristotle, Kant, and Hegel all broadly agreed on the normative importance of higher-order things.

In modern terms, Plato’s forms are higher-order things, as distinct from first-order things. Plato trusts higher-order things more than first-order ones, because he considers only higher-order things to be knowable in the sense of episteme, because only higher-order things contain an element of universality, and episteme applies only to universals, not particulars.

Aristotle agrees that higher-order things are ultimately more knowable, but believes it is possible to say more about first-order things, by relating them to each other and to higher-order things; that our initial rough, practical grasp of first-order things can help us to begin to grasp higher-order things by example; and that going up and down the ladder of abstraction successively can help us progressively enrich our understanding of both.

(Incidentally, I have always read the Platonic dialogues as emphasizing the normative importance of acquiring a practical grasp of forms more than specific existence claims about “the forms”. Aristotle’s criticisms make it clear that at least some in the Platonic Academy did understand Plato as making such existence claims, but particularly in what are regarded as later dialogues like Parmenides, Sophist, and Theaetetus, what is said about form seems relatively close to an Aristotelian view. It is even possible that these dialogues were influenced by the master’s even greater student.)

Early modern mechanism completely discarded Plato and Aristotle’s higher-order orientation. Descartes famously recommends that we start by analyzing everything into its simplest components. This temporarily played a role in many great scientific and technological advances, but was bad for philosophy and for people. Hegel calls this bottom-up approach Understanding, as distinct from Reason.

Early and mid-20th century logical foundationalism still adhered to this resolutely bottom-up view, but since the later 20th century, there has been an explosion in the use of higher-order formal concepts in mathematics, logic, and computer science. It turns out that even from an engineering point of view, higher-order representations are often more efficient, due to their much greater compactness.

Leibniz already tried to reconcile mechanistic science with a higher-order normative view. He also contributed to the early development of higher-order concepts in mathematics.

Kant and Hegel decisively revived an approach that is simultaneously higher-order and normative.