Archive for the ‘Metaphysics’ Category
I’ve been cleaning up my “Indefinite Divisibility” paper from last year. One of my arguments in it concerned supergunk: X is supergunk iff for every chain of parts of X, there is some y which is a proper part of each member of the chain. I claimed that supergunk was possible, and argued on that basis against absolutely unrestricted quantification. I even thought I had a kind of consistency proof for supergunk: in particular, a (proper class) model that satisfied the supergunk condition as long as the plural quantifier was restricted to set-sized collections. Call something like this set-supergunk.
Well, I was wrong. I’ve been suspicious for a while, and I finally proved it today: set-supergunk is impossible. So I thought I’d share my failure. In fact, an even stronger claim holds:
Theorem. If is atomless, then has a countable chain of parts such that nothing is a part of each of them.
Proof. Since is atomless, there is a (countable) sequence . For each positive integer , let be . Then let be the sum of . Note that the ’s are a countable chain. Note also that each is part of .
Now suppose that is a part of each . In that case, is part of each . But since is disjoint from , this means that is disjoint from each , and so by the definition of a mereological sum, is disjoint from each . This is a contradiction.
“The whole is nothing over and above the parts.” This is a nice thought, but it turns out to be difficult to make precise. One attempt is the “composition as identity” thesis: if the Xs compose y, the Xs are y.
This won’t work, at least not without great cost. The United States is composed of fifty states, and it is also composed of 435 congressional districts. If composition is identity, then the U.S. is the states, and the U.S. is the districts; thus the states are the districts. This is bad: composition-as-identity collapses mereologically coextensive pluralities, which means now your plural logic can be no more powerful than your mereology. So you lose the value of even having plural quantifiers. That’s a big sacrifice. (This argument is basically Ted Sider’s, in “Parthood”.)
But the problem here isn’t that the fusion of the Xs is something more than the mere Xs: rather, the fusion is something less. Mereological sums are less fine-grained than pluralities, so if we require each plurality to be identical to a particular sum, we lose the (important!) distinctions that plural logic makes.
This suggests a better way: mereological sums are abstractions from pluralities. Roughly speaking, sums are pluralities with some distinctions ignored. In particular, sums are what you get by abstracting from pluralities on the relation of being coextensive. (Analogously: colors are what you get when you abstract from objects on the same-color relation. Numbers are what you get when you abstract from pluralities on equinumerosity.)
Let’s polish this up a bit. Take overlap as primitive, and define parthood in the standard way:
- x is part of y iff everything that overlaps x overlaps y.
This has a natural plural generalization:
- The Xs are covered by the Ys iff everything that overlaps some X overlaps some Y.
Parthood is the limiting case of being covered when there’s just one X and one Y. (I’ll identify each object with its singleton plurality.) We can also define an equivalence relation:
- The Xs are coextensive with the Ys iff the Xs cover the Ys and the Ys cover the Xs.
Now we can state an abstraction principle. Let Fus be a new primitive function symbol taking one plural argument.
- Fus X = Fus Y iff the Xs are coextensive with the Ys.
(Compare Hume’s Principle: #X = #Y iff the Xs are equinumerous with the Ys.) This is the main principle governing composition. It isn’t the only principle we’ll need. For all I’ve said so far, fusions could live in Platonic heaven; but we need them to participate in mereological relations:
- The following are equivalent:
- The Ys cover the Xs.
- The Ys cover Fus X.
- Fus Y covers the Xs.
This guarantees that Fus X really is the fusion of the Xs by the standard definition of “fusion”. There is one final assumption needed to ensure that our mereology is standard:
- Parthood is antisymmetric. (If x is part of y and y is part of x, then x = y.)
Equivalently: Fus x = x. In the singular case, composition really is identity.
These three principles imply all of standard mereology. So just how innocent are they?
I think they’re fairly innocent, given the right conception of how abstraction works. I like a “tame” account of abstraction which doesn’t introduce any new ontological commitments. (This means tame abstraction is too weak for Frege arithmetic or for Frege’s Basic Law V—this is a good thing.) The basic idea is that abstract terms refer indefinitely to each of their instances. For example, the singular term “red” refers indefinitely to each red thing: we consider all red instances as if they were a single thing, without being specific as to which. (Semantically, you can understand indefinite reference in terms of supervaluations.) Red has the properties that all red things must share. E.g., if any red thing must be rosier than any taupe thing, then we can also say that red is rosier than taupe. Speaking of red doesn’t commit to any new entity—it’s just speaking of the old entities a new way.
As for colors, so for fusions. “The fusion of the Xs” doesn’t refer to some new exotic thing: it refers indefinitely to each plurality coextensive with the Xs. You could say it refers to the Xs, as long as you don’t mind the difference between coextensive pluralities. Furthermore, since whenever the Xs are coextensive with the Ys they stand in exactly the same covering relations, Principle 2 is justified.
Principle 3, on the other hand, is not entirely innocent. Given the definition of parthood, it amounts to extensionality: no distinct (singular) objects are coextensive. I think it’s right to consider this a separate, serious commitment, one that (unlike the rest of mereology) doesn’t flow from the mere conception of a mereological sum. It might, however, flow from the conception of an object. If you aren’t too worried about speaking completely fundamentally, antisymmetry can be had cheaply, by considering “objects” to be coextension-abstractions from the basic objects, in just the same way that sums are coextension-abstractions from the basic pluralities.
So, indeed, the whole is nothing more than its parts. It can’t be identified with any particular plurality of its parts, but it can be identified indefinitely with every plurality of its parts.
[There’s a technical issue for the semantics I’ve alluded to here. I’m treating Fus X as semantically plural (it refers indefinitely to pluralities), but it is syntactically singular. In particular, as a singular term it can be ascribed membership in pluralities. But this means that I need the semantics to allow pluralities to be members of pluralities—and so on—and this isn’t ordinarily allowed. So it looks like I’ll need to give the semantics in terms of “superplurals”. (See section 2.4 of the SEP article on plural quantifiers.) Whether this semantic richness should be reflected in the language is a separate issue—I’m inclined to think not, but I haven’t really thought it through.]
Does the universe come “facts first” or “laws first”? That is, in terms of metaphysical priority, do the non-nomic facts determine what the laws of nature are, or are the laws at the ground floor determining what the non-nomic facts are? (Or maybe neither grounds the other; I’ll ignore this view for now.) The best-known example of a facts-first theory is Lewis’s “best system” account: to be a law of nature is to be a member of the set of generalizations over the non-nomic facts that has the best balance of simplicity and strength. Here are two rough-and-ready arguments against an account like that. The first is the circularity argument I gestured at a few weeks ago:
- The laws explain the non-nomic facts.
- If Y explains X, then X does not explain Y.
- If X grounds Y, then X explains Y.
- So the non-nomic facts don’t ground the laws.
And this is the second:
- The non-nomic facts are many and disparate; the laws are simple and few.
- Prefer metaphysical theories that are simpler and more parsimonious at the fundamental level.
- So prefer laws-first to facts-first metaphysics.
The second premise is a methodological principle, rather than a general metaphysical claim (hence the imperative). It’s a ceteris paribus principle, and so the conclusion is a ceteris paribus conclusion: there is a presumption in favor of laws-first accounts.
In “The Nature of Laws” (1977), Michael Tooley claims that it is nomologically true that Fs are Gs just in case a certain relation R—the nomological relation—holds between the universal Fness and the universal Gness. He claims further that this relation-symbol “R” is a theoretical term whose referent is fixed in Ramsey-Lewis style: we specify some constraints C, and then stipulate that R is the unique relation that satisfies C. These are his constraints:
- R is a two-place relation on universals.
- R is contingent: there are universals Fness and Gness such that it is neither necessary that R(Fness, Gness) nor necessary that not R(Fness, Gness).
- R(Fness, Gness) logically entails that all Fs are Gs.
(I’ve left out the complications that are introduced to deal with laws that have different forms, such as “All non-F’s are G’s or H’s” (Tooley doesn’t think there are negative or disjunctive universals). But Tooley thinks there is a different nomological relation associated with each syntactic construction, so this doesn’t make a difference here.)
But it doesn’t look at all plausible to me that these constraints pick out a unique relation (assuming anything satisfies them at all). Look, here’s a non-nomological relation that satisfies conditions 1–4: the relation denoted by the two-place quantifier “All”—that is, the relation that holds between Fness and Gness just in case all Fs are Gs. Tooley hasn’t said anything that would distinguish his nomological relation from such run-of-the-mill categorical relations. This strikes me as a serious problem. Am I missing something?
(I’m working through David Armstrong’s What is a law of nature? now—I’ll see if he adds anything helpful.)
EDIT: I inadvertently left out one of Tooley’s constraints:
- R is irreducibly second-order.
You might think this might help. You might say in particular that the “All” relation is in fact reducible to less-than-second-order universals only—since, after all, “All(F, G)” holds iff for every x, if Fx then Gx. But this “reduction” involves the concept “for every”, which plausibly involves the “All”-relation in disguise. (Analogously, one might “reduce” a purported nomological relation R by pointing out that “R(F, G)” holds iff it is nomologically necessary that for every x, if Fx then Gx.) I guess I’m not really sure what the rules are for reducing universals.
Armstrong makes a conjecture along the same lines: “I speculate that the laws of nature constitute the only irreducibly second-order relations between universals” (84). So presumably he thinks that either there is no “All”-relation, or else that it is reducible to a lower order. Does anyone have an idea why he would think this?
One of my goals over spring break is to get familiar with some of the literature on laws of nature. I may blog some thoughts on it as I go.
This afternoon I read Michael Tooley’s “The Nature of Laws” (in the anthology edited by John Carroll). In the section on the epistemology of laws, Tooley shows how we could become confident that a certain law holds, in a Bayesian framework. He then argues that this confirmation story is a distinctive benefit of his account (the DTA account):
[T]here is a crucial assumption that seems reasonable if relations among universals are the truth-makers for laws, but not if facts about particulars are the truth-makers. This is the assumption that m and n [the prior probabilities of certain statements of laws] are not equal to zero. If one takes the view that it is facts about the particulars falling under a generalization that make it a law, then, if one is dealing with an infinite universe, it is hard to see how one can be justified in assigning any non-zero probability to a generalization, given evidence concerning only a finite number of instances. For surely there is some non-zero probability that any given particular will falsify the generalization, and this entails, given standard assumptions, that as the number of particulars becomes infinite, the probability that the generalization will be true is, in the limit, equal to zero.
In contrast, if relations among universals are the truth-makers for laws, the truth-maker for a given law is, in a sense, an “atomic” fact, and it would seem perfectly justified, given standard principles of confirmation theory, to assign some non-zero probability to this fact’s obtaining.
This can’t be right. If Tooley is right in the first paragraph that the probability of any universal generalization over particulars is zero, then appealing to the “atomicity” of nomological facts is no help. The problem is that, on his own view, the nomological relation between universals logically entails the corresponding universal generalization over particulars. But this means that, by monotonicity, the probability of the relation can be no greater than the probability of the generalization. So if the generalization has zero probability, so too does the relation.
The upshot is that if Tooley’s point in the first paragraph is right, then it’s devastating for just about any account of the epistemology of laws—because any account of laws will have it that a generalization being true-by-law entails it being plain-old-true. So we’d better figure out why Tooley’s point is wrong.
This distribution principle looks awfully plausible:
- A explains (B and C) iff (A explains B and A explains C).
But I think it might be false, at least in the right-to-left direction.
A potential counterexample comes from Aristotle. A bandit is lingering by a road, and a farmer is walking home down the same road. By chance (as we would say), they meet at place X at time T. There is a telic explanation in terms of the bandit’s purposes for his being at X at T, and there is a telic explanation in terms of the farmer’s purposes for his being at X at T. But there is no telic explanation for their meeting, even though I take it that their meeting just consists in both of them being at X at T. The meeting is just a coincidence.
Slightly more carefully: let BP be the bandit’s purposes and FP be the farmer’s purposes, and let BXT and FXT be the relevant location facts. Assuming that antecedent strengthening holds for “explains”, this means that (BP and FP) telically explains BXT, and (BP and FP) telically explains FXT. But since their meeting is coincidental, it seems plausible that (BP and FP) does not telically explain (BXT and FXT).
If distribution fails for telic explanation, then perhaps it fails for nomic explanations as well, for the same kind of reason.
Why does this matter? It’s relevant to my criticism last week of Lewis’s “best system” account of laws (not just my criticism). Briefly, I said: on Lewis’s account, the qualitative facts explain what the laws are, but the laws should explain the qualitative facts. That makes a very tight explanatory circle, and that’s bad.
A response might go: it’s true that the conjunction of the qualitative facts explain the laws. It’s also true that the laws explain each individual qualitative fact. But it doesn’t follow that the laws explain the conjunction of the qualitative facts—since the distributive principle fails—and so there is no bad explanatory circle.
I’ve been thinking more about the problem of fit I posted last week. Specifically, I’m trying to work out how a response appealing to reference magnetism would go.
Recall the puzzle: how is it that when we select hypotheses that best exemplify our theoretical values, we so often hit on the truth? A simple example: emeralds, even those we haven’t observed, are green, rather than grue. And lo, we believe they are green, rather than believing they are grue. It seems things could have been otherwise, in either of two ways:
- There might be people who project grue rather than green, in a world like ours.
- Or there might be people who (like us) project green, in a world where emeralds are grue.
Those people are in for a shock. Why are we so lucky?
A response in the Lewisian framework goes like this. Not all properties are created equal: green is metaphysically more natural than grue. In particular, it is semantically privileged: it is easier to have a term (or thought) about green than it is to have a term (or thought) about grue. This should take care of possibility 1. If there are people who theorize in terms of grue rather than green, their practices would have to be sufficiently perverse to overcome the force of green’s reference magnetism. There are details to fill in, but plausibly it would be hard for natural selection to produce creatures with such perverse practices.
But this still leaves possibility 2. Given that our theories are attracted to the natural properties, even so, why should a theory in terms of natural properties be true? The green-projectors in the world of grue emeralds have just as natural a theory as ours, to no avail.
But even though 2 is possible, we can still explain why it doesn’t obtain. What we need to explain is why emeralds are green—and we shouldn’t try to explain that by appeal to general metaphysics, but by something along these lines: the electrons in a chromium-beryllium crystal can only absorb photons with certain amounts of energy. That is, we explain why emeralds are green by appeal to the natural laws of our world.
Generalizing: “joint-carving” theories yield true predictions because their predictions are supported by natural laws. Why is this? On the Lewisian “best system” account of laws, it is partly constitutive of a natural law that it carve nature at the joints: naturalness is one of the features that distinguishes laws from mere accidental generalizations. So, much as reference magnetism makes it harder to have a theory that emeralds are grue than it is to have a theory that emeralds are green, so the best system account makes it harder to have a law that emeralds are grue than it is to have a law that emeralds are green. Then the idea is that, since our theories and our laws are both drawn to the same source, this makes it likely that they line up. Furthermore, since the laws explain the facts, this explains why our theories fit the facts.
Something isn’t right about this story; I’m having a hard time getting it clear, but here’s a stab. There’s a general tension in the best system account: on the one hand, the laws are supposed to explain the (non-nomic) facts; on the other hand, the (non-nomic) facts are metaphysically prior to the laws. But metaphysical priority is also an explanatory relation, and so it looks like we’re in a tight explanatory circle. (Surely this point has been made? I don’t know much of the literature on laws, so I’d welcome any pointers.)
This is relevant because the answer to the problem of fit relies on the explanatory role of laws—a role that seems difficult for the best systems account to bear up. But I feel pretty shaky on this, and would appreciate help.