There are many different scaling laws. At one extreme, there are simple scaling laws that are easy to learn, easy to use, and very useful in everyday life. Scaling laws can be and should be introduced at the elementaryschool level, and then reinforced and extended every year through middle school, high school, and beyond. Scaling laws are central to physics. This has been true since Day One of modern science. Galileo presented several important scaling results in 1638 (reference 1 or reference 2).
This document is meant to be a tutorial, covering the simplest and most broadly useful scaling laws.
At the other extreme, there are more subtle scaling laws that are used to solve very deep and complicated problems at the frontiers of scientific research. The importance of scaling continues to the present day. There are dozens of references to “scaling” at the Nobel Prize site (reference 3).
Don’t let this scare you away. To repeat: this document is meant to be a tutorial, covering the simplest and most broadly useful scaling laws. You don’t need to be a Nobel laureate to get a lot of value from scaling laws.
You may be familiar with a simple form of scaling in connection with scale models, such as in figure 1 and figure 2. (The two figures are the same, except that one has a larger scale than the other.) The same word shows up in connection with smallscale and largescale maps. For more about the terminology, see section 3.
Perhaps the best known scaling law pertains to the relationship between length and area. In figure 3, when it comes to length, every length in the large square is twice as great as the corresponding length in the small square. When it comes to area, you can see that the area of the large square is not twice as great, but rather four times as great as the area of the small square.
Now let’s see what happens if we scale up the lengths by a factor of 3 rather than 2. In figure 4, when it comes to length, every length in the large square is three times as great as the corresponding length in the small square. When it comes to area, the area of the large square is nine times as great as the area of the small square.
The general rule here is the socalled “square law”:
(ratio of areas) = (ratio of lengths)^{2} (1) 
In words, the ratio of areas goes like the second power of the ratio of lengths. Equivalently, you can say that the ratio of areas goes like the square of the ratio of lengths.
Continuing to examine the relationship between length and area, let’s see what happens with triangles. In figure 5, when it comes to length, every length in the large triangle is twice as great as the corresponding length in the small triangle. So the question is, how does the area of the large triangle compare to the area of the small triangle?
The right answer is that the area of the large triangle is four times the area of the small triangle, as you can see from figure 6.
So you can see that the “square law” in equation 1 is absolutely not limited to squares. In fact, squareness has got nothing to do with it.
The key idea here is that areas depend on width times height. So if the width is scaled up by a factor of K and the height is scaled up by a factor of K, the area necessarily gets scaled up by a factor of K^{2}.
We can extend this idea into three dimensions, as shown in figure 7. When it comes to length, every length in the large cube is two times as great as the corresponding length in the small cube. When it comes to area, the surface area of the large cube is four times as great as the surface area of the small cube. When it comes to volume, the volume of the large cube is eight times as great as the volume of the small cube.
The volume goes up by a factor of eight because the cube is twice as wide and twice as deep and twice as tall. That’s three factors of two.
Meanwhile the surface area of the cube went up by a factor of four, not a factor of eight. That’s because on each face of the cube, the surface goes like width times height; the surface does not have any thickness. Each face of the cube is locally two dimensional, even though if you put all six faces together the total surface extends across three dimensions. Reconciling these facts requires a sophisticated notion of dimensionality. We say that the surface is intrinsically twodimensional but is embedded in three dimensions.
The scaling laws for volume, area, and length can be expressed in terms of equations:
 (2) 
and if you want to get fancy, this can be expressed in one big equation
(ratio of volumes)^{(1/3)} = (ratio of areas)^{(1/2)} = (ratio of lengths)^{1} (3) 
In equation 3, if a quantity has extent in N spatial dimensions, we take the Nth root.
It is straightforward to scale things by factors that are not integers, as illustrated in figure 8. When it comes to length, every length in the large square is 1.5 times as great as the corresponding length in the original square. When it comes to area, the area of the large square is 2.25 times as great as the area of the original square.
You can confirm this graphically by counting squares on the far right of the diagram. Each of the tiny squares has 1/4th of the area of the original square. You can see that the large square has 9/4^{ths} of the area of the original square.
You can understand the factor of 2.25 in several ways. One way is to take the ratio of lengths (1.5) and multiply it by itself using long multiplication. You don’t even need to do long multiplication if you remember that 15 squared is 225. Another way is to recognize 1.5 as 3/2, which you can square in your head to get 9/4, which reduces to 2.25. This corresponds to scaling up the lengths by a factor of 3 and then scaling down by a factor of 2.
Of course the same thing works for triangles. When it comes to length, every length in the large triangle is 1.5 times as great as the corresponding length in the original triangle. When it comes to area, the area of the large triangle is 9/4^{ths} as great as the area of the original triangle.
The scaling law for triangles is even more useful than the scaling law for squares, because it is easy to divide any polygon into triangles. By applying the scaling law to each of the triangles, you can easily prove that the scaling law must apply to any polygon.
Going one step further down that road, note that the area and perimeter of any ordinary (nonfractal) plane figure can be approximated as closely as you want by a polygon. For example, a circle is (for most purposes) well approximated by a manysided regular polygon. In this way you can convince yourself that equation 1 applies to a very wide class of figures.
The same argument applies in three dimensions. First, convince yourself that equation 2 and equation 3 apply to any tetrahedron. Then note that any polyhedron can be divided into tetrahedra. Conclude that equation 2 and equation 3 apply to any polyhedron.
In principle, all of trigonometry is a collection of scaling laws, based on the idea of similar triangles. If you know triangle ABC has one right angle (as shown in figure 10) and you know one of the other angles (such as α), then you can infer the third angle. That triangle is similar to every other triangle with those three angles. All such triangles are related by a simple scale factor ... and lengthratios such as a/c, b/c, a/b are scaleinvariant.
We can give names to these ratios:
 a/c is called sin(α),
 b/c is called cos(α),
 a/b is called tan(α),
and so on.
We start with the triangle ABC as shown in figure 11. We construct line CD perpendicular to AB.
Now we have three triangles: the lower triangle ACD which has hypotenuse b, the upper triangle CBD which has hypotenuse a, and the whole triangle ABC which has hypotenuse c. You can easily show^{1} that these are all similar to each other, so scaling arguments apply.
The upper triangle is a scaleddown copy of the whole triangle, with each length scaled by a factor of a/c. Similarly the lower triangle has each length scaled by a factor of b/c. The whole triangle, naturally, is scaled relative to itself by a factor of c/c = 1.
Area scales like length squared.
You can also see that the area of the whole triangle is exactly covered by the two smaller triangles, which means:
 (4) 
Equation 4 says that the area of the upper triangle plus the area of the lower triangle is equal to the area of the whole thing. The first expresses this in dimensionless terms, using the area of the whole triangle as the unit of area. The second line expresses the same thing using conventional dimensions of area.
The second line of equation 4 is the conventional way of stating the Pythagorean theorem. The first line is another way of expressing the same idea, and can also be interpreted as a thinlydisguised version of the trigonometric identity:
sin^{2} + cos^{2} = 1 (5) 
Here is another way to look at this proof: Whereas the traditional pictorial version of the theorem erects a square on the outside of each side of the triangle, we erect similar right triangles on each side of the original triangle.
The scaling idea is the same no matter what shape we attach to the edges of the triangle, so we could use pentagons, as shown in figure 12: the area of the apentagon plus the area of the bpentagon is equal to the area of the cpentagon. We can generalize this to any shape, so long as the three attached shapes are all similar to each other.
This can be considered a generalization relative to the usual statement of the theorem. This point was made by Euclid, 2300 years ago. See Proof #7 in reference 4, which is a collection of proofs of the Pythagorean theorem. The theorem can be applied to any shape, but the theorem is most easily proved using triangles, as in figure 11.
I really like the proof shown in figure 11, because it is easy and wellnigh unforgettable, it yields additional insights beyond what was required, and the method is transferable to a host of other problems (although I rate the “dot product“ proof slightly higher, based on the same criteria).
Tangential remark: If you’ve never seen the dot product proof, here it is: For any nonzero vectors a and b in an Euclidean vector space, let c = a+b. Then
 (6) 
Two nonzero vectors are called perpendicular if and only if their dot product vanishes. This is the definition of “perpendicular” in this sort of vector space.
Consider the left half of figure 13. We see a circle inscribed in a square. The circle has area π a^{2} while the square has area 4 a^{2}. That means that compared to the square, the circle has π/4^{ths} of the area.
Now let us turn our attention to the right half of figure 13. We see an ellipse inscribed in a rectangle, in the ordinary symmetric way.
The right half of the figure is the same as the left half, except that the vertical lengths have been scaled by a factor of λ, such that b = λ a. (The horizontal lengths are unchanged.)
The question is, what about the area? If we had scaled all of the lengths (horizontal and vertical) by a factor of λ, then the area would be scaled by a factor of λ^{2} – but that is not the case here. In fact we scaled one length and not the other, so we pick up only one factor of λ. Therefore we expect the rectangle to have area λ 4 a^{2} i.e. 4 a b. This result can be verified by applying the elementary formula for the area of a rectangle.
As for the area of the ellipse, using the same scaling argument, we expect it to be scaled relative to the circle by one factor of λ. That is, we expect the area of the ellipse to be λ π a^{2} i.e. π a b. This result can be verified if you happen to remember the formula for the area of an ellipse (π times the semimajor axis times the semiminor axis).
Note that the dimensionless ratio π/4 is unchanged by the scaling as we go from left to right in figure 13. That means that compared to the square, the circle has π/4^{ths} of the area ... and by the same token, compared to the rectangle, the ellipse has π/4^{ths} of the area.
 = 
 = 
 (7) 
If you want an absolutely rigorous proof of this result, you can express the area in terms of an integral, and then do a change of variable. The resulting factor of b/a then pops out in front of the integral, since an integral is a linear operator. That means you don’t even need to evaluate the integral in order to obtain the scaling result.
A more intuitive explanation for why this works has two parts: First, there is an areatoarea scaling rule: We expect the area of one thing to scale like the first power of the area of another thing. So when we go from left to right in figure 13, we scale the area of the square by a factor of b/a, and also scale the area of the circle by a factor of b/a.
As a corollary, since we have pairs of areas, the ratio of areas is a dimensionless number (π/4^{ths} for these particular pairs) and this dimensionless ratio scales like the zeroth power of the scale factor. That is, whenever a scale factor appears in a numerator in equation 7, it is canceled by an identical scale factor in the corresponding denominator.
Tangential remark: Reference 5 uses figure 13 as the starting point for a discussion of the relationship between memorizing things and rederiving things.
Suppose we want to make a movie of an enormous pendulum clock, and we want the physics to look correct. The obvious way to do it would be to build a fullsized clock and film it using a normal camera.
However, there is another way to do it. We could build a scalemodel clock, and then film it in slow motion. In other words, we modify both the lengthscale and the timescale. If we do this correctly, the result looks entirely natural. Specifically, the scalefactor applied to the time must be the square root of the scalefactor applied to the length, as mentioned in item 8.
Let us focus on the specific question of whether the behavior will look correct. To answer this question, at first glance it seems we need to know two things, namely the timescale (t) and the lengthscale (L). However, because there exists a scaling law, we really only need to know one thing, namely the ratio t/√L.
As a more extreme example, consider the question of whether the flow past a sphere will be turbulent. To answer this question, at first glance it seems we need to know four things, namely the speed (V), the size (L), the fluid density (ρ), and the viscosity (µ). However, because there exists a scaling law, we really only need to know one thing, namely the Reynolds number, i.e. the ratio ρVL/µ, as mentioned in item 32.
Any scaling law fits this pattern: It reduces the number of variables that you need to worry about. Actually, any law of nature fits this pattern, if you want it to. Usually the name “scaling law” is restricted to cases where the law involves some product of powers of the original variables.
The idea of replacing many variables by fewer variables has been rediscovered multiple times in various contexts, and has been given multiple different names. One way to describe it is to say that the correct physics occupies a subspace within the full parameter space. In the context of statistics, the same idea goes by the name sufficient statistic. In the context of criticalpoint physics, the same idea goes by the name of universality.
Part of the skill in making good scaling arguments involves using good terminology and avoiding bad terminology.
Here is some bad terminology: In figure 6, for example, it is dangerous to say that the big triangle is twice as large, or twice as big, or has twice the size of the small square. That’s because depending on the context, largeness and bigness and size could refer to the lengths involved, or could equally well refer to areas or volumes. Example: In a very practical sense, a twoliter bottle of water is “twice as big” as a oneliter bottle, even though its height and diameter are only 1.26 times as big.
This comes a shock to some people, and it is certainly inconvenient, but if you want to be careful you must avoid any unqualified statement about largeness, bigness, size, et cetera. If you mean length, say length. If you mean area, say area. And so forth. As an example, note that in section 3.2, we do not compare object A to object B; we compare the volume of A to the volume of B.
Here is some good terminology: People who are adept at making scaling arguments say things like this routinely:
The term “scale factor” is treacherous. Sometimes it refers to the ratio of lengths, but not always. On maps, it refers to the inverse ratio of lengths. That means that a largescale map has a small scale factor (such as 50 miles to the inch), while a smallscale map has a large scale factor (such as 500 miles to the inch). This is a source of endless confusion among nonexperts, and even experts get it wrong sometimes. I try to avoid the term “scale factor” whenever possible.
The phrase “... k times greater than ...” is so problematic that it deserves detailed discussion.
Consider the following scenario: V_{A} denotes the volume of object A, while V_{B} denotes the volume of object B. We are given that V_{A} = 12 liters and V_{B} = 36 liters.
The first three of the following statements are good terminology, but then things go to pot:
Discussion: statement 1, statement 2, and statement 3 describe V_{B} in absolute terms as a multiple of V_{A}. In contrast, statement 4 describes V_{B} in relative terms, as an increment relative to the amount of V_{A} we started with; the increment, not the whole amount of V_{B}, is given as a multiple of V_{A}. In statement 5, the word “greater” (or “more”) is a comparative adjective, which might have hinted that we are describing V_{B} by means of a relative increment, but the word “factor” overrides this hint and makes it clear that we are describing V_{B} in absolute (not relative) terms. This may not be entirely logical from a grammatical point of view, but the meaning of this expression is reasonably well established. We deprecate both statement 6 and statement 7, because it is hard to know which of them is correct. The word “times” suggests that V_{B} is being described in absolute terms, while the comparative adjective “greater” suggests that V_{B} is being described by means of a relative increment. I can’t say which interpretation is correct; some authors assume one interpretation, while others assume the other. I recommend avoiding both versions entirely, and using something like statement 1 or statement 4 instead.
A different problem crops up in statement 8. Does the statement mean that every dimension of B (length, width, and height) is bigger by a factor of 3, or does it only mean that the volume is bigger by a factor of 3? When you are scaling some property, you need to be specific about which property you are scaling.
In Euclidean geometry, objects that are the same except for scale and except for isometries (such as translation, rotation, and reflection) are said to be geometrically similar. (Beware: This conflicts with the vernacular use of “similar” to mean merely approximately alike.) Objects that are geometrically similar and have the same scale are said to be congruent.
The word “scale” in this context has the same meaning as in terms like scale model, and largescale or smallscale maps. It is etymologically related to such things as the graduated scale on a burette. It is distantly related to the musical scale. It is unrelated to such things as the scales on a fish, or the scales used for weighing things.
The simplest forms of scaling are also known as proportional reasoning – but more generally, scaling laws go far beyond simple proportionality.
Stiffness is the inverse of compliance. The meaning of stiffness and compliance can be seen in figure 14. The left end of the beam is held fixed, while the right end is free to move.
When we apply a force F, the beam is deflected a distance d relative to its resting position. Then we say the stiffness (or spring constant) is k, where
k = 
 (8) 
and the compliance is 1/k.
Now the interesting thing is that if the beam has length L, width w, and thickness t, the stiffness scales like w (which is unsurprising), and also scales like the cube of t/L.
k ∝ w  ⎛ ⎜ ⎜ ⎝ 
 ⎞ ⎟ ⎟ ⎠ 
 (9) 
This cube law may comes as a surprise to some people. As discussed in section 5, breaking strength scales like crosssectional area (in this case, width times thickness), and you might expect that stiffness would scale the same way, but it doesn’t.
A beam that is twice as thick is vastly stiffer than two beams in parallel, as shown in figure 15.
Look what happens if you bend two beams that are not connected, just placed in parallel. The stiffness scales like the number of beams, i.e. like the total thickness. However, the free ends of the two beams don’t line up, as you can see if you look closely at figure 15.
In contrast, if you have a doublethickness beam – or a pair of beams glued together so that they can’t slide relative to each other – then the ends are forced to line up. That means when you bend it, the top half gets stretched and the bottom half gets compressed. This stretch and compression involves a lot of energy, and contributes greatly to the stiffness.
An Ibeam is designed to cultivate this effect, as shown in figure 16. If you try to wrap an Ibeam around a cylinder of radius R, the inner flange will try to follow a circle of diameter 2πR, while the outer flange will try to follow a circle of diameter 2π(R+h), where h is the height of the web. Since the two flanges started out the same length, and the web will keep the two flanges lined up, you can’t bend the Ibeam without compressing the inner flange and stretching the outer flange. This makes an Ibeam very much stiffer than a solid bar of comparable weight. See reference 6.
Breaking strength (unlike stiffness) scales simply like crosssectional area.
Following Galileo, let’s see what this implies about the bones of animals. The weight of the animal scales like linear size cubed. If we just scaled up the bones in proportion, the breaking strength would scale like linear size squared, which is not enough to keep up with the weight. Therefore the bones of a large animal must be not just thicker, but disproportionately thicker.
The thickness must scale like linear size to the threehalves power. Galileo published this result in 1638 (reference 1, page 170 of the National Edition). Figure 17 is a copy of his original drawing, showing a scaledup picture of a small animal’s bone, compared to an actual large animal’s bone. (He doesn’t say exactly which animals he is comparing.)
Galileo also says (ibid)
I believe that a little dog might carry on his back two or three dogs of the same size, whereas I doubt if a horse could carry even one horse of his own size.
If something scales like x^{N}, its derivative with respect to x will usually scale like X^{(N−1)}. (There is an exception for N=0.) Sometimes the notation suggests this (as in dF/dx), but sometimes it doesn’t (as in ∇F).
By the same token, if something scales like x^{N}, its integral with respect to x will usually scale like X^{(N+1)}. (There is an exception for N=−1.)
For example, the electrostatic potential of a point charge falls off line 1/r. The electric field falls off like 1/r^{2}. The field of an electric dipole falls off like 1/r^{3}. See also item 42.
Similarly, the gravitational potential falls of like 1/r, the gravitational acceleration falls off like 1/r^{2}, and the tideproducing stress falls off like 1/r^{3}. See reference 7.
The moment of inertia about a given axis is given quite generally by I = ∫r^{2} dm. For constant mass m, the moment of inertia scales like r^{2} times m. On the other hand, for constant density, the mass scales like r^{3} and the moment of inertia scales like r^{5} times density.
When we say something scales, or is scalable, we mean we can change the scale of the thing and it still makes sense. For example, a triangle is scalable, because if we change the size it is still a perfectly fine triangle.
There exist plenty of things that are not scalable. As pointed out in section 5, an elephant is not just a scaledup elephantshrew. That means the animal as whole is not scalable. If you look at the skeleton of the shrew, you know what the “natural” size of the animal must be; if you tried to rescale it substantially, the result wouldn’t make sense. There is a natural sizescale for this animal.
Skeletal strength is not the only issue; the metabolism of any creature will generate heat in proportion to volume, while the ability to dissipate heat into the environment scales like surface area. This causes problems for large creatures (too much heat buildup) and for small creatures (too much heat loss). For details, see reference 8 and reference 9.
If something is scalable, the thing makes sense on many different sizescales; if something is not scalable, there is only one sizescale that makes sense for it.
Often when a complex system is not scalable, it is because of a conflict between various subsystems. In the case of animals, the weight scales one way, and the strength of the bones scales another way. Either weight or strength is scalable separately, but when we put them together we get a conflict.
Atoms are not scalable. The Bohr radius is a natural sizescale for atoms that is fixed by the fundamental physics. You can’t make a scaledup hydrogen atom … unless you monkey with the fundamental physics, and nobody knows how to do that in practice. (The fundamental constants are called constants for a reason. That doesn’t prove they are constants, but it means you should think twice before assuming they are easily variable.)
When part of the system scales one way and part scales another way, we can often obtain useful scaling laws that work over some part of the domain, often a very large part. For example, as discussed in reference 10, we have a good scaling law for longwavelength waves and another good scaling law for shortwavelength waves. Each of these laws is valid within its part of the domain; it is only in the crossover region that we see nonscalable behavior.
Many scaling laws work exceedingly well over a wide range of practical situations, but any scaling law applied to anything made of atoms must break down eventually, if you push it far enough, because atoms are not scalable.
Scaling laws are intimately connected to dimensional analysis; see reference 10 for an introductory discussion of the capabilities and limitations of dimensional analysis.
Always look for the scaling law. It might be obvious, or it might not. For example, if you can’t find a scaling law for some quantity X, maybe there is a nifty scaling law for (1−X). Keep looking.
Note that many scaling laws take the form of a power law, but not all; see item 35 in section 7.
The rationale here is simple: If you find a scaling law, it greatly increases your understanding of what is going on. Scaling laws are easy to use, and are very powerful.
Consider the relationship between scaling laws and detailed formal analysis. Neither is a substitute for the other; rather, they reinforce each other. The scaling law may sometimes tell you the answer you need, but even if it doesn’t, it suggests how to do the analysis. The analysis, in turn, may reveal additional scaling laws.
This is similar to the relationship between diagrams and formal analysis. Neither is a substitute for the other; rather, they reinforce each other.
Here are some interesting scaling laws. Some of them are selfevident, but some are not.
Many people find this counterintuitive. They understand that the circumference of the ring will increase, and they understand that the outer diameter will expand outwards, but they erroneously expect that the inner diameter will move inwards. In fact the inner diameter increases along with all the other linear dimensions.
This is a fanciful example, but similar examples exist in the real world. Roads, bridges, and railroad tracks really do exhibit thermal expansion, and arrangements must be made to accommodate this. Also if you are welding something, you have to take thermal expansion into account. If it’s the right size when it’s hot, it won’t be when it cools off.
You need to scale the time by a factor of √N. That is, you shoot the explosion at a higher frame rate, so that when it is played back everything is seen in slow motion. In accordance with item 9, this means that each falling object falls through given (scaled) length in the appropriate amount of (scaled) time.
For simplicity, assume that the dominant cost is the energy required to lift the stone blocks into place, doing work against the gravitational field, starting from ground level. Assume the pyramid is a solid collection of blocks, not hollow inside.
It should come as no surprise that the fundamental constant є_{0} (the vacuum permittivity) has dimensions of farads per meter. This illustrates the connection between scaling laws and dimensional analysis.
This is an example of nondimensional scaling. For details, see section 8.5.
In nondimensional scaling, you already know the scaling law, and all you have to do is write it down and recognize it as a scaling law, as in section 8.5.
In other cases, we need to discover the scaling law. As we shall see, that involves two key steps:
Dimensional analysis will help you with step (a), but it cannot possibly help you with step (b). Sometimes people with a superficial understanding of dimensional analysis expect it to tell them everything they need to know. In fact, for all but the simplest problems, dimensional analysis is less than half the battle.
There are some fields such as fluid dynamics where you will never get past Square One unless you understand nondimensional scaling.
As mentioned in item 20, the distance to the apparent horizon scales like the square root of your height h above the surface (subject to mild restrictions).
This is a favorite introductory example of nondimensional scaling. It shows that scaling laws are more general than dimensional analysis. Here we have a nice scaling law that could not have been derived from dimensional analysis alone.
There are two physicallyrelevant lengths in the problem, namely the eyeheight (h) and the radius of curvature (R) of the surface. The distance (s) to the apparent horizon scales like the square root of h … and also like the square root of R (assuming h≪R). Indeed, using the methods of highschool geometry you can show that the exact result is s=√2Rh, so the distance to the apparent horizon scales like the geometric mean of height and diameter.
Dimensional analysis can tell you that s = h^{x}R^{(1−x)} = R (h/R)^{x}, for some x … but it can’t tell you the value of the exponent x. Dimensional analysis doesn’t care whether x=0.5 or anything else, over a wide range. The existence of a physicallyrelevant dimensionless group (h/R) makes this a nondimensional scaling problem, outside the reach of dimensional analysis.
As foreshadowed in item 19, the mean free path (λ) of a particle in a gas of similar particles scales like the molar volume (V) of the gas divided by the scattering crosssection (σ) of the particle.
This example is interesting, because it demonstrates that scaling laws often depend on understanding the physics.
If you tried to find a scaling law for the mean free path by mindless application of dimensional analysis, you wouldn’t get far. There is a natural length in the problem, namely the average spacing between particles, which scales like the cube root of the molar volume. You might guess that the mean free path would scale the same way. Such a guess would pass the test of dimensional analysis, but it does not pass the test of real physics.
You may have noticed that there is another length in the problem, namely the square root of the cross section. That allows us to form the dimensionless quantity
π_{1} = 
 (10) 
whereupon we could write the mean free path as
 (11) 
where the exponent a is (as yet) undetermined. Since π_{1} is dimensionless, dimensional analysis is powerless to tell us anything about the exponent a.
So let’s look at the physics. Let’s take a large region of the gas, roughly a cube of size D, and divide it into N thin layers. A view facing one such layer is shown in figure 20.
You can see that a particle incident on this layer would have a small probability of scattering against one of the particles in the layer. Call this probability p, with a lowercase p. The overall probability of passing through N such layers without scattering is
 (12) 
where P (with a capital P) denotes the probability that the particle will scatter somewhere in the N layers. Meanwhile, we also expect
 (13) 
which basically defines what we mean by the mean free path λ. Comparing these two equations, we expect p N to scale like D/λ.
We also expect p to scale like the number of particles in the layer, times the cross section of each particle. If we keep the molar volume constant, that means that p scales like σ D/N. Combining all the pieces, we find
 (14) 
That means we must have a=2 in equation 11, so the scaling law for the mean free path is
 (15) 
as advertised. This is a very simple, very useful result.
As another way of expressing the same result, we can approximate the average spacing between particles as
 (16) 
hence
 (17) 
We know that the density of a typical gas such as N_{2} or H_{2}O at STP is about a thousand times less than the density of the corresponding liquid. That tells us that x is on the order of ten times larger than the size of the molecule. We can plug that into equation 17 to obtain an estimate that says the mean free path is probably about 100 times larger than the interparticle spacing at STP.
In contrast, in the interplanetary medium in the inner solar system, the interparticle spacing is on the order of 1 cm or slightly less. That means the mean free path is verrry long indeed.
Let’s consider the ultrasimple chemical reaction
F_{2} ↔ 2F (18) 
and in particular let’s consider the equilibrium state in a vessel where that reaction is occurring in the gas phase. Let X denote the reaction coordinate, i.e. the degree to which the reaction has proceeded toward the right. Specifically, X=0 if we have 100% “reactants” (molecular F_{2}), and X=1 if we have 100% “products” (monatomic F).
We choose conditions of temperature and molar volume such that X is initially small but nonzero. We hold the temperature constant, and increase the system volume V by moving a piston. We predict that increasing the volume by a factor of Q increases X by a factor of √Q.
Let’s work out the numbers for a simple scenario. We adopt the convention that in this context (a gasphase reaction), square brackets [⋯] denote number density, i.e. the reciprocal of the molar volume. This is the convention used in introductorylevel chemistry courses and in many advanced, practical applications. (If you are interested in activity rather than density, see below.)
Let’s consider the scenario shown in equation 19. We perform a stepbystep analysis, proceeding from state A via B to C. We arrange the initial state such that the F_{2} molecules have a number density of 2 moles per cubic meter. We arrange the temperature etc. such that the number density of unbound F atoms is less by a factor of 100. This initial condition is shown in column A in equation 19.
 (19) 
As a first step, proceeding from state A to state B, we expand the system volume (V) by a factor of 2. We do this sufficiently quickly that we can temporarily ignore any chemical reactions; that is, we assume equation 18 takes place slowly whereas the expansion takes place quickly. The shortterm result of the expansion is shown in column B in equation 19. Both number densities have gone down by a factor of two, while the reaction coordinate X remains unchanged. This is is an outofequilibrium situation.
As the final step, we allow the chemical reaction to come to equilibrium under the new conditions. The number density of unbound F atoms will increase. (This will deplete the density of F_{2} molecules, but only by a small percentage, which we ignore.)
The final state is shown in column C in equation 19. You can see that the total number of unbound F atoms is √2 larger than it was initially. (The density is lower, but the system volume is bigger, so allinall the number is bigger.) To say the same thing more precisely, the reaction coordinate X has increased by a factor of √2, as expected.
There are at least two simple theoretical arguments why this must be so. One is a subset of the argument that leads to the Saha equation (equation 21). The other considers the balance of reaction rates for the forward and reverse reactions. In any case, theoretically or otherwise, the inescapable fact is that the equilibrium value of the reaction coordinate X changes as we move the piston.
Another thing you can see is that the ratio in the bottom row is the same in both equilibrium situations. This leads us to define
K_{d} := 
 (20) 
where the RHS is called the equilibrium ratio (aka equilibrium quotient) and where K_{d} is called the equilibrium “constant”.
We see that in equation 19, K_{d} has the same value in column C as it has in column A. That is, the equilibrium “constant” is constant under the conditions of this isothermalexpansion scenario. Alas, the equilibrium “constant” is not constant under other conditions. For starters, the equilibrium ratio – the RHS of equation 20 – is strongly dependent on temperature. So calling K_{d} the equilibrium “constant” is something of a misnomer.
There are good reasons why [F] appears squared in the numerator of the equilibrium ratio. There is a rule that says it must be squared because there are two units of F on the RHS of equation 18. The physical basis of this rule is connected to the fact that the recombination reaction F+F→F_{2} is a secondorder process, in contrast to the disintegration reaction F_{2}→F+F which is a firstorder process. So you expect the rates for the two reactions to depend on density in different ways.
In equation 19, we would like to be able to predict the scaling based on the dimensions ... but we can’t. The scaling is highly nontrivial. We see that the equilibrium quotient has units of density, as do [F] and [F_{2}]. Let’s consider the expansion and the expansion/reaction process separately:
There is no way you could have predicted this overall scaling behavior based on dimensional analysis.
This conflict between the dimensional analysis and the actual scaling behavior may look like a serious mistake, but it is not. It just shows the limitations of dimensional analysis. See section 8.3.2 for an explanation of how this comes about. As always, if the dimensions are telling you one thing and the scaling is telling you another, the scaling is incomparably more reliable because it is more closely connected to the physics. Dimensional analysis is only a hint as to how the scaling might go.
Here are the main points:
In equation 19, there are actually two volumes in the system. The obvious one is the system volume V. The lessobvious one is the quantumstatistical volume Λ^{3}. Explaining all there is to know about Λ^{3} is beyond the scope of the introductory course. However, it exists and is crucial to understanding the scaling of K_{d}.
The bottom line is that if you scale the volume of the system, it changes moreorless every density in the system, but it does not change K_{d}. Yes, K_{d} has dimensions of density, but it scales like the inverse of Λ^{3}, not like the inverse of the system volume V. That’s why the equilibrium ratio actually remains reasonably constant when you scale the system volume at constant temperature.
Note: Sometimes you may be required to take the logarithm of K_{d}. The units of log(K_{d}) (as defined by equation 20) are nontrivial and must be carried additively (not multiplicatively), as discussed in reference 10.
If all you want is an introductorylevel notion of what’s going on, stop here. The rest is beyond the scope of the introductory course.
Note: The example of a gasphase reaction was chosen to keep things as simple as possible. However, the same ideas can be applied to reactants in solution. An ideal solution is closely analogous to an ideal gas. Concentration is almost synonymous with number density, in the appropriate units. In the gasphase example we don’t need to worry about complications such as solventsolute interactions.
We can begin to see what’s really going on if we look at the Saha equation. It is usually applied to the ionization of interstellar hydrogen, in which case it takes the form
 = 
 exp( 
 ) (21) 
where [n_{p}] is the number density of free protons, [n_{e}] is the number density of free electrons, [n_{H}] is the number density of neutral hydrogen atoms, and E is the hydrogen ionization energy, i.e. the Rydberg.
As usual, Λ denotes the thermal de Broglie length,
 (22) 
Comparing equation 21 with equation 20, we see that K_{d} has legitimate dimensions of inverse volume … but it does not scale like the inverse volume of the system. It scales like the inverse de Broglie length cubed. That means it depends on temperature and on immutable physical constants such as the mass of the electron.
We can use this to construct a dimensionless equation
 (23) 
where this K is dimensionless (unlike K_{d}). In fact this K is just equal to the relevant Boltzmann factor. Things could hardly get any nicer or simpler than this.
The trick is that on the LHS of the top line of equation 23, rather than using some arbitrary “unit” of volume when expressing densities per unit volume, we have used the physicallyrelevant quantumstatistical volume Λ^{3}. (We slipped in the approximation that Λ_{H} is very nearly equal to Λ_{p}.)
This (moreorless) allows us to write
K := 
 (24) 
where {⋯} denotes the dimensionless activity (in contrast to [⋯] which still denotes number density, as in equation 20). (See reference 15 for an introduction to the notion of activity.)
In equation 24, K is dimensionless and (as expected) scales like the zeroth power of system volume. However, this did not solve any of the fundamental problems. We saw in connection with equation 20 and equation 19 that we had three quantities, all with the same dimensions. Of these, one scaled V to the 1 power, one scaled like V to the ½ power, and one scaled like V to the 0 power. When we make things dimensionless, equation 24 still has three quantities, and one scales V to the 1 power, one scales like V to the ½ power, and one scales like V to the 0 power. There is obviously no way you can predict all three of these results using dimensional analysis. (The best you can do is play whackamole, choosing dimensions so that one of the three scales the way its dimensions would suggest, while the other two conflict with their dimensions.)
We learned in high school algebra that it is OK to multiply both sides of an equation by the same thing. Therefore we can always convert equation 20 from dimensionful to dimensionless form by multiplying by some constant. This changes the dimensions of the equation, but it cannot possibly change the meaning! It’s just algebra. You should not imagine that equation 24 is in any way more correct or less correct than equation 20. They provably have exactly the same physical significance.
There is no escape from the fact that key elements of this problem are beyond the reach of dimensional analysis.
The key to understanding this is to note that there are two relevant volumes: the volume of the vessel, and the quantumstatistical volume Λ^{3}. We can construct perfectly good scaling laws, provided we base them on honesttogoodness physics, not mere dimensional analysis.
There is, alas, one huge fly in the ointment. In the context of chemistry, activity {⋯} is not conventionally defined using the quantumstatistical volume Λ^{3}. Believe it or not, it is conventional to stick in some completely arbitrary adhoc volume instead, namely the molar volume of air at STP or something like that. That makes activity an illegitimate dimensionless quantity as discussed in reference 10. (You can make anything appear dimensionless using dirty tricks like that.) That trick also throws away the temperature dependence of Λ^{3}, thereby disguising the temperature dependence of the equilibrium ratio. It’s awful.
Also beware that many authors denote activity by [⋯] as opposed to the {⋯} used here. This makes it difficult to tell at a glance whether density or activity is intended. Actually it is even worse than that, because you can find six separate definitions for [⋯], based on density, pressure, concentration, and the three corresponding activities.
Remark: The scaling properties of the equilibrium ratio provide a valuable lesson in the unity of statistical mechanics and thermodynamics. This is important because (believe it or not) there are some people who argue that the statistical definition of entropy (as discussed in reference 16) is somehow irrelevant to thermodynamics and to practical chemistry. Well, we see in this case stat mech is not just relevant, it is indispensable. Without it we would have no way to understand why activity is dimensionless yet scales like density. Classically there is only one relevant volume, namely the volume of the system. We need statistical mechanics to tell us about the other volume, namely Λ^{3}. It is only after we know about both of those volumes that we have any chance of understanding the dimensions and the scaling.
Suppose a wave (aka ray) encounters the interface between one medium and another medium with a slightly different index of refraction. There will always^{2} be some reflection from such an interface. The amount of reflection depends on the angle, and on the change in index, in accordance with the Fresnel equations.
Let the relative change in index be called є. That is, n_{2} / n_{1} = 1 + є, for some small є. It may be helpful to think of є in terms of the logarithmic derivative: є = Δ(ln(index)).
You can show that at any particular angle, the amount of reflection scales like є squared. This is a famous result, worth remembering. Perhaps more importantly, you don’t need to remember it, because you can rederive it whenever you need it, just by looking at the structure of the Fresnel equations. Expand the RHS as a Taylor series in є. The zerothorder term is zero, the firstorder term is zero, and the secondorder term is nonzero. This is an easy exercise in theoretical physics, involving little more than differentiating a polynomial. The coefficients of the polynomial are complicated functions of θ, but still it’s just a polynomial function of є. It’s even simpler as a function of n, and dn/dє = 1, so it comes to the same thing either way.
Depending on how you write the Fresnel equations, you may need to invoke Snell’s equation and/or the trig identity sin^{2} + cos^{2} = 1. I call that the “Pythagorean” trig identity, for obvious reasons.
We say that there is a universality property here. There is a universal curve for є^{2} R_{p}(є, θ), independent of є, when є is small.
We can immediately use this scaling law to understand that there should be no reflection from a gradual change in index. Divide the overall index change into N layers, with an indexchange є on the order of 1/N at each boundary. The reflection per layer scales like 1/N^{2}, so even if you add up all the layers, the overall reflection goes to zero in the largeN limit.
Suppose we randomly assign n items to d slots, and we care about the probability of a collision, i.e. two or more items in the same slot. For any given probability of collision, the number of items you can have scales like the square root of the number of slots.
A trivial example is the famous “birthday collision” question. Nonexperts tend to guess that n should scale in proportion to d. That’s what simple notions of dimensional analysis would suggest, but it’s nowhere near correct.
Nontrivial examples crop up in lots of places, including in computer science (e.g. compiler construction) and in cryptography.
In more detail: Provided n ≪ d, the probability of collision is given (to a good approximation) by:

Note that this applies to the entire probability curve, not just the point where the probability equals 50%. This is an example of universality, as defined in section 2. That is, by invoking the scaling law, we can plot the entire curve as a function of the single variable n(n−1)/2d.
Equation 25b has the advantage of being strictly a power law, as is customary for scaling laws, whereas equation 25a has the advantage of being slightly more accurate when n is not very large.
You may have noticed that all the exponents in equation 2 are integers. These integers reflect the intrinsic dimensionality of the objects involved. Until fairly recently, most things were considered onedimensional, twodimensional, or threedimensional … with nothing in between.
However, that’s not a hardandfast rule. There are some things that exhibit scaling exponents that are not integer, nor even rational numbers. Critical phenomena (as mentioned in item 40) are one example. Note that these are real, physical phenomena, not some perverse mathematical fantasy. Quasicrystals are another physical phenomenon with fractal properties.
The word “fractal” originally referred to fractional dimensionality. To make sense of that, you have to realize that it refers to Hausdorff dimension as opposed to the more familiar topological dimension.
According to the Hausdorff idea, we measure area – such as the area of a plane figure – according to how many disks of a given size it takes to cover the figure. Similarly we measure length – such as the length of the perimeter of the figure – according to how many disks it takes to cover the perimeter. We replace the idea of “scaling up the figure” with the idea of “using smaller disks”. Using smaller disks captures the same idea as “looking more closely” at the figure.
For ordinary (nonfractal) figures such as squares or triangles, the Hausdorff area scales like the square of the Hausdorff perimeter ... as expected.
In contrast, for fractals, if you look more closely you might see additional squiggly details on the perimeter, such that the perimeter grows more quickly than the square root of the area.
For more about fractals, see reference 17 and references therein.
The elementary type of scaling we are talking about today is part of a package of techniques for checking the work.
As always, some good advice is:
Check the units, check the dimensions, check the scaling more generally, check the symmetry, check that the vectors and scalars behave as they should, et cetera.
Following this advice may slightly increase the workload in the short run, but it greatly reduces the workload in the long run, especially when dealing with complex problems.
For example, in the neighborhood of equilibrium, to lowest order, the force had better be an odd function of x, and the energy had better be an even function. That is a symmetry check. It’s not exactly a scaling law, but it’s in the same spirit.
All such checks need to baked into the curriculum, like the oatmeal in oatmeal cookies; they can’t be sprinkled on afterwards. To say the same thing the other way, it does not make sense to have a National Scaling Day where everybody suddenly learns how to do scaling.
The scaling we are talking about here is not particularly tricky. If students are having a problem with it, the problem is almost guaranteed to not be a scaling problem per se. It is much more likely to be a symptom of a deeper problem, perhaps a poor grasp of basic algebra.
Some of the other things you can do with scaling – like conjuring up new equations out of thin air – involve serious industrialstrength wizardry. We can talk about that some other day.