Some Comments about Chapter 7 of Hollander & Wolfe



This is a long chapter, with a lot of sections. I think that it will make sense to cover the first part of the chapter (through Sec. 7.4) rather thoroughly, while also inserting Quade's test which we can do using StatXact, and then focus on only several of the remaining sections of the chapter,

As usual, H&W include an Assumption A3 which stipulates a shift model. Not only do they assume a shift model, but the model statement (given about midway between (7.1) and (7.2) on p. 272) imposes an additive structure --- with no allowance for interaction effects, the order of the distribution medians and the distances between distribution medians is the same for each block. My guess is that a lot of times, data in a two-way layout should not be assumed to follow a simple additive shift model.

The notation in the book isn't what I like to see --- they use rows (1st index) for the block, and columns (2nd index) for the treatment (which is opposite of what is typically done with two-way ANOVA models for mixed effects). I guess I'll try to follow H&W's system for the indexing in order to match the book, but I won't use their Rj and R.j notation --- instead I'm going to use the convention of the dot indicating summation, and then adding a bar to indicate an average. Also note that they use n to denote the number of blocks, instead of the number of observations per cell as is usually done with a two-way ANOVA. I guess I'll try to go with their notation here. (At least one other nonparametric statistics book uses n this way, and also reverses the rows and columns from the usual ANOVA presentation. It's annoying.)

The blocks can be viewed as being either a fixed effect or a random effect.


Section 7.1

Pages 272 and 273 give the basics about Friedman's test, which is the one nonparameteric test associated with a two-way layout which is typically covered in a basic course on applied statistics.

I'll offer some specific comments about the text below.
p. 274
H&W suggest that if there are ties, then compute the adjusted test statistic, S', and compare it to the values in Table A.22. Of course a better thing would be to use StatXact, which handles ties as described in Comment 9. If one doesn't have access to StatXact, an alternative to H&W's approximate scheme would be to break the ties conservatively (to minimize the value of the test statistic, and maximize the p-value), assigning only integer ranks, and then go to Table A.22.
p. 276, Example 7.1
With n = 22 blocks, the chi-square approximation ought to do okay, even if the p-value is in the neighborhood of 0.005. Still, rather than depend on Minitab's approximate p-value (Minitab uses the chi-square approximation regardless of what n is), you might as well enter the data into StatXact and obtain an exact p-value. StatXact yields an approximate p-value of 0.0038, which is in agreement with Minitab's 0.004. StatXact's exact p-value is 0.0031. So while the chi-square approximation isn't horrible in this case, it doesn't even get one significant digit correct. (Note: If one is going to use Minitab, there is a better way to get the values into C1 and C2. To get the values into C1 enter the set C1 command, and then enter 22(1) 22(2) 22(3) at the DATA> prompt (and then end at the next DATA> prompt). To get the values into C2 enter the set C2 command, and then enter 3(1:22) at the DATA> prompt (and then end at the next DATA> prompt).) I'll also point out that if one incorrectly treated the data like 3 independent samples and did a K-W test, the resulting p-value is about 0.40 (both Monte Carlo estimate of the exact value, and the chi-square approximation). Typically, the p-value which results from ignoring the two-way design and incorrectly doing a K-W test is larger (but whether it is larger or smaller, it is incorrect, because one doesn't have k independent samples).
p. 276, Comment 1
Note that no allowance for interactions gives us an additive model. For the Rounding First Base example (Example 7.1), this means that it should be believed that the same method is best for all of the players, and the same method is worst for all of the players. The inclusion of an interaction term in the model would allow for some players to do best with one method, and other players to do best with another method. Of course, nonzero interaction terms makes it more difficult to draw a conclusion about which method is best. But I think it makes sense to allow for interactions. If they were found to be nonsignificant (say with an ANOVA F test, based on an assumption of iid normal error terms), then one could feel better about adopting an additive model and concluding that there may be an overall best method. But if the test for interactions is significant, one should conclude that the same method need not be best for every player (and more careful testing should be done to help determine which method is actually the best one for each player).
p. 277, Comment 4
This is an important comment, since it explains the motivation for the test statistic. (In the first displayed equation, the (k-1)!/k! can be replaced by 1/k (which can be viewed as a simplification of the expression obtained using combinatorics, or which can be viewed as resulting from a simple symmetry argument --- under the null hypothesis, Rij is equally likely to take the values 1, 2, ..., k, and so each of these values will occur with probability 1/k).) Comment 5 also provides some motivation for the test statistic --- it's a rank version of the normal theory two-way ANOVA F test.
pp. 277-278, Comment 6
Assumption A3' indicates the more general null hypothesis --- that the k observations (one from each treatment) in each block arise from iid random variables, but that the distribution of the random variables can be different from block to block. So basically, we have a null hypothesis of no treatment differences (but there can be block differences). If a small p-value results, then there is strong evidence of differences in the treatment distributions for at least some of the blocks. (With this test applied to data from a randomized block design, it's hard to reach firm conclusions about the treatments --- even if one assumes a shift model (a common error term distribution for each cell), a small p-value only imples that not all of the treatment effects are the same, but the test doesn't provide any indication of which treatments are different from which other treatments --- it could be that they are all different, or it could be that only one differs from the rest, which are the same, with the one that is different being either smaller or larger than the others.)
p. 278, Comment 8
This is a good example of the type of thing that I refer to as a "brute force" derivation of a null sampling distribution.
p. 281, Comment 11
StatXact gives us some alternatives not included in H&W. One can consider Quade's test, and possibly stratified two-sample tests (if there are just two treatments).

Section 7.2

This is a two-way layout version of the test that I referred to in class as being a competitor to the J-T test --- the one which can be done using StatXact's Linear-by-linear Association test, applied to data in a one-way layout. Since that Ch. 6 type of test is so similar to Page's test for a monontone alternative in a two-way layout, I often refer to the one-way version as being a test similar to Page's test.

The test described in this section can be done using StatXact's Page test. (I'll post some comments about using StatXact for Page's test on my StatXact web page.)

I'll offer some specific comments about the text below.
p. 284
The alternative given by (7.9) is for a "one-sided" monotone alternative. One could also consider a "two-sided" monotone alternative, to be used in cases for which it is reasonable to think that if there are differences between treatments, then they will be of a monotone nature, but it isn't clear if the values will monotonically increase or decrease. (E.g., it might be thought that if a drug will have an effect, it will either monotonically increase or decrease with increasing dosage, but it is not known what the direction will be if differences are observed. Often the two-sided alternative makes sense when considering "side-effects." For example, if the drug is supposed to lower blood pressure, then a one-sided alternative could be considered to see if it is effective in doing so, and if greater dosages correspond to larger decreases (but not necessarily of a strictly monotone nature, or else a regression analysis may be more sensible). But one could also be concerned if the blood pressure medication affected something else (a side effect of sorts). For example, it could increase or decrease the amount of saliva in the mouth. In such a case, the desired outcome may be one of no change, but one could test for a difference, and anticipate that either a monotonically increasing or decreasing effect will be observed if there is any effect.)
p. 286, Large-Sample Approximation
Since L is integer-valued, it could be that a continuity correction of 1/2 will generally improve the normal approximation. If we consider Example 7.5, using the table in the back of H&W, one can only determine that the exact p-value is between 0.0014 and 0.0041. However, StatXact can be used to determine that the exact p-value is about 0.0025 (and you can use this value as a check to make sure that you're doing Page's test correctly using StatXact). The normal approximation without a continuity correction yields a p-value of about 0.0040, and the normal approximation with a continuity correction yields a p-value of about 0.0047. So, in this case, the continuity correction made the approximation worse. In another case that I considered in my course last summer, for which the p-value was also small, the continuity correction made the approximation worse. But I think it is the case that when the p-value isn't real small, a continuity correction can improve matters. (It can be noted that for k = 2, Page's test reduces to the sign test, and it is known that a continuity correction generally improves the normal approximation for a sign test (except for some cases for which the p-value is rather extreme).)
p. 186, Ties
StatXact is the best way to deal with ties --- it does things as is indicated by Comment 19 on pp. 289-290. For this test, H&W don't give the adjustment for ties for the null variance that makes the normal approximation perform better (see Comment 21 on p. 292 for more about the normal approximation in the presence of ties (in particular, how it's conservative when there are ties)), and of course the tables in the back of the book are derived for the case of no ties.
pp. 286-287, Example 7.2
From what I can gather from H&W, and also going to the original source (the classic book by Cochran and Cox), the data is from a randomized block experiment, which had three plots of land divided into five smaller subplots, with the 5 levels of potash randomly assigned to the five subplots within each large plot. (I suppose that it was thought that other characteristics of the the land could influence the strength of the cotton, and that a randomized block design would be better than a one-way layout in which 15 small plots of land were randomly assigned to the five treatment levels. I'll discuss this in class.) Four measurements were made for each of the 15 cells, but since Page's test only uses one observation per cell, the four measurements for each cell were averaged to yiled just one value per cell.
p. 287, Comment 14
This is an important comment, since it explains the motivation for the test statistic.
p. 287, Comment 15
This comment acknowledges that the test can still be a distribution-free test even if a shift model doesn't hold.
p. 288, Comment 17
Two facts contribute to the (k!)n = 36 possibilities indicated on p. 288 being equally-likely under the null hypothesis of no differences due to treatments: So for the case at hand, we have 3! * 3! = 6*6 = 36 equally-likely possibilities, and in general we have (k!)n equally-likely possibilities under the null hypothesis.
I'll also point out that to get an upper-tailed test p-value the "brute force" way, one can save some work by arranging the 36 equally-likely possibilities in a 6 by 6 table, and noting that (that when done in a sensible way) one has symmetry. Also, (when the table is constructed in a sensible way (I'll explain this in class)) one can note that the largest values of the test statistic occur in one corner of the table, and so it is typically not necessary to fill out the entire table to determine an upper-tail probability. (StatXact can obtain an upper-tail probability by using similar "tricks" to avoid consideration of all (k!)n equally-likely possibilities.)
p. 289-290, Comment 19
In class, I'll go over the determination of the p-value (the null probability that the test statistic exceeds the observed value of 23) for the example addressed on p. 290.
p. 292, Comment 22
We'll encounter Spearman's rank order correlation coefficient in Ch. 8.

Section 7.3

This is a two-way layout version of the one-way layout procedure described in Comment 63 in Sec. 6.5 (p. 247). Like the main procedure covered in Sec. 6.5 (the S-D-C-F multiple comparison procedure for one-way layouts), the procedure of Sec. 7.3, which I guess we can refer to an the W-N-M-T procedure, can be used to identify which treatments are significantly different from one another, and can be used as a test of the null hypothesis of no differences against the general alternative that there are some differences. But the procedure described in this section is more similar to the procedure described in Comment 63 on p. 247 because they both use joint ranks, whereas the S-D-C-F procedure does pairwise comparisons for which the other k-2 treatments have no influence in the determination of whether there is significant evidence that a given pair of treatments differ. (More on this is given in Comment 30 on p. 299.)

The first paragraph of Sec. 7.3 describes the procedure as one which would typically be applied after a rejection is obtained with Friedman's test (in order to determine which treatments differ from which other treatments), but as the last paragraph of Comment 24 indicates, the procedure can be used as a test (which is a competitor to Friedman's test).

I'll offer some specific comments about the text below.
p. 295, Procedure
Another way to describe the critical value, ralpha, that is equivalent to (7.26), is to say that the null hypothesis probability that at least one of the k choose 2 absolute differences of rank sums is at least ralpha is alpha.
p. 295, Procedure
StatXact isn't helpful in doing this procedure, and so we'll have to make use of Table A.24 of H&W or the asymptotic approximation based on the studentized range distribution. Unfortunately, Table A.24 only gives a few probabilities for each combination of k and n covered, and so it won't be (easily) possible to always produce exact p-values (and we may have to do something like report that 0.008 < p-value < 0.032). To obtain the needed rank sums, one can note (see p. 276) that Minitab's friedman command gives these, but at times a hand calculation may be quicker than putting the data into Minitab.
p. 296, Large-Sample Approximation
As usual, I don't agree with the otherwise decide part of statements like (7.27). Just because there is not strong evidence to conclude that there is a difference, it doesn't mean that there is no difference. There could be a somewhat mild difference. Also, it can be noted that a strict interpretation of the rule (7.27) can lead in inconsistencies. For example, as is the case with Example 7.3, one can conclude that methods 1 and 2 are equivalent, that methods 2 and 3 are equivalent, but that methods 1 and 3 differ.
p. 296, Ties
The use of midranks doesn't correspond to the assumptions underlying the table of the exact values (Table A.24). Also, it could produce a value, such as 21.5, which is not included in the table.
pp. 296-297, Example 7.3
To determine at what level of significance it can be concluded that method 2 differs from method 3, one can divide the absolute difference in the rank sums, 15, by the square root of 22(3)(4)/12 (i.e., the square root of 22). and compare the resulting value to the critical values of the studentized range distribution with 3 and infinity df (which are given in Table A.17). The resulting value is about 3.198, and using the 2nd row of entries in Table A.17 on p. 669, it can be concluded that methods 2 and 3 are different at level 0.10, but not at 0.05.
Using the last 2 lines of p. 296 and the first 2 lines of p. 297, it can be concluded that the p-value which results from a test of the null hypothesis of no differences dues to methods against the general alternative that there are some differences satisfies 0.001 < p-value < 0.005. The exact p-value from Friedman's test is about 0.0031, and so both tests give about the same result.
The last paragraph of the example considers the reduced data set comprised of just the first 15 cases. The largest absolute difference in rank sums is 15. Since Table A.24 includes exact results for the k = 3 and n = 15 case, it can be used to determine that the p-value of the test of the null hypothesis of no differences due to methods against the general alternative satisfies 0.010 < p-value < 0.028. To see what the large-sample approximation yields, we divide the largest absolute difference of 15 by the square root of nk(k+1)/12 (the square root of 15*3*4/12 = 15), obtaining a value of about 3.873, which can be compared to the entries in the 2nd row of Table A.17, which suggests that 0.010 < p-value < 0.025, in close agreement with the result from the table of the exact distribution.
p. 299, Comment 30
One way to avoid letting the k-2 other treatment results influence the determination of whether or not there is statistically significant evidence to conclude that a given pair differ is to do paired-sample tests on each k choose 2 pairs of samples, and then combine the results using Boole's inequality to obtain a conservative "overall" p-value of a test of the null hypothesis of no differences due to treatments against the general alternative. For small k and n an exact (nonconservative) test could be performed using this general scheme, but I don't know of any tables for such a test (and have never heard anyone propose such a test).

Section 7.4

This is a two-way layout version of the one-way layout procedure described in Sec. 6.7. Like the procedure covered in Sec. 6.7, the procedure of Sec. 7.4, which I guess we can refer to as the N-W-W-M procedure, can be used to identify which treatments are significantly better than the control (which could be a commonly used treatment that new treatments are compared to, or it could be no treatment, or even a placebo). It can also be used as a test of the null hypothesis of no differences against the alternative that there are some differences (see the 2nd paragraph of Comment 32 on p. 303), and if certain assumptions are made (like a shift model holds, or that if a treatment differs from the control it corresponds to an upward shifting of probability mass), then conclusions can be stated in terms of means and/or medians.

I'll offer some specific comments about the text below.
p. 301, Procedure
Another way to describe the critical value, r*alpha, that is equivalent to (7.29), is to say that the null hypothesis probability that at least one of the k - 1 differences of rank sums (of the form treatment minus control) is at least r*alpha is alpha.
p. 301, Procedure
StatXact isn't helpful in doing this procedure, and so we'll have to make use of Table A.25 of H&W or the asymptotic approximation. To obtain the needed rank sums, one can note (see p. 276) that Minitab's friedman command gives these, but at times a hand calculation may be quicker than putting the data into Minitab.
p. 301, Large-Sample Approximation
As usual, I don't agree with the otherwise decide part of statements like (7.30). Just because there is not strong evidence to conclude that there is a positive difference, it doesn't mean that there is no difference. There could be a somewhat mild positive difference, or there could be a negative difference.
Note that to use Table A.21, you use k - 1 for l.
p. 301, Ties
The use of midranks doesn't correspond to the assumptions underlying the table of the exact values (Table A.25). Also, it could produce a value, such as 21.5, which is not included in the table.
p. 302, Example 7.4
For the illustration of the large-sample method at the bottom of p. 302, one can divide the largest difference, 6, by the square root of nk(k+1)/6, and use that as a test statistic value to compare against the critical values given in the l = 2 column of the rho = 1/2 table on pp. 691-692. The test statistic value is 6 divided by the square root of 18*3*4/6 (the square root of 36, which equals 6). So the test statistic value is 1. Using the x = 1.00 value of the l = 2 column, the p-value is found to be 1 - 0.74520 = 0.25480, which is in the ballpark of the 0.2859 value from the table of the exact distribution. (Note: I have no clue as to why the k = 6 value is given at the bottom of p. 302 --- I suspect someone got mixed up at some point and thought that it was relevant.)

Section 7.5

I'm not going to cover this section in class. The estimates only make sense if one can assume an additive model with iid error terms (i.e., the error term distribution is the same for each cell, and the relationships between the medians is the same for each block, giving us a shift model), and I don't believe that this is usually the case.


Section 7.6

I think that this is a relatively important section, since such BIBD data is sometimes encountered, and it may be unwise to rely on normal theory ANOVA methods (since it's very hard to check all of the assumptions with such sparse data ... and I suspect that a lot of the time the assuptions are rather severely violated).

I'll offer some specific comments about the text below.
p. 310 (1st line)
In class, I'll establish the restriction that is given (solving Problem 59 on p. 315 of H&W), but I challenge you to try to establish it on your own first.
p. 310, Procedure
In class, I'll show the equality of the two expressions for d given in (7.43) (solving Problem 60 on p. 315 of H&W). We've encountered similar things previously in Ch. 6 and Ch. 7, so this one time I'll show you how to go about establishing something like (7.43). Also, I'll derive the mean of D under the null hypothesis.
p. 311, Large-Sample Approximation
Note that the chi-square approximation is conservative, particularly out in the tail. It can be shown that the null variance of D is less than 2(k-1), which is the variance of a chi-square distribution with k-1 df, indicating that large values of D, when the null hypothesis is true, occur less frequently than would be the case if D in fact had a chi-square distribution with k-1 df. As usual, the conservativeness of the approximation under the null hypothesis translates into larger than deserved approximate p-values which corresponds to reduced power when the alternative hypothesis is true. For many cases not covered by the exact tables, a good Monte Carlo estimate of the exact p-value would be the next best thing, but this would require quite a bit of effort. (StatXact does not do the test covered by this section.)
p. 312, Comment 45
Note that a shift model arrangement is not necessary --- the null hypothesis sampling distribution would be satisfied if the s observations in each block are iid, and so we can view the null hypothesis as one of no difference between treatments in any of the blocks, but allowing the common response distribution for the treatments to differ from block to block in ways other than a shift. Basically, the generalization of the shift model allows for the error term distribution to differ from block to block. In the more general setting, a small p-value should be interpretted as being strong evidence of some differences between at least some of the treatments in at least some of the blocks. (With a shift model assumption, the interpretation is that there are some differences between at least some of the treatments, but that the differences are the same no matter what block is considered. The shift model is addressed in Comment 48 on p. 313.)
p. 313, Comment 48
Note that A3' corresponds to the more general setting I referred to in my preceding comment: when the null hypothesis is true, the treatments have a common distribution in each block, but the distribution need not be the same in each block. But A3' imposes an additive structure on the medians under the alternative hypothesis. That is, if there are differences between treatments, the spacing of the medians is the same within each block (with only the error term distributions differing from block to block). I think that the additive model assumption isn't very realistic in a lot of cases, since it doesn't allow for larger differences in some blocks than in others (and even no differences in some of the blocks).
pp. 313-314, Comment 49
This is a good example of the type of thing that I refer to as a "brute force" derivation of a null sampling distribution.
pp. 314-315, Comment 51
This is a good example of the type of thing that I refer to as a "brute force" derivation of a null sampling distribution in the presence of ties. If StatXact included the test of this section (Durbin's test, for short), I feel safe in assuming that it would make use of this type of exact sampling distribution to obtain p-values. Given that StatXact doesn't do the test, and given that the chi-square approximation need not be good to small designs, and given that I worked through some small design examples of exact null sampling distributions in the presence of ties for some other tests in Ch. 7 and showed that ties can change the distribution quite a bit from the case with no ties, I think it would be wise to go to the extra trouble to obtain an exact p-value in some small design cases if there are ties (but perhaps not if the approximate p-value is not at all smallish).

Section 7.7

The first paragraph of Sec. 7.7 describes the procedure as one which would typically be applied after a rejection is obtained with the test of Sec. 7.6 (in order to determine which treatments differ from which other treatments), but it can also be viewed as a competitor to the test of Sec. 7.6. To perform a test of the null hypothesis of no differences between treatments against the general alternative, we can reject at level alpha if any of the k choose 2 pairs of treatments are deemed to be significantly different by the criterion given in (7.46) on p. 317.

I'll offer some specific comments about the text below.
p. 317, Procedure
Note that the procedure is approximate, instead of exact.
p. 318, Comment 55
The conservtive procedure may have low power to detect somewhat mild differences. Also, it may be rare that you can apply it due to the extremely limited number of designs covered by Table A.26.

Section 7.8

Since I think that other topics that we can cover this fall may be more useful, I'm not going to cover this section in class. (Ch. 7 is a rather long chapter, containing many procedures. I think that on the whole, it'll be better to not spend too much time on Ch. 7, so that we can have more time for other chapters, and for procedures that we can do on StatXact which are not covered in H&W. Performing the test of Sec. 7.8 is made difficult due to the lack of tables for the exact distribution (with the exception of a relatively small number of cases).)


Section 7.9

I will present the basics of this test in class, but I don't have anything to make note of here.


Section 7.10

I will present the basics of this procedure in class.

As usual, just because this method is a multiple comparison procedure, I don't think that one has to think of it as a follow-up procedure to be applied after the rejection of the null hypothesis by another procedure. Instead, it can be used to test the null hypothesis of no treatment differences within blocks against the general alternative that there are some differences between treatments (in at least some of the blocks).

I'll offer some specific comments about the text below.
p. 340, Procedure
The procedure described on p. 340 is an approximate one which should work okay as long as n isn't too small (but unfortunately, no guidelines are presented to help us decide if n is large enough). To perform a test of the null hypothesis of no treatment differences against the general alternative, one can consider all k choose two pairs of samples and see if any of them correspond to a significant difference at some specified level. Alternatively, to get a p-value, one can divide the largest absolute difference (the largest Su minus the smallest Su) by the square root factor on the right side of the inequality in (7.75), and compare the value of this test statistic to the quantiles given in Table A.17.
p. 342, Comment 82
The alternative presented is a conservative procedure which may be rather weak in power, which leads to the recommendation that the approximate procedure described on p. 340 be used "whenever the number of blocks is reasonably large." Unfortunately, no guidelines are presented to help us decide if n is large enough.

Section 7.11

I will spend a relatively long time discussing this section during class. Also, I'll work though a simple example, step by step. In addition, I'll give the results from the application of Friedman's test and Quade's test to the same simple example data set. Then, I'll give the results from applying all three tests (Friedman's test, Quade's test, and Doksum's test (the test of Sec. 7.11)) to Woody Wardward's base running data (on p. 274).

I'll offer some specific comments about the text below.
p. 344, Procedure
I'll stick with H&W's notation when I present this section during class. My presentation will be easier to follow if you spend some time getting comfortable with all of the notation prior to my lecture on this section. For example, it should be kept in mind that Hu. assumes a large value when the treatment u observations are generally smaller than the observations from the other treatments.
p. 344, Procedure
I don't think that the text makes it clear why the variance given by (7.82) and (7.83) corresponds to a suitable part of the denominator for the sum of the squared deviations that is the heart of the test statistic. (My advice is not to worry about this matter for now.)

Section 7.12

I will present this test in class, but I don't have anything to make note of here.


Section 7.13

Since I think that other topics that we can cover this fall may be more useful, I'm not going to cover this section in class. (Ch. 7 is a rather long chapter, containing many procedures. I think that on the whole, it'll be better to not spend too much time on Ch. 7, so that we can have more time for other chapters, and for procedures that we can do on StatXact which are not covered in H&W.)


Section 7.14

Since I think that other topics that we can cover this fall may be more useful, I'm not going to cover this section in class. (Ch. 7 is a rather long chapter, containing many procedures. I think that on the whole, it'll be better to not spend too much time on Ch. 7, so that we can have more time for other chapters, and for procedures that we can do on StatXact which are not covered in H&W.)


Section 7.15

Since I think that other topics that we can cover this fall may be more useful, I'm not going to cover this section in class. (Ch. 7 is a rather long chapter, containing many procedures. I think that on the whole, it'll be better to not spend too much time on Ch. 7, so that we can have more time for other chapters, and for procedures that we can do on StatXact which are not covered in H&W.)


Section 7.16

I will briefly mention some things from this section when I cover Sec. 7.11, but I don't intend to get into the details of this section.