Fitting Percentage of Body Fat to Simple Body Measurements: College Women

Abstract Percentage of body fat, age, weight, height, and 14 circumference measurements (e.g., waist) are given for 184 women aged 18–25. Body fat, one measure of health, was accurately determined by an underwater weighing technique which requires special equipment and training of the individuals conducting the process. Modeling body fat percentage using multiple regression provides a convenient method of estimating body fat percentage using measures collected using only a measuring tape and a scale. This dataset can be used to show students the utility of multiple regression and to provide practice in model building.


Introduction
As is well-documented in the medical literature, obesity is linked with high risk of disease (e.g., Finer 2015). For example, obesity is associated with high risk for both cancer (e.g., Pischon and Nimptsch 2016) and cardiovascular disease (e.g., Elagizi et al. 2018). Also, with respect to the COVID-19 pandemic, the Centers for Disease Control and Prevention (2021a) indicated that obesity increases the risk of both severe illness and hospitalization due to COVID-19 infection. While obesity may be quantified in a variety of ways (e.g., Deurenberg and Yap 1999), one such measure is percent body fat.
In this article, we suggest students predict percent body fat using least squares/multiple regression with the accompanying dataset. Students may be drawn to the body fat data of this article because of the uniform desire for good health. For younger students, the threat of COVID-19 infection will likely be of more immediate concern than that of either cancer or cardiovascular disease. Other articles in this journal discussing fitting a response to multiple predictor variables include Binnie (2004), Carter et al. (2014), Hartenian and Horton (2015), Kuiper (2008), Peterson and Ziegler (2021), and Witt (2013). In addition to containing datasets that are likely to be of interest to a wide range of students, these articles contain anecdotal information on pedagogy useful for improving student performance on least-squares/regression learning outcomes. Johnson (1996)  which requires special equipment and training of the individuals conducting the process. Building a predictive model for the response variable body fat in terms of the predictor variables age, weight, height, and the circumference measurements then allows men to use this model to estimate their own body fat without undergoing the inconvenient underwater weighing process. Only a tape measure and a scale then need to be used for such prediction. The dataset associated with Johnson (1996) has had some visibility in the statistical literature. Publications using this data for illustrative purposes and homework include, for example, Faraway (2014), Lock et al. (2017), Matloff (2017), Mills (2005), Mooney and Swift (1999), and Olive (2017). This dataset, in addition to being available at the URL listed in the references (Johnson 1996), may be found within a few R packages including mfp (with data(bodyfat)) and UsingR (with data(fat)).
The dataset accompanying this article is complementary in that, while similar measurements are taken, this time they are from young women. In particular, percentage of body fat (again, through underwater weighing), age, weight, height, and fourteen circumference measurements for 184 female students aged 18-25 from Brigham Young University are given. These data appear in the master's thesis of Dr. Jason V. Slack (Slack 1997). According to Slack (1997), "a total of 184 females ages 18-25 years were recruited from volunteers attending Brigham Young University. Subjects were notified through physical education activity classes, elementary education teaching classes, and word of mouth. " Dr. Slack (Slack 2021) believed these women to be "mostly Caucasian. " Dr. Fisher, listed above, served as the thesis advisor to Jason Slack. As was the case for the dataset on men, the dataset associated with this article provides a convenient way for young women to predict their percent body fat.
While a handful of different modeling approaches may be made to predicting body fat, including regression trees (Breiman et al. 1984) and neural nets (e.g., Chandra et al. 2016, Garcia andEspinosa 2013), we limit our discussion to prediction using least squares/multiple regression. Principal components analysis might be implemented to address the instability in model coefficients that result as highly correlated measurements are added or removed from a least-squares/regression model (e.g., Manly and Navarro-Alberto 2017). In particular, linear combinations of predictor variables may be used to form "new" predictor variables, but this is not detailed here.
I have used the women's body fat dataset in just a single offering of the second semester probability and statistics course at my university. The great majority of the roughly thirty students this class had either an industrial engineering or mathematics major. The first semester probability and statistics course in this two-semester sequence includes, among other topics, exposure to descriptive statistics and to hypothesis testing on a single population parameter using p-values. The statistical software package R is also introduced. The second semester course is split fairly evenly between more advanced hypothesis testing (e.g., several population samples) and least-squares/regression. While two semesters of calculus are required for this probability and statistics sequence, calculus should not be viewed as necessary for using the suggested student investigations in this article.
I assigned students to write a polished, word-processed leastsquares/regression report (see Appendix A) in which they fit percent body fat for women to the various predictors. They were to show a comprehensive set of approaches to developing such a model, clearly showing their reasoning, verifying any needed assumptions, and commenting on the likely error a female user in the 18-25 age range might encounter. Student learning outcomes were broadly classified as follows: • Implementing and explaining a variety of model building techniques using either a summary measure (such as adjusted R 2 ) or p-values, for example, • Best subsets • Forward selection • Backward elimination • Verifying any needed regression assumptions • Including, with discussion, supportive descriptive statistics (e.g., residual plots, normal probability plots) • Including, with discussion, hypothesis tests (e.g., formal tests of normality) • Quantifying the accuracy of their model for young women not used to build their least squares/regression model • Using the standard error • (Optional) Using cross-validation to get a more honest estimate of the typical error A rubric for evaluating these least squares/regression projects that addresses the above learning outcomes, as well as quality writing, appears in Appendix B. I did not, by the way, ask students to use any of the predicted body fat models to predict their own body fat. It is virtually certain that some students would be embarrassed to do so. My basic strategy in discussing model building with least squares/regression, that seemed to work well in previous offerings of this second-semester course, was to first discuss and demonstrate, using other datasets and especially the dataset on men (Johnson 1996), specific smaller topics (e.g., using backward elimination with adjusted R 2 to fit a response variable) within the broad category of least squares/regression model building. The demonstrations had a dual purpose-an aid to student learning of conceptual topics as well as their developing comfort with the R statistical software. Demonstration R code and relevant data was generally emailed to students prior to class so that students could execute such in class on their own (campus leased and required) computers. Understanding of these specific smaller topics was then assessed with homework exercises (given in Johnson 2020) with feedback prior to further assessment of several of such least squares/regression topics on exams. Graded exams with feedback were returned to students prior to the least squares/regression project for young women being assigned, which was due at the end of the term. The best student projects-and there were a few of these, seamlessly wove together the different aspects of building a model for body fat in a clearly written narrative. At least some of the weaker projects were by students who, as confessed during my office hours, did not allow sufficient time to complete the project.
While I did not conduct a formal survey of reactions to this project, students seemed to enjoy such based on the questions and general interaction I had with a subset of them. Not all students were especially keen on the importance I placed on writing well. (I believe few of the technical classes on our campus contain a significant writing component.) The biggest change I would make in any future least squares/regression project would be to include intermediate assignments to keep students from procrastinating. Such intermediate assignments could include (i) providing an outline (e.g., proposed sections), and (ii) a preliminary draft or drafts. Another obstacle concerned a couple of my students who transferred in credit for the first semester course. They found R to have a rather steep learning curve as their transferred courses did not make use of any statistical software. In the remainder of this article, concepts that tended to be difficult for students are pointed out in the relevant sections.
Numeric computation and graphical displays presented in this article are obtained using version 4.03 of R (R Core Team 2020). Selected R code associated with this article is listed in Appendix C. We often display numeric output to the complete number of decimal digits displayed by R for the sake of reproducibility.

Initial Inspection of Data
Percent body fat, Age, Weight (kg), and Height (meters) have already been mentioned as variables. Here is a list of the 14 circumference measurements on the women (all in cm): Ankle, Biceps, Calf, Chest, Elbow, Forearm, Hips, Knee, Neck, DThigh (distal thigh), MThigh (middle thigh), PThigh (proximal thigh), Waist, and Wrist. (Predictor variables are capitalized from here on.) Slack (1997) indicated that these circumference measurements were made "according to the procedures and locations recommended by Lohman, Roche and Martorell" (1988). The remaining variable in the dataset, Body Mass Index (BMI), is calculated as Weight in kilograms divided by squared Height, Height measured in meters. BMI may be viewed, then, as a weight measure with an adjustment for size of body frame. The Centers for Disease Control and Prevention (2021b) suggest "BMI can be used as a screening tool" to help categorize individuals as underweight (BMI < 18.5 kg/m 2 ), normal weight (18.5-24.9kg/m 2 ), overweight (25.0-29.9 kg/m 2 ) or obese (≥ 30 kg/m 2 ). BMI alone has been used to predict body fat (e.g., Mills 2005).
For the various numerical and graphical displays in this article, we retain the units used by Slack (1997). If students are more comfortable dealing with measurements in inches and pounds, then they may perform the appropriate conversions (1 inch = 2.54 cm, 1 kg ∼ = 2.205 lbs) prior to data exploration and analysis.
The bilateral circumferences Ankle, Biceps, Calf, Elbow, Forearm, Knee, DThigh, MThigh, PThigh, and Wrist each refer to measurements on just one side of the body. For researchers who take such measurements, it is standard to always take measurements on a fixed side of the body (Lohman, et al. 1988, chap. 4). Dr. Slack adhered to this practice and believes the measurements were taken on the right side, partly because he currently takes measurements on the right side. He admits, however, that he is not absolutely certain of such as these data were collected many years ago (Slack 2020). The Hips measurement, by the way, is taken across both hips.
Before having students diving into fitting body fat to the other variables, instructors may suggest students consider whether any data values could, in fact, be erroneous. Because of the redundancy of Height, Weight and BMI, the consistency of these three variables may be investigated. Doing so we find that for observation 34 that these three measurements are not consistent. The other observations, up to rounding, do seem to be consistent in these three measurements. Moving away from internal consistency of variables, other errors are apparent, or are at least likely, in the dataset. Boxplots, for example, may be used to identify (likely) errors. Figure 1, for example, shows a boxplot of Age. Slack (1997) indicates that the women in his study ranged in age from 18 through 25 years old. The boxplot shows two outliers: observation 137 concerns the only 25-year-old in the dataset, while observation 120 lists a 1-yearold woman-clearly an error. Table 1 lists other likely recording errors in the dataset; this list is not meant to be exhaustive. See, for example, the article by Holcomb and Spalsbury (2005) for further discussion of data integrity. For the sake of reproducibility, all of the numeric output that follows is done without modifying values in the dataset (without, for example, replacing values as NAs in R).
Helpful Hints: As already suggested, instructors may find it helpful to use the dataset on men (Johnson 1996) to demonstrate analyses they wish their students to be able to perform. Students may then, either individually or in groups, be tasked with similar analyses with the dataset on young women. To make analyses a bit easier, instructors may optionally provide their students with just a subset of the predictor variables. To help prevent copying of work, different subsets of the 184 cases may be used either within a class or across semesters.

Simple Numerical and Graphical Analysis
Producing pairwise plots of percent body fat against each of the remaining eighteen predictor variables provides some guidance as to whether particular predictors are useful as well as how predictors may be incorporated into a least squares/regression model. The plots of body fat against Age, Height, Hips, and Weight, for example, are shown in Figure 2. It seems that neither Age nor Height is useful in helping to predict percent body fat while both Hips and Weight should be useful. Examining the other pairwise plots suggests, other than Age and Height, each of the remaining sixteen predictor variables may be useful to at least some degree in predicting body fat. In each of these sixteen scatterplots there is nothing to suggest other than a linear relation between any predictor and body fat exists. So, while pairwise plots only give two-dimensional projections of what is happening in a higher-dimensional space, it is reasonable to try fitting body fat as linear in all the predictor variables excluding Age and Height.
To objectively quantify the strength of the linear associations in these pairwise graphical displays Pearson's correlation coefficient between body fat and the various predictors may be computed. These are shown in Table 2. Since at least some of the various (remaining) sixteen predictor variables are (intuitively) strongly associated, we hope to fit percent body fat by just a subset of these sixteen variables-giving a more parsimonious model. Students may be convinced of the likely redundancy of some of the predictors by examining additional pairwise plots and computing the correlation coefficients between predictors. Figure 3, for example, shows a plot of Weight versus Hips, these variables having a correlation of about 0.885. It seems likely, then, that any final predictive model for body fat would not need to incorporate both Weight and Hips.

Least Squares: Best Subsets
In this section we use the best subsets procedure to help build reasonable models for percent body fat. For simplicity, we focus on adjusted R 2 for measuring model performance.
Other measures of performance (e.g., Mallow's C p , Akaike's AIC), of course, may be used instead. Students generally need to be reminded that maximizing adjusted R 2 is equivalent to minimizing the standard error, or s.e., of regression. In particular, where s y is the standard deviation of percent body fat. The leaps package in R is convenient for finding the best adjusted R 2 models for a fixed number of predictor variables. In general, leaps() displays what is referred to as a "best subsets" collection of models. For p total predictor variables (for us, Figure 3. Weight versus Hips. ) and a fixed number of predictor variables, say k, the various p k subsets with k predictors are considered, and the best combination is displayed. To make this a bit more concrete, among the 16 3 = 560 subsets of 3 our 16 predictor variables, the subset with BMI, Hips, and Waist produces the best adjusted R 2 . This, along with the best subset for other fixed values of the number of predictors, is displayed in Table 3. In my experience, students tend to overly focus on single measures of model performance such as adjusted R 2 to the exclusion of other considerations, notably parsimony. That this was largely not an issue with the least squares/regression assignment may be due to the explicit instruction in this assignment (see Appendix A) to have a balance between accuracy and parsimony. The best adjusted R 2 model in our case has 9 predictor variables-but this a rather inconvenient number of measurements for a young woman to take to predict her body fat. Notice, from either Table 3 or Figure 4, that the highest values of adjusted R 2 plateau around about 0.66. It would be much easier for a user to use, say, just three measurementsnamely, BMI, Hips, and Waist giving an adjusted R 2 of about 0.63 which is not too far from the optimum. Using the lm() command in R, we obtain the fitted model The standard error with this model is 3.543%. The residuals range from about −9.35 to 9.43 with a first quartile of about −2.35 and a third quartile of about 2.21.
The above fitted coefficients are not strongly influenced by whether or not the observed values of BMI, Hips, and Waist for any particular woman are included or excluded from the model building process. Cook's distance, one such measure of influence, is small for each of the 184 observations. While there is lack of agreement as to what constitutes an observation having a large Cook's distance and high influence, it is not uncommon for practitioners to use a cutoff value of either one or one-half.
In the model just fit above, the largest Cook's distance for the 184 observations was just under 0.068.
While none of the 184 observations apparently have a strong influence on the fitted model coefficients, several of the women have unusual values of BMI, Hips, and Waist compared to the other women. The observations corresponding to these women are referred to as high leverage observations. Observation leverage values lie between 0 and 1 and, if k is the number of predictors (k = 3 in our case) and n is the number of observations (n = 184), the average leverage value is (k + 1)/n. According to one rule of thumb, observations with leverage values greater than three times this average are deemed to have high leverage. With the fitted model on BMI, Hips, and Waist, three of the 184 observations have high leverage (observations 104, 140, and 151 with leverages 0.188, 0.133, and 0.081, respectively). However, none of these high leverage observations are influential. For an observation to be influential in the Cook's distance sense, the observation must have both high leverage and a residual large in magnitude. The residuals for our three high leverage observations are all "small. " For further details on leverage and influence see, for example, Chatterjee and Hadi (2012).
For instructors using R who wish to discuss leverage and influence, once the output of R's lm() command is saved (e.g., fit <-lm()), one of the plots produced by a subsequent plot() command (e.g., plot(fit), alternatively par(mfrow=c(2,2)) followed by plot(fit)) will display a scatterplot of (standardized) residuals versus leverage values and flag, by observation number, observations with Cook's distance larger than one-half and larger than one. Leverage and Cook's distance are also among the diagnostics output with R's influence.measures() command.

Least Squares: Forward Selection and Backward Elimination Procedures
In this section we briefly examine the forward selection and backward elimination procedures to help build reasonable models for percent body fat. Again, for simplicity, we use adjusted R 2 for measuring model performance. Some elementary combinatorics gives a total of 2 p − 1 = 2 16 − 1 = 65,535 different models for body fat in terms of various linear combinations of our 16 predictors, containing at least one predictor. I argue in my least squares/regression classes that, because of this exponential explosion in the number of predictors, p, it might be good to have a backup plan to the best subsets procedure, in case our p is moderately large to large.
With forward selection on adjusted R 2 we start the process by determining the single predictor giving the best adjusted R 2 . This is the same as choosing the predictor with the highest absolute correlation coefficient. According to the correlations listed in Table 2, this is the predictor BMI. Keeping BMI as a predictor, the next forward selection step is adding a second predictor so as to maximize adjusted R 2 . This, after performing the corresponding 15 least-square fits using BMI and one other predictor, is accomplished by adding the predictor Hips. This process continues with, at each step, keeping the j predictors already selected at the previous step and then, after performing the corresponding p − j least-square fits, adding that variable that maximizes adjusted R 2 . Forward selection with our 16 predictors is summarized in Table 4. The best forward selection model is that which has the highest adjusted R 2 . In our case, this involves fitting percent body fat as linear in the nine variables BMI, Hips, Waist, Wrist, Knee, Ankle, Chest, Weight, and Biceps. (The forward selection procedure stops after trying to add Calf to our model as the adjusted R 2 decreases.) This is also the best subsets model from the previous section, but these two methods do not give the same result in general. Again, for a more Backward elimination starts by including all p predictor variables in the model. At each step one predictor variable is removed. With j predictors in the current model, after performing the corresponding j least-square fits with just one of these predictors dropped, we remove that variable resulting in the highest adjusted R 2 . The process stops if dropping another variable decreases the adjusted R 2 value. A summary is given in the two left-most columns of Table 5. The reader should verify that using backward elimination gives the same model for body fat as obtained before by both the best subsets and the forward selection process. In general, in terms of whatever optimization criterion is used (for us adjusted R 2 ), forward selection and backward elimination may not give models that perform as well as that obtained by the best subsets procedure. By only either adding or subtracting predictor variables (a nested sequence of models), we fail to examine all 2 p − 1 models considered by the best subsets procedure.
We omit discussion of general stepwise procedures that allow for either the addition of a predictor or the removal of a predictor at any given step. Details may be found in any number of regression textbooks (e.g., Draper and Smith 1998).

Regression Model Assumptions
Instead of using adjusted R 2 in the forward selection or backward elimination procedures, one may use p-values of the individual predictor variables (contingent on the other predictor variables in the model). With backward elimination, for example, body fat may be fit to the full model involving all 16 predictors. We then drop the variable with the highest p-value, specifically MThigh, with a p-value of 0.998412. Then body fat is fit to the remaining 15 predictor variables other than MThigh and we again eliminate that variable with the highest p-value. This time it is DThigh with a p-value of 0.937797. This removal process continues (see the left-and right-most columns of Table 5) until all individual coefficient variables have p-values below some threshold. Using 0.05 for this threshold (this is quite common) gives a predictive model for body fat in terms of the predictors Chest, Knee, Ankle, Waist, Hips and BMI (see Table 5). Using 0.01 for this threshold gives the model with only Waist, Hips, and BMI (again, see Table 5). A partial summary of the R output for this three-predictor model appears above in Section 4.
The two sequences of predictor variable deletions-one using adjusted R 2 and one using individual predictor p-values (nonequivalent measures), need not be the same. In this case they are, and this allows the common sequence to be displayed in Table 5. Typically, the threshold to remove using p-values, as suggested above, is fairly small. The reader should verify from Table 5 that any threshold below 0.205042 will lead to a smaller model using p-values than that optimizing adjusted R 2 .
Forward selection may also be performed using p-values. One problem in doing so using R is that several of our candidate predictors, when used singly, all produce a p-value which is stated as less than 2 × 10 −16 . In this case, the smallest p-value will correspond to the t value which is largest in magnitude. This occurs with BMI. Continuing, we find forward selection here using p-values produces the same sequence as listed in Table 4 using adjusted R 2 . This is not the case in general.
The use of p-values has come under greater scrutiny over the last few years (e.g., Wasserstein and Lazar 2016). Because of this, it is generally considered good practice to supplement their use with other approaches, such as using adjusted R 2 above. When prediction is the main goal of model fitting, the adjusted R 2 approach rather than the p-value approach is generally preferred when implementing stepwise procedures. With respect to our dataset, in fact, which contains some highly correlated (multicollinear) predictor variables, some practitioners suggest avoiding stepwise model building with p-values altogether. Multicollinearity leads to greater variability in the estimated regression model coefficients and the resulting weakened statistical power may make it difficult to identify statistically significant predictor variables. See, for example, Chatterjee andHadi (2012, p. 309), or Faraway (2014, p. 153), for more details.
The trustworthiness of p-values in our forward selection and backward elimination procedures above is contingent upon the multiple regression model assumptions being reasonably satisfied. Below, in brief outline form, is a list of traditional approaches to checking the necessary assumptions:  Consider the situation, for example, where we have fit percent body fat as linear in the predictors BMI, Hips, Waist, and Ankle. A portion of the R lm() output associated with this fitted model appears in Table 6. If we are implementing backward elimination using p-values with, say, a threshold of 0.01, then Ankle would be the next predictor to drop provided we can trust these p-values. In Appendix C under "Checking regression assumptions, " R code is provided to perform the suggested checks a, b, c above in this situation (further details are omitted). As these checks can be reasonably shown to be satisfied, we may drop Ankle. This gives the resulting model in the 3 predictors BMI, Hips, and Waist previously discussed toward the end of Section 4.
A not uncommon point of student confusion on the project was the failure to clearly indicate the need to check regression model assumptions at each iteration of stepwise procedures involving p-values.
The fitted model on BMI, Hips, and Waist Predicted Percent Body Fat = −32.43547 + 0.80741(BMI) + 0.22019(Hips) + 0.22170(Waist) may be used to produce point estimates of percent body fat for specified values of BMI, Hips, and Waist, but prediction intervals are also often of interest. For instructors using R, a prediction interval for the percent body fat for a particular woman may be obtained using R's predict() command. For example, the code sequence fit < -lm(Fat˜BMI + Hips + Waist, data=women) values < -data.frame(BMI=22, Waist=70, Hips=97) predict(fit, newdata=values, interval= 'prediction', level=.95) gives a 95% prediction interval of (15.2%, 29.2%) on percent body fat for a woman with BMI 22 kg/m 2 , Waist 70 cm, and Hips 97 cm. The midpoint of this interval, 22.2%, is the point estimate of percent body fat for a woman having these three measurements. (Bailey 1991 suggests a maximum of 22% body fat for women to have good health.) Changing the argument interval='prediction' to interval='confidence' within the above predict() command gives a 95% confidence interval of (21.7%, 22.7%) for the mean percent body fat of women with these measurements. The regression model assumptions must hold for these intervals to be trusted. For further details on these intervals see, for example, Faraway (2014).

Honest Estimates of Error
Once a fitted model for percent body fat has been chosen, it is appropriate to indicate to model users the typical error to be expected when they use such to predict their own percent body fat. For concreteness, suppose we consider the fitted model involving BMI, Hips, and Waist from Section 4. Here, the listed s.e. of regression is 3.543% body fat. In layperson's terms, this value is the typical size of a residual. This s.e. value, however, is likely to be optimistic as this measure is for a model tuned to the given data. Unfortunately, a significant minority of students failed to mention this shortcoming in their least squares/regression project.
This shortcoming of the s.e. may be overcome by performing the following for each j, j running from 1, 2, …, to n = 184: • Leaving out observation j, fit body fat to BMI, Hips, and Waist. • Compute the predicted body fat for the predictor values in observation j using the above fitted model. Also compute the associated residual, call it e (j) .
The second step gives a honest error for the jth observation as the jth observation was not used in the building of the associated fitted model in BMI, Hips, and Waist. (The fitted coefficients in the models will generally vary across the values of j). With this leave-out-one cross-validation approach, a more honest estimate of likely user error, is given by cross − validation error = 1 n − (k + 1) where e j is the (ordinary) residual corresponding to the jth observation when fitting the full model on all n observations. Here, k is the number of predictor variables (in a model that allows for a nonzero additive constant); k = 3 when using BMI, Hips, and Waist.
While it may appear that this cross-validation process requires n + 1 different least-square fits (see Appendix C for this approach), only one least-squares fit need be made. (The ordinary and cross-validated errors are linked through what is called the 'hat' matrix, the details of which may be found in some regression texts. See, for instance, Abraham and Ledolter 2006, pp. 188 and 189.) Omitting further details, the crossvalidation error for the model fitting percent body fat in terms of BMI, Hips, and Waist is very nearly 3.62%-a reasonably small estimated typical error.

Concluding Remarks
The women's body fat dataset provided contains the predictor variables Age, Weight, Height, and fourteen circumference measurements along with the response variable (percent body) Fat. These variables were measured on 184 women. This dataset may be used by instructors to illustrate a number of topics in least squares/multiple regression. These include initial numerical and graphical exploration of the data-including data integrity, variable selection/model building, verifying regression assumptions, and providing an honest estimate of the likely error for users. By way of reminder, the data provided are from a sample of convenience of young women who are mostly Caucasian. Fitted models to predict percent body fat should be limited to such users.
To assist instructors who wish to use this dataset in the classroom, a sample assignment and sample grading rubric are included. Relevant R code is also included.

Appendix A: Least Squares/Regression Assignment
Goal: Fit percent body fat to predictor variables for young college women. Details regarding these variables are provided in a separate (Word) file. Likewise, the actual data are provided in a separate (text) file. It is not enough to simply produce a fitted model. You are required to submit a word-processed report explaining your reasoning to justify your final fitted model. Also, your report should comment on the likely accuracy of using your final model for estimating percent body fat for young women not used to fit this model.
You may either work alone or with exactly one partner. (More than one partner is not allowed.) If you have a partner, make sure that both names appear on the report.
Any inanimate resources (notes, textbook, D2L page materialsincluding class videos, Internet) are allowed. Working with other people (other than your partner, if you have one) is not allowed.
Due Date: The last day of class. Submit either as a Word document or a pdf file (just one file).

Further Details:
• Report Components: • Include a Main Report containing reasoning/justification of final model and comments on likely accuracy for young women not used to build this model. A brief introduction and a brief conclusion should be included. Appropriate summary tables and graphics should be included. The main report should not contain any software (e.g., R) specific details (e.g., R commands). Be very selective on including R output. A good rule-of-thumb: If your R output contains information that you do not address in your exposition, then you have included too much R output. Summarizing R output in a different format, however, is most welcome and highly desirable. The main report should not be more than 20 pages (single spaced, 12-point font).
Audience: Write to an audience of students who have taken this course before, but need reminding of important details. Be selective on the inclusion of graphics. It is enough, for example, to give illustrative examples of the full analysis you conduct. In particular, I do not need to see every plot of the response (body fat) versus each of the possible predictors in an initial culling of predictors. Likewise, if/when verifying regression model assumptions, I do not need to see every analysis performed. Show me that you understand how to verify these assumptions by illustrative example (e.g., in just one step in the backward and/or forward processes in the transition from one model to the next model).
Divide your main report into sections, with headings (Introduction, …, Conclusion).
• Include, as appropriate, a Reference section (see Technical Matters below). • Include an Appendix containing some detail on what R commands were used. A subset of commands used is fine, an exhaustive list of R commands used is not necessary and is not appropriate. You may find it helpful to have headings and, possibly, sub-headings, to organize your listing of R commands. This appendix should not be more than 5 pages (single spaced, 10-point or 12-point font). • (Optional) Add, as an additional appendix, the variable descriptions Word file provided (e.g., this would make sense if you would like to refer to such in your main report).
• A Caution: • The final model, to some extent, should be a parsimonious model. (Don't overly obsess on optimizing any one particular measure of model performance.) • Grading Components: • Exposition (clarity, grammar, spelling).
• Reasoning-giving clear details on your thought process, including justifying any needed assumptions. • How comprehensive you are in demonstrating a mastery of the breadth of material presented in class without being verbose. • How well you follow the directions in this document.
• Your attention to appropriate fine detail (e.g., appropriate labels on figures, including units) • • (Minor Detail) You may find the use of a proportional font such as Courier New in Word to be useful in aligning R output in a more readable way. Panel of scatterplots, Figure 2: